Unnamed: 0
int64 0
7.24k
| id
int64 1
7.28k
| raw_text
stringlengths 9
124k
| vw_text
stringlengths 12
15k
|
---|---|---|---|
6,200 | 6,609 | Attentional Pooling for Action Recognition
Rohit Girdhar
Deva Ramanan
The Robotics Institute, Carnegie Mellon University
http://rohitgirdhar.github.io/AttentionalPoolingAction
Abstract
We introduce a simple yet surprisingly powerful model to incorporate attention
in action recognition and human object interaction tasks. Our proposed attention
module can be trained with or without extra supervision, and gives a sizable boost
in accuracy while keeping the network size and computational cost nearly the
same. It leads to significant improvements over state of the art base architecture
on three standard action recognition benchmarks across still images and videos,
and establishes new state of the art on MPII (12.5% relative improvement) and
HMDB (RGB) datasets. We also perform an extensive analysis of our attention
module both empirically and analytically. In terms of the latter, we introduce a
novel derivation of bottom-up and top-down attention as low-rank approximations
of bilinear pooling methods (typically used for fine-grained classification). From
this perspective, our attention formulation suggests a novel characterization of
action recognition as a fine-grained recognition problem.
1
Introduction
Human action recognition is a fundamental and well studied problem in computer vision. Traditional
approaches to action recognition relied on object detection [11, 19, 55], articulated pose [29, 33, 34,
53, 55], dense trajectories [50, 51] and part-based/structured models [9, 54, 56]. However, more
recently these methods have been surpassed by deep CNN-based representations [18, 30, 41, 46].
Interestingly, even video based action recognition has benefited greatly from advancements in image
based CNN models [20, 22, 42, 45]. With the exception of a few 3D-conv based methods [46, 47],
most approaches [12, 14, 15, 17, 52], including the current state of the art [52], use a variant of
discriminatively trained 2D-CNN [22] over the appearance (frames) and in some cases motion (optical
flow) modalities of the input video.
Attention: While using standard deep networks over the full image have shown great promise for
the task [52], it raises the question of whether action recognition can be considered as a general
classification problem. Some recent works have tried to generate more fine-grained representations
by extracting features around human pose keypoints [8] or on object/person bounding boxes [18, 30].
This form of ?hard-coded attention? helps improve performance, but requires labeling (or detecting)
objects or human pose. Moreover, these methods assume that focusing on the human or its parts is
always useful for discriminating actions. This might not necessarily be true for all actions; some
actions might be easier to distinguish using the background and context, like a ?basketball shoot? vs a
?throw?; while others might require paying close attention to objects being interacted by the human,
like in case of ?drinking from mug? vs ?drinking from water bottle?.
Our work: In this work, we propose a simple yet surprisingly powerful modification to standard
deep architectures to learn attention maps that automatically focus the attention on specific parts
of the input relevant to the task at hand. We show that our attention maps can be learned without
any additional supervision and automatically lead to significant improvements over the baseline
architecture. Moreover, our attention maps are easy to interpret and provide insight into where the
network should look when making a prediction, both in terms of bottom-up saliency and top-down
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
attention. We further experiment with adding human pose as an intermediate supervision to the
attention learning process to encourage the network to look for human object interactions. While
this makes little difference to the performance of image based recognition models, it leads to a larger
improvement on video datasets as videos consist of large number of ?non-iconic? frames where the
subject of object of actions may not be at the center of focus.
Our contributions: (1) An easy to use extension of state-of-the-art base architectures that incorporates attention to give significant improvement in action recognition performance at virtually
negligible increase in computation cost; (2) Extensive analysis of its performance on three action
recognition datasets across still images and videos, obtaining state of the art on MPII and HMDB-51
(RGB) and competitive on HICO; (3) Analysis of different base architectures for applicability of
our attention module; and (4) Mathematical analysis of our proposed attention module and showing
its equivalence to a rank-1 approximation of second order or bilinear pooling (typically used in fine
grained recognition methods [16, 26, 28]) suggesting a novel characterization of action recognition
as a fine grained recognition problem.
2
Related Work
Human action recognition is a well studied problem with various standard benchmarks spanning
across still images [7, 13, 33, 35, 56] and videos [24, 27, 40, 44]. The newer image based datasets such
as HICO [7] and MPII [33] are large and highly diverse, containing 600 and 393 classes respectively.
In contrast, collecting such diverse video based action datasets is hard, and hence existing popular
benchmarks like UCF101 [44] or HMDB51 [27] contain only 101 and 51 categories each. This in
turn has lead to much higher baseline performance on videos, eg. ? 94% [52] classification accuracy
on UCF101, compared to images, eg. ? 32% [30] mean average precision (mAP) on MPII.
Features: Video based action recognition methods focus on two main problems: action classification and (spatio-)temporal detection. While image based recognition problems, including action
recognition, have seen a large boost with the recent advancements in deep learning (e.g., MPII
performance went up from 5% mAP [33] to 27% mAP [18]), video based recognition still relies
on hand crafted features such as iDT [51] to obtain competitive performance. These features are
computed by extracting appearance and motion features along densely sampled point trajectories in
the video, aggregated into a fixed length representation by using fisher vectors [32]. Convolutional
neural network (CNN) based approaches to video action recognition have broadly followed two main
paradigms: (1) Multi-stream networks [41, 52] which split the input video into multiple modalities
such as RGB, optical flow, warped flow etc, train standard image based CNNs on top of those, and
late-fuse the predictions from each of the CNNs; and (2) 3D Conv Networks [46, 47] which represent
the video as a spatio-temporal blob and train a 3D convolutional model for action prediction. In terms
of performance, 3D conv based methods have been harder to scale and multi-stream methods [52]
currently hold state of the art performance on standard benchmarks. Our approach is complementary
to these paradigms and the attention module can be applied on top of either. We show results on
improving action classification over state of the art multi-stream model [52] in experiments.
Pose: There have also been previous works in incorporating human pose into action recognition [8,
10, 58]. In particular, P-CNN [8] computes local appearance and motion features along the pose
keypoints and aggregates those over the video for action prediction, but is not end-to-end trainable.
More recent work [58] adds pose as an additional stream in chained multi-stream fashion and shows
significant improvements. Our approach is complementary to these approaches as we use pose as
a regularizer in learning spatial attention maps to weight regions of the RGB frame. Moreover,
our method is not constrained by pose labels, and as we show in experiments, can show effective
performance with pose predicted by existing methods [4] or even without using pose.
Hard attention: Previous works in image based action recognition have shown impressive
performance by incorporating evidence from the human, context and pose keypoint bounding
boxes [8, 18, 30]. Gkioxari el al. [18] modified R-CNN pipeline to propose R*CNN, where they
choose an auxiliary box to encode context apart from the human bounding box. Mallya and Lazebnik [30] improve upon it by using the full image as the context and using multiple instance learning
(MIL) to reason over all humans present in the image to predict an action label for the image. Our
approach gets rid of the bounding box detection step and improves over both these methods by
automatically learning to attend to the most informative parts of the image for the task.
2
Soft attention: There has been relatively little work that explores unconstrained ?soft? attention for
action recognition, with the exception of [38, 43] for spatio-temporal and [39] for temporal attention.
Importantly, all these consider a video setting, where a LSTM network predicts a spatial attention
map for the current frame. Our method, however, uses a single frame to both predict and apply
spatial attention, making it amenable to both single image and video based use cases. [43] also uses
pose keypoints labeled in 3D videos to drive attention to parts of the body. In contrast, we learn an
unconstrained attention model that frequently learns to look around the human body for objects that
make it easier to classify the action.
Second-order pooling: Because our model uses a single set of appearance features to both predict
and apply an attention map, this makes the output quadratic in the features (Sec. 3.1). This observation
allows us to implement attention through second-order or bilinear pooling operations [28], made
efficient through low-rank approximations [16, 25, 26]. Our work is most related to [26], who point
out when efficiently implemented, low-rank approximations avoid explicitly computing second-order
features. We point out that a rank-1 approximation of second-order features is equivalent to an
attentional model sometimes denoted as ?self attention? [48]. Exposing this connection allows us
to explore several extensions, including variations of bottom-up and top-down attention, as well as
regularized attention maps that make use of additional supervised pose labels.
3
Approach
Our attentional pooling module is a trainable layer that plugs in after any convolutional layer of
a CNN. As most contemporary architectures [20, 22, 45] are fully convolutional with an average
pooling operation at the end, our module can be used to replace that operation with an attention
weighted pooling. We now derive the pooling layer as an efficient low-rank approximation to second
order pooling (Sec. 3.1). Then, we describe our network architecture that incorporates this attention
module and explore a pose regularized variant of the same (Sec. 3.2).
3.1
Attentional pooling as low-rank approximation of second-order pooling
Let us write the layer to be pooled as X ? Rn?f , where n is the number of spatial locations (e.g.,
n = 16 ? 16 = 256) and f is the number of channels (e.g., 2048). Standard sum (or max) pooling
would reduce this to vector in Rf ?1 , which could then be processed by a ?fully-connected? weight
vector w ? Rf ?1 to generate a classification score. We will denote matrices with upper case letters,
and vectors with lower-case bold letters. For the moment, assume we are training a binary classifier
(we generalize to more classes later in the derivation). We can formalize this pipeline with the
following notation:
scorepool (X) = 1T Xw,
X ? Rn?f , 1 ? Rn?1 , w ? Rf ?1
where
(1)
where 1 is a vector of all ones and x = 1T X ? R1?f is the (transposed) sum-pooled feature.
Second-order pooling: Following past work on second-order pooling [5], let us construct the feature
X T X ? Rf ?f . Prior work has demonstrated that such second-order statistics can be useful for
fine-grained classification [28]. Typically, one then ?vectorizes? this feature, and learns a f 2 vector
of weights to generate a score. If we write the vector of weights as a f ? f matrix, the inner product
between the two vectorized quantities can be succinctly written using the trace operator1 . The key
identity, T r(AB T ) = dot(A(:), B(:)) (using matlab notation), can easily be verified by plugging in
the definition of a trace operator. This allows us to write the classification score as follows:
scoreorder2 (X) = T r(X T XW T ),
where
X ? Rn?f , W ? Rf ?f
(2)
Low-rank second-order pooling: Let us approximate matrix W with a rank-1 approximation,
W = abT where a, b ? Rf ?1 . Plugging this into the above yields a novel formulation of attentional
1
https://en.wikipedia.org/wiki/Trace_(linear_algebra)
3
?1
?
?
?
?
?
?2
?
?
=
?
?
Bottom-up
Saliency
Attention
=
?
Method 1
?2
?
?1
?
2() order
pooling
?
Top-down
Attention
?
softmax
Pose Reg. Attention
Method 2
x-entropy
(a) Visualization of our approach to attentional pooling as a rank-1 approxiloss
nd
mation of 2 order pooling. By judicious ordering of the matrix multiplicaT
tions, one can avoid computing the second order feature X X and instead
compute the product of two attention maps. The top-down attentional map
Pose ? loss
is computed using class-specific weights ak , while the bottom-up map is
computed using class-agnostic weights b. We visualize the top-down and (b) We explore two architectures
bottom-up attention maps learned by our approach in Fig. 2.
in our work, explained in Sec. 3.2.
#
Figure 1: Visualization of our derivation and final network architectures.
pooling:
scoreattention (X) = T r(X T XbaT ),
T
where
X ? Rn?f , a, b ? Rf ?1
T
= T r(a X Xb)
T
(3)
(4)
T
= a X Xb
= aT X T (Xb)
(5)
(6)
where (4) makes use of the trace identity that T r(ABC) = T r(CAB) and (5) uses the fact that the
trace of a scalar is simply the scalar. The last line (6) gives efficient implementation of attentional
pooling: given a feature map X, compute an attention map over all n spatial locations with h =
Xb ? Rn?1 , that is then used to compute a weighted average of features x = X T h ? Rf ?1 . This
weighted-average feature is then pushed through a linear model aT x to produce the final score.
Interestingly, (6) can also be written as the following:
scoreattention (X) = (Xa)T X b
= (Xa)T (Xb)
(7)
(8)
The first line illustrates that the attentional heatmap can also be seen as Xa ? Rn?1 , with b being
the classifier of the attentionally-pooled feature. The second line illustrates that our formulation is
in fact symmetric, where the final score can be seen as the inner product between two attentional
heatmaps defined over all n spatial locations. Fig. 1a illustrates our approach.
Top-down attention: To generate prediction for multiple classes, we replace the weight matrix from
(2) with class-specific weights:
scoreorder2 (X, k) = T r(X T XWkT ),
where
X ? Rn?f , Wk ? Rf ?f
(9)
One could apply a similar derivation to produce class-specific vectors ak and bk , each of them
generating a class-specific attention map. Instead, we choose to distinctly model class-specific
?top-down? attention [3, 57] from bottom-up visual saliency that is class-agnostic [36]. We do so by
forcing one of the attention parameter vectors to be class-agnostic - e.g., bk = b. This makes our final
low-rank attentional model
scoreattention (X, k) = tTk h,
where
tk = Xak , h = Xb
(10)
equivalent to an inner product between top-down (class-specific) tk and bottom-up (saliency-based) h
attention maps. Our approach of combining top-down and botom-up attentional maps is reminiscent
of biologically-motivated schemes that modulate saliency maps with top-down cues [31]. This
suggests that our attentional model can also be implemented using a single, combined attention map
defined over all n spatial locations:
scoreattention (X, k) = 1T ck ,
where
ck = tk ? h,
(11)
where ? denotes elementwise multiplication and 1 is defined as before. We visualize the combined,
top-down, and bottom-up attention maps ck , tk , h ? Rn?1 in our experimental results.
4
3.2
Network Architecture
We now describe our network architecture to implement the attentional pooling described above. We
start from a state of the art base architecture, ResNet-101 [20]. It consists of a stack of ?modules?, each
of which contains multiple convolutional, pooling or identity mapping streams. It finally generates a
n1 ? n2 ? f spatial feature map, which is average pooled to get a f -dimensional vector and is then
classified using a linear classifier.
Our attention module plugs in at the last layer, after the spatial feature map. As shown in Fig. 1b
(Method 1), we predict a single channel bottom-up saliency map of same spatial resolution as the last
feature map, using a linear classifier on top of it (Xb). Similarly, we also generate the n1 ? n2 ? K
dimensional top-down attention map Xa, where K is number of classes. The two attention maps are
multiplied and spatially averaged to generate the K-dimensional output predictions ((Xa)T (Xb)).
These operations are equivalent to first multiplying the features with saliency (X T (Xb)) and then
passing through a classifier (a(X T (Xb))).
Pose: While this unconstrained attention module automatically learns to focus on relevant parts and
gives a sizable boost in accuracy, we take inspiration from previous work [8] and use human pose
keypoints to guide the attention. As shown in Fig. 1b (Method 2), we use a two-layer MLP on top
of the last layer to predict a 17 channel heatmap. The first 16 channels correspond to human pose
keypoints and incur a l2 loss against labeled (or detected, using [4]) pose> The final channel is used
as an unconstrained bottom-up attention map, as before. We refer to this method as pose-regularized
attention, and it can be thought of as a non-linear extension of previous attention map.
4
Experiments
Datasets: We experiment with three recent, large scale action recognition datasets, across still images
and videos, namely MPII, HICO and HMDB51. MPII Human Pose Dataset [33] contains 15205
images labeled with up to 16 human body keypoints, and classified into one of 393 action classes. It
is split into train, val (from authors of [18]) and test sets, with 8218, 6987 and 5708 images each. We
use the val set to compare with [18] and for ablative analysis while the final test results are obtained
by emailing our results to authors of [33]. The dataset is highly imbalanced and the evaluation is
performed using mean average precision (mAP) to equally weight all classes. HICO [7] is a recently
introduced dataset with labels for 600 human object interactions (HOI) combining 117 actions with
80 objects. It contains 38116 training and 9658 test images, with each image labeled with all the
HOIs active for that image (multi-label setting). Like MPII, this dataset is also highly unbalanced and
evaluation is performed using mAP over classes. Finally, to verify our method?s applicability to video
based action recognition, we experiment with a challenging trimmed action classification dataset,
HMDB51 [27]. It contains 6766 realistic and varied video clips from 51 action classes. Evaluation is
performed using average classification accuracy over three train/test splits from [23], each with 3570
train and 1530 test videos.
Baselines: Throughout the following sections, we compare our approach first to the standard base
architecture, mostly ResNet-101 [20], without the attention weighted pooling. Then we compare to
other reported methods and previous state of the art on the respective datasets.
MPII: We train our models for 393-way action classification on MPII with softmax cross-entropy
loss for both the baseline ResNet and our attentional model. We compare our performance in Tab. 1.
Our unconstrained attention model clearly out-performs the base ResNet model, as well as previous
state of the art methods involving detection of multiple contextual bounding boxes [18] and fusion
of full image with human bounding box features [30]. Our pose-regularized model performs best,
though the improvement is small. We visualize the attention maps learned in Fig. 2.
HICO: We train our model on HICO similar to MPII, and compare our performance in Tab. 2.
Again, we see a significant 5% boost over our base ResNet model. Moreover, we out-perform all
previous methods, including ones that use detection bounding boxes at test time except one [30],
when that is trained with a specialized weighted loss for this dataset. It is also worth noting that
the full image-only performance of VGG and ResNet were comparable in our experiments (29.4%
and 30.2%), suggesting that our approach shows larger relative improvement over a similar starting
baseline. Though we did not experiment with the same optimization setting as [30], we believe it
will give similar improvements there as well. Since this dataset also comes with labels decomposed
into actions and objects, we visualize what our attention model looks for, given images containing
5
GT Class
Test Image
Bottom Up
Other Class
Top Down
playing with animals
Combined
playing with animals
Top Down
resistance training
Combined
resistance training
garbage collector
garbage collector
video exercise workouts
video exercise workouts
forestry
forestry
rope skipping
rope skipping
violin
violin
bicycling
bicycling
marching band
marching band
yoga
yoga
basketball
basketball
skiing
skiing
chopping wood
chopping wood
yoga
yoga
calisthenics
calisthenics
fishing in stream
fishing in stream
Figure 2: Auto-generated (not hand-picked) visualization of bottom-up (Xb), top-down (Xak ) and combined
((Xak ) ? (Xb)) attention on validation images in MPII, that see largest improvement in softmax score for
correct class when trained with attention. Since the top-down/combined maps are class specific, we mention the
class name for which they are generated for on top left of those heatmaps. We consider 2 classes, the ground truth
(GT) for the image, and the class on which it gets lowest softmax score. The attention maps for GT class focus
aerobic
skiing
on the objects most useful for distinguishing aerobic
the class. Though
in many casesskiing
top-down and combined
maps
look similar, they capture different information, like in second example of garbage collector. While top-down
also focuses on the vehicles in background, combined map narrows it down to the garbage bags. (Best viewed
zoomed-in on screen)
bicycling
bicycling
sports ball
hotdog
bird
chopping wood
laptop
suitcase
donut
chopping wood
Figure 3: We crop a 100px patch around the attention peak for all images containing an HOI involving a given
object, and show 5 randomly picked patches for 6 object classes here. This suggests our attention model learns
to look for objects to improve HOI detection.
6
Table 1: Action classification performance on MPII dataset. Validation (Val) performance is reported on train set
split shared by authors of [18]. Test performance obtained from training on complete train set and submitting our
output file to authors of [33]. Note that even though our pose regularized model uses pose labels at training time
for regularizing attention, it does not require any pose input at test time. The top-half corresponds to a diagnostic
analysis of our approach with different base networks. Attention provides a strong 4% improvement for baseline
networks with larger spatial resolution (e.g., ResNet). Please see text for additional discussion. The bottom-half
reports prior work that makes use of object bounding boxes/pose. Our method performs slightly better with pose
annotations (on training data), but even without any pose or detection annotations, we outperform all prior work.
Method
Full Img
Inception-V2 (ours)
ResNet101 (ours)
Attn. Pool. (I-V2) (ours)
Attn. Pool. (R-101) (ours)
X
X
X
X
Dense Trajectory + Pose [33]
VGG16, RCNN [18]
VGG16, R*CNN [18]
VGG16, Fusion (best) [30]
VGG16, Fusion+MIL (best) [30]
Pose Reg. Attn. Pooling (ours)
X
X
X
X
Bbox
Pose
MIL
Val (mAP)
Test (mAP)
25.2
26.2
24.3
30.3
36.0
16.5
21.7
30.6
5.5
26.7
32.2
31.9
36.1
X
X
X
X
X
X
X
Table 2: Multi-label HOI classification performance on HICO dataset. The top-half compares our performance
to other full image based methods. The bottom-half reports methods that use object bounding boxes/pose. Our
model out-performs various approaches that need bounding boxes, multi-instance learning (MIL) or specialized
losses, and achieves performance competitive to state of the art. Note that even though our pose regularized
model uses computed pose labels at training time, it does not require any pose input at test time.
Method
Full Im.
Bbox/Pose
MIL
AlexNet+SVM [7]
VGG16, full image [30]
ResNet101, full image (ours)
ResNet101 with CBP [16] (impl. from [1])
Attentional Pooling (ours)
X
X
X
X
X
R*CNN [18] (reported in [30])
Scene-RCNN [18] (reported in [30])
Fusion (best reported) [30]
Pose Regularized Attentional Pooling (ours)
X
X
X
X
X
X
X
X
X
X
Fusion, weighted loss (best reported) [30]
X
X
X
Wtd Loss
mAP
19.4
29.4
30.2
26.8
35.0
28.5
29.0
33.8
34.6
X
36.1
interactions with a specific object. As Fig. 3 shows, the attention peak is typically close to the object
of interest, showing the importance of detecting objects in HOI detection tasks. Moreover, this
suggests that our attention maps can also function as weak-supervision for object detection.
HMDB51: Next, we apply our attentional method to the RGB stream of the current state of the art
deep model on this dataset, TSN [52]. TSN extends the standard two-stream [41] architecture by using
a much deeper base architecture [22] along with enforcing consensus over multiple frames from the
video at training time. For the purpose of this work, we focus on the RGB stream only but our method
is applicable to flow/warped-flow streams as well. We first train a TSN model using ResNet-101 as
base architecture after re-sizing input frames to 450px. This ensures larger spatial dimensions of
the output (14 ? 14), hence ensuring the last-layer features are amenable to attention. Though our
base ResNet model does worse than BN-inception TSN model, as Tab. 3 shows, using our attention
module improves the base model to do comparably well. Interestingly, on this dataset regularizing the
attention through pose gives a significant boost in performance, out-performing TSN and establishing
new state of the art on the RGB stream-only model for HMDB. We visualize the attention maps
with normal and pose-regularized attention in Fig. 4. The pose regularized attention are more peaky
near the human than their linear counterparts. This potentially explains the improvement using pose
on HMDB while it does not help as much on HICO or MPII; HICO and MPII, being image based
datasets typically have ?iconic? images, with the subjects and objects of action typically in the center
and focus of the image. Video frames in HMDB, on the other hand, may have the subject move all
7
Table 3: Action classification performance on HMDB51 dataset using only the RGB stream of a two-stream
model. Our base ResNet stream training is done over 480px rescaled images, same as used in our attention
model for comparison purposes. Our pose based attention model out-performs the base network by large margin,
and obtains state of the art RGB-only performance by out-performing TSN [52].
Method
Split 2
Split 3
Avg
54.4
48.2
51.1
54.4
49.5
46.5
51.6
51.1
49.2
46.7
49.7
50.9
51.0
48.9
46.7
47.1
50.8
52.2
Pose Reg. Attention
Attention
TSN, BN-inception (RGB) [52] (Via email with authors)
RGB Stream, ResNet50 (RGB) [14] (reported at [2])
RGB Stream, ResNet152 (RGB) [14] (reported at [2])
TSN, ResNet101 (RGB) (ours)
Linear Attentional Pooling (ours)
Pose regularized Attentional Pooling (ours)
Split 1
Figure 4: Attention maps with linear attention and pose regularized attention on a video from HMDB. Note the
pose-guided attention is better able to focus on regions of interest in the non-iconic frames.
across the frame throughout the video, and hence additional supervision through pose at training time
helps focus the attention at the right spot.
Full-rank pooling: Given our formulation of attention as low-rank second-order pooling, a natural
question is what would be the performance of a full-rank model? Explicitly computing the secondorder features of size f ? f for f = 2048 (and learning the associated classifier) is cumbersome.
Instead, we make use of the compact bilinear approach (CBP) of [16], which generates a lowdimensional approximation of full bilinear pooling [28] using the TensorSketch algorithm. To keep
the final output comparable to our attentional-pooled model, we project to f = 2048 dimensions.
We find it performs slightly worse than simple average pooling in Table 2. Note that we use an
existing implementation [1] with minimal hyper-parameter optimization, and leave a more rigorous
comparison to future work.
Rank-P approximation: While a full-rank model is cumbersome, we can still explore the effect of
using a higher, P -rank approximation. Essentially, a rank-P approximation generates P (1-channel)
bottom-up and (C channel) top-down attention maps, and the final prediction is the product of
corresponding heatmaps, summed over P . On MPII, we obtain mAP of 30.3, 29.9, 30.0 for P =1,
2 and 5 respectively, showing that the validation performance is relatively stable with P . We do
observe a drop in training loss with a higher P , indicating that a higher-rank approximation could be
useful for harder datasets and tasks.
Per-class attention maps: As we described in Sec. 3.1, our inspiration for combining class-specific
and class-agnostic classifiers (i.e. top-down and bottom-up attention respectively), came from the
Neuroscience literature on integrating top-down and bottom-up attention [31]. However, our model
can also be extended to learn completely class-specific attention maps, by predicting C bottom-up
attention maps, and combining each map with the corresponding softmax classifier for that class. We
experiment with this idea on MPII and obtain a mAP of 27.9 with 393 (=num-classes) attention maps,
compared to 30.3% with 1 map, and 26.2% without attention. On further analysis we observe that
both models achieve near perfect mAP on training data, implying that adding more parameters with
8
multiple attention maps leads to over-fitting on the relatively small MPII trainset. However, this may
be a viable approach for larger datasets.
Diagnostics: It is natural to consider variants of our model that only consider the bottom-up or
top-down attentional map. Interestingly, a top-down-only version of our model is equivalent to
standard average pooling from (1):
scoretop?down (X) = 1T Xak = 1T tk where tk = Xak
(12)
From this perspective, the above derivation gives the ability to generate attentional maps from averagepooling networks. Similar observations have been pointed out before [57]. Hence our comparison
with baseline models with average pooling essentially compares to ?top-down-only? attention models.
It is not clear how to construct a bottom-up only model, since it is class-agnostic, making it difficult to
produce class-specific scores. Rather, a reasonable approximation might be applying an off-the-shelf
(bottom-up) saliency method used to limit the spatial region that features are averaged over. Our
initial experiments with existing saliency-based methods [21] were not promising.
Base Network: Finally, we analyze the choice of base architecture for the effectiveness of our
proposed attentional pooling module. In Tab. 1, we compare the improvement using attention over
ResNet-101 (R-101) [20] and an BN-Inception (I-V2) [22]. Both models perform comparably when
trained for full image, however, while we see a 4% improvement on R-101 on using attention, we do
not see similar improvements for I-V2. This points to an important distinction in the two architectures,
i.e., Inception-style models are designed to be faster in inference and training by rapidly down
sampling input images in initial layers through max-pooling. While this reduces the computational
cost for later layers, it leads to most layers having very large receptive fields, and hence later neurons
have effective access to all of the image pixels. This suggests that all the spatial features at the last
layer could be highly similar. In contrast, R-101 downscales the spatial resolution gradually, allowing
the last layer features to specialize to different parts of the image, hence benefiting more from
attentional pooling. This effect was further corroborated by our experiments on HMDB, where using
the standard 224px input resolution showed no improvement with attention, while the same image
resized to 450px at input time did. This initial resize ensures the last-layer features are sufficiently
distinct to benefit from attentional pooling.
5
Discussion and Conclusion
An important distinction of our model from some previous works [18, 30] is that it does not explicitly
model action at an instance or bounding-box level. This, in fact, is a strength of our model; making it
capable of attending to objects outside of any person-instance bounding box (such as bags of garbage
for ?garbage collecting?, in Fig 2). In theory, our model can also be applied to instance-level action
recognition by applying attentional pooling over an instance?s RoI features. Such a model would learn
to look at different parts of human body and its interactions with nearby objects. However, it?s notable
that most existing action datasets, including [6, 7, 27, 33, 40, 44], come with only frame or video
level labels; and though [18, 30] are designed for instance-level recognition, they are not applied
as such. They either copy image level labels to instances or use multiple-instance learning, either
of which can be used in conjunction with our model. Another interesting connection that emerges
from our work is the relation between second-order pooling and attention. The two communities are
traditionally seen as distinct, and our work strongly suggests that they should mix: as newer action
datasets become more fine-grained, we should explore second-order pooling techniques for action
recognition. Similarly, second-order pooling can serve as a simple but strong baseline for the attention
community, which tends to focus on more complex sequential attention networks (based on RNNs or
LSTMs). It is also worth noting that similar ideas involving self attention and bilinear models have
recently also shown significant improvements in other tasks like image classification [49], language
translation [48] and visual question answering [37].
Conclusion: We have introduced a simple formulation of attention as low-rank second-order pooling, and illustrate it on the task of action classification from single (RGB) images. Our formulation
allows for explicit integration of bottom-up saliency and top-down attention, and can take advantage
of additional supervision when needed (through pose labels). Our model produces competitive or
state-of-the-art results on widely benchmarked datasets, by learning where to look when pooling
features across an image. Finally, it is easy to implement and requires few additional parameters,
making it an attractive alternative to standard pooling, which is a ubiquitous operation in nearly all
contemporary deep networks.
9
Acknowledgements: Authors would like to thank Olga Russakovsky for initial review. This
research was supported in part by the National Science Foundation (NSF) under grant numbers
CNS-1518865 and IIS-1618903, and the Defense Advanced Research Projects Agency (DARPA)
under Contract No. HR001117C0051. Additional support was provided by the Intel Science and
Technology Center for Visual Cloud Systems (ISTC-VCS). Any opinions, findings, conclusions or
recommendations expressed in this material are those of the authors and do not necessarily reflect the
view(s) of their employers or the above-mentioned funding sources.
References
[1] Compact bilinear pooling implementation.
compact_bilinear_pooling.
https://github.com/ronghanghu/tensorflow_
[2] Convolutional two-stream network fusion for video action recognition. http://www.robots.ox.ac.
uk/~vgg/software/two_stream_action/.
[3] F. Baluch and L. Itti. Mechanisms of top-down attention. Trends in Neurosciences, 2011.
[4] Z. Cao, T. Simon, S.-E. Wei, and Y. Sheikh. Realtime multi-person 2d pose estimation using part affinity
fields. In CVPR, 2017.
[5] J. Carreira, R. Caseiro, J. Batista, and C. Sminchisescu. Semantic segmentation with second-order pooling.
In ECCV, 2012.
[6] J. Carreira and A. Zisserman. Quo vadis, action recognition? a new model and the kinetics dataset. In
CVPR, 2017.
[7] Y.-W. Chao, Z. Wang, Y. He, J. Wang, and J. Deng. Hico: A benchmark for recognizing human-object
interactions in images. In ICCV, 2015.
[8] G. Ch?ron, I. Laptev, and C. Schmid. P-CNN: Pose-based CNN Features for Action Recognition. In ICCV,
2015.
[9] V. Delaitre, I. Laptev, and J. Sivic. Recognizing human actions in still images: a study of bag-of-features
and part-based representations. In BMVC, 2010.
[10] V. Delaitre, J. Sivic, and I. Laptev. Learning person-object interactions for action recognition in still images.
In NIPS, 2011.
[11] C. Desai, D. Ramanan, and C. Fowlkes. Discriminative models for static human-object interactions. In
CVPR-Workshops, 2010.
[12] J. Donahue, L. A. Hendricks, S. Guadarrama, S. V. M. Rohrbach, K. Saenko, and T. Darrell. Long-term
recurrent convolutional networks for visual recognition and description. In CVPR, 2015.
[13] M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, and A. Zisserman. The pascal visual object
classes (voc) challenge. IJCV, 2010.
[14] C. Feichtenhofer, A. Pinz, and R. Wildes. Spatiotemporal residual networks for video action recognition.
In NIPS, 2016.
[15] C. Feichtenhofer, A. Pinz, and A. Zisserman. Convolutional two-stream network fusion for video action
recognition. In CVPR, 2016.
[16] Y. Gao, O. Beijbom, N. Zhang, and T. Darrell. Compact bilinear pooling. In CVPR, 2016.
[17] R. Girdhar, D. Ramanan, A. Gupta, J. Sivic, and B. Russell. ActionVLAD: Learning spatio-temporal
aggregation for action classification. In CVPR, 2017.
[18] G. Gkioxari, R. Girshick, and J. Malik. Contextual action recognition with R*CNN. In ICCV, 2015.
[19] A. Gupta, A. Kembhavi, and L. S. Davis. Observing human-object interactions: Using spatial and
functional compatibility for recognition. PAMI, 2009.
[20] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. CVPR, 2016.
[21] X. Huang, C. Shen, X. Boix, and Q. Zhao. SALICON: Reducing the semantic gap in saliency prediction
by adapting deep neural networks. In ICCV, 2015.
[22] S. Ioffe and C. Szegedy. Batch normalization: Accelerating deep network training by reducing internal
covariate shift. ICML, 2015.
[23] Y. Jiang, J. Liu, A. Roshan Zamir, I. Laptev, M. Piccardi, M. Shah, and R. Sukthankar. THUMOS challenge:
Action recognition with a large number of classes. http://www.thumos.info/, 2013.
[24] W. Kay, J. Carreira, K. Simonyan, B. Zhang, C. Hillier, S. Vijayanarasimhan, F. Viola, T. Green, T. Back,
P. Natsev, et al. The kinetics human action video dataset. arXiv preprint arXiv:1705.06950, 2017.
10
[25] J.-H. Kim, K.-W. On, J. Kim, J.-W. Ha, and B.-T. Zhang. Hadamard product for low-rank bilinear pooling.
In ICLR, 2017.
[26] S. Kong and C. Fowlkes. Low-rank bilinear pooling for fine-grained classification. In CVPR, 2017.
[27] H. Kuehne, H. Jhuang, E. Garrote, T. Poggio, and T. Serre. HMDB: a large video database for human
motion recognition. In ICCV, 2011.
[28] T.-Y. Lin, A. RoyChowdhury, and S. Maji. Bilinear CNN models for fine-grained visual recognition. In
ICCV, 2015.
[29] S. Maji, L. Bourdev, and J. Malik. Action recognition from a distributed representation of pose and
appearance. In CVPR, 2011.
[30] A. Mallya and S. Lazebnik. Learning models for actions and person-object interactions with transfer to
question answering. In ECCV, 2016.
[31] V. Navalpakkam and L. Itti. An integrated model of top-down and bottom-up attention for optimizing
detection speed. In CVPR, 2006.
[32] F. Perronnin and C. Dance. Fisher kernels on visual vocabularies for image categorization. In CVPR, 2007.
[33] L. Pishchulin, M. Andriluka, and B. Schiele. Fine-grained activity recognition with holistic and pose based
features. In GCPR, 2014.
[34] D. Ramanan and D. A. Forsyth. Automatic annotation of everyday movements. In NIPS, 2003.
[35] M. Ronchi and P. Perona. Describing common human visual actions in images. In BMVC, 2015.
[36] U. Rutishauser, D. Walther, C. Koch, and P. Perona. Is bottom-up attention useful for object recognition?
In CVPR, 2004.
[37] A. Santoro, D. Raposo, D. G. Barrett, M. Malinowski, R. Pascanu, P. Battaglia, and T. Lillicrap. A simple
neural network module for relational reasoning. arXiv preprint arXiv:1706.01427, 2017.
[38] S. Sharma, R. Kiros, and R. Salakhutdinov. Action recognition using visual attention. ICLR-Workshops,
2016.
[39] Y. Shi, Y. Tian, Y. Wang, and T. Huang. Joint network based attention for action recognition. arXiv preprint
arXiv:1611.05215, 2016.
[40] G. A. Sigurdsson, G. Varol, X. Wang, A. Farhadi, I. Laptev, and A. Gupta. Hollywood in homes:
Crowdsourcing data collection for activity understanding. In ECCV, 2016.
[41] K. Simonyan and A. Zisserman. Two-stream convolutional networks for action recognition in videos. In
NIPS, 2014.
[42] K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition. In
ICLR, 2015.
[43] S. Song, C. Lan, J. Xing, W. Zeng, and J. Liu. An end-to-end spatio-temporal attention model for human
action recognition from skeleton data. In AAAI, 2017.
[44] K. Soomro, A. R. Zamir, and M. Shah. UCF101: A dataset of 101 human actions classes from videos in
the wild. CRCV-TR-12-01, 2012.
[45] C. Szegedy, S. Ioffe, and V. Vanhoucke. Inception-v4, inception-resnet and the impact of residual
connections on learning. 2016.
[46] D. Tran, L. Bourdev, R. Fergus, L. Torresani, and M. Paluri. Learning spatiotemporal features with 3d
convolutional networks. In ICCV, 2015.
[47] G. Varol, I. Laptev, and C. Schmid. Long-term temporal convolutions for action recognition. CoRR,
abs/1604.04494, 2016.
[48] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, and I. Polosukhin.
Attention is all you need. In NIPS, 2017.
[49] F. Wang, M. Jiang, C. Qian, S. Yang, C. Li, H. Zhang, X. Wang, and X. Tang. Residual attention network
for image classification. In CVPR, 2017.
[50] H. Wang, A. Kl?ser, C. Schmid, and L. Cheng-Lin. Action Recognition by Dense Trajectories. In CVPR,
2011.
[51] H. Wang and C. Schmid. Action recognition with improved trajectories. In ICCV, 2013.
[52] L. Wang, Y. Xiong, Z. Wang, Y. Qiao, D. Lin, X. Tang, and L. Van Gool. Temporal segment networks:
Towards good practices for deep action recognition. In ECCV, 2016.
[53] W. Yang, Y. Wang, and G. Mori. Recognizing human actions from still images with latent poses. In CVPR,
2010.
11
[54] B. Yao and L. Fei-Fei. Grouplet: A structured image representation for recognizing human and object
interactions. In CVPR, 2010.
[55] B. Yao and L. Fei-Fei. Modeling mutual context of object and human pose in human-object interaction
activities. In CVPR, 2010.
[56] B. Yao, X. Jiang, A. Khosla, A. Lin, L. Guibas, and L. Fei-Fei. Human action recognition by learning
bases of action attributes and parts. In ICCV, 2011.
[57] B. Zhou, A. Khosla, A. Lapedriza, A. Oliva, and A. Torralba. Learning deep features for discriminative
localization. In CVPR, 2016.
[58] M. Zolfaghari, G. L. Oliveira, N. Sedaghat, and T. Brox. Chained Multi-stream Networks Exploiting Pose,
Motion, and Appearance for Action Classification and Detection. In ICCV, 2017.
12
| 6609 |@word kong:1 cnn:14 version:1 nd:1 everingham:1 chopping:4 tried:1 rgb:16 bn:3 forestry:2 mention:1 tr:1 harder:2 moment:1 initial:4 liu:2 contains:4 score:8 quo:1 ours:11 interestingly:4 batista:1 past:1 existing:5 current:3 contextual:2 com:1 skipping:2 guadarrama:1 yet:2 written:2 reminiscent:1 exposing:1 parmar:1 realistic:1 informative:1 drop:1 designed:2 v:2 implying:1 cue:1 half:4 advancement:2 num:1 characterization:2 detecting:2 provides:1 location:4 cbp:2 ron:1 org:1 pascanu:1 zhang:5 mathematical:1 along:3 become:1 viable:1 walther:1 consists:1 specialize:1 ijcv:1 fitting:1 wild:1 introduce:2 paluri:1 frequently:1 kiros:1 multi:9 salakhutdinov:1 voc:1 decomposed:1 automatically:4 little:2 farhadi:1 conv:3 project:2 provided:1 moreover:5 notation:2 lapedriza:1 agnostic:5 laptop:1 lowest:1 what:2 alexnet:1 benchmarked:1 finding:1 temporal:8 impl:1 collecting:2 classifier:8 uk:1 ser:1 ramanan:4 grant:1 before:3 negligible:1 attend:1 local:1 tends:1 limit:1 io:1 bilinear:11 ak:2 jiang:3 establishing:1 pami:1 might:4 donut:1 bird:1 rnns:1 studied:2 equivalence:1 suggests:6 challenging:1 vaswani:1 tian:1 averaged:2 practice:1 implement:3 spot:1 thought:1 adapting:1 ucf101:3 integrating:1 get:3 close:2 operator:1 context:5 applying:2 vijayanarasimhan:1 sukthankar:1 www:2 gkioxari:2 map:56 equivalent:4 center:3 demonstrated:1 fishing:2 williams:1 attention:111 starting:1 shi:1 resolution:4 shen:1 qian:1 insight:1 attending:1 importantly:1 kay:1 variation:1 traditionally:1 us:6 distinguishing:1 secondorder:1 trend:1 recognition:55 peaky:1 predicts:1 labeled:4 corroborated:1 bottom:25 cloud:1 module:14 preprint:3 database:1 wang:11 capture:1 zamir:2 region:3 ensures:2 connected:1 sun:1 desai:1 went:1 ordering:1 russell:1 contemporary:2 rescaled:1 movement:1 mentioned:1 agency:1 schiele:1 skeleton:1 pinz:2 chained:2 trained:5 raise:1 deva:1 segment:1 laptev:6 around:3 incur:1 ablative:1 upon:1 serve:1 localization:1 completely:1 easily:1 darpa:1 joint:1 various:2 regularizer:1 maji:2 derivation:5 articulated:1 train:10 distinct:2 roshan:1 effective:2 describe:2 detected:1 labeling:1 aggregate:1 hyper:1 trainset:1 outside:1 larger:5 widely:1 cvpr:19 ability:1 statistic:1 simonyan:3 final:8 blob:1 xak:5 advantage:1 propose:2 tran:1 lowdimensional:1 interaction:12 product:6 zoomed:1 varol:2 relevant:2 combining:4 cao:1 rapidly:1 hadamard:1 holistic:1 achieve:1 benefiting:1 description:1 everyday:1 exploiting:1 interacted:1 darrell:2 r1:1 produce:4 generating:1 perfect:1 leave:1 categorization:1 object:34 help:3 derive:1 tions:1 tk:6 pose:58 resnet:12 illustrate:1 ac:1 recurrent:1 bourdev:2 paying:1 strong:2 throw:1 auxiliary:1 predicted:1 implemented:2 come:2 sizable:2 guided:1 correct:1 attribute:1 cnns:2 human:36 opinion:1 hoi:5 material:1 explains:1 require:3 sizing:1 im:1 extension:3 kinetics:2 drinking:2 hold:1 heatmaps:3 sufficiently:1 koch:1 considered:1 normal:1 guibas:1 great:1 hois:1 mapping:1 predict:5 visualize:5 roi:1 ground:1 achieves:1 torralba:1 purpose:2 battaglia:1 estimation:1 applicable:1 bag:3 label:12 currently:1 largest:1 hollywood:1 establishes:1 istc:1 weighted:6 ttk:1 suitcase:1 clearly:1 always:1 mation:1 modified:1 ck:3 rather:1 avoid:2 zhou:1 resized:1 shelf:1 mil:5 vectorizes:1 conjunction:1 encode:1 focus:11 improvement:17 iconic:3 rank:22 greatly:1 contrast:3 rigorous:1 baseline:8 kim:2 inference:1 polosukhin:1 el:1 feichtenhofer:2 perronnin:1 typically:6 integrated:1 santoro:1 perona:2 relation:1 pixel:1 compatibility:1 classification:20 pascal:1 denoted:1 bicycling:4 heatmap:2 art:15 spatial:16 constrained:1 brox:1 softmax:5 animal:2 construct:2 summed:1 resnet152:1 beach:1 sampling:1 having:1 field:2 look:8 icml:1 nearly:2 jones:1 future:1 others:1 idt:1 report:2 torresani:1 few:2 abt:1 randomly:1 densely:1 national:1 hmdb51:5 cns:1 n1:2 ab:2 detection:11 mlp:1 interest:2 highly:4 evaluation:3 diagnostics:1 xb:12 amenable:2 encourage:1 capable:1 poggio:1 respective:1 re:1 girshick:1 minimal:1 instance:9 classify:1 soft:2 modeling:1 cost:3 applicability:2 recognizing:4 reported:8 spatiotemporal:2 combined:8 person:5 st:1 fundamental:1 explores:1 discriminating:1 lstm:1 peak:2 caseiro:1 contract:1 off:1 v4:1 pool:2 yao:3 again:1 reflect:1 aaai:1 containing:3 choose:2 huang:2 worse:2 warped:2 zhao:1 style:1 itti:2 li:1 szegedy:2 suggesting:2 sec:5 pooled:5 bold:1 wk:1 forsyth:1 notable:1 explicitly:3 stream:22 later:3 performed:3 picked:2 vehicle:1 view:1 tab:4 analyze:1 observing:1 competitive:4 relied:1 start:1 aggregation:1 xing:1 annotation:3 simon:1 contribution:1 accuracy:4 convolutional:11 who:1 efficiently:1 yield:1 saliency:11 correspond:1 generalize:1 weak:1 mallya:2 comparably:2 ren:1 trajectory:5 multiplying:1 drive:1 worth:2 russakovsky:1 classified:2 cumbersome:2 email:1 definition:1 against:1 associated:1 transposed:1 static:1 sampled:1 dataset:15 popular:1 kuehne:1 emerges:1 improves:2 ubiquitous:1 segmentation:1 formalize:1 back:1 focusing:1 higher:4 supervised:1 zisserman:5 improved:1 wei:1 bmvc:2 formulation:6 done:1 box:13 though:7 strongly:1 ox:1 xa:5 inception:7 hand:4 lstms:1 zeng:1 believe:1 usa:1 effect:2 lillicrap:1 contain:1 true:1 verify:1 counterpart:1 name:1 aerobic:2 analytically:1 hence:6 spatially:1 symmetric:1 inspiration:2 semantic:2 eg:2 mug:1 attractive:1 basketball:3 self:2 please:1 davis:1 rope:2 complete:1 performs:6 motion:5 reasoning:1 image:55 lazebnik:2 shoot:1 novel:4 recently:3 funding:1 wikipedia:1 common:1 specialized:2 functional:1 empirically:1 uszkoreit:1 he:2 elementwise:1 interpret:1 mellon:1 significant:7 refer:1 automatic:1 unconstrained:5 similarly:2 pointed:1 language:1 dot:1 stable:1 access:1 supervision:6 impressive:1 robot:1 etc:1 base:17 add:1 gt:3 imbalanced:1 recent:4 wtd:1 perspective:2 showed:1 optimizing:1 apart:1 forcing:1 skiing:3 integration:1 binary:1 came:1 seen:4 additional:8 wilde:1 deng:1 aggregated:1 paradigm:2 sharma:1 vgg16:5 ii:1 full:14 multiple:8 keypoints:6 reduces:1 mix:1 faster:1 plug:2 cross:1 long:3 lin:4 equally:1 coded:1 plugging:2 ensuring:1 prediction:8 variant:3 involving:3 oliva:1 crop:1 vision:1 essentially:2 surpassed:1 impact:1 arxiv:6 represent:1 sometimes:1 normalization:1 kernel:1 robotics:1 background:2 fine:10 winn:1 girdhar:2 mutual:1 source:1 modality:2 extra:1 file:1 pooling:51 subject:3 virtually:1 flow:5 incorporates:2 effectiveness:1 extracting:2 near:2 noting:2 delaitre:2 intermediate:1 split:7 easy:3 yang:2 architecture:18 reduce:1 inner:3 idea:2 vgg:2 shift:1 whether:1 motivated:1 defense:1 accelerating:1 trimmed:1 soomro:1 song:1 resistance:2 passing:1 action:73 matlab:1 deep:12 garbage:6 useful:5 clear:1 malinowski:1 oliveira:1 band:2 clip:1 processed:1 category:1 http:5 generate:7 wiki:1 outperform:1 nsf:1 roychowdhury:1 diagnostic:1 neuroscience:2 per:1 diverse:2 broadly:1 carnegie:1 write:3 promise:1 key:1 lan:1 verified:1 ronchi:1 fuse:1 sum:2 wood:4 letter:2 powerful:2 you:1 extends:1 throughout:2 reasonable:1 employer:1 patch:2 realtime:1 home:1 garrote:1 resize:1 comparable:2 pushed:1 layer:14 followed:1 distinguish:1 gomez:1 cheng:1 quadratic:1 activity:3 strength:1 fei:6 scene:1 software:1 nearby:1 generates:3 speed:1 performing:2 optical:2 relatively:3 px:5 structured:2 ball:1 thumos:2 across:6 slightly:2 newer:2 sheikh:1 modification:1 making:5 biologically:1 explained:1 gradually:1 iccv:10 hillier:1 pipeline:2 mori:1 visualization:3 turn:1 describing:1 mechanism:1 needed:1 violin:2 end:5 operation:5 multiplied:1 apply:4 observe:2 v2:4 fowlkes:2 xiong:1 alternative:1 batch:1 shah:2 top:35 denotes:1 kembhavi:1 xw:2 move:1 malik:2 question:4 quantity:1 kaiser:1 receptive:1 traditional:1 affinity:1 iclr:3 attentional:27 thank:1 consensus:1 water:1 spanning:1 reason:1 enforcing:1 length:1 navalpakkam:1 difficult:1 mostly:1 potentially:1 info:1 trace:4 implementation:3 perform:3 allowing:1 upper:1 observation:2 neuron:1 datasets:14 convolution:1 benchmark:5 workout:2 viola:1 extended:1 relational:1 frame:11 rn:9 varied:1 stack:1 shazeer:1 community:2 salicon:1 bk:2 introduced:2 bottle:1 namely:1 kl:1 extensive:2 connection:3 trainable:2 andriluka:1 sivic:3 qiao:1 learned:3 narrow:1 distinction:2 beijbom:1 boost:5 nip:6 able:1 bbox:2 hendricks:1 challenge:2 rf:9 including:5 max:2 video:37 green:1 gool:2 natural:2 regularized:11 tsn:8 predicting:1 residual:4 advanced:1 scheme:1 improve:3 github:2 technology:1 pishchulin:1 keypoint:1 dance:1 auto:1 schmid:4 resnet50:1 chao:1 text:1 prior:3 literature:1 l2:1 acknowledgement:1 val:4 multiplication:1 rohit:1 relative:2 review:1 understanding:1 fully:2 loss:8 discriminatively:1 interesting:1 validation:3 rcnn:2 foundation:1 sedaghat:1 rutishauser:1 vanhoucke:1 vectorized:1 playing:2 translation:1 eccv:4 succinctly:1 jhuang:1 surprisingly:2 last:8 keeping:1 copy:1 supported:1 guide:1 attn:3 deeper:1 institute:1 distinctly:1 benefit:1 van:2 distributed:1 dimension:2 vocabulary:1 computes:1 author:7 made:1 avg:1 collection:1 approximate:1 obtains:1 compact:3 keep:1 active:1 ioffe:2 rid:1 img:1 serre:1 raposo:1 spatio:5 discriminative:2 fergus:1 latent:1 mpii:18 khosla:2 table:4 promising:1 learn:4 channel:7 transfer:1 ca:1 obtaining:1 improving:1 grouplet:1 sminchisescu:1 necessarily:2 complex:1 did:2 dense:3 main:2 bounding:12 n2:2 complementary:2 collector:3 body:4 benefited:1 crafted:1 fig:8 en:1 intel:1 screen:1 fashion:1 boix:1 precision:2 explicit:1 exercise:2 answering:2 late:1 learns:4 grained:10 donahue:1 tang:2 down:31 specific:12 covariate:1 showing:3 barrett:1 svm:1 submitting:1 gupta:3 evidence:1 fusion:7 consist:1 incorporating:2 workshop:2 adding:2 sequential:1 importance:1 corr:1 cab:1 illustrates:3 margin:1 gap:1 easier:2 marching:2 entropy:2 simply:1 appearance:6 explore:5 rohrbach:1 hmdb:8 visual:9 gao:1 expressed:1 sport:1 scalar:2 recommendation:1 ch:1 corresponds:1 truth:1 relies:1 abc:1 modulate:1 identity:3 viewed:1 towards:1 yoga:4 replace:2 fisher:2 shared:1 hard:3 judicious:1 carreira:3 except:1 reducing:2 olga:1 crcv:1 experimental:1 saenko:1 exception:2 indicating:1 internal:1 support:1 latter:1 unbalanced:1 incorporate:1 reg:3 regularizing:2 crowdsourcing:1 |
6,201 | 661 | Improving Convergence in Hierarchical
Matching Networks for Object
Recognition
Joachim Utans*
Gene Gindi t
Department of Electrical Engineering
Yale University
P. O. Box 2157 Yale Station
New Haven, CT 06520
Abstract
We are interested in the use of analog neural networks for recognizing visual objects. Objects are described by the set of parts
they are composed of and their structural relationship. Structural models are stored in a database and the recognition problem reduces to matching data to models in a structurally consistent way. The object recognition problem is in general very difficult in that it involves coupled problems of grouping, segmentation
and matching. We limit the problem here to the simultaneous labelling of the parts of a single object and the determination of
analog parameters. This coupled problem reduces to a weighted
match problem in which an optimizing neural network must minimize E(M, p) = LO'i MO'i WO'i(p), where the {MO'd are binary
match variables for data parts i to model parts a and {Wai(P)}
are weights dependent on parameters p . In this work we show that
by first solving for estimates p without solving for M ai , we may
obtain good initial parameter estimates that yield better solutions
for M and p.
*Current address: International Computer Science Institute, 1947 Center Street,
Suite 600, Berkeley, CA 94704, [email protected]
tCurrent address: SUNY Stony Brook, Department of Electrical Engineering, Stony
Brook, NY 11784
401
402
Utans and Gindi
Figure 1: Stored Model for a 3-Level Compositional Hierarchy (compare Figure 3) .
1
Recognition via Stochastic Forward Models
The Frameville object recognition system introduced by Mjolsness et al [5, 6, 1]
makes use of a compositional hierarchy to represent stored models. The recognition
problem is formulated as the minimization of an objective function. Mjolsness [3,4]
has proposed to derive the objective function describing the recognition problem
in a principled way from a stochastic model that describes the objects the system
is designed to recognize (stochastic visual grammar). The description mirrors the
data representation as a compositional hierarchy, at each stage the description of
the object becomes more detailed as parts are added.
The stochastic model assigns a probability distribution at each stage of that process.
Thus at each level of the hierarchy a more detailed description of parts in terms of
their subparts is given by specifying a probability distribution for the coordinates of
the subparts. Explicitly specifying these distributions allows for finer control over
individual part descriptions than the rather general parameter error terms used
before [1, 8]. The goal is to derive a joint probability distribution for an instance
of an object and its parts as it appears in the scene. This gives the probability of
observing such an object prior to the arrival of the data. Given an observed image,
the recognition problem can be stated as a Bayesian inference problem that the
neural network solves.
1.1
3-Level Stochastic Model
For example, consider the model shown in Figure 1 and 3. The object and its parts
are represented as line segments (sticks), the parameters were p = (x, y, I, ())T with
x , y denoting position, I the length of a stick and () its orientation. The model
considers only a rigid translation of an object in the image.
Only one model is stored. From a central position p = (x, y, I, ()), itself chosen
from a uniform density, the N{3 parts at the first level are placed. Their structural
relationships is stored as coordinates u{3 in an object-centered coordinate frame,
i.e. relative to p. While placing the parts, Gaussian distributed noise with mean 0
and is added to the position coordinates to capture the notion of natural variation
of the object's shape. The variance is coordinate specific, but we assume the same
distribution for the x and y coordinates, O"'ix; O"'~, is the variance for the length
Improving Convergence in Hierarchical Matching Networks for Object Recognition
component and UI9 for the relative angle. In addition, here we assume for simplicity
that all parts are independently distributed. Each of the parts {3 is composed of subparts. For simplicity of notation, we assume that each part {3 is composed from the
same number of subparts N m (note that the index 'Y in Figure 2 here corresponds
to the double index {3m to keep track of which part {3 subpart {3m belongs to on the
model side, i.e. the index (3m denotes the mth sub-part of part (3). The next step
models the unordering of parts in the image via a permutation matrix M, chosen
with probability P(M), by which their identity is lost. If this step were omitted,
the recognition problem would reduce to the problem of estimating part parameters
because the parts would already be labeled.
From the grammar we compute the final joint probability distribution (all constant
terms are collected in a constant C):
P(M, {P,3m}, {PtJ}, p) =
1.2
Frameville Architecture for Part Labelling within a single Object
The stochastic forward model for the part labelling problem with only a single object
present in the scene translates into a reduced Frameville architecture as depicted in
Figure 2. The compositional hierarchy parallels the steps in the stochastic model
as parts are added at each level. Match variables appear only at the lowest level,
corresponding to the permutation step of the grammar. Parts in the image must
be matched to model parts and parts found to belong to the stored object must be
grouped together.
The single match neuron Mai at the highest level can be set to unity since we assume
we know the object's identity and only a single object is present. Similarly, all terms
inaij from the first to the second level can be set to unity for the correct grouping
since the grouping is known at this point from the forward model description. In
addition, at the intermediate (second) level, we may set all M,3j = 1 for {3 = j
and MtJj = 0 otherwise with no loss of generality. These mid-level frames may
be matched ahead of time, but their parameters must be computed from data.
Introducing a part permutation at the intermediate levels thus is redundant. Given
this, an additional simplification ina grouping variables at the lowest (third) level
is possible. Since parts are pre-matched at all but the lowest level, inaj k can be
expressed in terms of the part match M"{k as inajk = M"{k1NA"{tJM,3j and explicitly
representing inaj k as variables is not necessary.
The input to the system are the {pk}, recognition involves finding the parameters
403
404
Utans and Gindi
Model
<l
Data
rame
?
?
?
?
x
y
9
I
Figure 2: Frameville Architecture for the Stochastic Model. The 3-level grammar leads to a reduced
"Frameville" style network architecture : a single model is stored on the model side and only one instance
of the model is present in the input data. The ovals on the model side represent the object, its parts
and subparts (compare Figure 1); the arcs INA represent their structural relationship . On the data side,
the triangles represent parameter vectors (or frames) describing an instance of the object in the scene.
At the lowest level the Pk represent the input data, parameters at higher levels in the hierarchy must be
computed by the network (represented as bold triangles) . ina represents the grouping of parts on the
data side (see text) . The horizontal lines represent assignments from frames on the data side to nodes
on the model side. At the intermediate level, frames are prematched to the corresponding parts on the
model side ; match variables are necessary only at the lowest level (represented as bold lines with circles).
P and {Pi} as well as the labelling of parts M. Thus, from Bayes Theorem
P( {pdIM, p, {Pi} )P(M, p, {Pi})
P({Pk} )
(2)
ex: P(M, p, {Pi}, {pd)
and recognition reduces to finding the most probable values for p, {Pi} and M
given the data:
arg max P(M, p, {Pi}, {pd)
(3)
M,P,{Pi}
Solving the inference problem involves finding the MAP estimate and is is equivalent
to minimizing the exponent in equation (1) with respect to M, P and {Pi}.
2
2.1
Bootstrap: Coarse Scale Hints to Initialize the Network
Compositional Hierarchy and Scale Space
In some labelling approaches found in the vision literature, an object is first labelled
at the coarse, low resolution, level and approximate parameters are found . In this
top-down approach the information at the higher, more abstract, levels is used
Improving Convergence in Hierarchical Matching Networks for Object Recognition
im---:.+-Human
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1L _ _ _ _ _ _ 1I
r---------'
1
II
1-- - , 1
iV
t
(I)
i
~
1_ _ _ _ _ _ _
Am
i
_ _ _ J1
III
spatial
scale
abstraction
Figure 3: Compositional Hierarchy vs . Scale Space Hierarchy_ A compositional hierarchy can represent a
scale space hierarchy. At successive levels in the hierarchy, more and more detail is added to the object_
to select initial values for the parts at the next lower level of abstraction. The
segmentation and labelling at this next lowest level is thus not done blindly; rather
it is strongly influenced contextually by the results at the level above.
In fact, in very general terms such a scheme was described by Marr and Nishihara [2].
They advocate in essence a hierarchical model base in which a shape is first matched
to the highest levels, and defaults in terms of relative object-based parameters of
parts at the next level are recalled from memory. These defaults then serve as initial
values in an unspecified segmentation algorithm that derives part parameters; this
step is repeated recursively until the lowest level is reached.
Note that the highest level of abstractions correspond to the coarsest levels of spatial
scale. There is nothing in the design of the model base that demands this, but invariably, elements at the top of a compositional hierarchy are of coarser scale since they
must both include the many subparts below, and summarize this inclusion with
relatively few parameters. Figure 3 illustrates the correspondence between these
representations. In this sense, the compositional hierarchy as applied to shapes
includes a notion of scale, but there is no "scale-space" operation of intentionally
blurring data. The notion of Scale Space as utilized here thus differs from the
application of the method to low-level computations in the visual domain where
auxiliary coarse scale representations are computed explicitly. The object representations in the Frameville system as described earlier combines both, bottom-up and
top-down elements. If the top-down aspects of the scheme described by Marr and
Nishihara [2] could be incorporated into the Frameville architecture, then our previous simulation results [8] suggest that much better performance can be expected
from the neural network. Two problems must be addressed: (1) How do we obtain,
from the observed raw data alone, a coarse estimate of the slot parameters at the
highest level and (2) given these crude estimates how do we utilize them to recall
default settings for the segmentation one level below?
405
406
Utans and Gindi
0,
Bootstrap
y
Model
Data
Figure 4: Bootstrap computation for a network from a 3- level grammar. Analog frame variables at the
top and intermediate level are initialized from data by a bootstrap computation (bold lines indicate the
flow of information)
2.2
Initialization of Coarse Scale Parameters
We propose to aid convergence by supplying initial values for the analog variables p
and {Pi}; these must be computed from data without making use of the labelling.
In general, it is not possible to solve for the analog parameters without knowledge
of the correct permutation matrix M. However, for the purpose of obtaining an
approximation f> one can derive a new objective function that does not depend on
M and the parameters {Pi} by integrating over the {Pi} and summing over all
possible permutation matrices M:
P(p,{pk})
J
L
=
d{pj}P(P,{Pi},{pd,M)
(4)
{M}IM
is a
permutation
This formulation leads to an Elastic Net type network [9, 7]. However, this implementation of a separate network for the bootstrap computations is expensive.
Here we use simpler computation where the coarse scale parameters are estimated
by computing sample averages, corresponding to finding the solution for the Elastic
Net in the high temperature limit [7]. For the position x we find, after integrating
over the {xi},
L
1
x
L{3m
1/(O"~xO"~mx)
_
1
L{3
1/ O"~x
13m k
L
(3
M{3mkXk
O"~xO"~mx
U{3x
O"~x
(5)
and similarly for y. Since the assignment M{3m k of subparts k on the data side
to subparts fJm on the model side is not known at this point, the first term in
equations (5) cannot be evaluated . After approximating the actual variance with
Improving Convergence in Hierarchical Matching Networks for Object Recognition
an average variance, these equations reduce to
1 1 1
x
N N
Xk - N N
uf3mx - N
13
m
L
k
13
m
L
13m
L uf3x
(6)
13 13
In terms of the objective function this translates into assuming that here the error
terms for all parts are weighted equally. Since these weights would depend on the
actual part match, this just corresponds to our ignorance regarding identity of the
parts. This approximation assumes that the variances do not differ by a large
amount, otherwise the approximation p will not be close to the true values. Since
the model can be designed such that the part primitives used at the lowest level
of the grammar are not highly specialized as would be the case for abstractions
at higher levels of the model, the approximation proved sufficient for the problems
studied here.
The neural network can be used to perform the calculation. The Elastic Net formulation assigns approximately equal weights to all possible assignments at high
temperatures. Thus, this behavior can be expressed in the original network with
match variables by choosing Mf3mk = l/{Nf3Nm ) V i,j. This leads to the following
two-pass bootstrap computation. Using this specific choice for M only the analog
variables need to be updated to compute the coarse scale estimates. The network
with constant M is just the neural network implementation for computing x from
equation (6). After these have converged, x can be used to compute Xj = x + uf3.
Thus, the parameters for intermediate levels can by hypothesized from the coarse
scale estimate x by adding the known transformation (recall that for intermediate
levels, the part identity is preserved and no permutation steps takes place (see Figure 2)). Then the network is restarted with random values for the match variables
to compute the correct labelling and the correct parameters.
2.3
Simulation Results
The bootstrap procedure has been implemented for a 3-level hierarchical model. The
model describes a "gingerbread man" as shown in Figure 3. The incorrect solutions
observed did not, in the vast majority of cases, violate the permutation matrix
constraint, i.e. the assignment was unique. However, even though the assignment
is unique, parts where not always assigned correctly. Most commonly, the identity
of neighboring parts was interchanged, in particular for cases with large variance.
The advantage of using the bootstrap initialization is clear from Figure 5. For
the simulation, cr~ = 2crt; the noise variance was identical for all parts. The network computed the solution reliably for large noise variances. In such cases the
performance of the network without initialization deteriorates rapidly. Only one
set of 10 experiments was used for the graph but in all simulations performed,
the network with initialization consistently outperformed the network without initialization. Figure 5(right) shows the time measured in the number of iterations
necessary for the network to converge; it is almost unaffected by the increase in the
noise variance. This is because the initial values derived from data are still close
to the final solution. While in some cases, the random starting point happens to
be close to the correct solution and the network without initialization converges
rapidly, Figure 5 reflect the typical behavior and demonstrate the advantage of
computing approximate initial values.
407
408
Utans and Gindi
Success Rate
Convergence Speed
100
300
'"
.-.,o-=
80
. oj
...
11)200
80
.oj
~
.....
o
...
.0
II)
~
'h .o
100
-=
20
0.2
0 .?
ott.
0.8
2
(122
0.8
I 0
o
0.0
0.2
0.4
a"1'
0.8
2
0.8
1.0
CT22
Figure 5 : Results Comparing the Network without and with Initialization (solid line) .
Left : The success rate indicates the rate at which the network converged to the correct solutions. /1~
denotes the noise variance at the intermediate level of the model and /1~ the noise variance at the lowest
level. Only one set of 10 experiments was used for the graph but in all simulations performed , the
network with initialization consistently outperformed the network without initialization .
Right: The graph shows the average time it takes for the network to converge (as measured by the
number of iterations) averaged over 10 experiments. Only simulations where the network converged to
the correct solution are used to compute the average time for convergence. The stopping criterion used
required all the match neurons to assume values M'j > 0.95 or M'J < 0 .05. The error bars denote the
standard deviation.
Acknowledgements
This work was supported in part by AFOSR grant AFOSR 90-0224 . Vie thank
E. Mjolsness and A. Rangarajan for many helpful discussions.
References
[1] G. Gindi, E. Mj~lsness, and P. Anandan. Neural networks for model based recognition. In Neural Networks: Concepts, Applications and Implementations, pages 144-173.
Prentice-Hall, 1991.
[2] David Marr. Vision. W. H. Freeman and Co., New York, 1982.
[3] E. Mjolsness. Bay~sian inference on visual grammars by neural nets that optimize.
Technical Report YALEU-DCS-TR-854, Yale University, Dept. of Computer Science,
1991.
[4] E. Mj~lsness. Visual grammars and their neural nets. In R.P. Lippmann J.E. Moody,
S.J. Hanson, editor, Advances in Neural Information Processing Systems 4. Morgan
Kaufmann Publishers, San Mateo, CA, 1992.
[5] Eric Mjolsness, Gene Gindi, and P. Anandan. Optimization in model matching and
perceptual organization: A first look. Research report yaleu/dcs/rr-634, Yale University, Department of Computer Science, 1988.
[6] Eric Mjolsness, Gene R. Gindi, and P. Anandan. Optimization in model matching and
perceptual organization. Neural Computation, vol. 1, no. 2, 1989.
[7] Joachim Utans. Neural Networks for Object Recognition within Compositional Hierarchies. PhD thesis, Department of Electrical Engineering, Yale University, New Haven,
CT 06520, 1992.
[8] Joachim Utans, Gene R. Gindi, Eric Mjolsness, and P. Anandan. Neural networks
for object recognition within compositional hierarchies: Initial experiments. Technical report 8903, Yale University, Center for Systems Science, Department Electrical
Engineering, 1989.
[9] A. L. Yuille. Generalized deformable models, statistical physics, and matching problems. Neural Computation, 2(2):1-24, 1990.
| 661 |@word simulation:6 tr:1 solid:1 recursively:1 yaleu:2 initial:7 denoting:1 current:1 comparing:1 stony:2 must:8 j1:1 shape:3 designed:2 v:1 alone:1 xk:1 supplying:1 coarse:8 node:1 successive:1 simpler:1 incorrect:1 advocate:1 combine:1 expected:1 behavior:2 freeman:1 actual:2 becomes:1 estimating:1 notation:1 matched:4 lowest:9 unspecified:1 finding:4 transformation:1 suite:1 berkeley:2 stick:2 control:1 grant:1 appear:1 before:1 vie:1 engineering:4 limit:2 approximately:1 initialization:9 studied:1 mateo:1 specifying:2 co:1 contextually:1 averaged:1 unique:2 lost:1 differs:1 bootstrap:8 procedure:1 matching:9 pre:1 integrating:2 suggest:1 cannot:1 close:3 prentice:1 optimize:1 equivalent:1 map:1 center:2 primitive:1 starting:1 independently:1 resolution:1 simplicity:2 assigns:2 marr:3 notion:3 coordinate:6 variation:1 updated:1 hierarchy:15 element:2 recognition:17 expensive:1 utilized:1 coarser:1 database:1 labeled:1 observed:3 bottom:1 electrical:4 capture:1 mjolsness:7 highest:4 icsi:1 principled:1 pd:3 depend:2 solving:3 segment:1 serve:1 yuille:1 eric:3 blurring:1 triangle:2 joint:2 represented:3 choosing:1 solve:1 otherwise:2 grammar:8 itself:1 final:2 advantage:2 rr:1 net:5 propose:1 neighboring:1 rapidly:2 deformable:1 description:5 convergence:7 double:1 rangarajan:1 converges:1 object:29 derive:3 measured:2 solves:1 auxiliary:1 implemented:1 involves:3 indicate:1 differ:1 correct:7 stochastic:8 centered:1 human:1 crt:1 probable:1 im:2 hall:1 mo:2 interchanged:1 omitted:1 purpose:1 ptj:1 outperformed:2 grouped:1 weighted:2 minimization:1 gaussian:1 always:1 rather:2 cr:1 derived:1 joachim:3 consistently:2 indicates:1 am:1 sense:1 helpful:1 inference:3 dependent:1 rigid:1 abstraction:4 stopping:1 mth:1 interested:1 arg:1 orientation:1 exponent:1 spatial:2 initialize:1 equal:1 identical:1 placing:1 represents:1 look:1 report:3 haven:2 hint:1 few:1 composed:3 recognize:1 individual:1 invariably:1 organization:2 highly:1 necessary:3 iv:1 mkxk:1 initialized:1 circle:1 instance:3 earlier:1 assignment:5 ott:1 introducing:1 deviation:1 uniform:1 recognizing:1 stored:7 density:1 international:1 physic:1 together:1 moody:1 thesis:1 central:1 reflect:1 inaj:2 style:1 bold:3 includes:1 explicitly:3 performed:2 nishihara:2 observing:1 reached:1 bayes:1 parallel:1 minimize:1 variance:11 kaufmann:1 yield:1 correspond:1 bayesian:1 raw:1 finer:1 unaffected:1 converged:3 simultaneous:1 influenced:1 wai:1 intentionally:1 proved:1 recall:2 knowledge:1 segmentation:4 appears:1 higher:3 formulation:2 done:1 box:1 strongly:1 generality:1 evaluated:1 just:2 stage:2 though:1 until:1 horizontal:1 hypothesized:1 concept:1 true:1 assigned:1 ignorance:1 essence:1 criterion:1 generalized:1 demonstrate:1 temperature:2 image:4 specialized:1 analog:6 belong:1 ai:1 similarly:2 inclusion:1 base:2 optimizing:1 belongs:1 binary:1 success:2 morgan:1 additional:1 anandan:4 converge:2 redundant:1 ii:2 violate:1 reduces:3 technical:2 match:10 determination:1 calculation:1 equally:1 vision:2 blindly:1 iteration:2 represent:7 preserved:1 addition:2 addressed:1 publisher:1 flow:1 structural:4 intermediate:7 iii:1 xj:1 architecture:5 reduce:2 regarding:1 pdim:1 translates:2 wo:1 york:1 compositional:11 detailed:2 clear:1 amount:1 mid:1 reduced:2 mai:1 utans:8 estimated:1 deteriorates:1 track:1 correctly:1 subpart:9 vol:1 suny:1 pj:1 utilize:1 vast:1 graph:3 angle:1 place:1 almost:1 ct:2 simplification:1 yale:6 correspondence:1 ahead:1 constraint:1 scene:3 aspect:1 speed:1 coarsest:1 relatively:1 department:5 describes:2 unity:2 making:1 happens:1 xo:2 equation:4 describing:2 know:1 operation:1 hierarchical:6 original:1 denotes:2 top:5 include:1 assumes:1 approximating:1 objective:4 added:4 already:1 gindi:9 mx:2 separate:1 thank:1 street:1 majority:1 considers:1 collected:1 assuming:1 length:2 index:3 relationship:3 minimizing:1 difficult:1 stated:1 design:1 implementation:3 reliably:1 perform:1 neuron:2 arc:1 incorporated:1 frame:6 dc:2 station:1 frameville:7 rame:1 introduced:1 david:1 required:1 hanson:1 recalled:1 brook:2 address:2 bar:1 below:2 inaij:1 summarize:1 fjm:1 max:1 memory:1 oj:2 natural:1 sian:1 representing:1 scheme:2 coupled:2 text:1 prior:1 literature:1 acknowledgement:1 relative:3 afosr:2 ina:3 loss:1 permutation:8 sufficient:1 consistent:1 editor:1 pi:12 translation:1 lo:1 placed:1 supported:1 side:10 institute:1 distributed:2 default:3 forward:3 commonly:1 san:1 approximate:2 lippmann:1 gene:4 keep:1 summing:1 xi:1 bay:1 mj:2 ca:2 elastic:3 obtaining:1 improving:4 domain:1 did:1 pk:4 noise:6 arrival:1 nothing:1 repeated:1 ny:1 aid:1 structurally:1 position:4 sub:1 crude:1 perceptual:2 third:1 ix:1 theorem:1 down:3 specific:2 grouping:5 derives:1 adding:1 mirror:1 phd:1 labelling:8 illustrates:1 demand:1 depicted:1 visual:5 expressed:2 restarted:1 corresponds:2 slot:1 goal:1 formulated:1 identity:5 labelled:1 man:1 typical:1 oval:1 pas:1 select:1 dept:1 ex:1 |
6,202 | 6,610 | On the Consistency of Quick Shift
Heinrich Jiang
Google Inc.
1600 Amphitheatre Parkway, Mountain View, CA 94043
[email protected]
Abstract
Quick Shift is a popular mode-seeking and clustering algorithm. We present finite
sample statistical consistency guarantees for Quick Shift on mode and cluster
recovery under mild distributional assumptions. We then apply our results to
construct a consistent modal regression algorithm.
1
Introduction
Quick Shift [16] is a clustering and mode-seeking procedure that has received much attention in
computer vision and related areas. It is simple and proceeds as follows: it moves each sample to its
closest sample with a higher empirical density if one exists in a ? radius ball, where the empirical
density is taken to be the Kernel Density Estimator (KDE). The output of the procedure can thus be
seen as a graph whose vertices are the sample points and a directed edge from each sample to its next
point if one exists. Furthermore, it can be seen that Quick Shift partitions the samples into trees which
can be taken as the final clusters, and the root of each such tree is an estimate of a local maxima.
Quick Shift was designed as an alternative to the better known mean-shift procedure [4, 5]. Mean-shift
performs a gradient ascent of the KDE starting at each sample until -convergence. The samples that
correspond to the same points of convergence are in the same cluster and the points of convergence
are taken to be the estimates of the modes. Both procedures aim at clustering the data points by
incrementally hill-climbing to a mode in the underlying density. Some key differences are that Quick
Shift restricts the steps to sample points and has the extra ? parameter. In this paper, we show that
Quick Shift can surprisingly attain strong statistical guarantees without the second-order density
assumptions required to analyze mean-shift.
We prove that Quick Shift recovers the modes of an arbitrary multimodal density at a minimax optimal
rate under mild nonparametric assumptions. This provides an alternative to known procedures with
similar statistical guarantees; however such procedures only recover the modes but fail to inform us
how to assign the sample points to a mode which is critical for clustering. Quick Shift on the other
hand recovers both the modes and the clustering assignments with statistical consistency guarantees.
Moreover, Quick Shift?s ability to do all of this has been extensively validated in practice.
A unique feature of Quick Shift is that it has a segmentation parameter ? which allows practioners
to merge clusters corresponding to certain less salient modes of the distribution. In other words, if
a local mode is not the maximizer of its ? -radius neighborhood, then its corresponding cluster will
become merged to that of another mode. Current consistent mode-seeking procedures [6, 12] fail to
allow one to control such segmentation. We give guarantees on how Quick Shift does this given an
arbitrary setting of ? .
We show that Quick Shift can also be used to recover the cluster tree. In cluster tree estimation, the
known procedures with the strongest statistical consistency guarantees include Robust Single Linkage
(RSL) [2] and its variants e.g. [13, 7]. We show that Quick Shift attains similar guarantees.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
Thus, Quick Shift, a simple and already popular procedure, can simultaneously recover the modes
with segmentation tuning, provide clustering assignments to the appropriate mode, and can estimate
the cluster tree of an unknown density f with the strong consistency guarantees. No other procedure
has been shown to have these properties.
Then we use Quick Shift to solve the modal regression problem [3], which involves estimating the
modes of the conditional density f (y|X) rather than the mean as in classical regression. Traditional
approaches use a modified version of mean-shift. We provide an alternative using Quick Shift which
has precise statistical consistency guarantees under much more mild assumptions.
Figure 1: Quick Shift example. Left: ? = ?. The procedure returns one tree, whose head is the
sample with highest empirical density. Right: ? set to a lower value. The edges with length greater
than ? are no longer present when compared to the left. We are left with three clusters.
2
Assumptions and Supporting Results
Algorithm 1 Quick Shift
Input: Samples X[n] := {x1 , ..., xn }, KDE bandwidth h, segmentation parameter ? > 0.
Initialize directed graph G with vertices {x1 , ..., xn } and no edges.
for i = 1 to n do
if there exists x ? X[n] such that fbh (x) > fbh (xi ) and ||x ? xi || ? ? then
Add to G a directed edge from xi to argminxj ?X[n] :fbh (xj )>fbh (xi ) ||xi ? xj ||.
end if
end for
return G.
2.1
Setup
Let X[n] = {x1 , ..., xn } be n i.i.d. samples drawn from distribution F with density f over the
uniform measure on Rd .
Assumption 1 (H?lder Density). f is H?lder continuous on compact support X ? Rd . i.e. |f (x) ?
f (x0 )| ? C? ||x ? x0 ||? for all x, x0 ? X and some 0 < ? ? 1 and C? > 0.
Definition 1 (Level Set). The ? level set of f is defined as Lf (?) := {x ? X : f (x) ? ?}.
Definition 2 (Hausdorff Distance). dH (A, A0 ) = max{supx?A d(x, A0 ), supx?A0 d(x, A)}, where
d(x, A) := inf x0 ?A ||x ? x0 ||.
The next assumption says that the level sets are continuous in the level ? w.r.t. the Hausdorff metric.
This ensures that there are no flat areas in which the procedure may get stuck at.
Assumption 2 (Uniform Continuity of Level Sets). For each > 0, there exists ? > 0 such that for
0 < ?, ?0 ? ||f ||? with |? ? ?0 | < ?, then dH (Lf (?), Lf (?0 )) < .
Remark 1. One of the difficulties is dealing with flat areas of the density. Procedures that try to
incrementally move points to nearby areas of higher density will have difficulties in regions where
there is no change in the density. The above assumption is a simple and mild formulation which
ensures there are no such flat regions.
Remark 2. Note that our assumptions are quite mild when compared to analyses of similar procedures like mean-shift, which require at least second-order smoothness assumptions. Interestingly, we
only require H?lder continuity.
2
2.2
KDE Bounds
We next give uniform bounds on KDE required to analyze Quick Shift.
d
Definition 3. Define
R kernel function K : R ? R?0 where R?0 denotes the non-negative real
numbers such that Rd K(u)du = 1.
We make the following mild regularity assumptions on K.
Assumption 3. (Spherically symmetric, non-increasing, and exponential decays) There exists nonincreasing function k : R?0 ? R?0 such that K(u) = k(|u|) for u ? Rd and there exists
?, C? , t0 > 0 such that for t > t0 , k(t) ? C? ? exp(?t? ).
Remark 3. These assumptions allow the popular kernels such as Gaussian, Exponential, Silverman,
uniform, triangular, tricube, Cosine, and Epanechnikov.
Definition 4 (Kernel Density Estimator). Given a kernel K and bandwidth h > 0 the KDE is defined
by
n
x ? Xi
1 X
b
K
.
fh (x) =
n ? hd i=1
h
Here we provide the uniform KDE bound which will be used for our analysis, established in [11].
Lemma 1. [`? bound for ?-H?lder continuous functions. Theorem 2 of [11]] There exists positive
constant C 0 depending on f and K such that the following holds with probability at least 1 ? 1/n
uniformly in h > (log n/n)1/d .
!
r
log
n
0
?
sup |fbh (x) ? f (x)| < C ? h +
.
n ? hd
x?Rd
3
Mode Estimation
In this section, we give guarantees about the local modes returned by Quick Shift. We make the
additional assumption that the modes are local maxima points with negative-definite Hessian.
Assumption 4. [Modes] A local maxima of f is a connected region M such that the density is
constant on M and decays around its boundaries. Assume that each local maxima of f is a point,
which we call a mode. Let M be the modes of f where M is a finite set. Then let f be twice
differentiable around a neighborhood of each x ? M and let f have a negative-definite Hessian at
each x ? M and those neighborhoods are disjoint.
This assumption leads to the following.
? C? > 0 such that the
Lemma 2 (Lemma 5 of [6]). Let f satisfy Assumption 4. There exists rM , C,
following holds for all x0 ? M simultaneously.
C? ? |x0 ? x|2 ? f (x0 ) ? f (x) ? C? ? |x0 ? x|2 ,
for all x ? Ax0 where Ax0 is a connected component of {x : f (x) ? inf x0 ?B(x0 ,rM ) f (x)} which
contains x0 and does not intersect with other modes.
The next assumption ensures that the level sets don?t become arbitrarily thin as long as we are
sufficiently away from the modes.
Assumption 5. [Level Set Regularity] For each ?, r > 0, there exists ? > 0 such that the following
holds for all connected components A of Lf (?) with ? > 0 and A 6? ?x0 ?M B(x0 , r). If x lies on
the boundary of A, then Vol(B(x, ?) ? A) > ? where Vol is volume w.r.t. the uniform measure on Rd .
We next give the result about mode recovery for Quick Shift. It says that as long as ? is small enough,
then as the number of samples grows, the roots of the trees returned by Quick Shift will bijectively
correspond to the true modes of f and the estimation errors will match lower bounds established by
Tsybakov [15] up to logarithmic factors. We defer the proof to Theorem 2 which is a generalization
of the following result.
Theorem 1 (Mode Estimation guarantees for Quick Shift). Let ? < rM /2 and Assumptions 1, 2, 3, 4,
c be the
and 5 hold. Choose h such that (log n)2/? ? h ? 0 and log n/(nhd ) ? 0 as n ? ?. Let M
3
heads of the trees in G (returned by Algorithm 1). There exists constant C depending on f and K
such that for n sufficiently large, with probability at least 1 ? 1/n,
!
r
log n
2
4/? 2
c
.
dH (M, M) < C (log n) h +
n ? hd
c In particular, taking h ? n?1/(4+d) optimizes the above rate to d(M, M)
c =
and |M| = |M|.
?1/(4+d)
?
O(n
). This matches the minimax optimal rate for mode estimation up to logarithmic factors.
We now give a stronger notion of mode that fits better for analyzing the role of ? . In the last result, it
was assumed that the practitioner wished to recover exactly the modes of the density f by taking ?
sufficiently small. Now, we analyze the case where ? is intentionally set to a particular value so that
Quick Shift produces segmentations that group modes together that are in close proximity to higher
density regions.
Definition 5. A mode x0 ? M is an (r, ?)+ -mode if f (x0 ) > f (x) + ? for all x ?
B(x0 , r)\B(x0 , rM ). A mode x0 ? M is an (r, ?)? -mode if f (x0 ) < f (x) ? ? for some
?
+
?
x ? B(x0 , r). Let M+
r,? ? M and Mr,? ? M denote the set of (r, ?) -modes and (r, ?) -modes
of f , respectively.
In other words, an (r, ?)+ -mode is a mode that is also a maximizer in a larger ball of radius r by at
least ? when outside of the region where there is quadratic decay and smoothness (B(x0 , rM )). An
(r, ?)? -mode is a mode that is not the maximizer in its radius r ball by a margin of at least ?.
The next result shows that Algorithm recovers the (? +, ?)+ -modes of f and excludes the (? ?, ?)? modes of f . The proof is in the appendix.
Theorem 2. (Generalization of Theorem 1) Let ?, > 0 and suppose Assumptions 1, 2, 3, 4, and 5
hold. Let h ? h(n) be chosen such that h ? 0 and log n/(nhd ) ? 0 as n ? ?. Then there exists
C > 0 depending on f and K such that the following holds for n sufficiently large with probability
?
c such that
at least 1 ? 1/n. For each x0 ? M+
??M
? +,? \M? ?,? , there exists unique x
!
r
log n
2
4/? 2
.
||x0 ? x
?|| < C (log n) h +
n ? hd
c ? |M| ? |M? |.
Moreover, |M|
? ?,?
In particular, taking ? 0 and ? ? 0 gives us an exact characterization of the asymptotic behavior
of Quick Shift in terms of mode recovery.
4
Assignment of Points to Modes
In this section, we give guarantees on how the points are assigned to their respective modes. We
first give the following definition which formalizes how two points are separated by a wide and deep
valley.
Definition 6. x1 , x2 ? X are (rs , ?)-separated if there exists a set S such that every path from x1
and x2 intersects with S and
sup
x?S+B(0,rs )
f (x) <
inf
f (x) ? ?.
x?B(x1 ,rs )?B(x2 ,rs )
Lemma 3. Suppose Assumptions 1, 2, 3, 4, and 5 hold. Let ? < rs /2 and choose h such that
(log n)2/? ? h ? 0 and log n/(nhd ) ? 0 as n ? ?. Let G be the output of Algorithm 1. The
following holds with probability at least 1 ? 1/n for n sufficiently large depending on f , K, ?, and ?
uniformly in all x1 , x2 ? X . If x1 and x2 are (rs , ?)-separated, then there cannot exist a directed
path from x1 to x2 in G.
Proof. Suppose that x1 and x2 are (rs , ?)-separated (with respect to set S) and there exists a
directed path from x1 to x2 in G. Given our choice of ? , there exists some point x ? G such that
x ? S+B(0, rs ) and x is on the path from x1 to x2 . We have f (x) < f (x1 )??. Choose n sufficiently
large such that by Lemma 1, supx?X |fbh (x) ? f (x)| < ?/2. Thus, we have fbh (x) < fbh (x1 ), which
means a directed path in G starting from x1 cannot contain x, a contradiction. The result follows.
4
Figure 2: Illustration of (rs , ?)-separation in 1 dimension. Here A and B are (rs , ?)-separated by S.
This is because the minimum density level of rs -radius balls around A and B (the red dotted line)
exceeds the maximum density level of the rs -radius ball around S by at least ? (golden dotted line).
In other words, there exists a sufficiently wide (controlled by rs and S) and deep (controlled by ?)
valley separating A and B. The results in this section will show that in such cases, these pairs of
points will not be assigned to the same cluster.
This leads to the following consequence about how samples are assigned to their respective modes.
Theorem 3. Assume the same conditions as Lemma 3. The following holds with probability at least
1 ? 1/n for n sufficiently large depending on f , K, ?, and ? uniformly in x ? X and x0 ? M. For
each x ? X and x0 ? M, if x and x0 are (rs , ?)-separated, then x will not be assigned to the tree
corresponding to x0 from Theorem 1.
Remark 4. In particular, taking ? ? 0 and rs ? 0 gives us guarantees for all points which have a
unique mode in which it can be assigned to.
We now give a more general version of (rs , ?)-separation, in which the condition holds if every path
between the two points dips down at some point. The same results as the above extend for this
definition in a straightforward manner.
Definition 7. x1 , x2 ? X are (rs , ?)-weakly-separated if there exists a set S, with x1 , x2 6? S +
B(0, rs ), such that every path P from x1 and x2 satisifes the following. (1) P ? S 6= ? and (2)
sup
x?P?S+B(0,rs )
f (x) <
inf
x?B(x01 ,rs )?B(x02 ,rs )
f (x) ? ?,
where x01 , x02 are defined as follows. Let P1 be the path obtained by starting at x1 and following
P until it intersects S, and P2 be the path obtained by following P starting from the last time it
intersects S until the end. Then x01 and x02 are points which respectively attain the highest values of f
on P1 and P2 .
5
Cluster Tree Recovery
The connected components of the level sets as the density level varies forms a hierarchical structure
known as the cluster tree.
Definition 8 (Cluster Tree). The cluster tree of f is given by
Cf (?) := connected components of {x ? X : f (x) ? ?}.
Definition 9. Let G(?) be the subgraph of G with vertices x ? X[n] such that fbh (x) > ? and edges
?
between pairs of vertices which have corresponding edges in G. Let G(?)
be the sets of vertices
corresponding to the connected components of G(?).
Definition 10. Suppose that A is a collection of sets of points in Rd . Then define Link(A, ?)
to be the result of repeatedly removing pairs A1 , A2 ? A from A (A1 6= A2 ) that satisfy
inf a1 ?A1 inf a2 ?A2 ||a1 ? a2 || < ? and adding A1 ? A2 to A until no such pairs exist.
Parameter settings for Algorithm 2: Suppose that ? ? ? (n) is chosen as a function of n such
such that ? ? 0 as n ? ?, ? (n) ? (log2 n/n)1/d and h ? h(n) is chosen such that h ? 0 and
log n/(nhd ) ? 0 as n ? ?.
The following is the main result of this section, the proof is in the appendix.
5
Algorithm 2 Quick Shift Cluster Tree Estimator
Input: Samples X[n] := {X1 , ..., Xn }, KDE bandwidth h, segmentation parameter ? > 0.
Let G be the output of Quick Shift (Algorithm 1) with above parameters.
bf (?) := Link(G(?),
?
For ? > 0, let C
? ).
b
return Cf
Theorem 4 (Consistency). Algorithm 2 converges in probability to the true cluster tree of f under
merge distortion (defined in [7]).
Remark 5. By combining the result of this section with the mode estimation result, we can obtain
the following interpretation. For any level ?, a component in G(?) estimates a connected component
of the ?-level set of f , and that further, the trees within that component in G(?) have a one-to-one
correspondence with the modes in the connected component.
Figure 3: Illustration on 1-dimensional density with three modes A, B, and C. When restricting Quick
Shift?s output to samples have empirical density above a certain threshold and connecting nearby
clusters, then this approximates the connected components of the true density level set. Moreover, we
give guarantees that such points will be assigned to clusters which correspond to modes within its
connected component.
6
Modal Regression
Suppose that we have joint density f (X, y) on Rd ? R w.r.t. to the Lebesgue measure. In modal
regression, we are interested in estimating the modes of the conditional f (y|X = x) given samples
from the joint distribution.
Algorithm 3 Quick Shift Modal Regression
Input: Samples D := {(x1 , y1 ), ..., (xn , yn )}, bandwidth h, ? > 0, and x ? X .
Let Y = {y1 , ..., yn } and fbh be the KDE computed w.r.t. D.
Initialize directed graph G with vertices Y and no edges.
for i = 1 to n do
if there exists yj ? [yi ? ?, yi + ? ] ? Y such that fbh (x, yj ) > fbh (x, yi ) then
Add to G an directed edge from yi to argminyi ?Y :fbh (x,yj )>fbh (x,yi ) ||yi ? yj ||.
end if
end for
return The roots of the trees of G as the estimates of the modes of f (y|X = x).
Theorem 5 (Consistency of Quick Shift Modal Regression). Suppose that ? ? ? (n) is chosen as a
function of n such such that ? ? 0 as n ? ?, ? (n) ? (log2 n/n)1/d and h ? h(n) is chosen such
that h ? 0 and log n/(nhd+1 ) ? 0 as n ? ?. Let Mx be the modes of the conditional density
cx be the output of Algorithm 3. Then with probability at least 1 ? 1/n uniformly
f (y|X = x) and M
in x such that f (y|X = x) and K satisfies Assumptions 1, 2, 3, 4, and 5,
cx ) ? 0 as n ? ?.
dH (Mx , M
6
7
Related Works
Mode Estimation. Perhaps the most popular procedure to estimate the modes is mean-shift; however,
it has proven quite difficult to analyze. Arias-Castro et al. [1] made much progress by utilizing
dynamical systems theory to show that mean-shift?s updates converge to the correct gradient ascent
steps. The recent work of Dasgupta and Kpotufe [6] was the first to give a procedure which recovers
the modes of a density with minimax optimal statistical guarantees in a multimodal density. They do
this by using a top-down traversal of the density levels of a proximity graph, borrowing from work
in cluster tree estimation. The procedure was shown to recover exactly the modes of the density at
minimax optimal rates.
In this work, we showed that Quick Shift attains the same guarantees while being a simpler approach
than known procedures that attain these guarantees [6, 12]. Moreover unlike these procedures, Quick
Shift also assigns the remaining samples to their appropriate modes. Furthermore, Quick Shift also
has a segmentation tuning parameter ? which allows us to merge the clusters of modes that are not
maximal in its ? -radius neighborhood into the clusters of other modes. This is useful as in practice,
one may not wish to pick up every single local maxima, especially when there are local maxima that
can be grouped together by proximity. We formalized the segmentation of such modes and identify
which modes get returned and which ones become merged into other modes? clusters by Quick Shift.
Cluster Tree Estimation. Work on cluster tree estimation has a long history. Some early work on
density based clustering by Hartigan [9] modeled the clusters of a density as the regions {x : f (x) ?
?} for some ?. This is called the density level-set of f at level ?. The cluster tree of f is the hierarchy
formed by the infinite collection of these clusters over all ?. Chaudhuri and Dasgupta [2] introduced
Robust Single Linkage (RSL) which was the first cluster tree estimation procedure with precise
statistical guarantees. Shortly after, Kpotufe and Luxburg [13] provided an estimator that ensured
false clusters were removed using used an extra pruning step. Interestingly, Quick Shift does not
require such a pruning step, since the points near cluster boundaries naturally get assigned to regions
with higher density and thus no spurious clusters are formed near these boundaries. Sriperumbudur
and Steinwart [14], Jiang [10], Wang et al. [17] showed that the popular DBSCAN algorithm [8] also
estimates these level sets. Eldridge et al. [7] introduced the merge distortion metric for cluster tree
estimates, which provided a stronger notion of consistency. We use their framework to analyze Quick
Shift and show that this simple estimator is consistent in merge distortion.
Figure 4: Density-based clusters discovered by level-set model {x : f (x) ? ?} (e.g. DBSCAN) vs
Quick Shift on a one dimensional density. Left two images: level sets for two density level settings.
Unassigned regions are noise and have no cluster assignment. Right two images: Quick Shift with
two different ? settings. The latter is a hill-climbing based clustering assignment.
Modal Regression. Nonparametric modal regression [3] is an alternative to classical regression,
where we are interested in estimating the modes of the conditional density f (y|X = x) rather
than the mean. Current approaches primarily use a modification of mean-shift; however analysis
for mean-shift require higher order smoothness assumptions. Using Quick Shift instead for modal
regression requires less regularity assumptions while having consistency guarantees.
8
Conclusion
We provided consistency guarantees for Quick Shift under mild assumptions. We showed that
Quick Shift recovers the modes of a density from a finite sample with minimax optimal guarantees.
The approach of this method is considerably different from known procedures that attain similar
guarantees. Moreover, Quick Shift allows tuning of the segmentation and we provided an analysis
of this behavior. We also showed that Quick Shift can be used as an alternative for estimating the
7
cluster tree which contrasts with current approaches which utilize proximity graph sweeps. We
then constructed a procedure for modal regression using Quick Shift which attains strong statistical
guarantees.
Appendix
Mode Estimation Proofs
Lemma 4. Suppose Assumptions 1, 2, 3, 4, and 5 hold. Let r? > 0 and h ? h(n) be chosen such
that h ? 0 and log n/(nhd ) ? 0 as n ? ?. Then the following holds for n sufficiently large with
probability at least 1 ? 1/n. Define
)
(
r
32C?
log n
2
4/? 2
0
r? := max
.
(log n) h , 17 ? C
n ? hd
C?
Suppose x0 ? M and x0 is the unique maximizer of f on B(x0 , r?).
argmaxx?B(x0 ,?r)?X[n] fbh (x), we have
Then letting x
? :=
||x0 ? x
?|| < r?.
Proof sketch. This follows from modifying the proof of Theorem 3 of [11] by replacing Rd \B(x0 , r?)
with B(x0 , r?)\B(x0 , r?). This leads us to
inf
x?B(x0 ,rn )
fbh (x) >
sup
fbh (x),
x?B(x0 ,?
r )\B(x0 ,?
r)
where rn := minx?X[n] |x0 ? x| and n is chosen sufficiently large such that r? < ? . Thus, |x0 ? x
?| ?
r?.
?
Proof of Theorem 2. Suppose that x0 ? M+
? := argmaxx?B(x0 ,? )?X[n] fbh (x).
? +,? \M? ?,? . Let x
c
We first show that x
? ? M.
q
?
n
By Lemma 4, we have |x0 ? x
?| ? r? where r?2 := max 32C?C (log n)4/? h2 , 17 ? C 0 log
. It
n?hd
remains to show that x
? = argmaxx?B(?x,? )?X[n] fbh (x). We have B(?
x, ? ) ? B(x0 , ? + r?). Choose
n sufficiently large such that (i) r? < , (ii) by Lemma 1, supx?X |fbh (x) ? f (x)| < ?/4 and (iii)
? Now, we have
r?2 < ?/(4C).
sup
fbh (x) ?
x?B(x0 ,? +?
r )\B(x0 ,? )
sup
f (x) + ?/4 ? f (x0 ) ? 3?/4
x?B(x0 ,? +?
r )\B(x0 ,? )
? f (?
x) + C? r?2 ? 3?/4 < f (?
x) ? ?/2 < fbh (?
x).
c
Thus, x
? = argmaxx?B(?x,? )?X[n] fbh (x). Hence, x
? ? M.
c such that ||?
Next, we show that it is unique. To do this, suppose that x
?0 ? M
x0 ? x0 || ? ? /2. Then
0
b
we have both x
? = argmaxx?B(?x,? )?X[n] fh (x) and x
? = argmaxx?B(?x0 ,? )?X[n] fbh (x). However,
choosing n sufficiently large such that r? < ? /2, we obtain x
? ? B(?
x0 , ? ). This implies that x
?=x
?0 , as
desired.
c ? |M| ? |M? |. Suppose that x
c Let ?0 := min{/3, ? /3, rM /2}. We
We now show |M|
? ? M.
? ?,?
show that B(?
x, ?0 ) ? M =
6 ?. Suppose otherwise. Let ? = f (?
x). By Assumptions 2 and 5, we
have that there exists ? > 0 and ? > 0 such that the following holds uniformly: Vol(B(?
x, ?0 ) ?
Lf (? + ?)) ? ?. Choose n sufficiently large such that (i) by Lemma 1, supx?X |fbh (x) ? f (x)| <
min ?/2, ?/4 and (ii) there exists a sample x ? B(?
x, /3) ? Lf (? + ?) ? X[n] by Lemma 7 of
b
Chaudhuri and Dasgupta [2]. Then fh (x) > ?+?/2 > fbh (?
x) but x ? B(?
x, ?0 ), a contradiction since
x
? is the maximizer of the KDE of the samples in its ? -radius neighborhood. Thus, B(?
x, ?0 ) ? M =
6 ?.
?
0
Now, suppose that there exists x0 ? B(?
x, ?0 ) ? M? ?,? . Then, there exists x ? B(x0 , ? ? 2?0 )
8
such that f (x0 ) ? f (x0 ) + ?. Then, if x
? is the closest sample point to x0 , we have for n sufficiently
0
large, |x ? x
?| ? ?0 and f (?
x) ? f (x0 ) + ?/2 and thus fbh (?
x) > f (?
x) ? ?/4 ? f (?
x) + ?/4 > fbh (?
x).
But x
? ? B(?
x, ? ) ? X[n] , contradicting the fact that x
? is the maximizer of the KDE over samples in
its ? -radius neighborhood. Thus, B(?
x, ?0 ) ? (M\M?
? ?,? ) 6= ?.
c such that x0 ? M\M?
Finally, suppose that there exists x
?, x
?0 ? M
x, ?0 ) ?
? ?,? and x0 ? B(?
B(?
x0 , ?0 ). Then, x
?, x
?0 ? B(x0 , ?0 ), thus |?
x?x
?0 | ? ? and thus x
?=x
?0 , as desired.
Cluster Tree Estimation Proofs
Lemma 5 (Minimality). The following holds with probability at least 1 ? 1/n. If A is a connected
bf (? ? )
component of {x ? X : f (x) ? ?}, then A ? X[n] is contained in the same component in C
for any > 0 as n ? ?.
Proof. It suffices to show that for each x ? A, there exists x0 ? B(x, ? /2) ? X[n] such that fbh (x0 ) >
? ? . Given our choice of ? , it follows by Lemma 7 of [2] that B(x, ? /2) ? X[n] is non-empty for n
sufficiently large. Let x0 ? B(x, ? /2) ? X[n] . Choose n sufficiently large such that by Lemma 1, we
have supx?X |fbh (x) ? f (x)| < /2. We have f (x0 ) ? inf B(x,? /2) f (x) ? ? ? C? (? /2)? > ? ? /2,
where the last inequality holds for n sufficiently large so that ? is sufficiently small. Thus, we have
fbh (x0 ) > ? ? , as desired.
Lemma 6 (Separation). Suppose that A and B are distinct connected components of {x ? X :
f (x) ? ?} which merge at {x ? X : f (x) ? ?}. Then A ? X[n] and B ? X[n] are separated in
bf (? + ) for any > 0 as n ? ?.
C
Proof. It suffices to assume that ? = ? + . Let A0 and B 0 be the connected components of
{x ? X : f (x) ? ? + /2} which contain A and B respectively. By the uniform continuity of f ,
there exists r? > 0 such that A + B(0, 3?
r) ? A0 . We have supx?A0 \(A+B(0,?r)) f (x) = ? + ? 0 for
0
some > 0.
Choose n sufficiently large such that by Lemma 1, we have supx?X |fbh (x) ? f (x)| < 0 /2.
bf (? + ) cannot belong to
Thus, supx?A0 \(A+B(0,?r)) fbh (x) < ? + ? 0 /2. Hence, points in C
A0 \ (A + B(0, r?)). Since A0 also contains A + B(0, 3?
r), it means that there cannot be a path from
A to B with points of empirical density at least ? + with all edges of length less than r?. The result
follows by taking n sufficiently large so that ? < r?, as desired.
Proof of Theorem 4. By the regularity assumptions on f and Theorem 2 of [7], we have that Algorithm 2 has both uniform minimality and uniform separation (defined in [7]), which implies
convergence in merge distortion.
Modal Regression Proofs
cx then y? is a consistent estimator
Proof of Theorem 5. There are two directions to show. (1) if y? ? M
c which estimates
of some mode y0 ? Mx . (2) For each mode, y0 ? M, there exists a unique y? ? M
it.
We first show (1). We show that [?
y ? ?, y? + ? ] ? Mx 6= ?. Suppose otherwise. Let ? = f (x, y?).
Choose ? < ? /4. Then by Assumptions 2 and 5, there exists ? > 0 such that taking = ? /2, we
have that there exists ? > 0 such that {(x, y 0 ) : y 0 ? [?
y ? ?, y? + ? ]} ? Lf (? + ?) contains connected
set A where Vol(A) > ?. Choose n sufficiently large such that (i) there exists y ? A ? Y , and (ii)
by Lemma 1, sup(x0 ,y0 ) |fbh (x0 , y 0 ) ? f (x0 , y 0 )| < ?/2. Then fbh (x, y) > ? + ?/2 > fbh (x, y?) but
y ? [?
y ? ?, y? + ? ], a contradiction since y? is the maximizer of the KDE in ? radius neighborhood
when restricted to X = x. Thus, there exists y0 ? Mx such that y0 ? [?
y ? ?, y? + ? ]. Moreover
this y0 ? Mx must be unique by Lemma 2. As n ? 0, we have ? ? 0 and thus consistency is
established for y? estimating y0 .
Now we show (2). Suppose that y0 ? Mx . From the above, for n sufficiently large, the maximizer of
the KDE in [y0 ? 2?, y0 + 2? ] ? Y is contained in [y0 ? ?, y0 + ? ]. Thus, there exists a root of the
tree contained in [y0 ? ?, y0 + ? ] and taking ? ? 0 gives us the desired result.
9
Acknowledgements
I thank the anonymous reviewers for their valuable feedback.
References
[1] Ery Arias-Castro, David Mason, and Bruno Pelletier. On the estimation of the gradient lines
of a density and the consistency of the mean-shift algorithm. Journal of Machine Learning
Research, 2015.
[2] Kamalika Chaudhuri and Sanjoy Dasgupta. Rates of convergence for the cluster tree. In
Advances in Neural Information Processing Systems, pages 343?351, 2010.
[3] Yen-Chi Chen, Christopher R Genovese, Ryan J Tibshirani, Larry Wasserman, et al. Nonparametric modal regression. The Annals of Statistics, 44(2):489?514, 2016.
[4] Yizong Cheng. Mean shift, mode seeking, and clustering. IEEE transactions on pattern analysis
and machine intelligence, 17(8):790?799, 1995.
[5] Dorin Comaniciu and Peter Meer. Mean shift: A robust approach toward feature space analysis.
IEEE Transactions on pattern analysis and machine intelligence, 24(5):603?619, 2002.
[6] Sanjoy Dasgupta and Samory Kpotufe. Optimal rates for k-nn density and mode estimation. In
Advances in Neural Information Processing Systems, pages 2555?2563, 2014.
[7] Justin Eldridge, Mikhail Belkin, and Yusu Wang. Beyond hartigan consistency: Merge distortion
metric for hierarchical clustering. In COLT, pages 588?606, 2015.
[8] Martin Ester, Hans-Peter Kriegel, J?rg Sander, and Xiaowei Xu. A density-based algorithm for
discovering clusters in large spatial databases with noise. In Kdd, volume 96, pages 226?231,
1996.
[9] John A Hartigan. Consistency of single linkage for high-density clusters. Journal of the
American Statistical Association, 76(374):388?394, 1981.
[10] Heinrich Jiang. Density level set estimation on manifolds with dbscan. In International
Conference on Machine Learning, pages 1684?1693, 2017.
[11] Heinrich Jiang. Uniform convergence rates for kernel density estimation. In International
Conference on Machine Learning, pages 1694?1703, 2017.
[12] Heinrich Jiang and Samory Kpotufe. Modal-set estimation with an application to clustering. In
International Conference on Artificial Intelligence and Statistics, pages 1197?1206, 2017.
[13] Samory Kpotufe and Ulrike V Luxburg. Pruning nearest neighbor cluster trees. In Proceedings
of the 28th International Conference on Machine Learning (ICML-11), pages 225?232, 2011.
[14] Bharath Sriperumbudur and Ingo Steinwart. Consistency and rates for clustering with dbscan.
In Artificial Intelligence and Statistics, pages 1090?1098, 2012.
[15] Aleksandr Borisovich Tsybakov. Recursive estimation of the mode of a multivariate distribution.
Problemy Peredachi Informatsii, 26(1):38?45, 1990.
[16] Andrea Vedaldi and Stefano Soatto. Quick shift and kernel methods for mode seeking. In
European Conference on Computer Vision, pages 705?718. Springer, 2008.
[17] Daren Wang, Xinyang Lu, and Alessandro Rinaldo. Optimal rates for cluster tree estimation
using kernel density estimators. arXiv preprint arXiv:1706.03113, 2017.
10
| 6610 |@word mild:7 version:2 stronger:2 rsl:2 bf:4 r:21 pick:1 contains:3 interestingly:2 xinyang:1 current:3 com:1 gmail:1 must:1 john:1 partition:1 kdd:1 designed:1 update:1 v:1 intelligence:4 discovering:1 epanechnikov:1 provides:1 characterization:1 simpler:1 constructed:1 become:3 prove:1 manner:1 x0:72 amphitheatre:1 andrea:1 p1:2 behavior:2 chi:1 increasing:1 provided:4 estimating:5 underlying:1 moreover:6 mountain:1 guarantee:23 formalizes:1 every:4 golden:1 exactly:2 ensured:1 rm:6 control:1 yn:2 positive:1 local:8 consequence:1 dbscan:4 aleksandr:1 analyzing:1 jiang:6 path:10 merge:8 twice:1 directed:8 unique:7 yj:4 practice:2 recursive:1 definite:2 lf:7 silverman:1 procedure:22 fbh:36 intersect:1 area:4 empirical:5 attain:4 vedaldi:1 word:3 get:3 cannot:4 close:1 valley:2 quick:47 reviewer:1 straightforward:1 attention:1 starting:4 formalized:1 recovery:4 assigns:1 wasserman:1 contradiction:3 estimator:7 utilizing:1 hd:6 meer:1 notion:2 annals:1 hierarchy:1 suppose:18 exact:1 distributional:1 database:1 role:1 preprint:1 wang:3 region:8 ensures:3 connected:14 highest:2 removed:1 valuable:1 alessandro:1 heinrich:5 traversal:1 weakly:1 multimodal:2 joint:2 intersects:3 separated:8 distinct:1 artificial:2 neighborhood:7 outside:1 choosing:1 whose:2 quite:2 larger:1 solve:1 say:2 distortion:5 otherwise:2 lder:4 triangular:1 ability:1 statistic:3 final:1 differentiable:1 maximal:1 combining:1 subgraph:1 chaudhuri:3 convergence:6 cluster:41 regularity:4 empty:1 produce:1 converges:1 depending:5 nearest:1 wished:1 received:1 progress:1 p2:2 strong:3 involves:1 implies:2 direction:1 radius:10 merged:2 correct:1 modifying:1 larry:1 require:4 assign:1 suffices:2 generalization:2 anonymous:1 ryan:1 hold:15 proximity:4 around:4 sufficiently:22 exp:1 early:1 a2:6 dorin:1 fh:3 estimation:20 grouped:1 gaussian:1 aim:1 modified:1 rather:2 unassigned:1 validated:1 contrast:1 attains:3 problemy:1 nn:1 a0:9 borrowing:1 spurious:1 interested:2 colt:1 spatial:1 initialize:2 construct:1 having:1 beach:1 icml:1 genovese:1 thin:1 primarily:1 belkin:1 simultaneously:2 lebesgue:1 nonincreasing:1 edge:9 respective:2 tree:29 desired:5 ax0:2 assignment:5 vertex:6 uniform:10 supx:9 varies:1 considerably:1 st:1 density:47 international:4 minimality:2 together:2 connecting:1 choose:9 ester:1 american:1 return:4 inc:1 satisfy:2 view:1 root:4 try:1 analyze:5 sup:7 red:1 ulrike:1 recover:5 ery:1 defer:1 yen:1 formed:2 correspond:3 identify:1 climbing:2 lu:1 history:1 bharath:1 strongest:1 inform:1 yusu:1 definition:12 sriperumbudur:2 intentionally:1 naturally:1 proof:14 recovers:5 popular:5 segmentation:9 higher:5 modal:13 formulation:1 furthermore:2 until:4 hand:1 steinwart:2 sketch:1 replacing:1 christopher:1 maximizer:8 google:1 incrementally:2 continuity:3 mode:74 perhaps:1 grows:1 usa:1 contain:2 true:3 hausdorff:2 hence:2 assigned:7 soatto:1 spherically:1 symmetric:1 comaniciu:1 cosine:1 yizong:1 hill:2 performs:1 stefano:1 image:2 volume:2 extend:1 interpretation:1 approximates:1 belong:1 association:1 smoothness:3 tuning:3 rd:9 consistency:16 bruno:1 han:1 longer:1 practioners:1 add:2 closest:2 multivariate:1 recent:1 showed:4 inf:8 optimizes:1 certain:2 inequality:1 arbitrarily:1 yi:6 seen:2 minimum:1 greater:1 additional:1 mr:1 argminxj:1 converge:1 x02:3 borisovich:1 ii:3 exceeds:1 match:2 long:4 a1:6 controlled:2 variant:1 regression:14 vision:2 metric:3 arxiv:2 kernel:8 extra:2 unlike:1 ascent:2 call:1 practitioner:1 near:2 iii:1 enough:1 sander:1 xj:2 fit:1 bandwidth:4 shift:59 t0:2 linkage:3 peter:2 returned:4 hessian:2 remark:5 repeatedly:1 deep:2 useful:1 nonparametric:3 tsybakov:2 extensively:1 exist:2 restricts:1 dotted:2 disjoint:1 tibshirani:1 dasgupta:5 vol:4 group:1 key:1 salient:1 threshold:1 drawn:1 hartigan:3 utilize:1 graph:5 excludes:1 luxburg:2 separation:4 appendix:3 bound:5 correspondence:1 cheng:1 quadratic:1 informatsii:1 x2:12 flat:3 nearby:2 min:2 martin:1 pelletier:1 ball:5 y0:14 modification:1 castro:2 restricted:1 taken:3 remains:1 fail:2 letting:1 end:5 apply:1 hierarchical:2 away:1 appropriate:2 bijectively:1 alternative:5 shortly:1 denotes:1 clustering:12 include:1 cf:2 top:1 remaining:1 log2:2 especially:1 classical:2 seeking:5 move:2 sweep:1 already:1 traditional:1 gradient:3 minx:1 mx:7 distance:1 link:2 thank:1 separating:1 manifold:1 toward:1 length:2 modeled:1 illustration:2 setup:1 difficult:1 kde:13 negative:3 unknown:1 kpotufe:5 ingo:1 finite:3 supporting:1 precise:2 head:2 y1:2 discovered:1 rn:2 arbitrary:2 introduced:2 david:1 pair:4 required:2 established:3 nip:1 justin:1 beyond:1 proceeds:1 dynamical:1 pattern:2 kriegel:1 max:3 critical:1 difficulty:2 minimax:5 acknowledgement:1 asymptotic:1 proven:1 h2:1 x01:3 consistent:4 surprisingly:1 last:3 allow:2 wide:2 neighbor:1 taking:7 mikhail:1 peredachi:1 boundary:4 dimension:1 xn:5 dip:1 feedback:1 stuck:1 collection:2 made:1 transaction:2 pruning:3 compact:1 dealing:1 nhd:6 parkway:1 assumed:1 xi:6 don:1 continuous:3 robust:3 ca:2 argmaxx:6 du:1 european:1 eldridge:2 main:1 noise:2 contradicting:1 x1:21 xu:1 samory:3 wish:1 exponential:2 lie:1 theorem:14 down:2 removing:1 mason:1 decay:3 exists:31 restricting:1 adding:1 false:1 kamalika:1 aria:2 margin:1 chen:1 rg:1 cx:3 logarithmic:2 rinaldo:1 contained:3 springer:1 xiaowei:1 satisfies:1 dh:4 conditional:4 change:1 infinite:1 uniformly:5 lemma:18 called:1 sanjoy:2 support:1 latter:1 |
6,203 | 6,611 | Breaking the Nonsmooth Barrier: A Scalable Parallel
Method for Composite Optimization
Fabian Pedregosa
INRIA/ENS?
Paris, France
R?emi Leblond
INRIA/ENS?
Paris, France
Simon Lacoste-Julien
MILA and DIRO
Universit?e de Montr?eal, Canada
Abstract
Due to their simplicity and excellent performance, parallel asynchronous variants
of stochastic gradient descent have become popular methods to solve a wide range
of large-scale optimization problems on multi-core architectures. Yet, despite their
practical success, support for nonsmooth objectives is still lacking, making them
unsuitable for many problems of interest in machine learning, such as the Lasso,
group Lasso or empirical risk minimization with convex constraints. In this work,
we propose and analyze P ROX A SAGA, a fully asynchronous sparse method inspired by S AGA, a variance reduced incremental gradient algorithm. The proposed
method is easy to implement and signi?cantly outperforms the state of the art on
several nonsmooth, large-scale problems. We prove that our method achieves a
theoretical linear speedup with respect to the sequential version under assumptions on the sparsity of gradients and block-separability of the proximal term.
Empirical benchmarks on a multi-core architecture illustrate practical speedups of
up to 12x on a 20-core machine.
1
Introduction
The widespread availability of multi-core computers motivates the development of parallel methods
adapted for these architectures. One of the most popular approaches is H OGWILD (Niu et al., 2011),
an asynchronous variant of stochastic gradient descent (S GD). In this algorithm, multiple threads run
the update rule of S GD asynchronously in parallel. As S GD, it only requires visiting a small batch
of random examples per iteration, which makes it ideally suited for large scale machine learning
problems. Due to its simplicity and excellent performance, this parallelization approach has recently
been extended to other variants of S GD with better convergence properties, such as S VRG (Johnson
& Zhang, 2013) and S AGA (Defazio et al., 2014).
Despite their practical success, existing parallel asynchronous variants of S GD are limited to smooth
objectives, making them inapplicable to many problems in machine learning and signal processing.
In this work, we develop a sparse variant of the S AGA algorithm and consider its parallel asynchronous variants for general composite optimization problems of the form:
arg min f (x) + h(x)
x?Rp
,
with f (x) :=
1
n
?n
i=1
fi (x)
,
(OPT)
where each fi is convex with L-Lipschitz gradient, the average function f is ?-strongly convex and
h is convex but potentially nonsmooth. We further assume that h is ?simple? in the sense that we
have access to its proximal operator,
? and that it is block-separable, that is, it can be decomposed
block coordinate-wise as h(x) = B?B hB ([x]B ), where B is a partition of the coef?cients into
?
?
DI Ecole
normale sup?erieure, CNRS, PSL Research University
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
subsets which will call blocks and hB only depends on coordinates in block B. Note that there is
no loss of generality in this last assumption as a unique block covering all coordinates is a valid
partition, though in this case, our sparse variant of the S AGA algorithm reduces to the original S AGA
algorithm and no gain from sparsity is obtained.
This template models a broad range of problems arising in machine learning and signal processing:
the ?nite-sum structure of f includes the least squares or logistic loss functions; the proximal term
h includes penalties such as the ?1 or group lasso penalty. Furthermore, this term can be extendedvalued, thus allowing for convex constraints through the indicator function.
Contributions. This work presents two main contributions. First, in ?2 we describe Sparse Proximal S AGA, a novel variant of the S AGA algorithm which features a reduced cost per iteration in the
presence of sparse gradients and a block-separable penalty. Like other variance reduced methods, it
enjoys a linear convergence rate under strong convexity. Second, in ?3 we present P ROX A SAGA, a
lock-free asynchronous parallel version of the aforementioned algorithm that does not require consistent reads. Our main results states that P ROX A SAGA obtains (under assumptions) a theoretical
linear speedup with respect to its sequential version. Empirical benchmarks reported in ?4 show that
this method dramatically outperforms state-of-the-art alternatives on large sparse datasets, while the
empirical speedup analysis illustrates the practical gains as well as its limitations.
1.1
Related work
Asynchronous coordinate-descent. For composite objective functions of the form (OPT), most of
the existing literature on asynchronous optimization has focused on variants of coordinate descent.
Liu & Wright (2015) proposed an asynchronous variant of (proximal) coordinate descent and proved
a near-linear speedup in the number of cores used, given a suitable step size. This approach has been
recently extended to general block-coordinate schemes by Peng et al. (2016), to greedy coordinatedescent schemes by You et al. (2016) and to non-convex problems by Davis et al. (2016). However,
as illustrated by our experiments, in the large sample regime coordinate descent compares poorly
against incremental gradient methods like S AGA.
Variance reduced incremental gradient and their asynchronous variants. Initially proposed in
the context of smooth optimization by Le Roux et al. (2012), variance reduced incremental gradient
methods have since been extended to minimize composite problems of the form (OPT) (see table
below). Smooth variants of these methods have also recently been extended to the asynchronous setting, where multiple threads run the update rule asynchronously and in parallel. Interestingly, none
of these methods achieve both simultaneously, i.e. asynchronous optimization of composite problems. Since variance reduced incremental gradient methods have shown state of the art performance
in both settings, this generalization is of key practical interest.
Objective
Smooth
Composite
Sequential Algorithm
S VRG (Johnson & Zhang, 2013)
S DCA (Shalev-Shwartz & Zhang, 2013)
S AGA (Defazio et al., 2014)
P ROX S DCA (Shalev-Shwartz et al., 2012)
S AGA (Defazio et al., 2014)
ProxS VRG (Xiao & Zhang, 2014)
Asynchronous Algorithm
S VRG (Reddi et al., 2015)
PASSCODE (Hsieh et al., 2015, S DCA variant)
A SAGA (Leblond et al., 2017, S AGA variant)
This work: P ROX A SAGA
On the dif?culty of a composite extension. Two key issues explain the paucity in the development of asynchronous incremental gradient methods for composite optimization. The ?rst issue
is related to the design of such algorithms. Asynchronous variants of S GD are most competitive
when the updates are sparse and have a small overlap, that is, when each update modi?es a small
and different subset of the coef?cients. This is typically achieved by updating only coef?cients for
which the partial gradient at a given iteration is nonzero,2 but existing schemes such as the lagged
updates technique (Schmidt et al., 2016) are not applicable in the asynchronous setting. The second
2
Although some regularizers are sparsity inducing, large scale datasets are often extremely sparse and leveraging this property is crucial for the ef?ciency of the method.
2
dif?culty is related to the analysis of such algorithms. All convergence proofs crucially use the Lipschitz condition on the gradient to bound the noise terms derived from asynchrony. However, in the
composite case, the gradient mapping term (Beck & Teboulle, 2009), which replaces the gradient
in proximal-gradient methods, does not have a bounded Lipschitz constant. Hence, the traditional
proof technique breaks down in this scenario.
Other approaches. Recently, Meng et al. (2017); Gu et al. (2016) independently proposed a doubly stochastic method to solve the problem at hand. Following Meng et al. (2017) we refer to it
as Async-P ROX S VRCD. This method performs coordinate descent-like updates in which the true
gradient is replaced by its S VRG approximation. It hence features a doubly-stochastic loop: at each
iteration we select a random coordinate and a random sample. Because the selected coordinate block
is uncorrelated with the chosen sample, the algorithm can be orders of magnitude slower than S AGA
in the presence of sparse gradients. Appendix F contains a comparison of these methods.
1.2
De?nitions and notations
By convention, we denote vectors and vector-valued functions in lowercase boldface (e.g. x) and
matrices in uppercase boldface (e.g. D). The proximal operator of a convex lower semicontinuous
function h is de?ned as proxh (x) := arg minz?Rp {h(z) + 12 ?x ? z?2 }. A function f is said to be
L-smooth if it is differentiable and its gradient is L-Lipschitz continuous. A function f is said to be
?-strongly convex if f ? ?2 ? ? ?2 is convex. We use the notation ? := L/? to denote the condition
number for an L-smooth and ?-strongly convex function.3
I p denotes the p-dimensional identity matrix, 1{cond} the characteristic function, which is?
1 if cond
n
evaluates to true and 0 otherwise. The average of a vector or matrix is denoted ? := n1 i=1 ?i .
We use ? ? ? for the Euclidean norm. For a positive semi-de?nite matrix D, we de?ne its associated
distance as ?x?2D := ?x, Dx?. We denote by [ x ]b the b-th coordinate in x. This notation is
overloaded so that for a collection of blocks T = {B1 , B2 , . . .}, [x]T denotes the vector x restricted
to the coordinates in the blocks of T . For convenience, when T consists of a single block B we use
[x]B as a shortcut of [x]{B} . Finally, we distinguish E, the full expectation taken with respect to
all the randomness in the system, from E, the conditional expectation of a random it (the random
feature sampled at each iteration by S GD-like algorithms) conditioned on all the ?past?, which the
context will clarify.
2
Sparse Proximal SAGA
Original S AGA algorithm. The original S AGA algorithm (Defazio et al., 2014) maintains two
moving quantities: the current iterate x and a table (memory) of historical gradients (?i )ni=1 . At
every iteration, it samples an index i ? {1, . . . , n} uniformly at random, and computes the next
iterate (x+ , ?+ ) according to the following recursion:
?
?
ui = ?fi (x) ? ?i + ? ; x+ = prox?h x ? ?ui ; ?+
(1)
i = ?fi (x) .
On each iteration, this update rule requires to visit all coef?cients even if the partial gradients ?fi are
sparse. Sparse partial gradients arise in a variety of practical scenarios: for example, in generalized
linear models the partial gradients inherit the sparsity pattern of the dataset. Given that large-scale
datasets are often sparse,4 leveraging this sparsity is crucial for the success of the optimizer.
Sparse Proximal S AGA algorithm. We will now describe an algorithm that leverages sparsity
in the partial gradients by only updating those blocks that intersect with the support of the partial
gradients. Since in this update scheme some blocks might appear more frequently than others, we
will need to counterbalance this undersirable effect with a well-chosen block-wise reweighting of
the average gradient and the proximal term.
In order to make precise this block-wise reweighting, we de?ne the following quantities. We denote
by Ti the extended support of ?fi , which is the set of blocks that intersect the support of ?fi ,
3
Since we have assumed that each individual fi is L-smooth, f itself is L-smooth ? but it could have a
smaller smoothness constant. Our rates are in terms of this bigger L/?, as is standard in the S AGA literature.
4
For example, in the LibSVM datasets suite, 8 out of the 11 datasets (as of May 2017) with more than a
million samples have a density between 10?4 and 10?6 .
3
formally de?ned as Ti := {B : supp(?fi ) ? B ?= ?, B ? B}. For totally separable penalties such
as the ?1 norm, the blocks are individual coordinates and so?
the extended support covers the same
coordinates as the support. Let dB := n/nB , where nB := i 1{B ? Ti } is the number of times
that B ? Ti . For simplicity we assume nB > 0, as otherwise the problem can be reformulated
without block B. The update rule in (1) requires computing the proximal operator of h, which
involves a full pass ?
on the coordinates. In our proposed algorithm, we replace h in (1) with the
function ?i (x) :=
B?Ti dB hB (x), whose form is justi?ed by the following three properties.
First, this function is zero outside Ti , allowing for sparse updates. Second, because of the block-wise
reweighting dB , the function ?i is an unbiased estimator of h (i.e., E ?i = h), property which will
be crucial to prove the convergence of the method. Third, ?i inherits the block-wise structure of h
and its proximal operator can be computed from that of h as [prox??i (x)]B = [prox(dB ?)hB (x)]B
if B ? Ti and [prox??i (x)]B = [x]B otherwise. Following Leblond et al. (2017), we will also
replace the dense gradient estimate ui by the sparse estimate v i := ?fi (x) ? ?i + D i ?, where
D i is the diagonal matrix de?ned block-wise as [D i ]B,B = dB 1{B ? Ti }I |B| . It is easy to verify
that the vector D i ? is a weighted projection onto the support of Ti and E D i ? = ?, making v i an
unbiased estimate of the gradient.
We now have all necessary elements to describe the Sparse Proximal S AGA algorithm. As the
original S AGA algorithm, it maintains two moving quantities: the current iterate x ? Rp and a
table of historical gradients (?i )ni=1 , ?i ? Rp . At each iteration, the algorithm samples an index
i ? {1, . . . , n} and computes the next iterate (x+ , ?+ ) as:
?
?
v i = ?fi (x) ? ?i + D i ? ; x+ = prox??i x ? ?v i ; ?+
i = ?fi (x) ,
(SPS)
where in a practical implementation the vector ? is updated incrementally at each iteration.
The above algorithm is sparse in the sense that it only requires to visit and update blocks in the
extended support: if B ?
/ Ti , by the sparsity of v i and prox?i , we have [x+ ]B = [x]B . Hence,
when the extended support Ti is sparse, this algorithm can be orders of magnitude faster than the
naive S AGA algorithm. The extended support is sparse for example when the partial gradients are
sparse and the penalty is separable, as is the case of the ?1 norm or the indicator function over a
hypercube, or when the the penalty is block-separable in a way such that only a small subset of the
blocks overlap with the support of the partial gradients. Initialization of variables and a reduced
storage scheme for the memory are discussed in the implementation details section of Appendix E.
Relationship with existing methods. This algorithm can be seen as a generalization of both the
Standard S AGA algorithm and the Sparse S AGA algorithm of Leblond et al. (2017). When the
proximal term is not block-separable, then dB = 1 (for a unique block B) and the algorithm defaults
to the Standard (dense) S AGA algorithm. In the smooth case (i.e., h = 0), the algorithm defaults to
the Sparse S AGA method. Hence we note that the sparse gradient estimate v i in our algorithm is the
same as the one proposed in Leblond et al. (2017). However, we emphasize that a straightforward
combination of this sparse update rule with the proximal update from the Standard S AGA algorithm
results in a nonconvergent algorithm: the block-wise reweighting of h is a surprisingly simple but
crucial change. We now give the convergence guarantees for this algorithm.
a
for any a ? 1 and f be ?-strongly convex (? > 0). Then Sparse Proximal
Theorem 1. Let ? = 5L
S AGA converges geometrically in expectation with a rate factor of at least ? = 15 min{ n1 , a ?1 }. That
is, for xt obtained after t updates, we have the following bound:
?n
E?xt ? x? ?2 ? (1 ? ?)t C0 , with C0 := ?x0 ? x? ?2 + 5L1 2 i=1 ??0i ? ?fi (x? )?2 .
Remark. For the step size ? = 1/5L, the convergence rate is (1 ? 1/5 min{1/n, 1/?}). We can thus
identify two regimes: the ?big data? regime, n ? ?, in which the rate factor is bounded by 1/5n, and
the ?ill-conditioned? regime, ? ? n, in which the rate factor is bounded by 1/5?. This rate roughly
matches the rate obtained by Defazio et al. (2014). While the step size bound of 1/5L is slightly
smaller than the 1/3L one obtained in that work, this can be explained by their stronger assumptions:
each fi is strongly convex whereas they are strongly convex only on average in this work. All proofs
for this section can be found in Appendix B.
4
Algorithm 1 P ROX A SAGA (analyzed)
1: Initialize shared variables x and (?i )n
i=1
2: keep doing in parallel
? = inconsistent read of x
x
3:
? = inconsistent read of ?
4:
?
5:
Sample i uniformly in {1, ..., n}
6:
Si := support of ?fi
7:
Ti := extended
?nsupport of ?fi in B
? j ] Ti
8:
[ ? ]Ti = 1/n j=1 [ ?
? i ] Si
9:
[ ?? ]Si = [?fi (?
x)]Si ? [?
? ]Ti = [ ?? ]Ti + [D i ? ]Ti
10:
[v
? )]Ti ? [?
11:
[ ?x ]Ti = [prox??i (?
x ? ?v
x]Ti
12:
for B in Ti do
13:
for b ? B do
14:
[ x ]b ? [ x ]b + [ ?x ]b
? atomic
15:
if b ? Si then
16:
[?i ]b ? [?fi (?
x)]b
17:
end if
18:
end for
19:
end for
20:
// (??? denotes shared memory update.)
21: end parallel loop
3
Algorithm 2 P ROX A SAGA (implemented)
1: Initialize shared variables x, (?i )n
i=1 , ?
2: keep doing in parallel
3:
Sample i uniformly in {1, ..., n}
4:
Si := support of ?fi
5:
Ti := extended support of ?fi in B
? ]Ti = inconsistent read of x on Ti
6:
[x
? i = inconsistent read of ?i
7:
?
8:
[ ? ]Ti = inconsistent read of ? on Ti
? i ] Si
9:
[ ?? ]Si = [?fi (?
x)]Si ? [?
? ]Ti = [?? ]Ti + [ D i ? ]Ti
10:
[v
? )]Ti ? [?
11:
[ ?x ]Ti = [prox??i (?
x ? ?v
x]Ti
12:
for B in Ti do
13:
for b in B do
14:
[ x ]b ? [ x ]b + [ ?x ]b
? atomic
15:
if b ? Si then
16:
[ ? ]b ? [?]b + 1/n[??]b ? atomic
17:
end if
18:
end for
19:
end for
20:
?i ? ?fi (?
x) (scalar update) ? atomic
21: end parallel loop
Asynchronous Sparse Proximal SAGA
We introduce P ROX A SAGA ? the asynchronous parallel variant of Sparse Proximal S AGA. In this
algorithm, multiple cores update a central parameter vector using the Sparse Proximal S AGA introduced in the previous section, and updates are performed asynchronously. The algorithm parameters
are read and written without vector locks, i.e., the vector content of the shared memory can potentially change while a core is reading or writing to main memory coordinate by coordinate. These
operations are typically called inconsistent (at the vector level).
The full algorithm is described in Algorithm 1 for its theoretical version (on which our analysis
is built) and in Algorithm 2 for its practical implementation. The practical implementation differs
from the analyzed agorithm in three points. First, in the implemented algorithm, index i is sampled
before reading the coef?cients to minimize memory access since only the extended support needs to
be read. Second, since our implementation targets generalized linear models, the memory ?i can be
compressed into a single scalar in L20 (see Appendix E). Third, ? is stored in memory and updated
incrementally instead of recomputed at each iteration.
The rest of the section is structured as follows: we start by describing our framework of analysis; we
then derive essential properties of P ROX A SAGA along with a classical delay assumption. Finally,
we state our main convergence and speedup result.
3.1
Analysis framework
As in most of the recent asynchronous optimization literature, we build on the hardware model introduced by Niu et al. (2011), with multiple cores reading and writing to a shared memory parameter
? t , the local copy of the
vector. These operations are asynchronous (lock-free) and inconsistent:5 x
parameters of a given core, does not necessarily correspond to a consistent iterate in memory.
?Perturbed? iterates. To handle this additional dif?culty, contrary to most contributions in this
?eld, we choose the ?perturbed iterate framework? proposed by Mania et al. (2017) and re?ned
by Leblond et al. (2017). This framework can analyze variants of S GD which obey the update rule:
xt+1 = xt ? ?v(xt , it ) , where v veri?es the unbiasedness condition E v(x, it ) = ?f (x)
5
This is an extension of the framework of Niu et al. (2011), where consistent updates were assumed.
5
and the expectation is computed with respect to it . In the asynchronous parallel setting, cores are
? t . As these inconsistent iterates are
reading inconsistent iterates from memory, which we denote x
affected by various delays induced by asynchrony, they cannot easily be written as a function of
their previous iterates. To alleviate this issue, Mania et al. (2017) choose to introduce an additional
quantity for the purpose of the analysis:
xt+1 := xt ? ?v(?
xt , it ) , the ?virtual iterate? ? which is never actually computed .
(2)
Note that this equation is the de?nition of this new quantity xt . This virtual iterate is useful for the
convergence analysis and makes for much easier proofs than in the related literature.
?After read? labeling. How we choose to de?ne the iteration counter t to label an iterate xt
matters in the analysis. In this paper, we follow the ?after read? labeling proposed in Leblond
et al. (2017), in which we update our iterate counter, t, as each core ?nishes reading its copy of
? t and ?
? t ). This means
the parameters (in the speci?c case of P ROX A SAGA, this includes both x
? t is the (t + 1)th fully completed read. One key advantage of this approach compared to the
that x
classical choice of Niu et al. (2011) ? where t is increasing after each successful update ? is that
? t are independent. This
it guarantees both that the it are uniformly distributed and that it and x
property is not veri?ed when using the ?after write? labeling of Niu et al. (2011), although it is still
implicitly assumed in the papers using this approach, see Leblond et al. (2017, Section 3.2) for a
discussion of issues related to the different labeling schemes.
Generalization to composite optimization. Although the perturbed iterate framework was designed for gradient-based updates, we can extend it to proximal methods by remarking that in the
sequential setting, proximal stochastic gradient descent and its variants can be characterized by the
following similar update rule:
?
?
xt+1 = xt ? ?g(xt , v it , it ) , with g(x, v, i) := ?1 x ? prox??i (x ? ?v) ,
(3)
where as before v veri?es the unbiasedness condition E v = ?f (x). The Proximal Sparse S AGA
iteration can be easily written within this template by using ?i and v i as de?ned in ?2. Using this
de?nition of g, we can de?ne P ROX A SAGA virtual iterates as:
? tit , it ) , with v
? tit = ?fit (?
? tit + D it ?t ,
xt+1 := xt ? ?g(?
xt , v
xt ) ? ?
(4)
? tit = ?fit (?
where as in the sequential case, the memory terms are updated as ?
xt ). Our theoretical
analysis of P ROX A SAGA will be based on this de?nition of the virtual iterate xt+1 .
3.2
Properties and assumptions
Now that we have introduced the ?after read? labeling for proximal methods in Eq. (4), we can
leverage the framework of Leblond et al. (2017, Section 3.3) to derive essential properties for the
analysis of P ROX A SAGA. We describe below three useful properties arising from the de?nition
of Algorithm 1, and then state a central (but standard) assumption that the delays induced by the
asynchrony are uniformly bounded.
? t for all r ? t. We
Independence: Due to the ?after read? global ordering, ir is independent of x
enforce the independence for r = t by having the cores read all the shared parameters before their
iterations.
? tit is an unbiased estimator of the gradient of f at x
? t . This property is a
Unbiasedness: The term v
? t.
consequence of the independence between it and x
Atomicity: The shared parameter coordinate update of [x]b on Line 14 is atomic. This means that
there are no overwrites for a single coordinate even if several cores compete for the same resources.
Most modern processors have support for atomic operations with minimal overhead.
Bounded overlap assumption. We assume that there exists a uniform bound, ? , on the maximum
number of overlapping iterations. This means that every coordinate update from iteration t is successfully written to memory before iteration t + ? + 1 starts. Our result will give us conditions on ?
to obtain linear speedups.
? t ? xt . The delay assumption of the previous paragraph allows to express the difference
Bounding x
? uiu , iu ) as:
between real and virtual iterate using the gradient mapping g u := g(?
xu , v
?
t
t
? t ?xt = ? t?1
x
u=(t?? )+ Gu g u , where Gu are p ? p diagonal matrices with terms in {0, +1}. (5)
6
? u and xu have received the corresponding updates. +1, on
0 represents instances where both x
? u has not yet received an update that is already in xu by
the contrary, represents instances where x
de?nition. This bound will prove essential to our analysis.
3.3
Analysis
In this section, we state our convergence and speedup results for P ROX A SAGA. The full details
of the analysis can be found in Appendix C. Following Niu et al. (2011), we introduce a sparsity
measure (generalized to the composite setting) that will appear in our results.
De?nition 1. Let ? := maxB?B |{i : Ti ? B}|/n. This is the normalized maximum number of
times that a block appears in the extended support. For example, if a block is present in all Ti , then
? = 1. If no two Ti share the same block, then ? = 1/n. We always have 1/n ? ? ? 1.
1
. For any step size
Theorem 2 (Convergence guarantee of P ROX A SAGA). Suppose ? ? 10?
?
a
1
6?
?
? = L with a ? a (? ) := 36 min{1, ? }, the inconsistent read iterates
of
Algorithm
1 converge
?1 1?
1
in expectation at a geometric rate factor of at least: ?(a) = 5 min n , a ? , i.e. E??
xt ? x ? ?2 ?
n?
t ?
?
(1 ? ?) C0 , where C0 is a constant independent of t (? a C0 with C0 as de?ned in Theorem 1).
This last result is similar to the original S AGA convergence result and our own Theorem 1, with both
an extra condition on ? and on?the maximum allowable step size. In the best sparsity case, ? = 1/n
and we get the condition ? ? n/10. We now compare the geometric rate above to the one of Sparse
Proximal S AGA to derive the necessary conditions under which P ROX A SAGA is linearly faster.
1
. If ? ? n, then using the step size ? = 1/36L, P ROX A S Corollary 1 (Speedup). Suppose ? ? 10?
?
1
AGA converges geometrically with rate factor ?( ?
). If ? < n, then using the step size ? = 1/36n?,
P ROX A SAGA converges geometrically with rate factor ?( n1 ). In both cases, the convergence rate
is the same as Sparse Proximal S AGA. Thus P ROX A SAGA is linearly faster than its sequential
counterpart up to a constant factor. Note that in both cases the step size does not depend on ? .
Furthermore, if ? ? 6?, we can use a universal step size of ?(1/L) to get a similar rate for P ROX A SAGA than Sparse Proximal S AGA, thus making it adaptive to local strong convexity since the
knowledge of ? is not required.
These speedup regimes are comparable with the best ones obtained in the smooth case, including Niu
et al. (2011); Reddi et al. (2015), even though unlike these papers, we support inconsistent reads
and nonsmooth objective functions. The one exception is Leblond et al. (2017), where the authors
prove that their algorithm, A SAGA, can obtain a linear speedup even without sparsity in the wellconditioned regime. In contrast, P ROX A SAGA always requires some sparsity. Whether this property
for smooth objective functions could be extended to the composite case remains an open problem.
Relative to A SY S PCD, in the best case scenario (where the components of the gradient are uncorre?
lated, a somewhat unrealistic setting), A SY S PCD can get a near-linear speedup for ? as big as 4 p.
?
1
1
Our result states that ? = O( / ?) is necessary for a linear speedup. This means in case ? ? /?p
our bound is better than the one obtained for A SY?
S PCD. Recalling that 1/n ? ? ? 1, it appears
that P ROX A SAGA is favored when n is bigger than p whereas A SY S PCD may have a better bound
otherwise, though this comparison should be taken with a grain of salt given the assumptions we
had to make to arrive at comparable quantities. An extended comparison with the related work can
be found in Appendix D.
4
Experiments
In this section, we compare P ROX A SAGA with related methods on different datasets. Although
P ROX A SAGA can be applied more broadly, we focus on ?1 + ?2 -regularized logistic regression, a
model of particular practical importance. The objective function takes the form
n
?
?
1?
log 1 + exp(?bi a?i x) +
n i=1
?1
2
2 ?x?2
+ ?2 ?x?1
,
(6)
where ai ? Rp and bi ? {?1, +1} are the data samples. Following Defazio et al. (2014), we set
?1 = 1/n. The amount of ?1 regularization (?2 ) is selected to give an approximate 1/10 nonzero
7
Table 1: Description of datasets.
Dataset
???????????????????????
KDD 2010 (Yu et al., 2010)
KDD 2012 (Juan et al., 2016)
Criteo (Juan et al., 2016)
p
19,264,097
149,639,105
45,840,617
1,163,024
54,686,452
1,000,000
?????????????
?????????????
???
??
?
??
?
??
?
??
??
?
??
??
??
?????????????????
??
??
?
??
?
??
?
??
???
??
?
??
????????????
?????????????
?
?
?
? ?? ?? ?? ?? ?? ??
???????????????
?????
??
??
??
??
??
??
?
?
?
?
??
??
?????????????????
??????????????????
????????????????????
??
??
??
??
??
??
?
?
?
?
density
n
??
????????????????
??????????????????
?
?
? ?? ?? ?? ?? ?? ??
???????????????
?????????
???????
28.12
1.25
1.25
10
2 ? 10?7
4 ? 10?5
??
?
??
?
??
?
??
??
?
0.15
0.85
0.89
??????????????
???
?????????????
?
L
?6
?
??
??
??
?????????????????
??
??????????????
????????????????
??
??
??
??
??
??
?
?
?
?
??????????????
?
?
?
? ?? ?? ?? ?? ?? ??
???????????????
?????
Figure 1: Convergence for asynchronous stochastic methods for ?1 + ?2 -regularized logistic
regression. Top: Suboptimality as a function of time for different asynchronous methods using 1
and 10 cores. Bottom: Running time speedup as function of the number of cores. P ROX A SAGA
achieves signi?cant speedups over its sequential version while being orders of magnitude faster than
competing methods. A SY S PCD achieves the highest speedups but it also the slowest overall method.
coef?cients. Implementation details are available in Appendix E. We chose the 3 datasets described
in Table 1
Results. We compare three parallel asynchronous methods on the aforementioned datasets: P ROX A SAGA (this work),6 A SY S PCD, the asynchronous proximal coordinate descent method of Liu &
Wright (2015) and the (synchronous) F ISTA algorithm (Beck & Teboulle, 2009), in which the gradient computation is parallelized by splitting the dataset into equal batches. We aim to benchmark
these methods in the most realistic scenario possible; to this end we use the following step size:
1/2L for P ROX A SAGA, 1/Lc for A SY S PCD, where Lc is the coordinate-wise Lipschitz constant
of the gradient, while F ISTA uses backtracking line-search. The results can be seen in Figure 1
(top) with both one (thus sequential) and ten processors. Two main observations can be made from
this ?gure. First, P ROX A SAGA is signi?cantly faster on these problems. Second, its asynchronous
version offers a signi?cant speedup over its sequential counterpart.
In Figure 1 (bottom) we present speedup with respect to the number of cores, where speedup is
computed as the time to achieve a suboptimality of 10?10 with one core divided by the time to
achieve the same suboptimality using several cores. While our theoretical speedups (with respect
to the number of iterations) are almost linear as our theory predicts (see Appendix F), we observe
a different story for our running time speedups. This can be attributed to memory access overhead,
which our model does not take into account. As predicted by our theoretical results, we observe
6
A reference C++/Python implementation of is available at https://github.com/fabianp/ProxASAGA
8
a high correlation between the ? dataset sparsity measure and the empirical speedup: KDD 2010
(? = 0.15) achieves a 11x speedup, while in Criteo (? = 0.89) the speedup is never above 6x.
Note that although competitor methods exhibit similar or sometimes better speedups, they remain
orders of magnitude slower than P ROX A SAGA in running time for large sparse problems. In fact,
our method is between 5x and 80x times faster (in time to reach 10?10 suboptimality) than F ISTA
and between 13x and 290x times faster than A SY S PCD (see Appendix F.3).
5
Conclusion and future work
In this work, we have described P ROX A SAGA, an asynchronous variance reduced algorithm with
support for composite objective functions. This method builds upon a novel sparse variant of the
(proximal) S AGA algorithm that takes advantage of sparsity in the individual gradients. We have
proven that this algorithm is linearly convergent under a condition on the step size and that it is
linearly faster than its sequential counterpart given a bound on the delay. Empirical benchmarks
show that P ROX A SAGA is orders of magnitude faster than existing state-of-the-art methods.
This work can be extended in several ways. First, we have focused on the S AGA method as the basic
iteration loop, but this approach can likely be extended to other proximal incremental schemes such
as S GD or ProxS VRG. Second, as mentioned in ?3.3, it is an open question whether it is possible to
obtain convergence guarantees without any sparsity assumption, as was done for A SAGA.
Acknowledgements
The authors would like to thank our colleagues Damien Garreau, Robert Gower, Thomas Kerdreux, Geoffrey Negiar and Konstantin Mishchenko for their feedback on this manuscript, and JeanBaptiste Alayrac for support managing the computational resources.
This work was partially supported by a Google Research Award. FP acknowledges support from the
?
chaire Economie
des nouvelles donn?ees with the data science joint research initiative with the fonds
AXA pour la recherche.
References
Bauschke, Heinz and Combettes, Patrick L. Convex analysis and monotone operator theory in
Hilbert spaces. Springer, 2011.
Beck, Amir and Teboulle, Marc. Gradient-based algorithms with applications to signal recovery.
Convex Optimization in Signal Processing and Communications, 2009.
Davis, Damek, Edmunds, Brent, and Udell, Madeleine. The sound of APALM clapping: faster
nonsmooth nonconvex optimization with stochastic asynchronous PALM. In Advances in Neural
Information Processing Systems 29, 2016.
Defazio, Aaron, Bach, Francis, and Lacoste-Julien, Simon. SAGA: A fast incremental gradient
method with support for non-strongly convex composite objectives. In Advances in Neural Information Processing Systems, 2014.
Gu, Bin, Huo, Zhouyuan, and Huang, Heng. Asynchronous stochastic block coordinate descent
with variance reduction. arXiv preprint arXiv:1610.09447v3, 2016.
Hsieh, Cho-Jui, Yu, Hsiang-Fu, and Dhillon, Inderjit S. PASSCoDe: parallel asynchronous stochastic dual coordinate descent. In ICML, 2015.
Johnson, Rie and Zhang, Tong. Accelerating stochastic gradient descent using predictive variance
reduction. In Advances in Neural Information Processing Systems, 2013.
Juan, Yuchin, Zhuang, Yong, Chin, Wei-Sheng, and Lin, Chih-Jen. Field-aware factorization machines for CTR prediction. In Proceedings of the 10th ACM Conference on Recommender Systems. ACM, 2016.
9
Le Roux, Nicolas, Schmidt, Mark, and Bach, Francis R. A stochastic gradient method with an exponential convergence rate for ?nite training sets. In Advances in Neural Information Processing
Systems, 2012.
Leblond, R?emi, Pedregosa, Fabian, and Lacoste-Julien, Simon. ASAGA: asynchronous parallel
SAGA. Proceedings of the 20th International Conference on Arti?cial Intelligence and Statistics
(AISTATS 2017), 2017.
Liu, Ji and Wright, Stephen J. Asynchronous stochastic coordinate descent: Parallelism and convergence properties. SIAM Journal on Optimization, 2015.
Mania, Horia, Pan, Xinghao, Papailiopoulos, Dimitris, Recht, Benjamin, Ramchandran, Kannan,
and Jordan, Michael I. Perturbed iterate analysis for asynchronous stochastic optimization. SIAM
Journal on Optimization, 2017.
Meng, Qi, Chen, Wei, Yu, Jingcheng, Wang, Taifeng, Ma, Zhi-Ming, and Liu, Tie-Yan. Asynchronous stochastic proximal optimization algorithms with variance reduction. In AAAI, 2017.
Nesterov, Yurii. Introductory lectures on convex optimization. Springer Science & Business Media,
2004.
Nesterov, Yurii. Gradient methods for minimizing composite functions. Mathematical Programming, 2013.
Niu, Feng, Recht, Benjamin, Re, Christopher, and Wright, Stephen. Hogwild: A lock-free approach
to parallelizing stochastic gradient descent. In Advances in Neural Information Processing Systems, 2011.
Peng, Zhimin, Xu, Yangyang, Yan, Ming, and Yin, Wotao. ARock: an algorithmic framework for
asynchronous parallel coordinate updates. SIAM Journal on Scienti?c Computing, 2016.
Reddi, Sashank J, Hefny, Ahmed, Sra, Suvrit, Poczos, Barnabas, and Smola, Alexander J. On
variance reduction in stochastic gradient descent and its asynchronous variants. In Advances in
Neural Information Processing Systems, 2015.
Schmidt, Mark, Le Roux, Nicolas, and Bach, Francis. Minimizing ?nite sums with the stochastic
average gradient. Mathematical Programming, 2016.
Shalev-Shwartz, Shai and Zhang, Tong. Stochastic dual coordinate ascent methods for regularized
loss minimization. Journal of Machine Learning Research, 2013.
Shalev-Shwartz, Shai et al.
arXiv:1211.2717, 2012.
Proximal stochastic dual coordinate ascent.
arXiv preprint
Xiao, Lin and Zhang, Tong. A proximal stochastic gradient method with progressive variance reduction. SIAM Journal on Optimization, 2014.
You, Yang, Lian, Xiangru, Liu, Ji, Yu, Hsiang-Fu, Dhillon, Inderjit S, Demmel, James, and Hsieh,
Cho-Jui. Asynchronous parallel greedy coordinate descent. In Advances In Neural Information
Processing Systems, 2016.
Yu, Hsiang-Fu, Lo, Hung-Yi, Hsieh, Hsun-Ping, Lou, Jing-Kai, McKenzie, Todd G, Chou, JungWei, Chung, Po-Han, Ho, Chia-Hua, Chang, Chun-Fu, Wei, Yin-Hsuan, et al. Feature engineering
and classi?er ensemble for KDD cup 2010. In KDD Cup, 2010.
Zhao, Tuo, Yu, Mo, Wang, Yiming, Arora, Raman, and Liu, Han. Accelerated mini-batch randomized block coordinate descent method. In Advances in neural information processing systems,
2014.
10
| 6611 |@word version:6 norm:3 stronger:1 c0:6 open:2 semicontinuous:1 crucially:1 hsieh:4 arti:1 eld:1 reduction:5 liu:6 contains:1 ecole:1 interestingly:1 outperforms:2 existing:5 past:1 current:2 com:1 leblond:11 si:10 yet:2 dx:1 written:4 grain:1 realistic:1 partition:2 kdd:5 cant:2 designed:1 update:29 greedy:2 selected:2 intelligence:1 amir:1 huo:1 core:18 recherche:1 gure:1 iterates:6 zhang:7 mathematical:2 along:1 become:1 initiative:1 prove:4 doubly:2 consists:1 overhead:2 introductory:1 paragraph:1 introduce:3 x0:1 peng:2 roughly:1 pour:1 frequently:1 multi:3 heinz:1 chaire:1 inspired:1 ming:2 decomposed:1 zhi:1 l20:1 totally:1 increasing:1 bounded:5 notation:3 medium:1 suite:1 guarantee:4 cial:1 every:2 ti:36 tie:1 universit:1 lated:1 appear:2 positive:1 before:4 engineering:1 local:2 todd:1 consequence:1 despite:2 meng:3 niu:8 inria:2 might:1 chose:1 initialization:1 dif:3 limited:1 factorization:1 range:2 bi:2 practical:10 unique:2 atomic:6 block:33 implement:1 differs:1 nite:4 intersect:2 empirical:6 universal:1 yan:2 composite:15 projection:1 jui:2 get:3 convenience:1 onto:1 cannot:1 operator:5 nb:3 risk:1 context:2 storage:1 writing:2 straightforward:1 independently:1 convex:17 focused:2 simplicity:3 roux:3 splitting:1 recovery:1 hsuan:1 passcode:2 rule:7 estimator:2 ogwild:1 handle:1 coordinate:31 updated:3 papailiopoulos:1 target:1 suppose:2 programming:2 us:1 element:1 updating:2 nitions:1 predicts:1 bottom:2 preprint:2 wang:2 ordering:1 counter:2 highest:1 mentioned:1 benjamin:2 convexity:2 ui:3 ideally:1 nesterov:2 barnabas:1 depend:1 tit:5 predictive:1 inapplicable:1 upon:1 gu:4 easily:2 joint:1 po:1 various:1 fast:1 describe:4 demmel:1 labeling:5 shalev:4 outside:1 whose:1 kai:1 solve:2 valued:1 otherwise:4 compressed:1 statistic:1 atomicity:1 economie:1 itself:1 asynchronously:3 advantage:2 differentiable:1 propose:1 cients:6 loop:4 culty:3 poorly:1 achieve:3 description:1 inducing:1 rst:1 convergence:16 jing:1 incremental:8 converges:3 yiming:1 illustrate:1 develop:1 derive:3 damien:1 received:2 eq:1 strong:2 implemented:2 predicted:1 signi:4 involves:1 convention:1 stochastic:20 virtual:5 bin:1 require:1 generalization:3 alleviate:1 opt:3 extension:2 clarify:1 wright:4 aga:35 exp:1 mapping:2 algorithmic:1 mo:1 achieves:4 optimizer:1 purpose:1 applicable:1 label:1 apalm:1 successfully:1 weighted:1 minimization:2 always:2 aim:1 normale:1 uiu:1 edmunds:1 corollary:1 derived:1 inherits:1 focus:1 slowest:1 contrast:1 criteo:2 chou:1 sense:2 damek:1 cnrs:1 lowercase:1 typically:2 initially:1 france:2 iu:1 arg:2 aforementioned:2 issue:4 ill:1 denoted:1 favored:1 overall:1 development:2 dual:3 art:4 initialize:2 equal:1 field:1 never:2 having:1 beach:1 aware:1 represents:2 broad:1 yu:6 icml:1 progressive:1 future:1 nonsmooth:6 others:1 modern:1 modi:1 simultaneously:1 individual:3 beck:3 replaced:1 n1:3 recalling:1 montr:1 interest:2 analyzed:2 uppercase:1 scienti:1 regularizers:1 fu:4 partial:8 necessary:3 euclidean:1 re:2 zhimin:1 theoretical:6 minimal:1 instance:2 eal:1 konstantin:1 teboulle:3 cover:1 cost:1 subset:3 uniform:1 delay:5 successful:1 johnson:3 bauschke:1 reported:1 stored:1 perturbed:4 proximal:32 gd:9 unbiasedness:3 recht:2 st:1 density:2 cho:2 international:1 siam:4 randomized:1 cantly:2 michael:1 ctr:1 central:2 aaai:1 choose:3 huang:1 juan:3 rox:32 brent:1 chung:1 sps:1 zhao:1 supp:1 account:1 prox:9 de:19 b2:1 availability:1 includes:3 matter:1 depends:1 performed:1 break:1 hogwild:1 analyze:2 sup:1 doing:2 competitive:1 start:2 maintains:2 parallel:19 francis:3 shai:2 simon:3 contribution:3 minimize:2 square:1 ir:1 ni:2 variance:11 characteristic:1 sy:8 correspond:1 identify:1 ensemble:1 none:1 randomness:1 processor:2 explain:1 ping:1 reach:1 coef:6 ed:2 against:1 evaluates:1 competitor:1 colleague:1 james:1 proof:4 di:1 associated:1 attributed:1 gain:2 sampled:2 proved:1 dataset:4 popular:2 knowledge:1 hilbert:1 hefny:1 actually:1 dca:3 proxh:1 appears:2 manuscript:1 follow:1 wei:3 rie:1 done:1 though:3 strongly:7 generality:1 furthermore:2 smola:1 correlation:1 hand:1 sheng:1 christopher:1 reweighting:4 overlapping:1 incrementally:2 widespread:1 google:1 logistic:3 asynchrony:3 arock:1 usa:1 effect:1 verify:1 true:2 unbiased:3 normalized:1 counterpart:3 hence:4 regularization:1 read:16 nonzero:2 dhillon:2 illustrated:1 covering:1 davis:2 suboptimality:4 generalized:3 allowable:1 chin:1 performs:1 l1:1 wise:8 novel:2 recently:4 fi:22 ef:1 ji:2 salt:1 million:1 discussed:1 extend:1 refer:1 cup:2 ai:1 smoothness:1 erieure:1 had:1 moving:2 access:3 han:2 mania:3 patrick:1 own:1 recent:1 scenario:4 axa:1 nonconvex:1 suvrit:1 success:3 yi:1 nition:6 seen:2 additional:2 somewhat:1 speci:1 parallelized:1 managing:1 converge:1 v3:1 signal:4 semi:1 stephen:2 multiple:4 full:4 sound:1 reduces:1 smooth:11 faster:10 match:1 characterized:1 vrg:6 chia:1 long:1 offer:1 divided:1 bach:3 lin:2 ahmed:1 visit:2 bigger:2 award:1 qi:1 prediction:1 scalable:1 variant:20 regression:2 basic:1 expectation:5 arxiv:4 iteration:18 sometimes:1 achieved:1 whereas:2 crucial:4 parallelization:1 rest:1 extra:1 veri:3 unlike:1 ascent:2 induced:2 db:6 contrary:2 leveraging:2 inconsistent:11 alayrac:1 call:1 reddi:3 ee:1 near:2 yang:1 presence:2 leverage:2 jordan:1 easy:2 maxb:1 hb:4 iterate:14 variety:1 fit:2 independence:3 architecture:3 lasso:3 competing:1 nishes:1 psl:1 synchronous:1 thread:2 whether:2 defazio:7 accelerating:1 penalty:6 sashank:1 reformulated:1 poczos:1 remark:1 dramatically:1 useful:2 amount:1 ten:1 hardware:1 reduced:8 http:1 async:1 arising:2 per:2 broadly:1 write:1 affected:1 express:1 group:2 key:3 recomputed:1 coordinatedescent:1 libsvm:1 lacoste:3 geometrically:3 monotone:1 sum:2 run:2 compete:1 you:2 arrive:1 almost:1 chih:1 raman:1 appendix:9 comparable:2 bound:8 distinguish:1 convergent:1 replaces:1 adapted:1 constraint:2 pcd:8 yangyang:1 yong:1 emi:2 min:5 extremely:1 separable:6 ned:6 speedup:25 structured:1 palm:1 according:1 combination:1 smaller:2 slightly:1 remain:1 separability:1 pan:1 making:4 nouvelles:1 explained:1 restricted:1 taken:2 equation:1 resource:2 remains:1 xiangru:1 describing:1 madeleine:1 overwrites:1 end:9 yurii:2 available:2 operation:3 clapping:1 xinghao:1 obey:1 observe:2 enforce:1 mckenzie:1 batch:3 alternative:1 schmidt:3 slower:2 rp:5 ho:1 original:5 thomas:1 denotes:3 top:2 running:3 completed:1 lock:4 unsuitable:1 paucity:1 gower:1 build:2 hypercube:1 classical:2 feng:1 objective:9 already:1 quantity:6 question:1 traditional:1 diagonal:2 visiting:1 said:2 gradient:49 exhibit:1 distance:1 thank:1 lou:1 boldface:2 kannan:1 index:3 relationship:1 mini:1 minimizing:2 robert:1 potentially:2 lagged:1 design:1 implementation:7 motivates:1 allowing:2 recommender:1 wotao:1 observation:1 datasets:9 benchmark:4 fabian:2 descent:17 extended:17 communication:1 precise:1 parallelizing:1 canada:1 tuo:1 overloaded:1 introduced:3 paris:2 required:1 nip:1 below:2 pattern:1 remarking:1 parallelism:1 dimitris:1 regime:6 sparsity:14 reading:5 fp:1 built:1 including:1 memory:14 unrealistic:1 suitable:1 overlap:3 business:1 regularized:3 indicator:2 recursion:1 counterbalance:1 scheme:7 github:1 zhuang:1 ne:4 julien:3 arora:1 acknowledges:1 naive:1 literature:4 geometric:2 python:1 acknowledgement:1 relative:1 lacking:1 fully:2 loss:3 lecture:1 limitation:1 proven:1 geoffrey:1 consistent:3 xiao:2 story:1 heng:1 uncorrelated:1 share:1 proxs:2 lo:1 surprisingly:1 last:2 free:3 asynchronous:37 copy:2 enjoys:1 supported:1 wide:1 template:2 barrier:1 sparse:35 distributed:1 feedback:1 default:2 valid:1 computes:2 author:2 collection:1 adaptive:1 made:1 historical:2 approximate:1 obtains:1 emphasize:1 implicitly:1 keep:2 global:1 b1:1 assumed:3 shwartz:4 continuous:1 search:1 table:5 donn:1 ca:1 sra:1 nicolas:2 taifeng:1 garreau:1 excellent:2 necessarily:1 marc:1 inherit:1 aistats:1 main:5 dense:2 linearly:4 asaga:1 big:2 noise:1 arise:1 bounding:1 mishchenko:1 ista:3 xu:4 mila:1 en:2 hsiang:3 tong:3 lc:2 combettes:1 saga:36 ciency:1 exponential:1 breaking:1 minz:1 justi:1 third:2 down:1 theorem:4 xt:22 nonconvergent:1 udell:1 jen:1 er:1 chun:1 essential:3 exists:1 sequential:10 importance:1 magnitude:5 ramchandran:1 illustrates:1 conditioned:2 fonds:1 chen:1 easier:1 suited:1 yin:2 backtracking:1 likely:1 partially:1 scalar:2 inderjit:2 chang:1 hua:1 springer:2 acm:2 ma:1 conditional:1 identity:1 lipschitz:5 replace:2 shortcut:1 change:2 shared:7 content:1 uniformly:5 classi:1 called:1 pas:1 e:3 la:1 cond:2 pedregosa:2 select:1 formally:1 exception:1 aaron:1 support:22 mark:2 alexander:1 accelerated:1 lian:1 hung:1 |
6,204 | 6,612 | Dual-Agent GANs for Photorealistic and Identity
Preserving Profile Face Synthesis
Jian Zhao1,2?? Lin Xiong3 Karlekar Jayashree3 Jianshu Li1 Fang Zhao1
Zhecan Wang4? Sugiri Pranata3 Shengmei Shen3
Shuicheng Yan1,5 Jiashi Feng1
1
3
National University of Singapore
2
National University of Defense Technology
Panasonic R&D Center Singapore 4 Franklin. W. Olin College of Engineering
5
Qihoo 360 AI Institute
{zhaojian90, jianshu}@u.nus.edu {lin.xiong, karlekar.jayashree, sugiri.pranata, shengmei.shen}@sg.panasonic.com
[email protected] {elezhf, eleyans, elefjia}@u.nus.edu
Abstract
Synthesizing realistic profile faces is promising for more efficiently training deep
pose-invariant models for large-scale unconstrained face recognition, by populating samples with extreme poses and avoiding tedious annotations. However,
learning from synthetic faces may not achieve the desired performance due to the
discrepancy between distributions of the synthetic and real face images. To narrow
this gap, we propose a Dual-Agent Generative Adversarial Network (DA-GAN)
model, which can improve the realism of a face simulator?s output using unlabeled
real faces, while preserving the identity information during the realism refinement. The dual agents are specifically designed for distinguishing real v.s. fake
and identities simultaneously. In particular, we employ an off-the-shelf 3D face
model as a simulator to generate profile face images with varying poses. DA-GAN
leverages a fully convolutional network as the generator to generate high-resolution
images and an auto-encoder as the discriminator with the dual agents. Besides
the novel architecture, we make several key modifications to the standard GAN
to preserve pose and texture, preserve identity and stabilize training process: (i)
a pose perception loss; (ii) an identity perception loss; (iii) an adversarial loss
with a boundary equilibrium regularization term. Experimental results show that
DA-GAN not only presents compelling perceptual results but also significantly
outperforms state-of-the-arts on the large-scale and challenging NIST IJB-A unconstrained face recognition benchmark. In addition, the proposed DA-GAN is
also promising as a new approach for solving generic transfer learning problems
more effectively. DA-GAN is the foundation of our submissions to NIST IJB-A
2017 face recognition competitions, where we won the 1st places on the tracks of
verification and identification.
1
Introduction
Unconstrained face recognition is a very important yet extremely challenging problem. In recent years,
deep learning techniques have significantly advanced large-scale unconstrained face recognition (8;
19; 27; 34; 29; 16), arguably driven by rapidly increasing resource of face images. However, labeling
huge amount of data for feeding supervised deep learning algorithms is undoubtedly expensive
and time-consuming. Moreover, as often observed in real-world scenarios, the pose distribution of
available face recognition datasets (e.g., IJB-A (15)) is usually unbalanced and has long-tail with
?
?
Homepage: https://zhaoj9014.github.io/
Jian Zhao and Zhecan Wang were interns at Panasonic R&D Center Singapore during this work.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
0.07
0.06
0.05
0.05
Frequency
Frequency
0.07
0.06
0.04
0.03
0.02
0.04
0.03
0.02
0.01
0.01
0
0
-100
-80
-60
-40
-20
0
20
40
60
80
-100
100
Pose
-80
-60
-40
-20
0
20
40
60
80
100
Pose
(a) Extremely unbalanced pose distribution.
(b) Well balanced pose distribution with DA-GAN.
Figure 1: Comparison of pose distribution in the IJB-A dataset (15) w/o and w/ DA-GAN.
large pose variations, as shown in Figure. 1a. This has become a main obstacle for further pushing
unconstrained face recognition performance. To address this critical issue, several research attempts
(32; 31; 35) have been made to employ synthetic profile face images as augmented extra data to
balance the pose variations.
However, naively learning from synthetic images can be problematic due to the distribution discrepancy between synthetic and real face images?synthetic data is often not realistic enough with
artifacts and severe texture losses. The low-quality synthesis face images would mislead the learned
face recognition model to overfit to fake information only presented in synthetic images and fail to
generalize well on real faces. Brute-forcedly increasing the realism of the simulator is often expensive
in terms of time cost and manpower, if possible.
In this work, we propose a novel Dual-Agent Generative Adversarial Network (DA-GAN) for profile
view synthesis, where the dual agents focus on discriminating the realism of synthetic profile face
images from a simulator using unlabled real data and perceiving the identity information, respectively.
In other words, the generator needs to play against a real?fake discriminator as well as an identity
discriminator simultaneously to generate high-quality faces that are really useful for unconstrained
face recognition.
In our method, a synthetic profile face image with a pre-specified pose is generated by a 3D morphable
face simulator. DA-GAN takes this synthetic face image as input and refines it through a conditioned
generative model. We leverage a Fully Convolutional Network (FCN) (17) that operates on the pixel
level as the generator to generate high-resolution face images and an auto-encoder network as the
discriminator. Different from vanilla GANs, DA-GAN introduces an auxiliary discriminative agent
to enforce the generator to preserve identity information of the generated faces, which is critical for
face recognition application. In addition, DA-GAN also imposes a pose perception loss to preserve
pose and texture. The refined synthetic profile face images present photorealistic quality with well
preserved identity information, which are used as augmented data together with real face images for
pose-invariant feature learning. For stabilizing the training process of such dual-agent GAN model,
we impose a boundary equilibrium regularization term.
Experimental results show that DA-GAN not only presents compelling perceptual results but also
significantly outperforms state-of-the-arts on the large-scale and challenging National Institute
of Standards and Technology (NIST) IARPA Janus Benchmark A (IJB-A) unconstrained face
recognition benchmark (15). DA-GAN leads us to further win the 1st places on verification and
identification tracks in the NIST IJB-A 2017 face recognition competitions. This strong evidence
shows that our ?recognition via generation" framework is effective and generic, and we expect that it
benefits for more face recognition and transfer learning applications in the real world.
Our contributions are summarized as follows.
? We propose a novel Dual-Agent Generative Adversarial Network (DA-GAN) for photorealistic and identity preserving profile face synthesis even under extreme poses.
? The proposed dual-agent architecture effectively combines prior knowledge from data distribution (adversarial training) and domain knowledge of faces (pose and identity perception
losses) to exactly recover the lost information inherent in projecting a 3D face into the 2D
image space.
? We present qualitative and quantitative experiments showing the possibility of a ?recognition
via generation" framework and achieve the top performance on the challenging NIST IJB-A
unconstrained face recognition benchmark (15) without extra human annotation efforts
2
3D Face Model
Simulator
?
Face RoI
Extraction
68-Point
Landmark Detection
Simulated Profile Face
Generator
+
Input
224?224?3
Conv 64?7?7
ReLU & BN
?
Lpp
+
Residual Block * 10
Conv 3?1?1
Output
224?224?3
Discriminator
Lip
Agent 1
?
Real
?
Synthetic
-
Ladv
Agent 2
Conv 3?3?3
ReLU
FC
784
Transition Down
Transition Up
Conv 3?1?1
ReLU
Figure 2: Overview of the proposed DA-GAN architecture. The simulator (upper panel) extracts
face RoI, localizes landmark points and produces synthesis faces with arbitrary poses, which are fed
to DA-GAN for realism refinement. DA-GAN uses a fully convolutional skip-net as the generator
(middle panel) and an auto-encoder as the discriminator (bottom panel). The dual agents focus on
both discriminating real v.s. fake (minimizing the loss Ladv ) and preserving identity information
(minimizing the loss Lip ). Best viewed in color.
by training deep neural networks on the refined face images together with real images.
To our best knowledge, our proposed DA-GAN is the first model that is effective for
automatically generating augmented data for face recognition in challenging conditions and
indeed improves performance. DA-GAN won the 1st places on verification and identification
tracks in the NIST IJB-A 2017 face recognition competitions.
2
Related works
As one of the most significant advancements on the research of deep generative models (14; 26),
GAN has drawn substantial attention from the deep learning and computer vision community since
it was first introduce by Goodfellow et al. (10). The GAN framework learns a generator network
and a discriminator network with competing loss. This min-max two-player game provides a simple
yet powerful way to estimate target distribution and to generate novel image samples. Mirza and
Osindero (21) introduce the conditional version of GAN, to condition on both the generator and
discriminator for effective image tagging. Berthelot et al. (2) propose a new Boundary Equilibrium
GAN (BE-GAN) framework paired with a loss derived from the Wasserstein distance for training
GAN, which derives a way of controlling the trade-off between image diversity and visual quality.
These successful applications of GAN motivate us to develop profile view synthesis methods based
on GAN. However, the generator of previous methods usually focus on generating images based on a
random noise vector or conditioned data and the discriminator only has a single agent to distinguish
real v.s. fake. Thus, in contrast to our method, the generated images do not have any discriminative
information that can be used for training a deep learning based recognition model. This separates us
well with previous GAN-based attempts.
3
Moreover, differnet from previous InfoGAN (5) which does not have the classification agent, and
Auxiliary Classifier GAN (AC-GAN) (22) which only performs classification, our propsoed DAGAN performs face verification with an intrigued data augmentation. DA-GAN is a novel and
practical model for efficient data augmentation and it is really effective in practice as proved in Sec.
4. DA-GAN generates the data in a completely different way from InfoGAN (5) and AC-GAN (22)
which generate images from a random noise input or abstract semantic labels. Therefore, inferior to
our model, those existing GAN-like models cannot exploit useful and rich prior information (e.g., the
shape, pose of faces) for effective data generation and augmentation. They cannot fully control the
generated images. In contrast, DA-GAN can fully control the generated images and adjust the face
pose (e.g., yaw angles) distribution which is extremely unbalanced in real-world scenarios. DA-GAN
can facilitate training more accurate face analysis models to solve the large pose variation problem
and other relevant problems in unconstrained face recognition.
Our proposed DA-GAN shares a similar idea with TP-GAN (13) that considers face synthesis based
on GAN framework, and Apple GAN (28) that considers learning from simulated and unsupervised
images through adversarial training. Our method differs from them in following aspects: 1) DA-GAN
aims to synthesize photorealistic and identity preserving profile faces to address the large variance
issue in unconstrained face recognition, whereas TP-GAN (13) tries to recover a frontal face from a
profile view and Apple GAN (28) is designed for much simpler scenarios (e.g., eye and hand image
refinement); 2) TP-GAN (13) and Apple GAN (28) suffer from categorical information loss which
limits their effectiveness in promoting recognition performance. In contrast, our proposed DA-GAN
architecture effectively overcomes this issue by introducing dual discriminator agents.
3
3.1
Dual-Agent GAN
Simulator
The main challenge for unconstrained face recognition lies in the large variation and few profile face
images for each subject, which is the main obstacle for learning a well-performed pose-invariant
model. To address this problem, we simulate face images with various pre-defined poses (i.e., yaw
angles), which explicitly augments the available training data without extra human annotation efforts
and balances the pose distribution.
In particular, as shown in Figure. 2, we first extracts the face Region of Interest (RoI) from each
available real face image, and estimate 68 facial landmark points using the Recurrent AttentiveRefinement (RAR) framework (31), which is robust to illumination changes and does not require
a shape model in advance. We then estimate a transformation matrix between the detected 2D
landmarks and the corresponding landmarks in the 3D Morphable Model (3D MM) using leastsquares fit (35). Finally, we simulate profile face images in various poses with pre-defined yaw
angles.
However, the performance of the simulator decreases dramatically under large poses (e.g., yaw angles
? {[?90o , ?60o ] ? [+60o , +90o ]}) due to artifacts and severe texture losses, misleading the network
to overfit to fake information only presented in synthetic images and fail to generalize well on real
data.
3.2
Generator
In order to generate photorealistic and identity preserving profile view face images which are truely
benefical for unconstrained face recognition, we further refine the above-mentioned simulated profile
face images with the proposed DA-GAN.
Inspired by the recent success of FCN-based methods on image-to-image applications (17; 9) and
the leading performance of skip-net on recognition tasks (12; 33), we modify a skip-net (ResNet
(12)) into a FCN-based architecture as the generator G? : RH?W ?C 7? RH?W ?C of DA-GAN to
learn a highly non-linear transformation for profile face image refinement, where ? are the network
parameters for the generator, and H, W , and C denote the image height, width, and channel number,
repectively.
Contextual information from global and local regions compensates each other and naturally benefits
face recognition. The hierarchical features within a skip-net are multi-scale in nature due to the
increasing receptive field sizes, which are combined together via skip connections. Such a combined representation comprehensively maintains the contextual information, which is crucial for
4
artifact removal, fragement stitching, and texture padding. Moreover, the FCN-based architecture is
advantageous for generating high-resolution image-level results. More details are provided in Sec. 4.
More formally, let the simulated profile face image be denoted by x and the refined face image be
denoted by x
?, then
x
? := G? (x).
(1)
The key requirements for DA-GAN are that the refined face image x
? should look like a real face
image in appearance while preserving the intrinsic identity and pose information from the simulator.
To this end, we propose to learn ? by minimizing a combination of three losses:
LG? = (?Ladv + ?1 Lip ) + ?2 Lpp ,
(2)
where Ladv is the adversarial loss for adding realism to the synthetic images and alleviating artifacts,
Lip is the identity perception loss for preserving the identity information, and Lpp is the pose
perception loss for preserving pose and texture information.
Lpp is a pixel-wise L1 loss, which is introduced to enforce the pose (i.e., yaw angle) consistency for
the synthetic profile face images before and after the refinement via DA-GAN:
Lpp =
W X
H
X
1
|xi,j ? x
?i,j |,
W ?H i j
(3)
where i, j traverse all pixels of x and x
?.
Although Lpp may lead some over smooth effects to the refined results, it is still an essential part for
both pose and texture information preserving and accelerated optimization.
To add realism to the synthetic images to really benefit face recognition performance, we need to
narrow the gap between the distributions of synthetic and real images. An ideal generator will make
it impossible to classify a given image as real or refined with high confidence. Meanwhile, preserving
the identity information is the essential and critical part for recognition. An ideal generator will
generate the refined face images that have small intra-class distance and large inter-class distance in
the feature space spanned by the deep neural networks for unconstrained face recognition. These
motivate the use of an adversarial pixel-wise discriminator with dual agents.
3.3
Dual-agent discriminator
To incorporate the prior knowledge from the profile faces? distribution and domain knowledge of
identities? distribution, we herein introduce a discriminator with dual agents for distinguishing real
v.s. fake and identities simultaneously. To facilitate this process, we leverage an auto-encoder as the
discriminator D? : RH?W ?C 7? RH?W ?C to be as simple as possible to avoid typical GAN tricks,
which first projects the input real / fake face image into high-dimensional feature space through
several Convolution (Conv) and Fully Connected (FC) layers of the encoder and then transformed
back to the image-level representation through several Deconvolution (Deconv) and Conv layers of
the decoder, as shown in Figure. 2. ? are the networks parameters for the discriminator. More details
are provided in Sec. 4.
One agent of D? is trained with Ladv to minimize the Wasserstein distance with a boundary equilibrium regularization term for maintaining a balance between the generator and discriminator losses as
first introduced in (2),
Ladv =
X
|yj ? D? (yj )| ? kt
j
X
|?
xi ? D? (?
xi )|,
(4)
i
where y denotes the real face image, kt is a boundary P
equilibrium regularizationP
term using Proportional Control Theory to maintain the equilibrium E[ i |?
xi ? D? (?
xi )|] = ?E[ j |yj ? D? (yj )|],
? is the diversity ratio.
Here kt is updated by
kt+1 = kt + ?(?
X
|yj ? D? (yj )| ?
j
X
|?
xi ? D? (?
xi )|),
(5)
i
where ? is the learning rate (proportional gain) for k. In essence, Eq.(5) can be thought of as a form
of close-loop feedback control in which kt is adjusted at each step.
5
Ladv serves as a supervision to push the refined face image to reside in the manifold of real images.
It can prevent the blurry effect, alleviate artifacts and produce visually pleasing results.
The other agent of D? is trained with Lip to preserve the identity discriminability of the refined face
images. Specially, we define Lip with the multi-class cross-entropy loss based on the output from the
bottleneck layer of D? .
Lip =
1 X
?(Yj log(D? (yj )) + (1 ? Yj )log(1 ? D? (yj )))
N j
+
1 X
?(Yi log(D? (?
xi )) + (1 ? Yi )log(1 ? D? (?
xi ))),
N i
(6)
where Y is the identity ground truth.
Thus, minimizing Lip would encourage deep features of the refined face images belonging to the same
identity to be close to each other. If one visualizes the learned deep features in the high-dimensional
space, the learned deep features of refined face image set form several compact clusters and each
cluster may be far away from others. Each cluster has a small variance. In this way, the refined
face images are enforced with well preserved identity information. We also conduct experiments for
illustration.
Using Lip alone makes the results prone to annoying artifacts, because the search for a local minimum
of Lip may go through a path that resides outside the manifold of natural face images. Thus, we
combine Lip with Ladv as the final objective function for D? to ensure that the search resides in that
manifold and produces photorealistic and identity preserving face image:
LD? = Ladv + ?1 Lip .
3.4
(7)
Loss functions for training
The goal of DA-GAN is to use a set of unlabeled real face images y to learn a generator G? that
adaptively refines a simulated profile face image x. The overall objective function for DA-GAN is:
(
LD? = Ladv + ?1 Lip ,
LG? = (?Ladv + ?1 Lip ) + ?2 Lpp .
(8)
We optimize DA-GAN by alternatively optimizing D? and G? for each training iteration. Similar
as in (2), we measure the convergence of DA-GAN by using the boundary
Pequilibrium concept:
we can frame the convergence process as finding the closest reconstruction j |yj ? D? (yj )| with
thePlowest absolute value
P of the instantaneous process error for the Proportion Control Theory
|? j |yj ? D? (yj )| ? i |?
xi ? D? (?
xi )||. This measurement can be formulated as:
Lcon =
X
|yj ? D? (yj )| + |?
j
X
|yj ? D? (yj )| ?
j
X
|?
xi ? D? (?
xi )||.
(9)
i
Lcon can be used to determine when the network has reached its final state or if the model has
collapsed. Detailed algorithm on the training procedures is provided in supplementary material Sec.
1.
4
4.1
Experiments
Experimental settings
Benchmark dataset: Except for synthesizing natural looking profile view face images, the proposed DA-GAN also aims to generate identity preserving face images for accurate face-centric
analysis with state-of-the-art deep learning models. Therefore, we evaluate the possibility of ?recognition via generation" of DA-GAN on the most challenging unconstrained face recognition benchmark
dataset IJB-A (15). IJB-A (15) contains both images and video frames from 500 subjects with
5,397 images and 2,042 videos that are split into 20,412 frames, 11.4 images and 4.2 videos per
subject, captured from in-the-wild environment to avoid the near frontal bias, along with protocols for
evaluation of both verification (1:1 comparison) and identification (1:N search) tasks. For training
and testing, 10 random splits are provided by each protocol, respectively. More details are provided
in supplementary material Sec. 2.
6
2.00
Convergence
1.75
1.50
1.25
1.00
0.50
0.25
0.05
0
25000
50000
75000
100000
125000
150000
175000
200000
225000
250000
Iterations
Figure 3: Quality of refined results w.r.t. the network convergence measurement Lcon .
Real
60?
70?
80?
90?
Simulated
Refined
Simulated
Refined
Simulated
Refined
(a) Refined results of DA-GAN.
Real Faces
Refined Synthetic Faces with DA-GAN
(b) Feature space of real faces and DA-GAN synthetic faces.
Figure 4: Qualitative analysis of DA-GAN.
Reproducibility: The proposed method is implementated by extending the Keras framework (6).
All networks are trained on three NVIDIA GeForce GTX TITAN X GPUs with 12GB memory for
each. Please refer to supplementary material Sec. 3 & 4 for full details on network architectures and
training procedures.
4.2
Results and discussions
Qualitative results ? DA-GAN: In order to illustrate the compelling perceptual results generated
by the proposed DA-GAN, we first visualize the quality of refined results w.r.t. the network convergence measurement Lcon , as shown in Figure. 3. As can be seen, our DA-GAN ensures a fast yet
stable convergence through the carefully designed optimization scheme and boundary equilibrium
regularization term. The network convergence measurement Lcon correlates well with image fidelity.
Most of the previous works (31; 32; 35) on profile view synthesis are dedicated to address this
problem within a pose range of ?60o . Because it is commonly believed that with a pose that is larger
than 60o , it is difficult for a model to generate faithful profile view images. Similarly, our simulator is
also good at normalizing small posed faces while suffers severe artifacts and texture losses under
large poses (e.g., yaw angles ? {[?90o , ?60o ] ? [+60o , +90o ]}), as shown in Figure. 4a the first
row for each subject. However, with enough training data and proper architecture and objective
function design of the proposed DA-GAN, it is in fact feasible to further refine such synthetic profile
face images under very large poses for high-quality natural looking results generation, as shown
in Figure. 4a the second row for each subject. Compared with the raw simulated faces, the refined
results by DA-GAN present a good photorealistic quality. More visualized samples are provided in
supplementary material Sec. 5.
To verify the superiority of DA-GAN as well as the contribution of each component, we also compare
the qualitative results produced by the vanilla GAN (10), Apple GAN (28), BE-GAN (2) and three
variations of DA-GAN in terms of w/o Ladv , Lip , Lpp in each case, repectively. Please refer to
supplementary material Sec. 5 for details.
7
Table 1: Performance comparison of DA-GAN with state-of-the-arts on IJB-A verification protocol.
For all metrics, a higher number means better performance. The results are averaged over 10 testing
splits. Symbol ?-" implies that the result is not reported for that method. Standard deviation is not
available for some methods. The results offered by our proposed method are highlighted in bold.
Method
OpenBR (15)
GOTS (15)
Pooling faces (11)
LSFS (30)
Deep Multi-pose (1)
DCNNmanual (4)
Triplet Similarity (27)
VGG-Face (23)
PAMs (19)
DCNNf usion (3)
Masi et al. (20)
Triplet Embedding (27)
All-In-One (25)
Template Adaptation (8)
NAN (34)
L2 -softmax (24)
b-1
b-2
DA-GAN (ours)
TAR @
FAR=0.10
0.433 ? 0.006
0.627 ? 0.012
0.631
0.895 ? 0.013
0.911
0.947 ? 0.011
0.945 ? 0.002
0.652 ? 0.037
0.967 ? 0.009
0.964 ? 0.005
0.976 ? 0.004
0.979 ? 0.004
0.978 ? 0.003
0.984 ? 0.002
0.989 ? 0.003
0.978 ? 0.003
0.991 ? 0.003
Face verification
TAR @
FAR=0.01
0.236 ? 0.009
0.406 ? 0.014
0.309
0.733 ? 0.034
0.787
0.787 ? 0.043
0.790 ? 0.030
0.805 ? 0.030
0.826 ? 0.018
0.838 ? 0.042
0.886
0.900 ? 0.010
0.922 ? 0.010
0.939 ? 0.013
0.941 ? 0.008
0.970 ? 0.004
0.963 ? 0.007
0.950 ? 0.009
0.976 ? 0.007
TAR @
FAR=0.001
0.104 ? 0.014
0.198 ? 0.008
0.514 ? 0.060
0.590 ? 0.050
0.725
0.813 ? 0.020
0.823 ? 0.020
0.836 ? 0.027
0.881 ? 0.011
0.943 ? 0.005
0.920 ? 0.006
0.901 ? 0.008
0.930 ? 0.005
Table 2: Performance comparison of DA-GAN with state-of-the-arts on IJB-A identification protocol.
For FNIR metric, a lower number means better performance. For the other metrics, a higher number
means better performance. The results offered by our proposed method are highlighted in bold.
Method
OpenBR (15)
GOTS (15)
B-CNN (7)
LSFS (30)
Pooling faces (11)
Deep Multi-pose (1)
DCNNmanual (4)
Triplet Similarity (27)
VGG-Face (23)
PAMs (19)
DCNNf usion (3)
Masi et al. (20)
Triplet Embedding (27)
Template Adaptation (8)
All-In-One (25)
NAN (34)
L2 -softmax (24)
b-1
b-2
DA-GAN (ours)
FNIR @
FPIR=0.10
0.851 ? 0.028
0.765 ? 0.033
0.659 ? 0.032
0.387 ? 0.032
0.250
0.246 ? 0.014
0.33 ? 0.031
0.210 ? 0.033
0.137 ? 0.014
0.118 ? 0.016
0.113 ? 0.014
0.083 ? 0.009
0.044 ? 0.006
0.068 ? 0.010
0.108 ? 0.008
0.051 ? 0.009
Face identification
FNIR @
Rank1
FPIR=0.01
0.934 ? 0.017 0.246 ? 0.011
0.953 ? 0.024 0.433 ? 0.021
0.857 ? 0.027 0.588 ? 0.020
0.617 ? 0.063 0.820 ? 0.024
0.846
0.480
0.846
0.852 ? 0.018
0.444 ? 0.065 0.880 ? 0.015
0.539 ? 0.077 0.913 ? 0.011
0.840 ? 0.012
0.423 ? 0.094 0.903 ? 0.012
0.906
0.247 ? 0.030 0.932 ? 0.010
0.226 ? 0.049 0.928 ? 0.010
0.208 ? 0.020 0.947 ? 0.008
0.183 ? 0.041 0.958 ? 0.005
0.085 ? 0.041
0.973 ? 0.005
0.125 ? 0.035 0.966 ? 0.006
0.179 ? 0.042 0.960 ? 0.007
0.110 ? 0.039
0.971 ? 0.007
Rank5
0.375 ? 0.008
0.595 ? 0.020
0.796 ? 0.017
0.929 ? 0.013
0.933
0.927
0.937 ? 0.010
0.950 ? 0.007
0.925 ? 0.008
0.965 ? 0.008
0.962
0.977 ? 0.004
0.980 ? 0.005
0.987 ? 0.003
0.982 ? 0.004
0.989 ? 0.003
To gain insights into the effectivenss of identity preserving quality of our DA-GAN, we further use
t-SNE (18) to visualize the deep features of both refined profile faces and real faces in a 2D space in
Figure. 4b. As can be seen, the refined profile face images present small intra-class distance and large
inter-class distance, which is similar to those of real faces. This reveals that DA-GAN ensures well
preserved identity information with the auxiliary agent for Lip .
Quantitative results ? ?recognition via generation": To quantitatively verify the superiority of
?recognition via generation" of DA-GAN, we conduct unconstrained face recognition (i.e., verification
and identification) on IJB-A benchmark dataset (15) with three different settings. In the three settings,
8
the pre-trained deep recognition models are respectively fine-tuned on the original training data of
each split without extra data (baseline 1: b-1), the original training data of each split with extra
synthetic faces by our simulator (baseline 2: b-2), and the original training data of each split with
extra refined faces by our DA-GAN (our method: ?recognition via generation" framework based on
DA-GAN, DA-GAN for short). The performance comparison of DA-GAN with the two baselines
and other state-of-the-arts on IJB-A (15) unconstrained face verification and identification protocols
are given in Table. 1 and Table. 2.
We can observe that even with extra training data, b-2 presents inferior performance than b-1 for all
metrics of both face verification and identification. This demonstrates that naively learning from
synthetic images can be problematic due to a gap between synthetic and real image distributions
? synthetic data is often not realistic enough with artifacts and severe texture losses, misleading
the network to overfit to fake information only presented in synthetic images and fail to generalize
well on real data. In contrast, with the injection of photorealistic and identity preserving faces
generated by DA-GAN without extra human annotation efforts, our method outperforms b-1 by
1.00% for TAR @ FAR=0.001 of verification and 1.50% for FNIR @ FPIR=0.01, 0.50% for Rank1 of identification. Our method achieves comparable performance with L2 -softmax (24), which
employ a much more computational complex recognition model even without fine-tuning or template
adaptation procedures as we do. Moreover, DA-GAN outperforms NAN (34) by 4.90% for TAR @
FAR=0.001 of verification and 7.30% for FNIR @ FPIR=0.01, 1.30% for Rank-1 of identification.
These results won the 1st places on verification and identification tracks in NIST IJB-A 2017 face
recognition competitions3 . This well verified the promissing potential of synthetic face images by
our DA-GAN on the large-scale and challenging unconstrained face recognition problem.
Finally, we visualize the verification and identification closed set results for IJB-A (15) split1 to gain
insights into unconstrained face recognition with the proposed ?recognition via generation" framework
based on DA-GAN. For fully detailed visualization results in high resolution and corresponding
analysis, please refer to supplementary material Sec. 6 & 7.
5
Conclusion
We propose a novel Dual-Agent Generative Adversarial Network (DA-GAN) for photorealistic and
identity preserving profile face synthesis. DA-GAN combines prior knowledge from data distribution
(adversarial training) and domain knowledge of faces (pose and identity perception loss) to exactly
recover the lost information inherent in projecting a 3D face into the 2D image space. DA-GAN can
be optimized in a fast yet stable way with an imposed boundary equilibrium regularization term that
balances the power of the discriminator against the generator. One promissing potential application of
the proposed DA-GAN is for solving generic transfer learning problems more effectively. Qualitative
and quantitative experiments verify the possibility of our ?recognition via generation" framework,
which achieved the top performance on the large-scale and challenging NIST IJB-A unconstrained
face recognition benchmark without extra human annotation efforts. Based on DA-GAN, we won the
1st places on verification and identification tracks in NIST IJB-A 2017 face recognition competitions.
It would be interesting to apply DA-GAN for other transfer learning applications in the future.
Acknowledgement
The work of Jian Zhao was partially supported by China Scholarship Council (CSC) grant
201503170248.
The work of Jiashi Feng was partially supported by National University of Singapore startup grant
R-263-000-C08-133, Ministry of Education of Singapore AcRF Tier One grant R-263-000-C21-112
and NUS IDS grant R-263-000-C67-646.
We would like to thank Junliang Xing (Institute of Automation, Chinese Academy of Sciences),
Hengzhu Liu, Xucan Chen, and Yongping Zhai (National University of Defense Technology) for
helpful discussions.
3
We submitted our results for both verification and identification protocols to NIST IJB-A 2017 face
recognition competition committee on 29th, March, 2017. We received the official notification on our top
performance on both tracks on 26th, Apirl, 2017. The IJB-A benchmark dataset, relevant information and
leaderboard can be found at https://www.nist.gov/programs-projects/face-challenges.
9
References
[1] W. AbdAlmageed, Y. Wu, S. Rawls, S. Harel, T. Hassner, I. Masi, J. Choi, J. Lekust, J. Kim, P. Natarajan,
et al. Face recognition using deep multi-pose representations. In Applications of Computer Vision (WACV),
2016 IEEE Winter Conference on, pages 1?9. IEEE, 2016.
[2] D. Berthelot, T. Schumm, and L. Metz. Began: Boundary equilibrium generative adversarial networks.
arXiv preprint arXiv:1703.10717, 2017.
[3] J.-C. Chen, V. M. Patel, and R. Chellappa. Unconstrained face verification using deep cnn features. In
Applications of Computer Vision (WACV), 2016 IEEE Winter Conference on, pages 1?9. IEEE, 2016.
[4] J.-C. Chen, R. Ranjan, A. Kumar, C.-H. Chen, V. M. Patel, and R. Chellappa. An end-to-end system for
unconstrained face verification with deep convolutional neural networks. In Proceedings of the IEEE
International Conference on Computer Vision Workshops, pages 118?126, 2015.
[5] X. Chen, Y. Duan, R. Houthooft, J. Schulman, I. Sutskever, and P. Abbeel. Infogan: Interpretable
representation learning by information maximizing generative adversarial nets. In Advances in Neural
Information Processing Systems, pages 2172?2180, 2016.
[6] F. Chollet. keras. https://github.com/fchollet/keras, 2015.
[7] A. R. Chowdhury, T.-Y. Lin, S. Maji, and E. Learned-Miller. One-to-many face recognition with bilinear
cnns. In Applications of Computer Vision (WACV), 2016 IEEE Winter Conference on, pages 1?9. IEEE,
2016.
[8] N. Crosswhite, J. Byrne, O. M. Parkhi, C. Stauffer, Q. Cao, and A. Zisserman. Template adaptation for
face verification and identification. arXiv preprint arXiv:1603.03958, 2016.
[9] K. Gong, X. Liang, X. Shen, and L. Lin. Look into person: Self-supervised structure-sensitive learning
and a new benchmark for human parsing. arXiv preprint arXiv:1703.05446, 2017.
[10] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio.
Generative adversarial nets. In Advances in neural information processing systems, pages 2672?2680,
2014.
[11] T. Hassner, I. Masi, J. Kim, J. Choi, S. Harel, P. Natarajan, and G. Medioni. Pooling faces: template based
face recognition with pooled face images. In Proceedings of the IEEE Conference on Computer Vision and
Pattern Recognition Workshops, pages 59?67, 2016.
[12] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. In Proceedings of the
IEEE Conference on Computer Vision and Pattern Recognition, pages 770?778, 2016.
[13] R. Huang, S. Zhang, T. Li, and R. He. Beyond face rotation: Global and local perception gan for
photorealistic and identity preserving frontal view synthesis. arXiv preprint arXiv:1704.04086, 2017.
[14] D. P. Kingma and M. Welling. Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114, 2013.
[15] B. F. Klare, B. Klein, E. Taborsky, A. Blanton, J. Cheney, K. Allen, P. Grother, A. Mah, M. Burge, and
A. K. Jain. Pushing the frontiers of unconstrained face detection and recognition: Iarpa janus benchmark a.
In 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 1931?1939. IEEE,
2015.
[16] J. Li, J. Zhao, F. Zhao, H. Liu, J. Li, S. Shen, J. Feng, and T. Sim. Robust face recognition with deep
multi-view representation learning. In Proceedings of the 2016 ACM on Multimedia Conference, pages
1068?1072. ACM, 2016.
[17] J. Long, E. Shelhamer, and T. Darrell. Fully convolutional networks for semantic segmentation. In
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 3431?3440,
2015.
[18] L. v. d. Maaten and G. Hinton. Visualizing data using t-sne. Journal of Machine Learning Research,
9(Nov):2579?2605, 2008.
[19] I. Masi, S. Rawls, G. Medioni, and P. Natarajan. Pose-aware face recognition in the wild. In Proceedings
of the IEEE Conference on Computer Vision and Pattern Recognition, pages 4838?4846, 2016.
[20] I. Masi, A. T. Tran, J. T. Leksut, T. Hassner, and G. Medioni. Do we really need to collect millions of faces
for effective face recognition? arXiv preprint arXiv:1603.07057, 2016.
[21] M. Mirza and S. Osindero. Conditional generative adversarial nets. arXiv preprint arXiv:1411.1784, 2014.
[22] A. Odena, C. Olah, and J. Shlens. Conditional image synthesis with auxiliary classifier gans. arXiv preprint
arXiv:1610.09585, 2016.
[23] O. M. Parkhi, A. Vedaldi, and A. Zisserman. Deep face recognition.
[24] R. Ranjan, C. D. Castillo, and R. Chellappa. L2-constrained softmax loss for discriminative face verification.
arXiv preprint arXiv:1703.09507, 2017.
[25] R. Ranjan, S. Sankaranarayanan, C. D. Castillo, and R. Chellappa. An all-in-one convolutional neural
network for face analysis. arXiv preprint arXiv:1611.00851, 2016.
[26] D. J. Rezende, S. Mohamed, and D. Wierstra. Stochastic backpropagation and approximate inference in
deep generative models. arXiv preprint arXiv:1401.4082, 2014.
[27] S. Sankaranarayanan, A. Alavi, C. D. Castillo, and R. Chellappa. Triplet probabilistic embedding for
face verification and clustering. In Biometrics Theory, Applications and Systems (BTAS), 2016 IEEE 8th
International Conference on, pages 1?8. IEEE, 2016.
[28] A. Shrivastava, T. Pfister, O. Tuzel, J. Susskind, W. Wang, and R. Webb. Learning from simulated and
unsupervised images through adversarial training. arXiv preprint arXiv:1612.07828, 2016.
[29] Y. Taigman, M. Yang, M. Ranzato, and L. Wolf. Deepface: Closing the gap to human-level performance in
face verification. In Proceedings of the Conference on Computer Vision and Pattern Recognition, pages
1701?1708, 2014.
10
[30] D. Wang, C. Otto, and A. K. Jain. Face search at scale: 80 million gallery. arXiv preprint arXiv:1507.07242,
2015.
[31] S. Xiao, J. Feng, J. Xing, H. Lai, S. Yan, and A. Kassim. Robust facial landmark detection via recurrent
attentive-refinement networks. In European Conference on Computer Vision, pages 57?72. Springer, 2016.
[32] S. Xiao, L. Liu, X. Nie, J. Feng, A. A. Kassim, and S. Yan. A live face swapper. In Proceedings of the
2016 ACM on Multimedia Conference, pages 691?692. ACM, 2016.
[33] S. Xie, R. Girshick, P. Doll?r, Z. Tu, and K. He. Aggregated residual transformations for deep neural
networks. arXiv preprint arXiv:1611.05431, 2016.
[34] J. Yang, P. Ren, D. Chen, F. Wen, H. Li, and G. Hua. Neural aggregation network for video face recognition.
arXiv preprint arXiv:1603.05474, 2016.
[35] X. Zhu, J. Yan, D. Yi, Z. Lei, and S. Z. Li. Discriminative 3d morphable model fitting. In Automatic Face
and Gesture Recognition (FG), 2015 11th IEEE International Conference and Workshops on, volume 1,
pages 1?8. IEEE, 2015.
11
| 6612 |@word cnn:2 version:1 middle:1 advantageous:1 proportion:1 tedious:1 annoying:1 shuicheng:1 bn:1 lpp:8 ld:2 liu:3 contains:1 tuned:1 ours:2 franklin:1 outperforms:4 existing:1 com:2 contextual:2 yet:4 parsing:1 refines:2 realistic:3 csc:1 shape:2 designed:3 interpretable:1 alone:1 generative:11 advancement:1 realism:7 short:1 ijb:20 provides:1 cheney:1 traverse:1 simpler:1 zhang:2 height:1 wierstra:1 along:1 olah:1 become:1 qualitative:5 combine:3 wild:2 fitting:1 introduce:3 inter:2 tagging:1 indeed:1 simulator:12 multi:6 inspired:1 automatically:1 gov:1 duan:1 increasing:3 conv:6 provided:6 project:2 moreover:4 panel:3 homepage:1 finding:1 transformation:3 truely:1 quantitative:3 exactly:2 classifier:2 demonstrates:1 brute:1 control:5 grant:4 superiority:2 arguably:1 before:1 engineering:1 local:3 modify:1 limit:1 io:1 bilinear:1 encoding:1 id:1 path:1 discriminability:1 china:1 collect:1 challenging:8 range:1 averaged:1 c21:1 practical:1 faithful:1 yj:18 testing:2 lost:2 block:1 practice:1 differs:1 backpropagation:1 susskind:1 procedure:3 tuzel:1 lsfs:2 yan:3 significantly:3 thought:1 vedaldi:1 word:1 pre:4 confidence:1 cannot:2 unlabeled:2 close:2 collapsed:1 impossible:1 live:1 optimize:1 www:1 imposed:1 center:2 ranjan:3 maximizing:1 go:1 attention:1 shen:3 mislead:1 resolution:4 stabilizing:1 pouget:1 insight:2 spanned:1 shlens:1 fang:1 embedding:3 variation:5 updated:1 target:1 play:1 controlling:1 alleviating:1 distinguishing:2 us:1 goodfellow:2 trick:1 synthesize:1 recognition:62 expensive:2 natarajan:3 submission:1 observed:1 bottom:1 preprint:15 wang:4 region:2 ensures:2 connected:1 sun:1 ranzato:1 trade:1 decrease:1 balanced:1 substantial:1 mentioned:1 environment:1 nie:1 warde:1 motivate:2 trained:4 solving:2 completely:1 various:2 maji:1 jain:2 fast:2 effective:6 chellappa:5 detected:1 labeling:1 startup:1 outside:1 refined:23 supplementary:6 solve:1 larger:1 posed:1 cvpr:1 encoder:5 compensates:1 otto:1 highlighted:2 final:2 mah:1 net:7 propose:6 reconstruction:1 tran:1 adaptation:4 tu:1 relevant:2 loop:1 repectively:2 rapidly:1 cao:1 reproducibility:1 achieve:2 academy:1 competition:5 sutskever:1 convergence:7 cluster:3 requirement:1 extending:1 darrell:1 produce:3 generating:3 resnet:1 illustrate:1 develop:1 ac:2 gong:1 pose:42 recurrent:2 received:1 sim:1 eq:1 strong:1 auxiliary:4 skip:5 implies:1 cnns:1 stochastic:1 human:6 material:6 rar:1 education:1 require:1 hassner:3 feeding:1 abbeel:1 really:4 alleviate:1 leastsquares:1 adjusted:1 frontier:1 mm:1 ground:1 roi:3 visually:1 equilibrium:9 visualize:3 achieves:1 qihoo:1 label:1 council:1 sensitive:1 aim:2 avoid:2 shelf:1 tar:5 varying:1 derived:1 focus:3 rezende:1 eleyans:1 rank:1 contrast:4 adversarial:15 stauffer:1 baseline:3 kim:2 helpful:1 inference:1 transformed:1 pixel:4 issue:3 dual:16 classification:2 overall:1 denoted:2 fidelity:1 art:6 softmax:4 constrained:1 field:1 aware:1 extraction:1 beach:1 look:2 unsupervised:2 yaw:6 fcn:4 discrepancy:2 future:1 mirza:3 others:1 quantitatively:1 inherent:2 employ:3 few:1 wen:1 winter:3 harel:2 simultaneously:3 national:5 preserve:5 maintain:1 attempt:2 pleasing:1 detection:3 undoubtedly:1 huge:1 interest:1 possibility:3 highly:1 intra:2 chowdhury:1 evaluation:1 severe:4 adjust:1 introduces:1 extreme:2 farley:1 usion:2 accurate:2 kt:6 encourage:1 facial:2 biometrics:1 conduct:2 desired:1 girshick:1 classify:1 compelling:3 obstacle:2 tp:3 cost:1 introducing:1 deviation:1 jiashi:2 successful:1 osindero:2 reported:1 synthetic:26 combined:2 adaptively:1 st:6 person:1 international:3 discriminating:2 probabilistic:1 off:2 synthesis:11 together:3 gans:3 augmentation:3 huang:1 zhao:4 leading:1 li:5 potential:2 diversity:2 student:1 stabilize:1 summarized:1 sec:9 bold:2 titan:1 automation:1 pooled:1 explicitly:1 performed:1 view:9 try:1 closed:1 reached:1 xing:2 recover:3 maintains:1 rank1:2 metz:1 annotation:5 bayes:1 aggregation:1 alavi:1 contribution:2 minimize:1 convolutional:6 variance:2 efficiently:1 miller:1 kassim:2 generalize:3 identification:16 raw:1 populating:1 produced:1 ren:2 apple:4 visualizes:1 submitted:1 suffers:1 notification:1 against:2 attentive:1 frequency:2 geforce:1 mohamed:1 naturally:1 gain:3 photorealistic:10 dataset:5 proved:1 knowledge:7 color:1 improves:1 segmentation:1 carefully:1 back:1 centric:1 masi:6 higher:2 supervised:2 xie:1 zisserman:2 overfit:3 hand:1 acrf:1 quality:9 artifact:8 lei:1 usa:1 facilitate:2 effect:2 concept:1 gtx:1 verify:3 byrne:1 regularization:5 semantic:2 visualizing:1 during:2 game:1 width:1 inferior:2 essence:1 please:3 self:1 won:4 ladv:12 performs:2 l1:1 dedicated:1 allen:1 image:81 wise:2 instantaneous:1 novel:6 variational:1 began:1 rotation:1 overview:1 volume:1 million:2 tail:1 he:3 berthelot:2 significant:1 measurement:4 refer:3 ai:1 tuning:1 unconstrained:22 vanilla:2 consistency:1 similarly:1 automatic:1 closing:1 stable:2 supervision:1 similarity:2 morphable:3 add:1 closest:1 recent:2 optimizing:1 driven:1 scenario:3 nvidia:1 success:1 yi:3 preserving:17 minimum:1 wasserstein:2 captured:1 impose:1 seen:2 ministry:1 determine:1 aggregated:1 ii:1 full:1 smooth:1 panasonic:3 gesture:1 cross:1 long:3 lin:4 believed:1 lai:1 paired:1 vision:12 metric:4 arxiv:30 iteration:2 promissing:2 achieved:1 preserved:3 addition:2 whereas:1 fine:2 jian:3 crucial:1 extra:9 specially:1 subject:5 pooling:3 effectiveness:1 near:1 yang:2 leverage:3 ideal:2 iii:1 enough:3 split:6 kera:3 bengio:1 relu:3 fit:1 li1:1 architecture:8 competing:1 idea:1 vgg:2 bottleneck:1 defense:2 gb:1 padding:1 effort:4 suffer:1 deep:24 dramatically:1 useful:2 fake:9 detailed:2 amount:1 visualized:1 jianshu:2 augments:1 generate:10 http:3 problematic:2 singapore:5 track:6 per:1 klein:1 key:2 drawn:1 prevent:1 verified:1 schumm:1 karlekar:2 chollet:1 year:1 houthooft:1 enforced:1 grother:1 angle:6 taigman:1 powerful:1 place:5 wu:1 maaten:1 comparable:1 layer:3 nan:3 distinguish:1 courville:1 refine:2 unlabled:1 generates:1 aspect:1 simulate:2 extremely:3 min:1 kumar:1 injection:1 gpus:1 combination:1 march:1 belonging:1 modification:1 projecting:2 invariant:3 tier:1 resource:1 visualization:1 rawls:2 fail:3 committee:1 fed:1 stitching:1 end:3 serf:1 available:4 doll:1 promoting:1 observe:1 hierarchical:1 away:1 generic:3 enforce:2 blurry:1 apply:1 xiong:1 original:3 top:3 denotes:1 ensure:1 clustering:1 gan:96 maintaining:1 pushing:2 exploit:1 intrigued:1 scholarship:1 chinese:1 feng:4 objective:3 deconv:1 receptive:1 win:1 distance:6 separate:1 thank:1 simulated:10 landmark:6 decoder:1 manifold:3 considers:2 gallery:1 ozair:1 besides:1 illustration:1 ratio:1 balance:4 minimizing:4 zhai:1 liang:1 lg:2 difficult:1 webb:1 sne:2 zhao1:2 synthesizing:2 design:1 proper:1 upper:1 convolution:1 datasets:1 benchmark:11 nist:11 c08:1 hinton:1 looking:2 frame:3 arbitrary:1 community:1 introduced:2 specified:1 connection:1 discriminator:17 optimized:1 learned:4 narrow:2 herein:1 kingma:1 nu:3 nip:1 address:4 beyond:1 usually:2 perception:8 pattern:6 challenge:2 program:1 max:1 memory:1 video:4 medioni:3 power:1 critical:3 odena:1 natural:3 residual:3 olin:2 zhu:1 advanced:1 localizes:1 scheme:1 improve:1 technology:3 github:2 eye:1 misleading:2 categorical:1 auto:5 extract:2 prior:4 sg:1 l2:4 removal:1 acknowledgement:1 schulman:1 fully:8 loss:24 expect:1 generation:10 interesting:1 proportional:2 wacv:3 generator:17 leaderboard:1 foundation:1 shelhamer:1 agent:24 offered:2 verification:22 imposes:1 xiao:2 feng1:1 share:1 row:2 prone:1 fchollet:1 supported:2 bias:1 institute:3 dagan:1 face:148 comprehensively:1 template:5 absolute:1 fg:1 benefit:3 boundary:9 feedback:1 world:3 transition:2 rich:1 resides:2 reside:1 made:1 refinement:6 commonly:1 far:6 welling:1 correlate:1 janus:2 compact:1 patel:2 nov:1 approximate:1 overcomes:1 global:2 reveals:1 consuming:1 discriminative:4 xi:13 alternatively:1 search:4 triplet:5 table:4 lip:16 promising:2 nature:1 transfer:4 robust:3 ca:1 learn:3 channel:1 shrivastava:1 european:1 complex:1 meanwhile:1 domain:3 da:68 yan1:1 protocol:6 official:1 main:3 rh:4 noise:2 profile:29 iarpa:2 xu:1 augmented:3 shen3:1 burge:1 lie:1 perceptual:3 infogan:3 learns:1 down:1 c67:1 choi:2 showing:1 symbol:1 abadie:1 evidence:1 derives:1 sankaranarayanan:2 naively:2 intrinsic:1 essential:2 deconvolution:1 adding:1 effectively:4 normalizing:1 workshop:3 texture:9 illumination:1 conditioned:2 push:1 gap:4 chen:6 entropy:1 fc:2 appearance:1 intern:1 visual:1 partially:2 hua:1 deepface:1 springer:1 wolf:1 truth:1 acm:4 conditional:3 identity:32 viewed:1 goal:1 formulated:1 feasible:1 change:1 parkhi:2 specifically:1 perceiving:1 operates:1 typical:1 except:1 multimedia:2 castillo:3 pfister:1 experimental:3 player:1 formally:1 college:1 unbalanced:3 frontal:3 accelerated:1 incorporate:1 evaluate:1 avoiding:1 |
6,205 | 6,613 | Dilated Recurrent Neural Networks
Shiyu Chang1?, Yang Zhang1?, Wei Han2?, Mo Yu1 , Xiaoxiao Guo1 , Wei Tan1 ,
Xiaodong Cui1 , Michael Witbrock1 , Mark Hasegawa-Johnson2 , Thomas S. Huang2
1
IBM Thomas J. Watson Research Center, Yorktown, NY 10598, USA
2
University of Illinois at Urbana-Champaign, Urbana, IL 61801, USA
{shiyu.chang, yang.zhang2, xiaoxiao.guo}@ibm.com,
{yum, wtan, cuix, witbrock}@us.ibm.com,
{weihan3, jhasegaw, t-huang1}@illinois.edu
Abstract
Learning with recurrent neural networks (RNNs) on long sequences is a notoriously difficult task. There are three major challenges: 1) complex dependencies, 2)
vanishing and exploding gradients, and 3) efficient parallelization. In this paper,
we introduce a simple yet effective RNN connection structure, the D ILATED RNN,
which simultaneously tackles all of these challenges. The proposed architecture is
characterized by multi-resolution dilated recurrent skip connections, and can be
combined flexibly with diverse RNN cells. Moreover, the D ILATED RNN reduces
the number of parameters needed and enhances training efficiency significantly,
while matching state-of-the-art performance (even with standard RNN cells) in
tasks involving very long-term dependencies. To provide a theory-based quantification of the architecture?s advantages, we introduce a memory capacity measure,
the mean recurrent length, which is more suitable for RNNs with long skip
connections than existing measures. We rigorously prove the advantages of the
D ILATED RNN over other recurrent neural architectures. The code for our method
is publicly available1 .
1
Introduction
Recurrent neural networks (RNNs) have been shown to have remarkable performance on many
sequential learning problems. However, long sequence learning with RNNs remains a challenging
problem for the following reasons: first, memorizing extremely long-term dependencies while
maintaining mid- and short-term memory is difficult; second, training RNNs using back-propagationthrough-time is impeded by vanishing and exploding gradients; And lastly, both forward- and
back-propagation are performed in a sequential manner, which makes the training time-consuming.
Many attempts have been made to overcome these difficulties using specialized neural structures, cells,
and optimization techniques. Long short-term memory (LSTM) [10] and gated recurrent units (GRU)
[6] powerfully model complex data dependencies. Recent attempts have focused on multi-timescale
designs, including clockwork RNNs [12], phased LSTM [17], hierarchical multi-scale RNNs [5], etc.
The problem of vanishing and exploding gradients is mitigated by LSTM and GRU memory gates;
other partial solutions include gradient clipping [18], orthogonal and unitary weight optimization
[2, 14, 24], and skip connections across multiple timestamps [8, 30]. For efficient sequential training,
WaveNet [22] abandoned RNN structures, proposing instead the dilated causal convolutional neural
network (CNN) architecture, which provides significant advantages in working directly with raw
audio waveforms. However, the length of dependencies captured by a dilated CNN is limited by its
kernel size, whereas an RNN?s autoregressive modeling can, in theory, capture potentially infinitely
?
1
Denotes equal contribution.
https://github.com/code-terminator/DilatedRNN
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
*+
*+
*+
*+5
,.
,/
,3
,0
,1
,2
,-
,4
!"
!#
!(
!$
!%
!&
!'
!)
!"
!#
!(
!$
!%
!&
!'
!)
Figure 1: (left) A single-layer RNN with recurrent skip connections. (mid) A single-layer RNN with
dilated recurrent skip connections. (right) A computation structure equivalent to the second graph,
which reduces the sequence length by four times.
long dependencies with a small number of parameters. Recently, Yu et al. [27] proposed learningbased RNNs with the ability to jump (skim input text) after seeing a few timestamps worth of data;
although the authors showed that the modified LSTM with jumping provides up to a six-fold speed
increase, the efficiency gain is mainly in the testing phase.
In this paper, we introduce the D ILATED RNN, a neural connection architecture analogous to the
dilated CNN [22, 28], but under a recurrent setting. Our approach provides a simple yet useful
solution that tries to alleviate all challenges simultaneously. The D ILATED RNN is a multi-layer, and
cell-independent architecture characterized by multi-resolution dilated recurrent skip connections.
The main contributions of this work are as follows. 1) We introduce a new dilated recurrent skip
connection as the key building block of the proposed architecture. These alleviate gradient problems
and extend the range of temporal dependencies like conventional recurrent skip connections, but in
the dilated version require fewer parameters and significantly enhance computational efficiency. 2)
We stack multiple dilated recurrent layers with hierarchical dilations to construct a D ILATED RNN,
which learns temporal dependencies of different scales at different layers. 3) We present the mean
recurrent length as a new neural memory capacity measure that reveals the performance difference
between the previously developed recurrent skip-connections and the dilated version. We also verify
the optimality of the exponentially increasing dilation distribution used in the proposed architecture.
It is worth mentioning that, the recent proposed Dilated LSTM [23] can be viewed as a special case
of our model, which contains only one dilated recurrent layer with fixed dilation. The main purpose
of their model is to reduce the temporal resolution on time-sensitive tasks. Thus, the Dilated LSTM
is not a general solution for modeling at multiple temporal resolutions.
We empirically validate the D ILATED RNN in multiple RNN settings on a variety of sequential
learning tasks, including long-term memorization, pixel-by-pixel classification of handwritten digits
(with permutation and noise), character-level language modeling, and speaker identification with
raw audio waveforms. The D ILATED RNN improves significantly on the performance of a regular
RNN, LSTM, or GRU with far fewer parameters. Many studies [6, 14] have shown that vanilla RNN
cells perform poorly in these learning tasks. However, within the proposed structure, even vanilla
RNN cells outperform more sophisticated designs, and match the state-of-the-art. We believe that the
D ILATED RNN provides a simple and generic approach to learning on very long sequences.
2
Dilated Recurrent Neural Networks
The main ingredients of the D ILATED RNN are its dilated recurrent skip connection and its use of
exponentially increasing dilation; these will be discussed in the following two subsections respectively.
2.1
Dilated Recurrent Skip Connection
Denote c(l)
t as the cell in layer l at time t. The dilated skip connection can be represented as
?
?
(l)
(l) (l)
ct = f xt , ct s(l) .
(1)
?
?
(l)
(l) (l)
(l)
ct = f xt , ct 1 , ct s(l) .
(2)
This is similar to the regular skip connection[8, 30], which can be represented as
s(l) is referred to as the skip length, or dilation of layer l; xt as the input to layer l at time t; and
(l)
f (?) denotes any RNN cell and output operations, e.g. Vanilla RNN cell, LSTM, GRU etc. Both
skip connections allow information to travel along fewer edges. The difference between dilated and
2
Output
Hidden Layer
Dilation = 4
Hidden Layer
Dilation = 2
Hidden Layer
Dilation = 1
Input
Figure 2: (left) An example of a three-layer D ILATED RNN with dilation 1, 2, and 4. (right) An
example of a two-layer D ILATED RNN, with dilation 2 in the first layer. In such a case, extra
embedding connections are required (red arrows) to compensate missing data dependencies.
regular skip connection is that the dependency on c(l)
t 1 is removed in dilated skip connection. The left
and middle graphs in figure 1 illustrate the differences between two architectures with dilation or skip
length s(l) = 4, where Wr0 is removed in the middle graph. This reduces the number of parameters.
More importantly, computational efficiency of a parallel implementation (e.g., using GPUs) can be
greatly improved by parallelizing operations that, in a regular RNN, would be impossible. The middle
and right graphs in figure 1 illustrate the idea with s(l) = 4 as an example. The input subsequences
(l)
(l)
(l)
(l)
(l)
{x4t }, {x4t+1 }, {x4t+2 } and {x4t+3 } are given four different colors. The four cell chains, {c4t },
(l)
(l)
(l)
{c4t+1 }, {c4t+2 } and {c4t+3 }, can be computed in parallel by feeding the four subsequences into a
regular RNN, as shown in the right of figure 1. The output can then be obtained by interweaving
among the four output chains. The degree of parallelization is increased by s(l) times.
2.2
Exponentially Increasing Dilation
To extract complex data dependencies, we stack dilated recurrent layers to construct D ILATED RNN.
Similar to settings that were introduced in WaveNet [22], the dilation increases exponentially across
layers. Denote s(l) as the dilation of the l-th layer. Then,
s(l) = M l
1
(3)
, l = 1, ? ? ? , L.
The left side of figure 2 depicts an example of D ILATED RNN with L = 3 and M = 2. On one
hand, stacking multiple dilated recurrent layers increases the model capacity. On the other hand,
exponentially increasing dilation brings two benefits. First, it makes different layers focus on different
temporal resolutions. Second, it reduces the average length of paths between nodes at different
timestamps, which improves the ability of RNNs to extract long-term dependencies and prevents
vanishing and exploding gradients. A formal proof of this statement will be given in section 3.
To improve overall computational efficiency, a generalization of our standard D ILATED RNN is also
proposed. The dilation in the generalized D ILATED RNN does not start at one, but M l0 . Formally,
s(l) = M (l
1+l0 )
, l = 1, ? ? ? , L and l0
0,
(4)
where
is called the starting dilation. To compensate for the missing dependencies shorter than
M l0 , a 1-by-M (l0 ) convolutional layer is appended as the final layer. The right side of figure 2
illustrates an example of l0 = 1, i.e. dilations start at two. Without the red edges, there would be
no edges connecting nodes at odd and even time stamps. As discussed in section 2.1, the computational efficiency can be increased by M l0 by breaking the input sequence into M l0 downsampled
subsequences, and feeding each into a L l0 -layer regular D ILATED RNN with shared weights.
M0l
3
The Memory Capacity of D ILATED RNN
In this section, we extend the analysis framework in [30] to establish better measures of memory
capacity and parameter efficiency, which will be discussed in the following two sections respectively.
3.1
Memory Capacity
To facilitate theoretical analysis, we apply the cyclic graph Gc notation introduced in [30].
Definition 3.1 (Cyclic Graph). The cyclic graph representation of an RNN structure is a directed
multi-graph, GC = (VC , EC ). Each edge is labeled as e = (u, v, ) 2 EC , where u is the origin
3
node, v is the destination node, and is the number of time steps the edge travels. Each node is
labeled as v = (i, p) 2 VC , where i is the time index of the node modulo m, m is the period of the
graph, and p is the node index. GC must contain at least one directed cycle. Along the edges of any
directed cycle, the summation of must not be zero.
Define di (n) as the length of the shortest path from any input node at time i to any output node at
time i + n. In [30], a measure of the memory capacity is proposed that essentially only looks at
di (m), where m is the period of the graph. This is reasonable when the period is small. However,
when the period is large, the entire distribution of di (n), 8n ? m makes a difference, not just the
one at span m. As a concrete example, suppose there is an RNN architecture of period m = 10, 000,
implemented using equation (2) with skip length s(l) = m, so that di (n) = n for n = 1, ? ? ? , 9, 999
and di (m) = 1. This network rapidly memorizes the dependence on inputs at time i of the outputs
at time i + m = i + 10, 000, but shorter dependencies 2 ? n ? 9, 999 are much harder to learn.
Motivated by this, we proposed the following additional measure of memory capacity.
Definition 3.2 (Mean Recurrent Length). For an RNN with cycle m, the mean recurrent length is
m
1 X
d? =
max di (n).
m n=1 i2V
(5)
Mean recurrent length studies the average dilation across different time spans within a cycle. An
architecture with good memory capacity should generally have a small recurrent length for all
time spans. Otherwise the network can only selectively memorize information at a few time spans.
Also, we take the maximum over i, so as to punish networks that have good length only for a few
starting times, which can only well memorize information originating from those specific times. The
proposed mean recurrent length has an interesting reciprocal relation with the short-term memory
(STM) measure proposed in [11], but mean recurrent length emphasizes more on long-term memory
capacity, which is more suitable for our intended applications.
With this, we are ready to illustrate the memory advantage of D ILATED RNN . Consider two RNN
architectures. One is the proposed D ILATED RNN structure with d layers and M = 2 (equation (1)).
The other is a regular d-layer RNN with skip connections (equation (2)). If the skip connections are
of skip s(l) = 2l 1 , then connections in the RNN are a strict superset of those in the D ILATED RNN ,
and the RNN accomplishes exactly the same d? as the D ILATED RNN , but with twice the number of
trainable parameters (see section 3.2). Suppose one were to give every layer in the RNN the largest
possible skip for any graph with a period of m = 2d 1 : set s(l) = 2d 1 in every layer, which is the
regular skip RNN setting. This apparent advantage turns out to be a disadvantage, because time spans
of 2 ? n < m suffer from increased path lengths, and therefore
d? = (m
1)/2 + log2 m + 1/m + 1,
(6)
which grows linearly with m. On the other hand, for the proposed D ILATED RNN,
d? = (3m
1)/2m log2 m + 1/m + 1,
(7)
where d? only grows logarithmically with m, which is much smaller than that of regular skip RNN.
It implies that the information in the past on average travels along much fewer edges, and thus
undergoes far less attenuation. The derivation is given in appendix A in the supplementary materials.
3.2
Parameter Efficiency
The advantage of D ILATED RNN lies not only in the memory capacity but also the number of
parameters that achieves such memory capacity. To quantify the analysis, the following measure is
introduced.
Definition 3.3 (Number of Recurrent Edges per Node). Denote Card{?} as the set cardinality. For
an RNN represented as GC = (VC , EC ), the number of recurrent edges per node, Nr , is defined as
Nr = Card {e = (u, v, ) 2 EC :
6= 0} / Card{VC }.
(8)
Ideally, we would want a network that has large recurrent skips while maintaining a small number of
recurrent weights. It is easy to show that Nr for D ILATED RNN is 1 and that for RNNs with regular
skip connections is 2. The D ILATED RNN has half the recurrent complexity as the RNN with regular
skip RNN because of the removal of the direct recurrent edge. The following theorem states that the
D ILATED RNN is able to achieve the best memory capacity among a class of connection structures
with Nr = 1, and thus is among the most parameter efficient RNN architectures.
4
Theorem 3.1 (Parameter Efficiency of D ILATED RNN). Consider a subset of d-layer RNNs with
period m = M d 1 that consists purely of dilated skip connections (hence Nr = 1). For the RNNs in
this subset, there are d different dilations, 1 = s1 ? s2 ? ? ? ? ? sd = m, and
si = n i si
1,
(9)
where ni is any arbitrary positive integer. Among this subset, the d-layer D ILATED RNN with dilation
?
rate {M 0 , ? ? ? , M d 1 } achieves the smallest d.
The proof is motivated by [4], and is given in appendix B.
3.3
Comparing with Dilated CNN
Since D ILATED RNN is motivated by dilated CNN [22, 28], it is useful to compare their memory
capacities. Although cyclic graph, mean recurrent length and number of recurrent edges per node
are designed for recurrent structures, they happen to be applicable to dilated CNN as well. What?s
more, it can be easily shown that, compared to a D ILATED RNN with the same number of layers and
dilation rate of each layer, a dilated CNN has exactly the same number of recurrent edges per node,
and a slightly smaller (by log2 m) mean recurrent length. Hence both architectures have the same
model complexity, and it looks like a dilated CNN has a slightly better memory capacity.
However, mean recurrent length only measures the memory capacity within a cycle. When going
beyond a cycle, it is already shown that the recurrent length grows linearly with the number of cycles
[30] for RNN structures, including D ILATED RNN, whereas for a dilated CNN, the receptive field
size is always finite (thus mean recurrent length goes to infinity beyond the receptive field size). For
example, with dilation rate M = 2l 1 and d layers l = 1, ? ? ? , d, a dilated CNN has a receptive
field size of 2d , which is two cycles. On the other hand, the memory of a D ILATED RNN can go far
beyond two cycles, particularly with the sophisticated units like GRU and LSTM. Hence the memory
capacity advantage of D ILATED RNN over a dilated CNN is obvious.
4
Experiments
In this section, we evaluate the performance of D ILATED RNN on four different tasks, which include
long-term memorization, pixel-by-pixel MNIST classification [15], character-level language modeling
on the Penn Treebank [16], and speaker identification with raw waveforms on VCTK [26]. We also
investigate the effect of dilation on performance and computational efficiency.
Unless specified otherwise, all the models are implemented with Tensorflow [1]. We use the default
nonlinearities and RMSProp optimizer [21] with learning rate 0.001 and decay rate of 0.9. All
weight matrices are initialized by the standard normal distribution. The batch size is set to 128.
Furthermore, in all the experiments, we apply the sequence classification setting [25], where the
output layer only adds at the end of the sequence. Results are reported for trained models that achieve
the best validation loss. Unless stated otherwise, no tricks, such as gradient clipping [18], learning
rate annealing, recurrent weight dropout [20], recurrent batch norm [20], layer norm [3], etc., are
applied. All the tasks are sequence level classification tasks, and therefore the ?gridding? problem
[29] is irrelevant. No ?degridded? module is needed.
Three RNN cells, Vanilla, LSTM and GRU cells, are combined with the D ILATED RNN , which
we refer to as dilated Vanilla, dilated LSTM and dilated GRU, respectively. The common baselines
include single-layer RNNs (denoted as Vanilla RNN, LSTM, and GRU), multi-layer RNNs (denoted
as stack Vanilla, stack LSTM, and stack GRU), and Vanilla RNN with regular skip connections
(denoted as Skip Vanilla). Additional baselines will be specified in the corresponding subsections.
4.1
Copy memory problem
This task tests the ability of recurrent models in memorizing long-term information. We follow
a similar setup in [2, 24, 10]. Each input sequence is of length T + 20. The first ten values are
randomly generated from integers 0 to 7; the next T 1 values are all 8; the last 11 values are all
9, where the first 9 signals that the model needs to start to output the first 10 values of the inputs.
Different from the settings in [2, 24], the average cross-entropy loss is only measured at the last 10
timestamps. Therefore, the random guess yields an expected average cross entropy of ln(8) ? 2.079.
5
Figure 3: Results of the copy memory problem with T = 500 (left) and T = 1000 (right). The dilatedRNN converges quickly to the perfect solutions. Except for RNNs with dilated skip connections, all
other methods are unable to improve over random guesses.
The D ILATED RNN uses 9 layers with hidden state size of 10. The dilation starts from one to 256 at
the last hidden layer. The single-layer baselines have 256 hidden units. The multi-layer baselines use
the same number of layers and hidden state size as the D ILATED RNN . The skip Vanilla has 9 layers,
and the skip length at each layer is 256, which matches the maximum dilation of the D ILATED RNN.
The convergence curves in two settings, T = 500 and 1, 000, are shown in figure 3. In both settings,
the D ILATED RNN with vanilla cells converges to a good optimum after about 1,000 training iterations,
whereas dilated LSTM and GRU converge slower. It might be because the LSTM and GRU cells are
much more complex than the vanilla unit. Except for the proposed models, all the other models are
unable to do better than the random guess, including the skip Vanilla. These results suggest that the
proposed structure as a simple renovation is very useful for problems requiring very long memory.
4.2
Pixel-by-pixel MNIST
Sequential classification on the MNIST digits [15] is commonly used to test the performance of RNNs.
We first implement two settings. In the first setting, called the unpermuted setting, we follow the same
setups in [2, 13, 14, 24, 30] by serializing each image into a 784 x 1 sequence. The second setting,
called permuted setting, rearranges the input sequence with a fixed permutations. Training, validation
and testing sets are the default ones in Tensorflow. Hyperparameters and results are reported in table
1. In addition to the baselines already described, we also implement the dilated CNN. However, the
receptive fields size of a nine-layer dilated CNN is 512, and is insufficient to cover the sequence
length of 784. Therefore, we added one more layer to the dilated CNN, which enlarges its receptive
field size to 1,024. It also forms a slight advantage of dilated CNN over the D ILATED RNN structures.
In the unpermuted setting, the dilated GRU achieves the best evaluation accuracy of 99.2. However,
the performance improvements of dilated GRU and LSTM over both the single- and multi-layer
ones are marginal, which might be because the task is too simple. Further, we observe significant
performance differences between stack Vanilla and skip vanilla, which is consistent with the findings
in [30] that RNNs can better model long-term dependencies and achieves good results when recurrent
skip connections added. Nevertheless, the dilated vanilla has yet another significant performance gain
over the skip Vanilla, which is consistent with our argument in section 3, that the D ILATED RNN has
a much more balanced memory over a wide range of time periods than RNNs with the regular skips.
The performance of the dilated CNN is dominated by dilated LSTM and GRU, even when the latter
have fewer parameters (in the 20 hidden units case) than the former (in the 50 hidden units case).
In the permuted setting, almost all performances are lower. However, the D ILATED RNN models
maintain very high evaluation accuracies. In particular, dilated Vanilla outperforms the previous
RNN-based state-of-the-art Zoneout [13] with a comparable number of parameters. It achieves
test accuracy of 96.1 with only 44k parameters. Note that the previous state-of-the-art utilizes the
recurrent batch normalization. The version without it has a much lower performance compared to
all the dilated models. We believe the consistently high performance of the D ILATED RNN across
different permutations is due to its hierarchical multi-resolution dilations. In addition, the dilated
CNN is able the achieve the best performance, which is in accordance with our claim in section 3.3
that dilated CNN has a slightly shorter mean recurrent length than D ILATED RNN architectures, when
sequence length fall within its receptive field size. However, note that this is achieved by adding one
additional layer to expand its receptive field size compared to the RNN counterparts. When the useful
information lies outside its receptive field, the dilated CNN might fail completely.
6
Figure 4: Results of the noisy MNIST task with T = 1000 (left) and 2000 (right). RNN models
without skip connections fail. D ILATED RNN significant outperforms regular recurrent skips and
on-pars with the dilated CNN.
Table 1: Results for unpermuted and permuted pixel-by-pixel MNIST. Italic numbers indicate the
results copied from the original paper. The best results are bold.
Method
Vanilla RNN
LSTM [24]
GRU
IRNN [14]
Full uRNN [24]
Skipped RNN [30]
Zoneout [13]
Dilated CNN [22]
Dilated Vanilla
Dilated LSTM
Dilated GRU
#
layers
1/9
1/9
1/9
1
1
1/9
1
10
9
9
9
hidden /
layer
256 / 20
256 / 20
256 / 20
100
512
95 / 20
100
20 / 50
20 / 50
20 / 50
20 / 50
# parameters
(?, k)
68 / 7
270 / 28
200 / 21
12
270
16 / 11
42
7 / 46
7 / 44
28 / 173
21 / 130
Max
dilations
1
1
1
1
1
21 / 256
1
512
256
256
256
Unpermuted
test accuracy
- / 49.1
98.2 / 98.7
99.1 / 98.8
97.0
97.5
98.1 / 85.4
98.0 / 98.3
97.7 / 98.0
98. 9 / 98.9
99.0 / 99.2
Permunted
test accuracy
71.6 / 88.5
91.7 / 89.5
94.1 / 91.3
?82.0
94.1
94.0 / 91.8
93.1 / 95.92
95.7 / 96.7
95.5 / 96.1
94.2 / 95.4
94.4 / 94.6
In addition to these two settings, we propose a more challenging task called the noisy MNIST, where
we pad the unpermuted pixel sequences with [0, 1] uniform random noise to the length of T . The
results with two setups T = 1, 000 and T = 2, 000 are shown in figure 4. The dilated recurrent
models and skip RNN have 9 layers and 20 hidden units per layer. The number of skips at each layer
of skip RNN is 256. The dilated CNN has 10 layers and 11 layers for T = 1, 000 and T = 2, 000,
respectively. This expands the receptive field size of the dilated CNN to the entire input sequence.
The number of filters per layer is 20. It is worth mentioning that, in the case of T = 2, 000, if
we use a 10-layer dilated CNN instead, it will only produce random guesses. This is because the
output node only sees the last 1, 024 input samples which do not contain any informative data. All
the other reported models have the same hyperparameters as shown in the first three row of table 1.
We found that none of the models without skip connections is able to learn. Although skip Vanilla
remains learning, its performance drops compared to the first unpermuted setup. On the contrary,
the D ILATED RNN and dilated CNN models obtain almost the same performances as before. It is
also worth mentioning that in all three experiments, the D ILATED RNN models are able to achieve
comparable results with an extremely small number of parameters.
4.3
Language modeling
We further investigate the task of predicting the next character on the Penn Treebank dataset [16]. We
follow the data splitting rule with the sequence length of 100 that are commonly used in previous
studies. This corpus contains 1 million words, which is small and prone to over-fitting. Therefore
model regularization methods have been shown effective on the validation and test set performances.
Unlike many existing approaches, we apply no regularization other than a dropout on the input layer.
Instead, we focus on investigating the regularization effect of the dilated structure itself. Results are
shown in table 2. Although Zoneout, LayerNorm HM-LSTM and HyperNetowrks outperform the
D ILATED RNN models, they apply batch or layer normalizations as regularization. To the best of our
knowledge, the dilated GRU with 1.27 BPC achieves the best result among models of similar sizes
2
with recurrent batch norm [20].
7
Table 2: Character-level language modeling on the Penn Tree Bank dataset.
Method
LSTM
GRU
Recurrent BN-LSTM [7]
Recurrent dropout LSTM [20]
Zoneout [13]
LayerNorm HM-LSTM [5]
HyperNetworks [9]
Dilated Vanilla
Dilated LSTM
Dilated GRU
Method
MFCC
Raw
#
layers
1/5
1/5
1
1
1
3
1/2
5
5
5
hidden
/ layer
1k / 256
1k / 256
1k
1k
1k
512
1k
256
256
256
# parameters
(?, M)
4.25 / 1.9
3.19 / 1.42
4.25
4.25
4.91 / 14.41
0.6
1.9
1.42
Max
dilations
1
1
1
1
1
1
1
64
64
64
Table 3: Speaker identification on the VCTK dataset.
GRU
Fused GRU
Dilated GRU
#
layers
5/1
1
6/8
hidden
/ layer
20 / 128
256
50
# parameters
(?, k)
16 / 68
225
103 / 133
Min
dilations
1
32 / 8
32 / 8
Max
dilations
1
32 /8
1024
Evaluation
BPC
1.31 / 1.33
1.32 / 1.33
1.32
1.30
1.27
1.24
1.26 / 1.223
1.37
1.31
1.27
Evaluation
accuracy
0.66 / 0.77
0.45 / 0.65
0.64 / 0.74
without layer normalizations. Also, the dilated models outperform their regular counterparts, Vanilla
(didn?t converge, omitted), LSTM and GRU, without increasing the model complexity.
4.4
Speaker identification from raw waveform
We also perform the speaker identification task using the corpus from VCTK [26]. Learning audio
models directly from the raw waveform poses a difficult challenge for recurrent models because of
the vastly long-term dependency. Recently the CLDNN family of models [19] managed to match
or surpass the log mel-frequency features in several speech problems using waveform. However,
CLDNNs coarsen the temporal granularity by pooling the first-layer CNN output before feeding it
into the subsequent RNN layers, so as to solve the memory challenge. Instead, the D ILATED RNN
directly works on the raw waveform without pooling, which is considered more difficult.
To achieve a feasible training time, we adopt the efficient generalization of the D ILATED RNN as
proposed in equation (4) with l0 = 3 and l0 = 5 . As mentioned before, if the dilations do not start
at one, the model is equivalent to multiple shared-weight networks, each working on partial inputs,
and the predictions are made by fusing the information using a 1-by-M l0 convolutional layer. Our
baseline GRU model follows the same setting with various resolutions (referred to as fused-GRU),
with dilation starting at 8. This baseline has 8 share-weight GRU networks, and each subnetwork
works on 1/8 of the subsampled sequences. The same fusion layer is used to obtain the final prediction.
Since most other regular baselines failed to converge, we also implemented the MFCC-based models
on the same task setting for reference. The 13-dimensional log-mel frequency features are computed
with 25ms window and 5ms shift. The inputs of MFCC models are of length 100 to match the
input duration in the waveform-based models. The MFCC feature has two natural advantages: 1)
no information loss from operating on subsequences; 2) shorter sequence length. Nevertheless, our
dilated models operating directly on the waveform still offer a competitive performance (Table 3).
4.5
Discussion
In this subsection, we first investigate the relationship between performance and the number of
dilations. We compare the D ILATED RNN models with different numbers of layers on the noisy
MNIST T = 1, 000 task. All models use vanilla RNN cells with hidden state size 20. The number of
dilations starts at one. In figure 5, we observe that the classification accuracy and rate of convergence
increases as the models become deeper. Recall the maximum skip is exponential in the number of
layers. Thus, the deeper model has a larger maximum skip and mean recurrent length.
Second, we consider maintaining a large maximum skip with a smaller number of layers, by increasing
the dilation at the bottom layer of D ILATED RNN . First, we construct a nine-layer D ILATED RNN
3
with layer normalization [3].
8
Figure 5: Results for dilated vanilla with different numbers of layers on the noisy MNIST dataset.
The performance and convergent speed increase as the number of layers increases.
Figure 6: Training time (left) and evaluation performance (right) for dilated vanilla that starts at
different numbers of dilations at the bottom layer. The maximum dilations for all models are 256.
model with vanilla RNN cells. The number of dilations starts at 1, and hidden state size is 20. This
architecture is referred to as ?starts at 1? in figure 6. Then, we remove the bottom hidden layers
one-by-one to construct seven new models. The last created model has three layers, and the number
of dilations starts at 64. Figure 6 demonstrates both the wall time and evaluation accuracy for 50,000
training iterations of noisy MNIST dataset. The training time reduces by roughly 50% for every
dropped layer (for every doubling of the minimum dilation). Although the testing performance
decreases when the dilation does not start at one, the effect is marginal with s(0) = 2, and small with
4 ? s(0) ? 16. Notably, the model with dilation starting at 64 is able to train within 17 minutes by
using a single Nvidia P-100 GPU while maintaining a 93.5% test accuracy.
5
Conclusion
Our experiments with D ILATED RNN provide strong evidence that this simple multi-timescale
architectural choice can reliably improve the ability of recurrent models to learn long-term dependency
in problems from different domains. We found that the D ILATED RNN trains faster, requires less
hyperparameter tuning, and needs fewer parameters to achieve the state-of-the-arts. In complement
to the experimental results, we have provided a theoretical analysis showing the advantages of
D ILATED RNN and proved its optimality under a meaningful architectural measure of RNNs.
Acknowledgement
Authors would like to thank Tom Le Paine ([email protected]) and Ryan Musa
([email protected]) for their insightful discussions.
9
References
[1] Mart?n Abadi, Ashish Agarwal, Paul Barham, Eugene Brevdo, Zhifeng Chen, Craig Citro, Greg S
Corrado, Andy Davis, Jeffrey Dean, Matthieu Devin, et al. Tensorflow: Large-scale machine learning on
heterogeneous distributed systems. arXiv preprint arXiv:1603.04467, 2016.
[2] Martin Arjovsky, Amar Shah, and Yoshua Bengio. Unitary evolution recurrent neural networks. In
International Conference on Machine Learning, pages 1120?1128, 2016.
[3] Jimmy Lei Ba, Jamie Ryan Kiros, and Geoffrey E Hinton. Layer normalization. arXiv preprint
arXiv:1607.06450, 2016.
[4] Eduardo R Caianiello, Gaetano Scarpetta, and Giovanna Simoncelli. A systemic study of monetary systems.
International Journal Of General System, 8(2):81?92, 1982.
[5] Junyoung Chung, Sungjin Ahn, and Yoshua Bengio. Hierarchical multiscale recurrent neural networks.
arXiv preprint arXiv:1609.01704, 2016.
[6] Junyoung Chung, Caglar Gulcehre, KyungHyun Cho, and Yoshua Bengio. Empirical evaluation of gated
recurrent neural networks on sequence modeling. arXiv preprint arXiv:1412.3555, 2014.
[7] Tim Cooijmans, Nicolas Ballas, C?sar Laurent, ?a?glar G?l?ehre, and Aaron Courville. Recurrent batch
normalization. arXiv preprint arXiv:1603.09025, 2016.
[8] Salah El Hihi and Yoshua Bengio. Hierarchical recurrent neural networks for long-term dependencies. In
Nips, volume 409, 1995.
[9] David Ha, Andrew Dai, and Quoc V Le. Hypernetworks. arXiv preprint arXiv:1609.09106, 2016.
[10] Sepp Hochreiter and J?rgen Schmidhuber. Long short-term memory. Neural computation, 9(8):1735?1780,
1997.
[11] Herbert Jaeger. Short term memory in echo state networks, volume 5. GMD-Forschungszentrum Informationstechnik, 2001.
[12] Jan Koutnik, Klaus Greff, Faustino Gomez, and Juergen Schmidhuber. A clockwork rnn. arXiv preprint
arXiv:1402.3511, 2014.
[13] David Krueger, Tegan Maharaj, J?nos Kram?r, Mohammad Pezeshki, Nicolas Ballas, Nan Rosemary Ke,
Anirudh Goyal, Yoshua Bengio, Hugo Larochelle, Aaron Courville, et al. Zoneout: Regularizing rnns by
randomly preserving hidden activations. arXiv preprint arXiv:1606.01305, 2016.
[14] Quoc V Le, Navdeep Jaitly, and Geoffrey E Hinton. A simple way to initialize recurrent networks of
rectified linear units. arXiv preprint arXiv:1504.00941, 2015.
[15] Yann LeCun, L?on Bottou, Yoshua Bengio, and Patrick Haffner. Gradient-based learning applied to
document recognition. Proceedings of the IEEE, 86(11):2278?2324, 1998.
[16] Mitchell P Marcus, Mary Ann Marcinkiewicz, and Beatrice Santorini. Building a large annotated corpus
of english: The penn treebank. Computational linguistics, 19(2):313?330, 1993.
[17] Daniel Neil, Michael Pfeiffer, and Shih-Chii Liu. Phased LSTM: accelerating recurrent network training
for long or event-based sequences. arXiv preprint arXiv:1610.09513, 2016.
[18] Razvan Pascanu, Tomas Mikolov, and Yoshua Bengio. On the difficulty of training recurrent neural
networks. ICML (3), 28:1310?1318, 2013.
[19] Tara N Sainath, Ron J Weiss, Andrew Senior, Kevin W Wilson, and Oriol Vinyals. Learning the speech frontend with raw waveform cldnns. In Sixteenth Annual Conference of the International Speech Communication
Association, 2015.
[20] Stanislau Semeniuta, Aliaksei Severyn, and Erhardt Barth. Recurrent dropout without memory loss. arXiv
preprint arXiv:1603.05118, 2016.
[21] Tijmen Tieleman and Geoffrey Hinton. Lecture 6.5-rmsprop: Divide the gradient by a running average of
its recent magnitude. COURSERA: Neural networks for machine learning, 4(2), 2012.
[22] A?ron van den Oord, Sander Dieleman, Heiga Zen, Karen Simonyan, Oriol Vinyals, Alex Graves, Nal
Kalchbrenner, Andrew Senior, and Koray Kavukcuoglu. Wavenet: A generative model for raw audio.
CoRR abs/1609.03499, 2016.
10
[23] Alexander Sasha Vezhnevets, Simon Osindero, Tom Schaul, Nicolas Heess, Max Jaderberg, David Silver,
and Koray Kavukcuoglu. Feudal networks for hierarchical reinforcement learning. arXiv preprint
arXiv:1703.01161, 2017.
[24] Scott Wisdom, Thomas Powers, John Hershey, Jonathan Le Roux, and Les Atlas. Full-capacity unitary
recurrent neural networks. In Advances in Neural Information Processing Systems, pages 4880?4888,
2016.
[25] Zhengzheng Xing, Jian Pei, and Eamonn Keogh. A brief survey on sequence classification. ACM Sigkdd
Explorations Newsletter, 12(1):40?48, 2010.
[26] Junichi Yamagishi. English multi-speaker corpus for cstr voice cloning toolkit. http://homepages.inf.
ed.ac.uk/jyamagis/page3/page58/page58.html, 2012.
[27] Adams W Yu, Hongrae Lee, and Quoc V Le. Learning to skim text. arXiv preprint arXiv:1704.06877,
2017.
[28] Fisher Yu and Vladlen Koltun. Multi-scale context aggregation by dilated convolutions. arXiv preprint
arXiv:1511.07122, 2015.
[29] Fisher Yu, Vladlen Koltun, and Thomas Funkhouser.
arXiv:1705.09914, 2017.
Dilated residual networks.
arXiv preprint
[30] Saizheng Zhang, Yuhuai Wu, Tong Che, Zhouhan Lin, Roland Memisevic, Ruslan R Salakhutdinov, and
Yoshua Bengio. Architectural complexity measures of recurrent neural networks. In Advances in Neural
Information Processing Systems, pages 1822?1830, 2016.
11
| 6613 |@word cnn:26 middle:3 version:3 norm:3 bn:1 harder:1 cyclic:4 contains:2 liu:1 daniel:1 document:1 past:1 existing:2 outperforms:2 com:4 comparing:1 si:2 yet:3 activation:1 must:2 gpu:1 john:1 devin:1 timestamps:4 happen:1 informative:1 subsequent:1 remove:1 designed:1 drop:1 atlas:1 half:1 fewer:6 guess:4 generative:1 reciprocal:1 vanishing:4 short:5 provides:4 pascanu:1 node:14 ron:2 zhang:1 along:3 direct:1 become:1 koltun:2 abadi:1 prove:1 consists:1 yu1:1 fitting:1 manner:1 introduce:4 notably:1 interweaving:1 expected:1 roughly:1 kiros:1 multi:13 wavenet:3 salakhutdinov:1 window:1 cardinality:1 increasing:6 stm:1 provided:1 moreover:1 mitigated:1 notation:1 didn:1 homepage:1 what:1 yamagishi:1 developed:1 proposing:1 finding:1 eduardo:1 temporal:6 every:4 attenuation:1 expands:1 tackle:1 exactly:2 demonstrates:1 uk:1 unit:8 penn:4 zhang1:1 positive:1 before:3 dropped:1 accordance:1 sd:1 punish:1 laurent:1 path:3 shiyu:2 might:3 rnns:20 twice:1 challenging:2 mentioning:3 limited:1 range:2 systemic:1 phased:2 directed:3 lecun:1 testing:3 learningbased:1 block:1 implement:2 goyal:1 razvan:1 digit:2 jan:1 rnn:101 empirical:1 significantly:3 matching:1 word:1 regular:16 seeing:1 downsampled:1 suggest:1 context:1 impossible:1 memorization:2 equivalent:2 conventional:1 dean:1 center:1 missing:2 clockwork:2 go:2 sepp:1 flexibly:1 starting:4 hypernetworks:2 focused:1 ke:1 resolution:7 impeded:1 splitting:1 duration:1 jimmy:1 matthieu:1 tomas:1 roux:1 rule:1 importantly:1 zoneout:5 embedding:1 analogous:1 sar:1 suppose:2 modulo:1 us:1 origin:1 jaitly:1 trick:1 logarithmically:1 recognition:1 particularly:1 labeled:2 bottom:3 module:1 preprint:15 capture:1 cycle:9 coursera:1 decrease:1 removed:2 balanced:1 mentioned:1 complexity:4 rmsprop:2 ideally:1 rigorously:1 caianiello:1 trained:1 purely:1 powerfully:1 efficiency:10 completely:1 easily:1 represented:3 various:1 aliaksei:1 pezeshki:1 derivation:1 train:2 effective:2 eamonn:1 klaus:1 kevin:1 outside:1 kalchbrenner:1 apparent:1 saizheng:1 huang1:1 supplementary:1 solve:1 larger:1 otherwise:3 enlarges:1 ability:4 simonyan:1 neil:1 amar:1 timescale:2 noisy:5 itself:1 final:2 echo:1 sequence:21 advantage:10 propose:1 jamie:1 huang2:1 monetary:1 rapidly:1 glar:1 poorly:1 achieve:6 sixteenth:1 schaul:1 guo1:1 validate:1 convergence:2 optimum:1 jaeger:1 produce:1 perfect:1 converges:2 silver:1 adam:1 tim:1 illustrate:3 recurrent:67 andrew:3 pose:1 ac:1 measured:1 odd:1 strong:1 implemented:3 skip:49 memorize:2 implies:1 quantify:1 indicate:1 larochelle:1 waveform:10 survey:1 annotated:1 filter:1 vc:4 exploration:1 material:1 cstr:1 require:1 feeding:3 beatrice:1 generalization:2 wall:1 alleviate:2 marcinkiewicz:1 ryan:2 summation:1 keogh:1 zhang2:1 considered:1 normal:1 dieleman:1 mo:1 claim:1 rgen:1 major:1 achieves:6 optimizer:1 smallest:1 adopt:1 omitted:1 purpose:1 ruslan:1 travel:3 applicable:1 faustino:1 yuhuai:1 sensitive:1 largest:1 always:1 modified:1 wilson:1 coarsen:1 serializing:1 l0:12 focus:2 rosemary:1 improvement:1 consistently:1 mainly:1 cloning:1 greatly:1 skipped:1 sigkdd:1 maharaj:1 baseline:8 el:1 entire:2 chang1:1 pad:1 hidden:17 relation:1 originating:1 going:1 expand:1 pixel:9 overall:1 classification:7 among:5 html:1 i2v:1 denoted:3 art:5 special:1 initialize:1 marginal:2 wtan:1 equal:1 construct:4 field:9 beach:1 koray:2 yu:4 look:2 icml:1 yoshua:8 few:3 randomly:2 simultaneously:2 anirudh:1 subsampled:1 phase:1 intended:1 jeffrey:1 maintain:1 attempt:2 ab:1 investigate:3 evaluation:7 bpc:2 chain:2 andy:1 rearranges:1 edge:12 partial:2 jumping:1 shorter:4 orthogonal:1 unless:2 tree:1 divide:1 initialized:1 causal:1 theoretical:2 increased:3 modeling:7 stanislau:1 c4t:4 cover:1 disadvantage:1 juergen:1 clipping:2 yum:1 witbrock:1 stacking:1 subset:3 fusing:1 uniform:1 osindero:1 too:1 reported:3 dependency:18 koutnik:1 combined:2 cho:1 st:1 lstm:27 international:3 oord:1 memisevic:1 destination:1 lee:1 zhouhan:1 michael:2 enhance:1 connecting:1 concrete:1 quickly:1 fused:2 ashish:1 vastly:1 zen:1 severyn:1 chung:2 nonlinearities:1 bold:1 dilated:72 performed:1 try:1 memorizes:1 red:2 start:11 competitive:1 xing:1 parallel:2 aggregation:1 simon:1 contribution:2 appended:1 il:1 publicly:1 ni:1 convolutional:3 accuracy:9 greg:1 yield:1 wisdom:1 chii:1 raw:9 handwritten:1 identification:5 kavukcuoglu:2 emphasizes:1 craig:1 none:1 notoriously:1 worth:4 mfcc:4 rectified:1 ed:1 xiaoxiao:2 definition:3 frequency:2 obvious:1 proof:2 di:6 gain:2 dataset:5 proved:1 mitchell:1 recall:1 subsection:3 color:1 improves:2 knowledge:1 sophisticated:2 back:2 barth:1 follow:3 tom:2 hershey:1 wei:3 improved:1 furthermore:1 just:1 lastly:1 layernorm:2 working:2 hand:4 multiscale:1 propagation:1 undergoes:1 brings:1 lei:1 believe:2 grows:3 mary:1 xiaodong:1 building:2 usa:3 verify:1 facilitate:1 contain:2 effect:3 requiring:1 hence:3 former:1 counterpart:2 regularization:4 evolution:1 kyunghyun:1 managed:1 funkhouser:1 davis:1 speaker:6 mel:2 yorktown:1 m:2 generalized:1 junichi:1 mohammad:1 newsletter:1 greff:1 image:1 recently:2 krueger:1 common:1 specialized:1 permuted:3 empirically:1 hugo:1 vezhnevets:1 exponentially:5 ballas:2 million:1 extend:2 discussed:3 slight:1 salah:1 hihi:1 volume:2 association:1 significant:4 refer:1 tuning:1 vanilla:27 illinois:3 language:4 toolkit:1 operating:2 ahn:1 etc:3 add:1 patrick:1 recent:3 showed:1 irrelevant:1 inf:1 schmidhuber:2 nvidia:1 watson:1 captured:1 minimum:1 additional:3 arjovsky:1 dai:1 herbert:1 preserving:1 accomplishes:1 converge:3 shortest:1 period:8 corrado:1 exploding:4 signal:1 multiple:6 full:2 simoncelli:1 reduces:5 champaign:1 match:4 characterized:2 faster:1 cross:2 long:21 compensate:2 offer:1 page3:1 lin:1 roland:1 prediction:2 involving:1 paine:1 heterogeneous:1 essentially:1 navdeep:1 arxiv:30 iteration:2 kernel:1 normalization:6 agarwal:1 achieved:1 cell:16 hochreiter:1 whereas:3 want:1 addition:3 annealing:1 jian:1 parallelization:2 extra:1 unlike:1 strict:1 pooling:2 contrary:1 integer:2 unitary:3 yang:2 granularity:1 bengio:8 superset:1 easy:1 sander:1 variety:1 architecture:16 reduce:1 idea:1 barham:1 haffner:1 shift:1 six:1 motivated:3 urnn:1 accelerating:1 suffer:1 karen:1 speech:3 nine:2 heess:1 useful:4 generally:1 mid:2 ten:1 gmd:1 http:2 outperform:3 per:6 diverse:1 sainath:1 hyperparameter:1 key:1 four:6 vctk:3 nevertheless:2 shih:1 nal:1 graph:12 unpermuted:6 almost:2 reasonable:1 family:1 architectural:3 yann:1 utilizes:1 wu:1 appendix:2 comparable:2 dropout:4 layer:80 ct:5 nan:1 gomez:1 convergent:1 copied:1 fold:1 courville:2 annual:1 infinity:1 feudal:1 alex:1 dominated:1 speed:2 argument:1 extremely:2 optimality:2 span:5 min:1 mikolov:1 martin:1 gpus:1 vladlen:2 across:4 smaller:3 slightly:3 character:4 page58:2 s1:1 quoc:3 memorizing:2 den:1 ln:1 equation:4 remains:2 previously:1 turn:1 fail:2 needed:2 end:1 gulcehre:1 brevdo:1 operation:2 apply:4 observe:2 hierarchical:6 generic:1 batch:6 voice:1 shah:1 gate:1 gridding:1 slower:1 thomas:4 abandoned:1 denotes:2 original:1 include:3 linguistics:1 running:1 log2:3 maintaining:4 establish:1 already:2 added:2 receptive:9 dependence:1 nr:5 italic:1 che:1 enhances:1 gradient:9 subnetwork:1 unable:2 card:3 thank:1 capacity:18 seven:1 reason:1 marcus:1 length:32 code:2 index:2 relationship:1 insufficient:1 tijmen:1 difficult:4 setup:4 potentially:1 statement:1 hasegawa:1 stated:1 ba:1 design:2 implementation:1 reliably:1 pei:1 gated:2 perform:2 convolution:1 urbana:2 finite:1 caglar:1 hinton:3 santorini:1 communication:1 gc:4 stack:6 arbitrary:1 parallelizing:1 introduced:3 complement:1 david:3 gru:26 required:1 specified:2 connection:30 trainable:1 tensorflow:3 heiga:1 nip:2 able:5 beyond:3 kram:1 scott:1 challenge:5 including:4 memory:30 max:5 power:1 suitable:2 event:1 available1:1 quantification:1 difficulty:2 predicting:1 natural:1 residual:1 pfeiffer:1 improve:3 github:1 brief:1 created:1 ready:1 hm:2 extract:2 text:2 eugene:1 acknowledgement:1 removal:1 graf:1 loss:4 par:1 permutation:3 lecture:1 interesting:1 geoffrey:3 remarkable:1 ingredient:1 validation:3 degree:1 consistent:2 treebank:3 bank:1 share:1 ibm:4 row:1 prone:1 ehre:1 last:5 copy:2 english:2 side:2 allow:1 formal:1 deeper:2 senior:2 wide:1 fall:1 benefit:1 distributed:1 overcome:1 default:2 curve:1 van:1 autoregressive:1 forward:1 made:2 jump:1 author:2 commonly:2 sungjin:1 reinforcement:1 far:3 ec:4 jaderberg:1 reveals:1 investigating:1 corpus:4 cooijmans:1 forschungszentrum:1 consuming:1 scarpetta:1 subsequence:4 dilation:43 table:7 learn:3 ca:1 nicolas:3 bottou:1 complex:4 terminator:1 domain:1 main:3 linearly:2 arrow:1 s2:1 noise:2 hyperparameters:2 paul:1 referred:3 junyoung:2 depicts:1 ny:1 tong:1 erhardt:1 exponential:1 sasha:1 lie:2 stamp:1 breaking:1 learns:1 zhifeng:1 theorem:2 minute:1 xt:3 specific:1 showing:1 insightful:1 decay:1 evidence:1 fusion:1 irnn:1 mnist:9 sequential:5 adding:1 corr:1 frontend:1 magnitude:1 illustrates:1 chen:1 entropy:2 infinitely:1 informationstechnik:1 failed:1 prevents:1 vinyals:2 tegan:1 doubling:1 chang:1 tieleman:1 semeniuta:1 acm:1 mart:1 viewed:1 ann:1 shared:2 fisher:2 feasible:1 except:2 surpass:1 called:4 experimental:1 meaningful:1 citro:1 aaron:2 formally:1 selectively:1 tara:1 mark:1 guo:1 latter:1 jonathan:1 alexander:1 oriol:2 skim:2 evaluate:1 audio:4 regularizing:1 |
6,206 | 6,614 | Hunt For The Unique, Stable, Sparse And Fast
Feature Learning On Graphs
Saurabh Verma
Department of Computer Science
University of Minnesota, Twin Cities
[email protected]
Zhi-Li Zhang
Department of Computer Science
University of Minnesota, Twin Cities
[email protected]
Abstract
For the purpose of learning on graphs, we hunt for a graph feature representation
that exhibit certain uniqueness, stability and sparsity properties while also being
amenable to fast computation. This leads to the discovery of family of graph
spectral distances (denoted as F GSD) and their based graph feature representations,
which we prove to possess most of these desired properties. To both evaluate
the quality of graph features produced by F GSD and demonstrate their utility, we
apply them to the graph classification problem. Through extensive experiments, we
show that a simple SVM based classification algorithm, driven with our powerful
F GSD based graph features, significantly outperforms all the more sophisticated
state-of-art algorithms on the unlabeled node datasets in terms of both accuracy
and speed; it also yields very competitive results on the labeled datasets ? despite
the fact it does not utilize any node label information.
1
Introduction
In the past decade, there has been tremendous interests in learning on collection of graphs for
various purposes, in particular for solving graph classification problem. Several applications of graph
classification can be found in the domain of bioinformatics, or chemoinformatics, or social networks.
A fundamental question inherent in graph classification is determining whether two graph structures
are identical, i.e., the graph isomorphism problem, which was not known to belong either P or NP
until recently. In the seminal paper [2], Babai shows that the graph isomorphism can be solved in
quasipolynomial time; while of enormous theoretical signficance, the implication of this result in
developing practical algorithms is still unclear. Fortunately, in graph classification problems one
is more interested in whether two graphs have ?similar? (as opposed to identical) structures. This
allows for potentially much faster (yet not fully explored) algorithms to be successfully applied to the
graph classification while also accounting for graph isomorphism. One approach to get around both
these intimately tied problems together is to learn an explicit graph representation that is invariant
under graph isomorphism1 but also useful for extracting graph features.
More specifically, given a graph G, we are interested in learning a graph representation (or spectrum),
R : G ? (g1 , g2 , ..., gr ), that captures certain inherent ?atomic? (unique) sub-structures of the graph
and is invariant under graph isomorphism (i.e., two isomorphic graphs yield the same representation).
Subsequently, we want to learn a feature function F : R ? (f1 , f2 , ..., fd ) from R such that the
graph features {fi }di=1 can be employed for solving the graph classification problem. However, in
machine learning, not much attention has been given towards learning R and most of the previous
studies have focused on designing graph kernels and thus bypasses computing any explicit graph
representation. The series of papers (19, 20, 22) by Kondor et al. are some of the first (and few) that
are concerned with constructing explicit graph features ? using a group theoretic approach ? that are
invariant to graph isomorphism and can be successfully applied to the graph classification problem.
1
That is, invariant under permutation of graph vertex labels.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
Pairs of graph nodes
are generated from an
unknown distribution
Figure 1: Graph Generation Model: Graph
spectrum is assumed to be encoded in pairwise
node distances which are generated from some
distribution. Nodes connect together to form
a graph in such a way that pairwise node distances are preserved (eg. ( ? ) node-pair with
distance 0.75 is preserved even though they are
not directly connected).
Nodes connect together
to form a graph such that
the pairwise distances are
preserved
0.75
0.80
0.55
0.80
0.75
0.55
Inspired by such an approach, we also explicitly deal with learning a graph representation R and
show how to derive graph features F from R.
Our approach is quite novel and builds upon the following assumption: Graph atomic structure (or
spectrum) is encoded in the multiset2 of all node pairwise distances. Figure 1 shows the complete
graph generation model based on this premise. The origin of our assumption can be traced back to
the study of homometric structure, i.e, structures with the same multiset of interatomic distances [28].
On graphs, two vertex sets are called non-homometric if the multisets of distances determined by
them are different. (It is an unexplored problem whether there exists any distance metric on the graph
for which two vertex sets of non-isomorphic graphs are always non-homometric; but the converse
is not true, an example is the shortest path distance.) This argument provides the validity of our
assumption that the graph atomic structure is being encoded in pairwise distances. Further, we have
empirically found that the biharmonic distance [23] multisets are unique for at-least upto 10-vertex
size simple connected graphs (? 11 million graphs) and it remains as an open problem to show a
contradictory example. Moreover, we show that for a certain distance function Sf on the graph, one
can uniquely recover all the graph intrinsic properties while also being able to capture both local &
global information about the graph. Thus, we define R as the multiset of node pairwise distances
based on some distance function Sf , which will be the main focus of this paper.
We hunt for such a family of distances on graphs and its core members for which most of the properties
of an ideal graph spectrum (see Section 3) hold, including invariance under graph isomorphism and
the uniqueness property. This hunt leads us to the discovery of a family of graph spectral distance
(F GSD) and one would find harmonic (effective resistance) and biharmonic distance on graphs as the
suitable members of this family for graph representation R. Finally, for solving graph classification
(where graphs can be of different nodes sizes), we simply construct F feature vector from the
histogram of R (a multiset) and feed it to a standard classification algorithm.
Our current work focuses only on unlabeled graphs but can be extended to labeled graphs using the
same strategy as in shortest path kernel [4]. Nevertheless, our comprehensive results show that F GSD
graph features are powerful enough to significantly outperform the current state-of-art algorithms on
unlabeled datasets and are very competitive on labeled datasets ? despite the fact that they do not
utilize any node label information. In summary, the major contributions of our paper are:
? Introducing a novel & conceptually simple yet powerful graph feature representation (or spectrum)
based on the multiset of node pairwise distances.
? Discovering F GSD as a well-suited candidate for our proposed graph spectrum.
? Proving that F GSD based graph features exhibit certain uniqueness, stability, sparsity properties
and can be computationally fast with O (N 2 ) complexity, where N is the number of graph nodes
in a graph.
? Showing the superior performance of F GSD based graph features on graph classification tasks.
2
Related Work
Previous studies on graph classification can be grouped into three main categories. The first category
is concerned with constructing explicit graph features such as the skew spectrum 20 and its successor,
graphlet spectrum [22] based on group-theoretic approaches. Both are computational expensive. The
second and more popular category deals with designing graph kernels, among which, strong ones
are graphlets [30], random walks or shortest paths [4], neighborhood subgraph pairwise distance
2
A set in which an element can occur multiple of times.
2
kernel [9], Weisfeiler-Lehman kernel [31], deep graph kernels [34], graph invariant kernels [27] and
multiscale Laplacian graph kernel [21]. A tangential work [24] related to constructing features based
on atoms 3D space coordinates rather than operating on a graph structure, can be also considered in
this category. Our effort on learning R from F GSD can be seen as a part of first category, since we
explicitly investigate numerous properties of our proposed graph spectrum. While, extracting F from
R is more inspired from the work of graph kernels.
The third category involves developing convolutional neural networks (CNNs) for graphs, where
several models have been proposed to define convolution networks on graphs. The most common
model is based on generalizing convolutional networks through the graph Fourier transform via
a graph Laplacian [7, 16]. Defferrard et al. [11] extend this model by constructing fast localized
spectral filters for efficient graph coarsening as a pooling operation for CNNs on graphs. Some
variants of these models were considered in [18, 1], where the output of each neural network layer is
computed using a propagation rule that takes the graph adjacency matrix and node feature vectors into
account while updating the network weights. In [12], the convolution operation is defined by hashing
of local graph node features along with the local structure information. Likewise, in [26] local node
sequences are ?canonicalized? to create receptive fields and then fed into a 1D convolutional neural
network for classification. Among the aforementioned graph CNNs models, only those in [26, 1, 12]
are relevant to this work since they are designed to account for graphs of different sizes, while others
assume a global structure where the one-to-one correspondence of input vertices are already known.
3
Family of Graph Spectral Distances and Graph Spectrum
Basic Setup and Notations: Consider a weighted, undirected (and connected) graph G = (V, E, W )
of size N = |V |, where V is the vertex set, E the edge set (with no self-loops) and W = [wxy ] the
nonnegative weighted adjacency matrix. The standard graph Laplacian is defined as L = D ? W ,
where D is the degree matrix. It is semi-definite and admits an eigen-decomposition of the form
L = ???T , where ? = diag[?k ] is the diagonal matrix formed by the eigenvalues ?0 = 0 <
?1 ? ? ? ? ? ?N ?1 , and ? = [?0 , ..., ?N ?1 ] is an orthogonal matrix formed by the corresponding
eigenvectors ?k ?s. For x ? V , we use ?k (x) to denote the x-entry value of ?k . Let f be an
arbitrary nonnegative (real-analytical) function on R+ with f (0) = 0, 1 = [1, .., 1]T is the all-one
vector and J = 11T . Then, using slight abuse of notion, we define f (L) := ?f (?)?T and
f (?) := diag[f (?k )]. Also, f (L)xy represent xy-entry value in f (L) matrix. Lastly, I is identity
matrix and L+ is Moore-Penrose Pseudoinverse of L.
F GSD Definition: For x, y ? V , we define the f -spectral distance between x and y on G as follows:
Sf (x, y) =
N
?1
X
f (?k )(?k (x) ? ?k (y))2
(1)
k=0
We will refer to {Sf (x, y)|f }, as the family of graph spectral distances. Without loss of generality,
we assume that the derivative f 0 (?) 6= 0 for ? > 0, and then by Lagrange Inversion Theorem [33], f
is invertible and thus bijective. For reasons that will be clear shortly, we are particularly interested in
two sub-families of F GSD, where f is monotonic function (increasing or decreasing) of ?. Depending
on the sub-family, the f -spectral distance can capture different type of information in a graph.
F GSD Elements Encode Local Structure Information: For f (?) = ?p (p ? 1), one can show
that Sf (x, y) = (Lp )xx + (Lp )yy ? 2(Lp )xy . If the shortest path from x to y is larger than p, then
(Lp )xy = 0. This is based on the fact (Lp )xy captures only p-hop local neighborhood information [32]
on the graph. Hence, broadly for an increasing function of f (e.g., a polynomial function of degree
atleast p ? 1), Sf (x, y) captures the local structure information.
F GSD Elements Encode Global Structure Information: On the other hand, f as a decreasing
function yields Sf (x, y) = ((L+ )p )xx + (((L+ )p )yy ? 2((L+ )p )xy . This captures the global
J ?1
J
information, since the xy-entry of L+ = (L + N
) ?N
accounts for all paths from node x to y
+ p
(and so does (L ) ). Several known globally aware graph distances can be derived from this F GSD
sub-family. For f (?) = 1/? where ? > 0, Sf (x, y) is the harmonic (or effective resistance) distance.
More generally, for f (?) = 1/?p , p ? 1, Sf (x, y) is the polyharmonic distance (p = 2 is biharmonic
distance). Lastly f (?k ) = e?2t?k yields Sf (x, y) that is equivalent to the heat diffusion distance.
3
F GSD Graph Signal Processing Point of View: From graph signal processing perspective, Sf (x, y)
is a distance computed based on spectral filter properties [32], where f (?) act as a band-pass filter. Or,
it can be viewed in terms of spectral graph wavelets [15] as: Sf (x, y) = ?f,x (x)+?f,y (y)?2?f,x (y),
PN ?1
where ?f,x (y) = k=0 f (?k )?k (x)?k (y) (and ?f,x (x), ?f,y (y) are similarly defined) is a spectral
graph wavelet of scale 1, centered at node x and f (?) act as a graph wavelet kernel.
F GSD Based Graph Spectrum: Using the F GSD based distance matrix Sf = [Sf (x, y)] directly,
e.g., for graph classification, requires us being able to solve the graph isomorphism problem efficiently.
But no known polynomial time algorithm is available; the best algorithm today theoretically takes
quasipolynomial time [2]. However, motivated from the study of homometric structure and the fact
that each element of F GSD encodes some local or global sub-structure information of the graph,
inspired us to define the graph spectrum as R = {Sf (x, y)|?(x, y) ? V }. Thus, comparing two
R?s implicitly evaluates the sub-structural similarity between two graphs. For instance, R based on
harmonic distance contains sub-structural properties related to the spanning trees of a graph [29].
Our main concern in this paper would be choosing an appropriate f (?) function in order to generate
R which can exhibit ideal graph spectrum properties as discuss below. Also, we want F to inherent
these properties directly from R, which is made possible by defining F as the histogram of R.
Finally, we lay down those important fundamental properties of an ideal graph spectrum that one
would like R & F to obey on a graph G = (V, E, W ).
1. R & F must be invariant under any permutation ? of vertex labels. That is, R(G) = R(G? ) or
R(W ) = R(P W P T ) for any permutation matrix P .
2. R & F must have a unique representation for non-isomorphic graphs. That is, R(G1 ) 6= R(G2 )
for any two non-isomorphic graphs G1 and G2 .
3. R & F must be stable under small perturbation. That is, if graph G2 (W2 ) = G1 (W1 + ?), for a
small perturbation norm matrix k?k, then the norm of kF (G2 ) ? F (G1 )k should also be small
or bounded in order to maintain the stability.
4. F must be sparse (if high-dimensional) for all the sparsity reasons desirable in machine learning.
5. R & F must be computationally fast for efficiency and scalability purposes.
4
Uniqueness of Family of Graph Spectral Distances and Graph Spectrum
We first start with exploring the graph invariance and uniqueness properties of R & F based on F GSD.
Uniqueness is a very important (desirable) property, since it will determine whether the elements of
R set are complete (i.e., how good they are), in the sense whether R is sufficient enough to recover
all the intrinsic structural properties of a graph. We state the following important uniqueness theorem.
Theorem 1 (Uniqueness of F GSD) 3 The f -spectral distance matrix Sf = [Sf (x, y)] uniquely
determines the underlying graph (up to graph isomorphism), and each graph has a unique Sf (up to
permutation). More precisely, two undirected, weighted (and connected) graphs G1 and G2 have the
same F GSD based distance matrix up to permutation, i.e., SG1 = P SG2 P T for some permutation
matrix P , if and only if the two graphs are isomorphic.
Implications: Our proof is based on establishing the following key relationship: f (L) =
? 21 (I ? N1 J)Sf (I ? N1 J). Since f is bijective, one can uniquely recover ? from f (?). One
of the consequence of Theorem 1 is that the R based on multiset of F GSD is invariant under the
permutation of graph vertex labels and thus, satisfies the graph invariance property. Also, F will
inherent this property since R remains the same. Unfortunately, it is possible that the multiset of
some F GSD members can be same for non-isomorphic graphs (otherwise, we would have a O (N 2 )
polynomial time algorithm for solving graph isomorphism problem!). However, it is known that
all non-isomorphic graphs with less than nine vertices have unique multisets of harmonic distance.
While, for nine & ten vertex (simple) graphs, we have exactly 11 & 49 pairs of non-isomorphic
graphs (out of total 274,668 & 12,005,168 graphs) with the same harmonic spectra. These examples
show that there are significantly very low numbers of non-unique harmonic spectrums. Moreover, we
empirically found that the biharmonic distance has all unique multisets for at-least upto ten vertices
(? 11 million graphs) and we couldn?t find any non-isomorphic graphs with the same biharmonic
multisets. Further, we have the following theorem regarding the uniqueness of R.
3
1
1
Variant of Theorem 1 also hold true for the normalized graph Laplacian Lnorm = D? 2 LD? 2 .
4
Theorem 2 (Uniqueness of Graph Harmonic Spectrum) Let G = (V, E, W ) be a graph of size
|V | with an unweighted adjacency matrix W . Then, if two graphs G1 and G2 have the same number
of nodes but different number of edges, i.e, |V1 | = |V2 | but |E1 | =
6 |E2 |, then with respect to the
harmonic distance multiset, R(G1 ) 6= R(G2 ).
Implications: Our proof relies on the fact that the effective resistance distance is a monotone function
with respect to adding or removing edges. It shows that R based on some F GSD members specially
harmonic distance is atleast theoretically known to be unique to a certain degree. F also inherent this
property, fully under the condition h ? 0 (or for small enough h), where h is the histogram binwidth.
Overall the certain uniqueness of R along with containing local or global structural properties in its
each element dictate that the R is capable enough to serve as the complete powerful Graph Spectrum.
4.1
Unifying Relationship Between F SGD and Graph Embedding and Dimension Reduction
Before delving into other properties, we uncover an essential relationship between F GSD and
p Graph
1
Embedding in Euclidean space and Dimension Reduction techniques. Let f (?) 2 = diag[ f (?k )]
1
and define ? = ?f (?) 2 . Then, the f -spectral distance can be expressed as Sf (x, y) = ||?(x) ?
?(y)||22 , where ?(x) is the xth row of ?. Thus, ? represents an Euclidean embedding of G where
each node x is represented by the vector ?(x). Now for instance, if f (?) = 1, then by taking the first
p columns of ? yields embedding exactly equal to Laplacian Eigenmap (LE) [3] based on random
walk graph Laplacian (Lrw = D?1 L). For f (?) = ?2t and L = D?1 W , we get the Diffusion
Map [25]. Thus, f (?) function has one-to-one correspondence relationship with spectral dimension
reduction techniques. We have the following theorem concerning Graph Embedding based on F GSD.
Theorem 3 (Uniqueness of F GSD Graph Embedding) Each graph G can be isometrically embedded into a Euclidean space using F GSD as an isometric measure. This isometric embedding is unique,
0
if all the eigenvalues of G Laplacian
are distinct and there does not exist any other graph G with
q
0
0
Laplacian eigenvectors ?k = f (?j )/f (?j )?k , ?k ? [1, N ? 1].
Implications: The above theorem shows that F GSD provides a unique way to embed the graph
vertices into Euclidean space possibly without loosing any structural information of the graph. This
could potentially serve as a cogent tool to convert an unstructured data into a structure data (similar to
structure2vec 10 or node2vec 14 tool) which can enable us to perform standard inference tasks
in Euclidean space. Note that the uniqueness condition is quite strict and holds for co-spectral graphs.
In short, we have following uniqueness relationship, where ? is the Euclidean embedding of G graph.
Sf
5
f (LG )
LG
f (LG )
?G
Stability of Family of Graph Spectral Distances and Graph Spectrum
Next, we hunt for the stable members of the F GSD that are robust against the perturbation or noise
in the datasets. Specifically, we will look at the stability of R and F based on F GSD from f (?)
perspective by first analyzing its influence on a single edge perturbation (or in other words analyzing
rank one modification of Laplacian matrix). This will lead us to find the stable members and what
restrictions we need to impose on f (?) function for stability. We will further show that f -spectral
distance function also satisfies the notion of uniform stability [6] in a certain sense. For our analysis,
we will restrict f (?) as a monotone function of ?, for ? > 0. Let 4w ? 0 be the change after
0
0
modifying w weight on any single edge to w on the graph, where 4w = w ? w.
Theorem 4 (Eigenfunction Stability of F GSD) Let 4Sxy be the change in Sf (x, y) distance with
respect to 4w change in weight of any single edge on the graph. Then, 4Sxy for any vertex pair
(x, y) is bounded with respect to the function of eigenvalue as follows,
4Sxy ? 2 |f (?N ?1 + 24w) ? f (?1 )|
Implications: Since, R = {Sf (x, y)|?(x, y) ? V }, then each element of R is itself bounded
by 4Sxy . Now, recall that F is a histogram of R, then F won?t change, if binwdith is large
enough to accommodate the perturbation i.e., h ? 24Sxy ?(x, y) assuming all elements of R are
5
at the center of their respective histogram bins. Besides h, the other way to make R robust is by
choosing a suitable f (?) function. Lets
consider the behavior 4Sxy on f (?) = ?p for p > 0. Then,
p
p
4Sxy ? 2 (?N ?1 + 24w) ? ?1 and as a result, 4Sxy is an increasing function with respect
to p which implies that stability decreases with increase
in p. For p = 0, stability does not change
|p|
with respect to ?. While, for p < 0, 4Sxy ? 2 1/?1 ? 1/(?N ?1 + 24w)|p| . Here, 4Sxy is
a decreasing function with respect to |p|, which implies that stability increases with decrease in p.
The results conforms with the reasoning that eigenvectors corresponding to smaller eigenvalues are
smoother (i.e., oscillates slowly) than large eigenvectors (corresponding to large eigenvalues) and
decreasing p will attenuate the contribution of large eigenvectors, making the f -spectral distance
more stable and less susceptible towards perturbation or noise. However, decreasing p too much could
result in lost of local information contained in eigenvectors with larger eigenvalues and therefore, a
balance needs to be maintained. Overall, Theorem 4 shows that either through suitable h or decreasing
f (?) function, stability of R & F can be controlled to satisfy the Ideal Spectrum Property 3.
Infact, we can further show that Sf (x, y) between any two vertex (x, y) on a graph, with 0 < ? ?
w ? ? bounded weights, is tightly bounded to a certain expected value.
Theorem 5 (Uniform Stability of F GSD) Let E[Sf (x, y)] be the expected value of Sf (x, y) between vertex pair (x, y), over all possible graphs with fixed ordering of N vertices. Then we have,
with probability 1 ? ?, where ? ? (0, 1) and ? depends upon ?, ?, N .
r
p
Sf (x, y) ? E[Sf (x, y)] ? f (?) N (N ? 1) log 1
?
Implications: The above theorem is based on the fact 4Sxy can itself be upper bounded over all
possible graphs generated on a fixed ordering of N vertices. This is a very similar condition needed
for a learning algorithm to satisfy the notion of uniform stability in order to give generalization
guarantees. The f -spectral distance function can itself be thought of as a learning algorithm which
admits uniform stability (precise definition in supplementary) and indicates a strong stability behavior
over all possible graphs and further act as a generalizable learning algorithm on the graph. Theorem 5
also reveals that the deviation can bepminimized by
choosing decreasing f (?) function and it would
be suitable, if f (?) grow with O 1/ N (N ? 1) rate in order to maintain stability for large graphs.
So far, we have narrow down our interest to R & F based on the bijective and decreasing f (?) function for achieving both uniqueness and stability. This eliminates all forms of increasing polynomial
functions as a good choice of f (?). As a result, we can focus on inverse (or rational) form of polynomial functions such as polyharmonic distances. A by-product of our analysis results in revealing a
new class of stable dimension reduction techniques, possible by scaling Laplacian eigenvectors with
decreasing function of f (?), although such connections have already been known before.
6
Sparsity of Family of Graph Spectral Distances and Graph Spectrum
40
Figure 2: Figure shows the number of unique
elements present in R formed by different f spectral distance on all graphs (of |V | = 9,
total 261, 080 graphs). Graph
in enumeration
dices are sorted according to R( ?1 )G . We can
observe that f (?) = ?1 increases in form of a
step function and lower bounds all other f (?)
with an addition constant. (Best viewed in color
and when zoom-in.)
30
-R(f (6))G
f (6) =
f (6) =
f (6) =
f (6) =
f (6) =
f (6) =
20
10
0
50000
100000
150000
200000
Graph Enumeration Index
1
6
1
60:2
1
60:5
1
61:5
1
62
1
63
250000
Sparsity is desirable for both computational and statistical efficiency. In this section, we investigate
the sparsity produced in F by choosing different f (?) functions. Here, sparsity refers to its usual
definition of ?how many zero features are present in F graph feature vector?. Since F is a histogram
of R, number of non-zero elements in F will always be less than equal to number of unique (or
distinct) elements in R. However, due to the lack of any theoretical support, we rely on empirical
evidence and conjecture the following statement.
6
0
Data Sample Index
Data Sample Index
0
50
100
150
0
200
400
600
800
1000
1200
1400
1600
1800
50
100
150
0
2000
200
400
600
800
1000
1200
1400
1600
1800
2000
Feature Index
Feature Index
a. Harmonic distance based graph feature
matrix (matrix sparsity= 97.12%). Presence
of blue dot ( ) indicates feature value > 0.
b. Biharmonic distance based graph feature
matrix (matrix sparsity= 94.28%). Presence
of blue dot ( ) indicates feature value > 0.
c. Harmonic distance based feature matrix
sparsity shown with respect to per class label.
d. Biharmonic distance based feature matrix
sparsity shown with respect to per class label.
Figure 3: Feature space for MUTAG (composed of two class sizes 125 & 63): Both harmonic &
biharmonic based graph spectrum encodes a sparse high dimensional feature representation F for
graphs which can clearly distinguish the two classes as depicted in above sub-figures.
Conjecture (Sparsity of F GSD Graph Spectrum) For any graph G, let R(f (?))G represents the
number of unique elements present in the multiset of R, computed on an unweighted graph G based
on some monotonic decreasing f (?) function. Then, the following holds,
1
+2
R(f (?)) ? R
? G
G
The conjecture is based on the observation that, in the Figure 2, R ?1 + 2 lower bounds all
given monotonic decreasing f (?) along with an addition constant of 2. Same trends are observed
for different
graph
sizes |V |. Interestingly, when graph enumeration indices are sorted according
1
to size R ? , we further observe that f (?) = ?1 increases in the form of a step function. From
this conjecture, we can directly conclude that the F based on f (?) = ?1 produce the most sparse
features because number of unique elements in its R is always less than any other R. Figure 3,
further supports this conjecture which shows the feature space computed for MUTAG dataset in
case of harmonic and biharmonic spectrums. However, this raises a question of trade-off between
maintaining uniqueness and sparsity, since biharmonic distance multisets are found to be unique
for more number of graphs than harmonic distance. Nonetheless, some preliminary experiments
measuring harmonic vs. biharmonic performance on graph classification (in supplementary), suggest
that the sparsity is more favorable than uniqueness since it results in higher classification accuracy.
7
Fast Computation of Family of Graph Spectral Distances and Spectrum
Finally, we provide the general recipe of computing any member of F GSD in fast manner. In order to
avoid direct eigenvalue decomposition, we can either perform approximation or leverage structural
properties and sparsity of f (L) for efficient exact computation of Sf and thus, R.
Approximation: Inspired from the spectral graph wavelet work [32], the recipe for approximating
F GSD is to decompose f (?) possibly
Pr into an approximate polynomial series (for example, chebyshev
polynomials) as follows: f (?) = i=0 ai Ti (?) such that Ti (x) can be computed in recursive
manner
Pr
from few lower order terms (Ti?1 (x), Ti?2 (x), ..., Ti?c (x)). Then it follows, f (L) = i=0 ai Ti (L).
In this case, the cost of computing will reduce to O (r|E|) for sparse L which is very less expensive,
since O (r|E|) O (N 2 ). But, if f (?) is an inverse polynomial form of function, then computing
P
?1
r
f (L) =
a
T
(L)
= f (L+
i
i
r ), boils down to efficiently computing (a single) Moore
i=0
Penrose Pseudo inverse of a matrix.
Efficient Exact Computation: By leveraging f (L) structural properties and its sparsity, we can
efficiently perform exact computation of f (L+ ) in much more better way than the eigenvalue
7
decomposition. We propose such a method 1 which is the generalization of [23] work. We can show
J
that, f (L)f (L+ )?1 = I ? N
. Therefore, f (L)lk+ = Bk , where lk+ and Bk are the k th column of
J
+
f (L ) and B = I ? N matrices, respectively. So, first we can find a particular solution of following
T
(sparse) linear system: f (L)x = Bk and then obtain lk+ = x ? 11T x1 x. The particular solution x can
be obtained by replacing any single row and corresponding column of f (L) by zeros, and setting
diagonal entry at their intersection to one, and replacing corresponding row of B by zeros. This gives
a (non-singular) sparse linear system which can be solved very efficiently by performing cholesky
factorization and back-substitution, resulting in overall O (N 2 ) complexity as shown in [5]. Beside
this, there are few other fast methods to compute Pseudo inverse, particularly given by [17].
Complexity
SP [4]
GK[34](k ?
{3, 4, 5}) (d ? N )
SGS[20]
GS [22](k ?
[2, 6])
DCNN[1]
MLG[21]
e < N)
(N
F GSD
Approximate
?
O (N dk?1 )
?
?
?
e3)
O (N
O (r|E|)
Worst-Case
3
O (N )
k
O (N )
3
O (N )
O (N
2+k
)
2
O (N )
3
O (N )
O (N 2 )
Table 1: F GSD complexity comparison with few strong state-of-art algorithms (showing variables
that are only dependent on N & |E|). It reveals that the F GSD complexity is better than the most.
As a result, it leads to a very efficient
O (r|E|) complexity through approximation with the worst-case O (N 2 )
complexity in exact computation of R.
Table 1, shows the complexity comparison with other state-of-art methods. Since, number of elements in R
are O (N 2 ), then F is also bounded
by O (N 2 ) and thus satisfies the ideal
graph spectrum Property 5. Finally,
Algorithm 1 summarizes the complete
procedure of computing R & F .
8
Algorithm 1 Computing R and F based on F GSD.
Input: Given graphs {Gi = (Vi , Ei , Wi )}M
i=1 , f (?),
number of bins b, binwidth h.
Output: Ri and Fi ?i ? [1, M ].
for i = 1 to M do
Compute f (Li ) using approx. or exact method 1.
Compute Si = diag(f (Li ))J + Jdiag(f (Li )) ?
2f (Li ).
Set Ri = {Sxy |?(x, y) ? |Vi |}.
Compute Fi = histogram(Ri , b, h).
end for
Experiments and Results
F GSD Graph Spectrum Settings: We chose harmonic distance as an ideal candidate for F . For fast
computation, we adopted our proposed efficient exact computation method 1. And for computing
histogram, we fix binwidth size and set the number of bins such that its range covers all {Ri }M
1
elements of M number of graphs. Therefore, we had only one parameter, binwidth size, chosen from
the set {0.001, 0.0001, 0.00001}. This results in F feature vector dimension in range 100?1000, 000
with feature matrix sparsity > 90% in all cases. Our F GSD code is available at github4 .
Datasets: We employed wide variety of datasets considered as benchmark [1, 34, 21, 26] in graph
classification task to evaluate the quality of produce F GSD graph features. We adopted 7 bioinformatics datasets: Mutag, PTC, Proteins, NCI1, NCI109, D&D, MAO and 5 social network datasets:
Collab, REDDIT-Binary, REDDIT-Multi-5K, IMDB-Binary, IMDB-Multi. D&D dataset contains
691 enzymes and 587 non-enzymes proteins structures. While, MAO dataset contains 38 molecules
that are antidepressant drugs and 30 do not. For other datasets, details can be found in [34].
Experimental Set-up: All experiments were performed on a single Intel-Core [email protected]
and 64GB RAM machine. We compare our method with 6 state-of-art Graphs Kernels: Random
Walk (RW) [13], Shortest Path Kernel (SP) [4], Graphlet Kernel (GK) [30], Weisfeiler-Lehman Kernel
(WL) [31], Deep Graph Kernels (DGK) [34], Multiscale Laplacian Graph Kernels (MLK) [21]. And
proposed, 2 recent state-of-art Graph Convolutional Networks: PATCHY-SAN (PSCN) [26],
Diffusion CNNs (DCNN) [1]. And, 2 strong Graph Spectrums: the Skew Spectrum (SGS) [20],
Graphlet Spectrum (GS) [22]. We adopt the same procedure from previous works [26, 34] to make a
fair comparison and used 10-fold cross validation with LIBSVM [8] library to test the classification
performance. Parameters of SVM are independently tuned using training folds data and best average
classification accuracies is reported for each method. We provide node degree as the labeled data for
algorithms that do not operate directly on unlabeled data. Further details about parameters selection
for baseline methods are present in supplementary materials.
4
https://github.com/vermaMachineLearning/FGSD
8
F GSD
(Wall-Time)
Dataset (No. Graphs,
Max. Nodes)
RW
[2003]
SP
[2005]
GK
[2009]
WL
[2011]
DGK
[2015]
MLG
(WallTime) [2016]
DCNN
[2016]
SGS
[2008]
MUTAG (188, 28)
83.50
87.23
84.04
87.28
86.17
87.23(5s)
66.51
88.61
92.12(0.3s)
PTC (344, 109)
55.52
58.72
60.17
55.61
59.88
62.20(18s)
55.79
?
62.80(0.07s)
PROTEINS (1113, 620)
68.46
72.14
71.78
70.06
71.69
71.35(277s)
65.22
?
73.42(5s)
NCI1 (4110, 111)
>D
68.15
62.07
77.23
64.40
77.57(620s)
63.10
62.72
79.80(31s)
NCI109 (4127, 111)
>D
68.30
62.04
78.43
67.14
75.91(600s)
60.67
62.62
78.84(35s)
D & D (1178, 5748)
>D
>D
75.05
73.76
72.75
77.02(7.5hr)
OMR
?
77.10(25s)
MAO (68, 27)
83.52
90.35
80.88
89.79
87.76
91.17(13s)
76.10
?
95.58(0.1s)
Table 2: Classification accuracy on unlabeled bioinformatics datasets. Results in bold indicate all
methods with accuracy within range 2.0 from the top result and blue color (for range > 2.0), indicates
the new state-of-art result. Green color highlights the best time computation, if it?s 5?faster (among
the mentioned). ?OMR? is out of memory error, ?> D? is computation exceed 24hrs.
Dataset
(Graphs)
GK
[2009]
DGK
[2015]
PSCN
[2016]
F GSD
Dataset
MLG
[2016]
DCNN
[2016]
PSCN
[2016]
GS
[2009]
F GSD *
COLLAB
(5000)
72.84
73.09
72.60
80.02
MUTAG
87.94
(4s)
66.98
92.63
(3s)
88.11
92.12
(0.3s)
REDDIT-B
(2000)
77.34
78.04
86.30
86.50
PTC
63.26
(21s)
56.60
62.90
(6s)
?
62.80
(0.07s)
REDDIT-M
(5000)
41.01
41.27
49.10
47.76
NCI1
81.75
(621s)
62.61
78.59
(76s)
65.0
79.80
(31s)
IMDB-B
(1000)
65.87
66.96
71.00
73.62
D&D
78.18
(7.5hr)
OMR
77.12
(154s)
?
77.10
(25s)
IMDB-M
(1500)
43.89
44.55
45.23
52.41
MAO
88.29
(12s)
75.14
?
?
95.58
(0.1s)
Table 3: Classification accuracy on social
network datasets. F GSD significantly outperforms other methods.
Table 4: Classification accuracy on labeled bioinformatics datasets. * emphasize that F GSD did not
utilize any node labels.
.
Classification Results: From Table 2, it is clear that F GSD consistently outperforms every other
state-of-art algorithms on unlabeled bioinformatics datasets and that too significantly in many cases.
F GSD even performs better for social network graphs as shown in Table 3 and achieves a very
significant 7% ? 8% more accuracy than the current state-of-art PSCNs on COLLAB and IMDB-M
datasets. Also from run-time perspective (excluding any data loading or classification time for all
algorithms), it is pretty fast (2x?1000x times faster) as compare to others. These appealing results
further motivated us to compare F GSD on the labeled datasets (even though, it is not a complete
fair comparison). Table 4 shows that F GSD is still very competitive with all other strong (recent)
algorithms that utilize node labeled data. Infact on MAO dataset, F GSD sets a new state-of-art result
and stays within 0% ? 2% range of accuracy from the best on all labeled datasets. On few labeled
datasets, we found MLG to have slightly better performance than the others, but it is 1000 times
slower than F GSD when graph size jumps to few thousand nodes (see D&D Results). Altogether,
F GSD shows very promising results in both accuracy & speed on all type of datasets and over all the
more sophisticated algorithms. These results also point out the fact that there is untapped hidden
potential in the graph structure which current algorithms are not harnessing despite having labeled
data at their disposal.
9
Conclusion
We present a conceptually simple yet powerful and theoretically motivated graph representation. In
particular, our graph representation based on the discovery of family of graph spectral distances can
exhibits uniqueness, stability, sparsity and are computationally fast. Moreover, our hunt specifically
leads to the harmonic and next to it, biharmonic distances as an ideal members of this family for
extracting graph features. Finally, our extensive results show that F GSD based graph features are
powerful enough to dominate the unlabeled graph classification task over all the more sophisticated
algorithms and competitive enough to yield high classification accuracy on labeled data even without
utilizing any node labels. In our future work, we plan to generalize the F GSD for labeled dataset in
order to utilize the useful node and edge label information in the graph representation.
9
10
Acknowledgments
This research was supported in part by ARO MURI Award W911NF-12-1-0385, DTRA grant
HDTRA1-14-1-0040, and NSF grants CNS-1618339, CNS-1618339 and CNS-1617729.
References
[1] J. Atwood and D. Towsley. Diffusion-convolutional neural networks. In Advances in Neural Information
Processing Systems, pages 1993?2001, 2016.
[2] L. Babai. Graph isomorphism in quasipolynomial time. CoRR, abs/1512.03547, 2015.
[3] M. Belkin and P. Niyogi. Laplacian eigenmaps for dimensionality reduction and data representation.
Neural computation, 15(6):1373?1396, 2003.
[4] K. M. Borgwardt and H.-P. Kriegel. Shortest-path kernels on graphs. In Data Mining, Fifth IEEE
International Conference on, pages 8?pp. IEEE, 2005.
[5] M. Botsch, D. Bommes, and L. Kobbelt. Efficient linear system solvers for mesh processing. In Mathematics
of Surfaces XI, pages 62?83. Springer, 2005.
[6] O. Bousquet and A. Elisseeff. Stability and generalization. Journal of Machine Learning Research, 2
(Mar):499?526, 2002.
[7] J. Bruna, W. Zaremba, A. Szlam, and Y. LeCun. Spectral networks and locally connected networks on
graphs. arXiv preprint arXiv:1312.6203, 2013.
[8] C.-C. Chang and C.-J. Lin. Libsvm: a library for support vector machines. ACM Transactions on Intelligent
Systems and Technology (TIST), 2(3):27, 2011.
[9] F. Costa and K. De Grave. Fast neighborhood subgraph pairwise distance kernel. In Proceedings of the
26th International Conference on Machine Learning, pages 255?262. Omnipress, 2010.
[10] H. Dai, B. Dai, and L. Song. Discriminative embeddings of latent variable models for structured data. In
Proceedings of the 33rd International Conference on International Conference on Machine Learning Volume 48, 2016.
[11] M. Defferrard, X. Bresson, and P. Vandergheynst. Convolutional neural networks on graphs with fast
localized spectral filtering. In Advances in Neural Information Processing Systems, pages 3837?3845,
2016.
[12] D. K. Duvenaud, D. Maclaurin, J. Iparraguirre, R. Bombarell, T. Hirzel, A. Aspuru-Guzik, and R. P. Adams.
Convolutional networks on graphs for learning molecular fingerprints. In Advances in neural information
processing systems, pages 2224?2232, 2015.
[13] T. G?rtner, P. Flach, and S. Wrobel. On graph kernels: Hardness results and efficient alternatives. In
Learning Theory and Kernel Machines, pages 129?143. Springer, 2003.
[14] A. Grover and J. Leskovec. node2vec: Scalable feature learning for networks. In Proceedings of the 22nd
ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2016.
[15] D. K. Hammond, P. Vandergheynst, and R. Gribonval. Wavelets on graphs via spectral graph theory.
Applied and Computational Harmonic Analysis, 30(2):129?150, 2011.
[16] M. Henaff, J. Bruna, and Y. LeCun. Deep convolutional networks on graph-structured data. arXiv preprint
arXiv:1506.05163, 2015.
[17] V. N. Katsikis, D. Pappas, and A. Petralias. An improved method for the computation of the moore?penrose
inverse matrix. Applied Mathematics and Computation, 217(23):9828?9834, 2011.
[18] T. N. Kipf and M. Welling. Semi-supervised classification with graph convolutional networks. arXiv
preprint arXiv:1609.02907, 2016.
[19] R. Kondor. A complete set of rotationally and translationally invariant features for images. CoRR,
abs/cs/0701127, 2007.
[20] R. Kondor and K. M. Borgwardt. The skew spectrum of graphs. In Proceedings of the 25th international
conference on Machine learning, pages 496?503. ACM, 2008.
[21] R. Kondor and H. Pan. The multiscale laplacian graph kernel. In Advances in Neural Information
Processing Systems, pages 2982?2990, 2016.
[22] R. Kondor, N. Shervashidze, and K. M. Borgwardt. The graphlet spectrum. In Proceedings of the 26th
Annual International Conference on Machine Learning, pages 529?536. ACM, 2009.
[23] Y. Lipman, R. M. Rustamov, and T. A. Funkhouser. Biharmonic distance. ACM Transactions on Graphics
(TOG), 29(3):27, 2010.
[24] G. Montavon, K. Hansen, S. Fazli, M. Rupp, F. Biegler, A. Ziehe, A. Tkatchenko, A. V. Lilienfeld, and
K.-R. M?ller. Learning invariant representations of molecules for atomization energy prediction. In
Advances in Neural Information Processing Systems, pages 440?448, 2012.
[25] B. Nadler, S. Lafon, R. Coifman, and I. Kevrekidis. Diffusion maps, spectral clustering and eigenfunctions
of fokker-planck operators. In NIPS, pages 955?962, 2005.
[26] M. Niepert, M. Ahmed, and K. Kutzkov. Learning convolutional neural networks for graphs. In Proceedings
of the 33rd annual international conference on machine learning. ACM, 2016.
[27] F. Orsini, P. Frasconi, and L. De Raedt. Graph invariant kernels. In IJCAI, pages 3756?3762, 2015.
[28] J. Rosenblatt and P. D. Seymour. The structure of homometric sets. SIAM Journal on Algebraic Discrete
Methods, 3(3):343?350, 1982.
10
[29] L. W. Shapiro. An electrical lemma. Mathematics Magazine, 60(1):36?38, 1987.
[30] N. Shervashidze, S. Vishwanathan, T. Petri, K. Mehlhorn, and K. M. Borgwardt. Efficient graphlet kernels
for large graph comparison. In AISTATS, volume 5, pages 488?495, 2009.
[31] N. Shervashidze, P. Schweitzer, E. J. v. Leeuwen, K. Mehlhorn, and K. M. Borgwardt. Weisfeiler-lehman
graph kernels. Journal of Machine Learning Research, 12(Sep):2539?2561, 2011.
[32] D. I. Shuman, S. K. Narang, P. Frossard, A. Ortega, and P. Vandergheynst. The emerging field of signal
processing on graphs: Extending high-dimensional data analysis to networks and other irregular domains.
IEEE Signal Processing Magazine, 30(3):83?98, 2013.
[33] E. T. Whittaker and G. N. Watson. A course of modern analysis. Cambridge university press, 1996.
[34] P. Yanardag and S. Vishwanathan. Deep graph kernels. In Proceedings of the 21th ACM SIGKDD
International Conference on Knowledge Discovery and Data Mining, pages 1365?1374. ACM, 2015.
11
| 6614 |@word kondor:5 inversion:1 polynomial:8 norm:2 loading:1 flach:1 nd:1 open:1 sg2:1 accounting:1 decomposition:3 elisseeff:1 kutzkov:1 sgd:1 mlk:1 accommodate:1 ld:1 reduction:5 substitution:1 series:2 contains:3 tist:1 tuned:1 interestingly:1 outperforms:3 past:1 current:4 comparing:1 com:1 si:1 yet:3 must:5 mesh:1 designed:1 graphlets:1 v:1 discovering:1 core:2 short:1 gribonval:1 provides:2 multiset:8 node:28 wxy:1 zhang:2 mehlhorn:2 along:3 schweitzer:1 direct:1 prove:1 manner:2 coifman:1 theoretically:3 node2vec:2 pairwise:9 hardness:1 expected:2 frossard:1 behavior:2 multi:2 inspired:4 globally:1 decreasing:11 zhi:1 enumeration:3 solver:1 increasing:4 xx:2 moreover:3 notation:1 bounded:7 kevrekidis:1 underlying:1 what:1 reddit:4 emerging:1 generalizable:1 guarantee:1 pseudo:2 every:1 unexplored:1 act:3 ti:6 isometrically:1 exactly:2 oscillates:1 walltime:1 zaremba:1 szlam:1 converse:1 grant:2 planck:1 before:2 local:10 seymour:1 consequence:1 despite:3 analyzing:2 establishing:1 path:7 abuse:1 chose:1 co:1 hunt:6 factorization:1 range:5 unique:16 practical:1 acknowledgment:1 lecun:2 atomic:3 lost:1 graphlet:5 definite:1 recursive:1 procedure:2 dice:1 empirical:1 drug:1 significantly:5 dictate:1 thought:1 revealing:1 word:1 refers:1 suggest:1 protein:3 get:2 unlabeled:7 selection:1 operator:1 influence:1 seminal:1 restriction:1 equivalent:1 map:2 center:1 attention:1 independently:1 focused:1 unstructured:1 rule:1 utilizing:1 dominate:1 stability:20 proving:1 notion:3 coordinate:1 embedding:8 today:1 magazine:2 exact:6 guzik:1 designing:2 origin:1 element:15 trend:1 expensive:2 particularly:2 updating:1 lay:1 aware:1 muri:1 labeled:12 observed:1 preprint:3 solved:2 capture:6 worst:2 thousand:1 electrical:1 connected:5 ordering:2 decrease:2 trade:1 mentioned:1 complexity:8 raise:1 solving:4 serve:2 upon:2 tog:1 f2:1 efficiency:2 imdb:5 sep:1 various:1 represented:1 heat:1 fast:13 effective:3 distinct:2 couldn:1 shervashidze:3 neighborhood:3 choosing:4 harnessing:1 quite:2 encoded:3 larger:2 solve:1 supplementary:3 grave:1 narang:1 otherwise:1 niyogi:1 gi:1 g1:8 transform:1 itself:3 sequence:1 eigenvalue:8 analytical:1 propose:1 aro:1 product:1 relevant:1 loop:1 subgraph:2 scalability:1 recipe:2 defferrard:2 ijcai:1 extending:1 produce:2 adam:1 derive:1 depending:1 strong:5 c:3 involves:1 implies:2 indicate:1 cnns:4 subsequently:1 filter:3 centered:1 modifying:1 successor:1 enable:1 material:1 adjacency:3 bin:3 premise:1 f1:1 generalization:3 fix:1 preliminary:1 wall:1 decompose:1 exploring:1 hold:4 around:1 considered:3 duvenaud:1 maclaurin:1 nadler:1 major:1 achieves:1 adopt:1 purpose:3 uniqueness:18 favorable:1 dgk:3 label:10 hansen:1 grouped:1 wl:2 create:1 city:2 successfully:2 interatomic:1 weighted:3 tool:2 clearly:1 dcnn:4 always:3 rather:1 pn:1 avoid:1 encode:2 derived:1 focus:3 consistently:1 rank:1 indicates:4 sigkdd:2 baseline:1 sense:2 inference:1 dependent:1 hidden:1 interested:3 overall:3 classification:28 among:3 aforementioned:1 denoted:1 nci1:3 plan:1 art:10 saurabh:1 construct:1 field:2 frasconi:1 beach:1 atom:1 hop:1 identical:2 represents:2 having:1 look:1 lipman:1 future:1 hdtra1:1 np:1 others:3 intelligent:1 inherent:5 few:6 tangential:1 modern:1 belkin:1 composed:1 tightly:1 babai:2 comprehensive:1 zoom:1 translationally:1 cns:3 maintain:2 n1:2 ab:2 interest:2 fd:1 investigate:2 mining:3 umn:2 amenable:1 implication:6 edge:7 capable:1 xy:7 respective:1 conforms:1 orthogonal:1 tree:1 euclidean:6 walk:3 desired:1 theoretical:2 leskovec:1 leeuwen:1 instance:2 column:3 cover:1 patchy:1 w911nf:1 measuring:1 bresson:1 raedt:1 cost:1 introducing:1 vertex:17 entry:4 deviation:1 uniform:4 eigenmaps:1 gr:1 too:2 graphic:1 reported:1 connect:2 st:1 borgwardt:5 fundamental:2 international:9 siam:1 stay:1 off:1 invertible:1 together:3 w1:1 atwood:1 opposed:1 containing:1 possibly:2 slowly:1 fazli:1 weisfeiler:3 derivative:1 li:5 account:3 potential:1 de:2 twin:2 bold:1 lehman:3 satisfy:2 untapped:1 bombarell:1 explicitly:2 biharmonic:13 depends:1 vi:2 performed:1 view:1 towsley:1 hirzel:1 competitive:4 recover:3 start:1 contribution:2 formed:3 accuracy:11 convolutional:10 likewise:1 efficiently:4 yield:6 conceptually:2 generalize:1 produced:2 hammond:1 shuman:1 definition:3 mlg:4 evaluates:1 against:1 nonetheless:1 energy:1 pp:1 e2:1 proof:2 di:1 boil:1 rational:1 costa:1 dataset:8 popular:1 recall:1 color:3 knowledge:2 dimensionality:1 lilienfeld:1 sophisticated:3 uncover:1 back:2 feed:1 disposal:1 hashing:1 higher:1 isometric:2 supervised:1 improved:1 though:2 mar:1 generality:1 pappa:1 niepert:1 lastly:2 until:1 hand:1 replacing:2 ei:1 multiscale:3 mutag:5 propagation:1 lack:1 iparraguirre:1 quality:2 usa:1 validity:1 normalized:1 true:2 hence:1 moore:3 funkhouser:1 eg:1 deal:2 self:1 uniquely:3 maintained:1 won:1 ortega:1 bijective:3 theoretic:2 demonstrate:1 complete:6 performs:1 omnipress:1 reasoning:1 image:1 harmonic:18 novel:2 recently:1 fi:3 superior:1 common:1 empirically:2 volume:2 million:2 belong:1 extend:1 slight:1 refer:1 significant:1 cambridge:1 ai:2 attenuate:1 approx:1 rd:2 mathematics:3 similarly:1 had:1 dot:2 minnesota:2 bruna:2 stable:6 kipf:1 similarity:1 operating:1 surface:1 fingerprint:1 enzyme:2 recent:2 perspective:3 henaff:1 driven:1 certain:8 collab:3 binary:2 watson:1 gsd:57 seen:1 rotationally:1 fortunately:1 dai:2 impose:1 dtra:1 employed:2 determine:1 shortest:6 ller:1 signal:4 semi:2 smoother:1 multiple:1 desirable:3 faster:3 ahmed:1 pscn:3 long:1 cross:1 lin:1 concerning:1 e1:1 award:1 molecular:1 laplacian:13 controlled:1 prediction:1 variant:2 basic:1 scalable:1 metric:1 arxiv:6 histogram:8 kernel:25 represent:1 irregular:1 preserved:3 addition:2 want:2 tkatchenko:1 grow:1 singular:1 w2:1 eliminates:1 specially:1 posse:1 equal:2 strict:1 operate:1 pooling:1 eigenfunctions:1 undirected:2 member:8 leveraging:1 coarsening:1 kobbelt:1 extracting:3 structural:7 presence:2 ideal:7 leverage:1 exceed:1 enough:7 concerned:2 embeddings:1 variety:1 restrict:1 reduce:1 regarding:1 chebyshev:1 i7:1 whether:5 motivated:3 utility:1 isomorphism:10 gb:1 effort:1 song:1 resistance:3 algebraic:1 e3:1 nine:2 deep:4 useful:2 generally:1 clear:2 eigenvectors:7 band:1 ten:2 locally:1 category:6 rw:2 generate:1 http:1 outperform:1 exist:1 shapiro:1 nsf:1 per:2 yy:2 blue:3 broadly:1 xth:1 rosenblatt:1 discrete:1 group:2 key:1 nevertheless:1 enormous:1 traced:1 achieving:1 ptc:3 libsvm:2 diffusion:5 utilize:5 v1:1 ram:1 graph:222 monotone:2 convert:1 run:1 inverse:5 powerful:6 family:15 summarizes:1 scaling:1 layer:1 bound:2 distinguish:1 correspondence:2 nci109:2 fold:2 nonnegative:2 g:3 annual:2 occur:1 precisely:1 vishwanathan:2 ri:4 encodes:2 bousquet:1 fourier:1 speed:2 argument:1 performing:1 conjecture:5 department:2 developing:2 according:2 structured:2 smaller:1 slightly:1 pan:1 intimately:1 wi:1 lp:5 appealing:1 yanardag:1 modification:1 making:1 invariant:10 pr:2 computationally:3 remains:2 rtner:1 skew:3 discus:1 structure2vec:1 needed:1 fed:1 end:1 adopted:2 available:2 operation:2 cogent:1 apply:1 obey:1 observe:2 v2:1 spectral:28 upto:2 appropriate:1 alternative:1 shortly:1 eigen:1 slower:1 altogether:1 top:1 clustering:1 maintaining:1 unifying:1 build:1 approximating:1 question:2 already:2 strategy:1 receptive:1 sxy:12 usual:1 diagonal:2 unclear:1 exhibit:4 distance:61 reason:2 spanning:1 rupp:1 assuming:1 besides:1 code:1 index:6 relationship:5 balance:1 setup:1 unfortunately:1 lg:3 susceptible:1 potentially:2 statement:1 gk:4 unknown:1 perform:3 upper:1 convolution:2 observation:1 datasets:19 benchmark:1 defining:1 extended:1 excluding:1 precise:1 perturbation:6 arbitrary:1 bk:3 pair:5 extensive:2 connection:1 narrow:1 tremendous:1 nip:2 eigenfunction:1 able:2 kriegel:1 below:1 sparsity:18 including:1 max:1 green:1 memory:1 suitable:4 rely:1 petri:1 hr:3 github:1 technology:1 library:2 numerous:1 lk:3 multisets:6 sg:3 discovery:5 kf:1 determining:1 embedded:1 fully:2 loss:1 permutation:7 beside:1 highlight:1 generation:2 filtering:1 grover:1 localized:2 vandergheynst:3 validation:1 degree:4 sufficient:1 verma:2 bypass:1 atleast:2 row:3 course:1 summary:1 sg1:1 supported:1 wide:1 aspuru:1 taking:1 fifth:1 sparse:7 ghz:1 dimension:5 canonicalized:1 lafon:1 unweighted:2 collection:1 made:1 san:1 jump:1 far:1 social:4 transaction:2 welling:1 approximate:2 emphasize:1 implicitly:1 global:6 pseudoinverse:1 reveals:2 assumed:1 conclude:1 xi:1 discriminative:1 chemoinformatics:1 spectrum:33 infact:2 biegler:1 latent:1 decade:1 pretty:1 table:8 promising:1 learn:2 delving:1 ca:1 robust:2 molecule:2 constructing:4 domain:2 diag:4 sp:3 did:1 main:3 aistats:1 noise:2 fair:2 rustamov:1 x1:1 intel:1 atomization:1 sub:8 mao:5 binwidth:4 explicit:4 sf:29 candidate:2 tied:1 third:1 wavelet:5 montavon:1 theorem:15 down:3 removing:1 embed:1 wrobel:1 showing:2 explored:1 dk:1 svm:2 admits:2 concern:1 evidence:1 exists:1 intrinsic:2 essential:1 adding:1 corr:2 suited:1 generalizing:1 depicted:1 intersection:1 simply:1 penrose:3 lagrange:1 expressed:1 contained:1 g2:8 chang:1 monotonic:3 springer:2 fokker:1 determines:1 satisfies:3 relies:1 acm:8 whittaker:1 identity:1 viewed:2 sorted:2 loosing:1 towards:2 change:5 specifically:3 determined:1 lemma:1 contradictory:1 called:1 total:2 isomorphic:9 invariance:3 pas:1 experimental:1 lrw:1 ziehe:1 omr:3 support:3 cholesky:1 bioinformatics:5 eigenmap:1 evaluate:2 |
6,207 | 6,615 | Scalable Generalized Linear Bandits:
Online Computation and Hashing
Kwang-Sung Jun
UW-Madison
[email protected]
Aniruddha Bhargava
UW-Madison
[email protected]
Robert Nowak
UW-Madison
[email protected]
Rebecca Willett
UW-Madison
[email protected]
Abstract
Generalized Linear Bandits (GLBs), a natural extension of the stochastic linear
bandits, has been popular and successful in recent years. However, existing GLBs
scale poorly with the number of rounds and the number of arms, limiting their
utility in practice. This paper proposes new, scalable solutions to the GLB problem
in two respects. First, unlike existing GLBs, whose per-time-step space and time
complexity grow at least linearly with time t, we propose a new algorithm that
performs online computations to enjoy a constant space and time complexity. At
its heart is a novel Generalized Linear extension of the Online-to-confidence-set
Conversion (GLOC method) that takes any online learning algorithm and turns it
into a GLB algorithm. As a special case, we apply GLOC to the online Newton
step algorithm, which results in a low-regret GLB algorithm with much lower
time and memory complexity than prior work. Second, for the case where the
number N of arms is very large, we propose new algorithms in which each next
arm is selected via an inner product search. Such methods can be implemented
via hashing algorithms (i.e., ?hash-amenable?) and result in a time complexity
sublinear in N . While a Thompson sampling extension of GLOC is hash-amenable,
its regret bound for d-dimensional arm sets scales with d3/2 , whereas GLOC?s
regret bound scales with d. Towards closing this gap, we propose a new hashamenable algorithm whose regret bound scales with d5/4 . Finally, we propose a
fast approximate hash-key computation (inner product) with a better accuracy than
the state-of-the-art, which can be of independent interest. We conclude the paper
with preliminary experimental results confirming the merits of our methods.
1
Introduction
This paper considers the problem of making generalized linear bandits (GLBs) scalable. In the
stochastic GLB problem, a learner makes successive decisions to maximize her cumulative rewards.
Specifically, at time t the learner observes a set of arms Xt ? Rd . The learner then chooses an arm
?
xt ? Xt and receives a stochastic reward yt that is a noisy function of xt : yt = ?(x>
t ? ) + ?t , where
?
? ? Rd is unknown, ?:R?R is a known nonlinear mapping, and ?t ? R is some zero-mean noise.
This reward structure encompasses generalized linear models [29]; e.g., Bernoulli, Poisson, etc.
The key aspect of the bandit problem is that the learner does not know how much reward she would
have received, had she chosen another arm. The estimation on ? ? is thus biased by the history of the
selected arms, and one needs to mix in exploratory arm selections to avoid ruling out the optimal
arm. This is well-known as the exploration-exploitation dilemma. The performance of a learner is
evaluated by its regret that measures how much cumulative reward she would have gained additionally
if she had known the true ? ? . We provide backgrounds and formal definitions in Section 2.
A linear case of the problem above (?(z) = z) is called the (stochastic) linear bandit problem. Since
the first formulation of the linear bandits [7], there has been a flurry of studies on the problem [11,
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
34, 1, 9, 5]. In an effort to generalize the restrictive linear rewards, Filippi et al. [15] propose the
GLB problem and provide a low-regret algorithm, whose Thompson sampling version appears later
in Abeille & Lazaric [3]. Li et al. [27] evaluates GLBs via extensive experiments where GLBs exhibit
lower regrets than linear bandits for 0/1 rewards. Li et al. [28] achieves a smaller regret bound when
the arm set Xt is finite, though with an impractical algorithm.
However, we claim that all existing GLB algorithms [15, 28] suffer from two scalability issues that
limit their practical use: (i) under a large time horizon and (ii) under a large number N of arms.
First, existing GLBs require storing all the arms and rewards appeared so far, {(xs , ys )}ts=1 , so
the space complexity grows linearly with t. Furthermore, they have to solve a batch optimization
problem for the maximum likelihood estimation (MLE) at each time step t whose per-time-step time
complexity grows at least linearly with t. While Zhang et al. [41] provide a solution whose space
and time complexity do not grow over time, they consider a specific 0/1 reward with the logistic link
function, and a generic solution for GLBs is not provided.
Second, existing GLBs have linear time complexities in N . This is impractical when N is very large,
which is not uncommon in applications of GLBs such as online advertisements, recommendation
systems, and interactive retrieval of images or documents [26, 27, 40, 21, 25] where arms are items in
a very large database. Furthermore, the interactive nature of these systems requires prompt responses
as users do not want to wait. This implies that the typical linear time in N is not tenable. Towards a
sublinear time in N , locality sensitive hashings [18] or its extensions [35, 36, 30] are good candidates
as they have been successful in fast similarity search and other machine learning problems like active
learning [22], where the search time scales with N ? for some ? < 1 (? is usually optimized and
often ranges from 0.4 to 0.8 depending on the target search accuracy). Leveraging hashing in GLBs,
however, relies critically on the objective function used for arm selections. The function must take a
form that is readily optimized using existing hashing algorithms.1 For example, algorithms whose
objective function (a function of each arm x ? Xt ) can be written as a distance or inner product
between x and a query q are hash-amenable as there exist hashing methods for such functions.
To be scalable to a large time horizon, we propose a new algorithmic framework called Generalized
Linear Online-to-confidence-set Conversion (GLOC) that takes in an online learning (OL) algorithm
with a low ?OL? regret bound and turns it into a GLB algorithm with a low ?GLB? regret bound. The
key tool is a novel generalization of the online-to-confidence-set conversion technique used in [2]
(also similar to [14, 10, 16, 41]). This allows us to construct a confidence set for ? ? , which is then
used to choose an arm xt according to the well-known optimism in the face of uncertainty principle.
By relying on an online learner, GLOC inherently performs online computations and is thus free from
the scalability issues in large time steps. While any online learner equipped with a low OL regret
bound can be used, we choose the online Newton step (ONS) algorithm and prove a tight OL regret
bound, which results in a practical GLB algorithm with almost the same regret bound as existing
inefficient GLB algorithms. We present our proposed algorithms and their regret bounds in Section 3.
For large number N of arms, our proposed algorithm
Algorithm
Regret
Hash-amenable
?
GLOC is not hash-amenable, to our knowledge, due
?
GLOC
O(d
T
)
7
?
to its nonlinear criterion for arm selection. As the first
? 3/2 T )
GLOC-TS O(d
3
?
attempt, we derive a Thompson sampling [5, 3] exten? 5/4 T )
QGLOC
O(d
3
sion of GLOC (GLOC-TS), which is hash-amenable
due to its linear criterion. However, its regret bound Table 1: Comparison of GLBs algorithms for
scales with d3/2 for d-dimensional arm sets, which d-dimensional arm sets T is the time horizon.
is far from d of GLOC. Towards closing this gap, we QGLOC achieves the smallest regret among
propose a new algorithm Quadratic GLOC (QGLOC) hash-amenable algorithms.
with a regret bound that scales with d5/4 . We summarize the comparison of our proposed GLB
algorithms in Table 1. In Section 4, we present GLOC-TS, QGLOC, and their regret bound.
Note that, while hashing achieves a time complexity sublinear in N , there is a nontrivial overhead
of computing the projections to determine the hash keys. As an extra contribution, we reduce this
overhead by proposing a new sampling-based approximate inner product method. Our proposed
sampling method has smaller variance than the state-of-the-art sampling method proposed by [22, 24]
when the vectors are normally distributed, which fits our setting where projection vectors are indeed
normally distributed. Moreover, our method results in thinner tails in the distribution of estimation
1
Without this designation, no currently known bandit algorithm achieves a sublinear time complexity in N .
2
error than the existing method, which implies a better concentration. We elaborate more on reducing
the computational complexity of QOFUL in Section 5.
2
Preliminaries
We review relevant backgrounds here. A refers to a GLB algorithm, and B refers to an online
learning algorithm. Let Bd (S) be the d-dimensional Euclidean ball of radius S, which
? overloads
the notation B. Let A?i be the i-th column vector of a matrix A. Define ||x||A := x> Ax and
2
vec(A) := [A?1 ; A?2 ; ? ? ? ; A?d ] ? Rd Given a function f : R ? R, we denote by f 0 and f 00 its
first and second derivative, respectively. We define [N ] := {1, 2, . . . , N }.
Generalized Linear Model (GLM) Consider modeling the reward y as one-dimensional exponential family such as Bernoulli or Poisson. When the feature vector x is believed to correlate with
y, one popular modeling assumption is the generalized linear model (GLM) that turns the natural
parameter of an exponential family model into x> ? ? where ? ? is a parameter
[29]:
yz
?
m(z)
+ h(y, ? ) ,
(1)
P(y | z = x> ? ? ) = exp
g(? )
where ? ? R+ is a known scale parameter and m, g, and h are normalizers. It is known that
m0 (z) = E[y | z] =: ?(z) and m00 (z) = Var(y | z). We call ?(z) the inverse link function.
Throughout, we assume that the exponential family being used in a GLM has a minimal representation,
which ensures that m(z) is strictly convex [38, Prop. 3.1]. Then, the negative log likelihood (NLL)
`(z, y) := ?yz + m(z) of a GLM is strictly convex. We refer to such GLMs as the canonical GLM.
In the case of Bernoulli rewards y ? {0, 1}, m(z) = log(1 + exp(z)), ?(z) = (1 + exp(?z))?1 ,
?
0
and the NLL can be written as the logistic loss: log(1 + exp(?y 0 (x>
t ? ))), where y = 2y ? 1.
Generalized Linear Bandits (GLB) Recall that xt is the arm chosen at time t by an algorithm.
We assume that the arm set Xt can be of an infinite cardinality, although we focus on finite arm sets in
hashing part of the paper (Section 4). One can write down the reward model (1) in a different form:
?
(2)
yt = ?(x>
t ? ) + ?t ,
t?1
where ?t is conditionally R-sub-Gaussian given xt and {(xs , ?s )}s=1 . For example, Bernoulli
?
> ?
> ?
reward model has ?t as 1 ? ?(x>
t ? ) w.p. ?(xt ? ) and ??(xt ? ) otherwise. Assume that
?
||? ||2 ? S, wherepS is known.
? One can show that the sub-Gaussian scale R is determined by ?:
0
R = supz?(?S,S) ? (z) ? L, where L is the Lipschitz constant of ?. Throughout, we assume
that each arm has `2 -norm at most 1: ||x||2 ? 1, ?x ? Xt , ?t. Let xt,? := maxx?Xt x> ? ? . The
performance of a GLB algorithm A is analyzed by the expected cumulative regret (or simply regret):
PT
> ?
A > ?
A
RegretA
T :=
t=1 ?(xt,? ? ) ? ?((xt ) ? ), where xt makes the dependence on A explicit.
We remark that our results in this paper hold true for a strictly larger family of distributions than the
canonical GLM, which we call the non-canonical GLM and explain below. The condition is that the
reward model follows (2) where the R is now independent from ? that satisfies the following:
Assumption 1. ? is L-Lipschitz on [?S, S] and continuously differentiable on (?S, S). Furthermore,
inf z?(?S,S) ?0 (z) = ? for some finite ? > 0 (thus ? is strictly increasing).
Define ?0 (z) at ?S as their limits. Under Assumption 1, m is defined to be an integral of ?. Then,
one can show that m is ?-strongly convex on B1 (S). An example of the non-canonical GLM is the
probit model for 0/1 reward where ? is the Gaussian CDF, which is popular and competitive to the
Bernoulli GLM as evaluated by Li et al. [27]. Note that canonical GLMs satisfy Assumption 1.
3
Generalized Linear Bandits with Online Computation
We describe and analyze a new GLB algorithm called Generalized Linear Online-to-confidence-set
Conversion (GLOC) that performs online computations, unlike existing GLB algorithms.
GLOC employs the optimism in the face of uncertainty principle, which dates back to [7]. That is, we
maintain a confidence set Ct (defined below) that traps the true parameter ? ? with high probability
(w.h.p.) and choose the arm with the largest feasible reward given Ct?1 as a constraint:
? t ) := arg
(xt , ?
max
hx, ?i
(3)
x?Xt ,??Ct?1
The main difference between GLOC and existing GLBs is in the computation of the Ct ?s. Prior
methods involve ?batch" computations that involve all past observations, and so scale poorly with
3
t. In contrast, GLOC takes in an online learner B, and uses B as a co-routine instead of relying on
a batch procedure to construct a confidence set. Specifically, at each time t GLOC feeds the loss
function `t (?) := `(x>
t ?, yt ) into the learner B which then outputs its parameter prediction ? t . Let
t?d
Xt ? R
be the design matrix consisting of x1 , . . . , xt . Define Vt := ?I + X>
t Xt , where ?
bt := V?1 X> zt be the ridge
is the ridge parameter. Let zt := x>
?
and
z
:=
[z
;
?
?
?
;
z
].
Let
?
t
t
1
t
t
t
t
regression estimator taking zt as responses. Theorem 1 below is the key result for constructing our
confidence set Ct , which is a function of the parameter predictions {? s }ts=1 and the online (OL)
regret bound Bt of the learner B. All the proofs are in the supplementary material (SM).
Theorem 1. (Generalized Linear Online-to-Confidence-Set Conversion) Suppose we feed loss functions {`s (?)}ts=1 into online learner B. Let ? s be the parameter predicted at time step s by B.
Assume that B has an OL regret bound
Pt Bt : ?? ? Bd (S), ?t ? 1,
(4)
`s (? s ) ? `s (?) ? Bt .
qs=1
2
4
2
Let ?(Bt ) := 1 + ?4 Bt + 8R
1 + ?2 Bt + ?4R
4 ? 2 ). Then, with probability (w.p.) at least 1 ? ?,
?2 log( ?
?
2
bt || ? ?(Bt ) + ?S 2 ? ||zt ||2 ? ?
b> X> zt =: ?t .
?t ? 1, ||? ? ?
(5)
2
t
t
Vt
Note that the center of the ellipsoid is the ridge regression estimator on the predicted natural
parameters zs = x>
s ? s rather than the rewards. Theorem 1 motivates the following confidence set:
bt ||2 ? ?t }
Ct := {? ? Rd : ||? ? ?
(6)
Vt
?
which traps ? for all t ? 1, w.p. at least 1 ? ?. See Algorithm 1 for pseudocode. One way to solve
the optimization problem (3) is to define the function ?(x) := max??Ct?1 x> ?, and then use the
Lagrangian method to write:
p
bt?1 + ?t?1 ||x|| ?1 .
xGLOC
:= arg max x> ?
(7)
t
Vt?1
x?X
t
We prove the regret bound of GLOC in the following theorem.
Theorem 2. Let {? t } be a nondecreasing sequence
such that ? t ??t . Then, w.p. at least 1 ? ?,
q
RegretGLOC
= O L ? T dT log T
T
Although any low-regret online learner can be
Algorithm 1 GLOC
combined with GLOC, one would like to ensure
1: Input: R > 0, ? ? (0, 1), S > 0, ? > 0, ? > 0,
that ? T is O(polylog(T )) in which
? case the total
an online learner B with known regret bounds
?
regret can be bounded by O( T ). This means
{Bt }t?1 .
that we must use online learners whose OL regret 2: Set V = ?I.
0
grows logarithmically in T such as [20, 31]. In 3: for t = 1, 2, . . . do
this work, we consider the online Newton step 4: Compute xt by solving (3).
(ONS) algorithm [20].
5:
Pull xt and then observe yt .
6:
Receive ? from B.
5:
Output ? t .
t
Online Newton Step (ONS) for Generalized 7: Feed into B
the loss `t (?) = `(x>
t ?, yt ).
Linear Models Note that ONS requires the loss 8: Update V = V
>
>
t
t?1 + xt xt and zt = xt ? t
functions to be ?-exp-concave. One can show
?1 >
b
9:
Compute ? t = Vt Xt zt and ?t as in (5).
that `t (?) is ?-exp-concave [20, Sec. 2.2]. Then, 10: Define C as
in (6).
t
GLOC can use ONS and its OL regret bound to 11: end for
solve the GLB problem. However, motivated by
the fact?
that the OL regret bound Bt appears in the Algorithm 2 ONS-GLM
radius ?t of the confidence set while a tighter 1: Input: ? > 0, > 0, S > 0.
confidence set tends to reduce the bandit regret 2: A0 = I.
in practice, we derive a tight data-dependent OL 3: Set ?1 ? Bd (S) arbitrarily.
4: for t = 1, 2, 3, . . . do
regret bound tailored to GLMs.
We present our version of ONS for GLMs (ONS- 6: Observe xt and yt .
GLM) in Algorithm 2. `0 (z, y) is the first deriva- 7: Incur loss `(x>
t ? t , yt ) .
tive w.r.t. z and the parameter is for inverting 8: At = At?1 + xt x>
t
matrices conveniently (usually = 1 or 0.1). The
`0 (x>
t ? t ,yt )
9:
? 0t+1 = ? t ?
A?1
t xt
?
only difference from the original ONS [20] is that
10:
? t+1 = arg min??Bd (S) ||? ? ? 0t+1 ||2At
we rely on the strong convexity of m(z) instead
11: end for
of the ?-exp-concavity of the loss thanks to the
2
GLM structure. Theorem 3 states that we achieve the desired polylogarithmic regret in T .
2
A similar change to ONS has been applied in [16, 41].
4
Theorem 3. Define gs := `0 (x>
s ? s , ys ). The regret of ONS-GLM satisfies, for any > 0 and t ? 1,
Pt
Pt
?
1
2
2
2
ONS
,
s=1 `s (? s ) ? `s (? ) ? 2?
s=1 gs ||xs ||A?1 + 2?S =: Bt
s
2
2
where BtONS = O( L +R? log(t) d log t), ?t
2
?2
R
BtONS = O( L +
d log t).
?
? w.p. 1,
? 1 w.h.p. If maxs?1 |?s | is bounded by R
We emphasize that the OL regret bound is data-dependent. A confidence set constructed by combining
Theorem 1 and Theorem 3 directly implies the following regret bound of GLOC with ONS-GLM.
Corollary 1. Define ?tONS by replacing Bt with BtONS in (5). With probability at least 1 ? 2?,
n
o
bt ||2 ? ? ONS .
?t ? 1, ? ? ? C ONS := ? ? Rd : ||? ? ?
(8)
t
Vt
t
Corollary
2. Run GLOC with
Then, w.p. at least 1 ? 2?, ?T ? 1, RegretGLOC
=
T
?
L(L+R)
3/2
GLOC
?
?
?
=
O
d T log (T ) where O ignores log log(t). If |?t | is bounded by R, RegretT
? ? ?
? L(L+R) d T log(T ) .
O
?
CtONS .
We make regret bound comparisons ignoring log log T factors. For generic arm sets, our dependence
on d is optimal for linear rewards [34]. For the Bernoulli GLM, our regret has the same order as Zhang
et al. [41]. One can show that the regret of Filippi et al. [15] has the same order as ours if we use their
assumption that
by Rmax . For unbounded noise, Li et al. [28] have regret
? the reward yt is bounded
?
O((LR/?)d T log T ), which is log T factor smaller than ours and has LR in place of L(L + R).
While L(L + R) could be an artifact of our analysis, the gap
? is not too large for canonical GLMs.
Let L be the smallest Lipschitz constant of ?. Then, R = L. If L ? 1, R satisfies R > L, and so
L(L + R) = O(LR). If L > 1, then L(L + R) = O(L2 ), which is larger than LR = O(L3/2 ). For
the Gaussian GLM with known ?
variance ? 2 , L = R = 1.3 For finite arm sets, SupCB-GLM of Li
? dT log N ) that has a better scaling with d but is not a practical
et al. [28] achieves regret of O(
algorithm as it wastes a large number of arm pulls. Finally, we remark that none of the existing GLB
algorithms are scalable to large T . Zhang et al. [41] is scalable to large T , but is restricted to the
Bernoulli GLM; e.g., theirs does not allow the probit model (non-canonical GLM) that is popular and
shown to be competitive to the Bernoulli GLM [27].
Discussion The trick of obtaining a confidence set from an online learner appeared first in [13, 14]
for the linear model, and then was used in [10, 16, 41]. GLOC is slightly different from these studies
and rather close to Abbasi-Yadkori et al. [2] in that the confidence set is a function of a known regret
bound. This generality frees us from re-deriving a confidence set for every online learner. Our result
is essentially a nontrivial extension of Abbasi-Yadkori et al. [2] to GLMs.
One might have notice that Ct does not use ? t+1 that is available before pulling xt+1 and has the
most up-to-date information. This is inherent to GLOC as it relies on the OL regret bound directly.
One can modify the proof of ONS-GLM to have a tighter confidence set Ct that uses ? t+1 as we
show in SM Section E. However, this is now specific to ONS-GLM, which looses generality.
4
Hash-Amenable Generalized Linear Bandits
We now turn to a setting where the arm set is finite but very large. For example, imagine an interactive
retrieval scenario [33, 25, 6] where a user is shown K images (e.g., shoes) at a time and provides
relevance feedback (e.g., yes/no or 5-star rating) on each image, which is repeated until the user is
satisfied. In this paper, we focus on showing one image (i.e., arm) at a time.4 Most existing algorithms
require maximizing an objective function (e.g., (7)), the complexity of which scales linearly with the
number N of arms. This can easily become prohibitive for large numbers of images. Furthermore,
the system has to perform real-time computations to promptly choose which image to show the user
in the next round. Thus, it is critical for a practical system to have a time complexity sublinear in N .
One naive approach is to select a subset of arms ahead of time, such as volumetric spanners [19].
However, this is specialized for an efficient exploration only and can rule out a large number of
good arms. Another option is to use hashing methods. Locality-sensitive hashing and Maximum
3
The reason why R is not ? here is that the sufficient statistic of the GLM is y/?, which is equivalent to
dealing with the normalized reward. Then, ? appears as a factor in the regret bound.
4
One image at a time is a simplification of the practical setting. One can extend it to showing multiple
images at a time, which is a special case of the combinatorial bandits of Qin et al. [32].
5
Inner Product Search (MIPS) are effective and well-understood tools but can only be used when the
objective function is a distance or an inner product computation; (7) cannot be written in this form.
In this section, we consider alternatives to GLOC which are compatible with hashing.
Thompson Sampling We present a Thompson sampling (TS) version of GLOC called GLOC-TS
bt?1 , ?t?1 V?1 ). TS is known to
that chooses an arm xt = arg maxx?Xt x> ?? t where ?? t ? N (?
t?1
perform well in practice [8] and can solve the polytope arm set case in polynomial time5 whereas
algorithms that solve an objective function like (3) (e.g., [1]) cannot since they have to solve an
NP-hard problem [5]. We present the regret bound of GLOC-TS below. Due to space constraints, we
present the pseudocode and the full version of the result in SM.
bt?1 , ? ONS V?1 ), RegretGLOC-TS =
Theorem
4. (Informal) If we run GLOC-TS with ?? t ? N (?
t?1 t?1
?
? T
?
L(L+R) 3/2
3/2
?
?
?
O
T log (T ) w.h.p. If ?t is bounded by R, then O L(L+R) d3/2 T log(T ) .
d
?
?
3/2
Notice that the regret now scales with d
as expected from the analysis of linear TS [4], which
is higher than scaling with d of GLOC. This is concerning in the interactive retrieval or product
recommendation scenario since the relevance of the shown items is harmed, which makes us wonder
if one can improve the regret without loosing the hash-amenability.
Quadratic GLOC We now propose a new hash-amenable algorithm called Quadratic GLOC
(QGLOC). Recall that GLOC chooses the arm xGLOC by (7). Define r = minx?X ||x||2 and
mt?1 :=
min
||x||V?1 ,
(9)
x:||x||2 ?[r,1]
t?1
?1
which is r times the square root of the smallest eigenvalue of Vt?1 . It is easy to see that mt?1 ?
?
||x||V?1 for all x ? X and that mt?1 ? r/ t + ? using the definition of Vt?1 . There is an
t?1
alternative way to define mt?1 without relying on r, which we present in SM.
Let c0 > 0 be the exploration-exploitation tradeoff parameter (elaborated upon later). At time t,
QGLOC chooses the arm
1/4
bt?1 , xi + ?t?1 ||x||2 ?1 = arg max hqt , ?(x)i ,
(10)
xQGLOC
:=
arg
max
h
?
t
Vt?1
x?Xt
x?Xt
4c0 mt?1
?
1/4
?1
2
bt?1 ; vec( t?1 V )] ? Rd+d and ?(x) := [x; vec(xx> )]. The key property of
where qt = [?
t?1
4c0 mt?1
QGLOC is that the objective function is now quadratic in x, thus the name Quadratic GLOC, and
can be written as an inner product. Thus, QGLOC is hash-amenable. We present the regret bound of
QGLOC (10) in Theorem 5. The key step of the proof is that the QGLOC objective function (10)
plus c0 ? 3/4 mt?1 is a tight upper bound of the GLOC objective function (7).
at least 1 ? 2?, RegretQGLOC
=
Theorem
5. Run QGLOC with CtONS . Then, w.p.
T
?
2
1 L+R 1/2
L+R 3/2
L+R ?1/2
5/4
O
+
c
Ld
T
log
(T
)
.
By
setting
c
=
,
the
regret
0
0
c0
?
?
?
?
L(L+R) 5/4
2
bound is O( ? d
T log (T )).
Note that one can have a better dependence on log T when ?t is bounded (available in the proof).
The regret bound of QGLOC is a d1/4 factor improvement over that of GLOC-TS; see Table 1.
Furthermore, in (10) c0 is a free parameter that adjusts the balance between the exploitation (the first
term) and exploration (the second term). Interestingly, the regret guarantee does not break down when
adjusting c0 in Theorem 5. Such a characteristic is not found in existing algorithms but is attractive
to practitioners, which we elaborate in SM.
Maximum Inner Product Search (MIPS) Hashing While MIPS hashing algorithms such as [35,
36, 30] can solve (10) in time sublinear in N , these necessarily introduce an approximation error.
Ideally, one would like the following guarantee on the error with probability at least 1 ? ?H :
0
? ? X is called cH -MIPS w.r.t. a given
Definition 1. Let X ? Rd satisfy |X | < ?. A data point x
? i ? cH ? maxx?X hq, xi for some cH < 1. An algorithm is called cH -MIPS
query q if it satisfies hq, x
0
if, given a query q ? Rd , it retrieves x ? X that is cH -MIPS w.r.t. q.
Unfortunately, existing MIPS algorithms do not directly offer such a guarantee, and one must build a
series of hashing schemes with varying hashing parameters like Har-Peled et al. [18]. Under the fixed
budget setting T , we elaborate our construction that is simpler than [18] in SM.
5
ConfidenceBall1 algorithm of Dani et al. [11] can solve the problem in polynomial time as well.
6
Time and Space Complexity Our construction involves saving Gaussian projection vectors that
are used for determining hash keys and saving the buckets containing pointers to the actual arm
vectors. The time complexity for retrieving a cH -MIPS solution involves determining hash keys
and evaluating inner products with the arms in the retrieved buckets. Let ?? < 1 be an opti0
mized value
for
the hashing
?(see [35] for
detail). The time complexity for d -dimensional veclog(dT )
tors is O log log(c?1 ) N ? log(N )d0 , and the space complexity (except the original data) is
H
log(dT )
??
0
O log(c
. While the time and space complexity grows with the time horiN
(N
+
log(N
)d
)
?1
)
H
zon T , the dependence is mild; log log(T ) and log(T ), respectively. QGLOC uses d0 = d + d2 ,6 and
GLOC-TS uses d0 = d0 . While both achieve a time complexity sublinear in N , the time complexity
of GLOC-TS scales with d that is better than scaling with d2 of QGLOC. However, GLOC-TS has a
d1/4 -factor worse regret bound than QGLOC.
Discussion While it is reasonable to incur small errors in solving the arm selection criteria like (10)
and sacrifice some regret in practice, the regret bounds of QGLOC and GLOC-TS do not hold
anymore. Though not the focus of our paper, we prove a regret bound under the presence of the
hashing error in the fixed budget setting for QGLOC; see SM. Although the result therein has an
inefficient space complexity that is linear in T , it provides the first low regret bound with time
sublinear in N , to our knowledge.
5
Approximate Inner Product Computations with L1 Sampling
While hashing allows a time complexity sublinear in N , it performs an additional computation for determining the hash keys. Consider a hashing with U tables and length-k hash
keys. Given a query q and projection vectors
a(1) , . . . , a(U k) , the hashing computes q> a(i) ,
?i ? [U k] to determine the hash key of q. To
reduce such an overhead, approximate inner
product methods like [22, 24] are attractive
since hash keys are determined by discretizing
the inner products; small inner product errors
often do not alter the hash keys.
1
5
0.9
0
0.8
0.7
-5
100
L2
L1
(a)
101
102
103
d
(b)
Figure 1: (a) A box plot of estimators. L1 and L2
have the same variance, but L2 has thicker tails. (b)
The frequency of L1 inducing smaller variance than
L2 in 1000 trials. After 100 dimensions, L1 mostly
has smaller variance than L2.
In this section, we propose an improved approximate inner product method called L1 sampling which
we claim is more accurate than the sampling proposed by Jain et al. [22], which we call L2 sampling.
Consider an inner product q> a. The main idea is to construct an unbiased estimate of q> a. That is,
let p ? Rd be a probability vector. Let
i.i.d.
(11)
ik ? Multinomial(p) and Gk := qik aik /pik , k ? [m] .
Pm
1
>
>
It is easy to see that EGk = q a. By taking m k=1 Gk as an estimate of q a, the time complexity
is now O(mU k) rather than O(d0 U k). The key is to choose the right p. L2 sampling uses p(L2) :=
[qi2 /||q||22 ]i . Departing from L2, we propose p(L1) that we call L1 sampling and define as follows:
p(L1) := [|q1 |; ? ? ? ; |qd0 |]/||q||1 .
(12)
We compare L1 with L2 in two different point of view. Due to space constraints, we summarize the
key ideas and defer the details to SM.
The first is on their concentration of measure. Lemma 1 below shows an error bound of L1 whose
failure probability decays exponentially in m. This is in contrast to decaying polynomially of L2 [22],
which is inferior.7
Lemma 1. Define Gk as in (11) with p = p(L1) . Then, given a target error > 0,
Pm
m2
P 1
Gk ? q> a ? ? 2 exp ?
(13)
2
2
m
k=1
2||q||1 ||a||max
To illustrate such a difference, we fix q and a in 1000 dimension and apply L2 and L1 sampling 20K
times each with m = 5 where we scale down the L2 distribution so its variance matches that of L1.
Note that this does not mean we need to store vec(xx> ) since an inner product with it is structured.
In fact, one can show a bound for L2 that fails with exponentially-decaying probability. However, the bound
introduces a constant that can be arbitrarily large, which makes the tails thick. We provide details on this in SM.
6
7
7
Algorithm
QGLOC
QGLOC-Hash
GLOC-TS
GLOC-TS-Hash
(a)
(b)
Cum. Regret
266.6 (?19.7)
285.0 (?30.3)
277.0 (?36.1)
289.1 (?28.1)
(c)
Figure 2: Cumulative regrets with confidence intervals under the (a) logit and (b) probit model. (c)
Cumulative regrets with confidence intervals of hash-amenable algorithms.
Figure 1(a) shows that L2 has thicker tails than L1. Note this is not a pathological case but a typical
case for Gaussian q and a. This confirms our claim that L1 is safer than L2.
Another point of comparison is the variance of L2 and L1. We show that the variance of L1 may or
may not be larger than L2 in SM; there is no absolute winner. However, if q and a follow a Gaussian
distribution, then L1 induces smaller variances than L2 for large enough d; see Lemma 9 in SM.
Figure 1(b) confirms such a result. The actual gap between the variance of L2 and L1 is also nontrivial
under the Gaussian assumption. For instance, with d = 200, the average variance of Gk induced by
L2 is 0.99 whereas that induced by L1 is 0.63 on average. Although a stochastic assumption on the
vectors being inner-producted is often unrealistic, in our work we deal with projection vectors a that
are truly normally distributed.
6
Experiments
We now show our experiment results comparing GLB algorithms and hash-amenable algorithms.
GLB Algorithms We compare GLOC with two different algorithms: UCB-GLM [28] and Online
Learning for Logit Model (OL2M) [41].8 For each trial, we draw ? ? ? Rd and N arms (X ) uniformly
at random from the unit sphere. We set d = 10 and Xt = X , ?t ? 1. Note it is a common practice to
scale the confidence set radius for bandits [8, 27]. Following Zhang et al. [41], for OL2M we set the
squared radius ?t = c log(det(Zt )/det(Z1 )), where c is a tuning parameter. For UCB-GLM, we set
?
Pt
the radius as ? = cd log t. For GLOC, we replace ?tONS with c s=1 gs2 ||xs ||2A?1 . While parameter
s
tuning in practice is nontrivial, for the sake of comparison we tune c ? {101 , 100.5 , . . . , 10?3 } and
report the best one. We perform 40 trials up to time T = 3000 for each method and compute
confidence bounds on the regret.
We consider two GLM rewards: (i) the logit model (the Bernoulli GLM) and (ii) the probit model
(non-canonical GLM) for 0/1 rewards that sets ? as the probit function. Since OL2M is for the
logit model only, we expect to see the consequences of model mismatch in the probit setting. For
GLOC and UCB-GLM, we specify the correct reward model. We plot the cumulative regret under the
logit model in Figure 2(a). All three methods perform similarly, and we do not find any statistically
significant difference based on paired t test. The result for the probit model in Figure 2(b) shows that
OL2M indeed has higher regret than both GLOC and UCB-GLM due to the model mismatch in the
probit setting. Specifically, we verify that at t = 3000 the difference between the regret of UCB-GLM
and OL2M is statistically significant. Furthermore, OL2M exhibits a significantly higher variance in
the regret, which is unattractive in practice. This shows the importance of being generalizable to any
GLM reward. Note we observe a big increase in running time for UCB-GLM compared to OL2M
and GLOC.
Hash-Amenable GLBs To compare hash-amenable GLBs, we use the logit model as above but now
with N =100,000 and T =5000. We run QGLOC, QGLOC with hashing (QGLOC-Hash), GLOC-TS,
and GLOC-TS with hashing (GLOC-TS-Hash), where we use the hashing to compute the objective
function (e.g., (10)) on just 1% of the data points and save a significant amount of computation.
Details on our hashing implementation is found in SM. Figure 2(c) summarizes the result. We observe
that QGLOC-Hash and GLOC-TS-Hash increase regret from QGLOC and GLOC-TS, respectively,
but only moderately, which shows the efficacy of hashing.
7
Future Work
In this paper, we have proposed scalable algorithms for the GLB problem: (i) for large time horizon
T and (ii) for large number N of arms. There exists a number of interesting future work. First,
8
We have chosen UCB-GLM over GLM-UCB of Filippi et al. [15] as UCB-GLM has a lower regret bound.
8
we would like to extend the GLM rewards to the single index models [23] so one does not need to
know the function ? ahead of time under mild assumptions. Second, closing the regret bound gap
between QGLOC and GLOC without loosing
? hash-amenability would be interesting: i.e., develop
a hash-amenable GLB algorithm with O(d T ) regret. In this direction, a first attempt could be to
design a hashing scheme that can directly solve (7) approximately.
Acknowledgments This work was partially supported by the NSF grant IIS-1447449 and the
MURI grant 2015-05174-04. The authors thank Yasin Abbasi-Yadkori and Anshumali Shrivastava
for providing constructive feedback and Xin Hunt for her contribution at the initial stage.
References
[1] Abbasi-Yadkori, Yasin, Pal, David, and Szepesvari, Csaba. Improved Algorithms for Linear
Stochastic Bandits. Advances in Neural Information Processing Systems (NIPS), pp. 1?19,
2011.
[2] Abbasi-Yadkori, Yasin, Pal, David, and Szepesvari, Csaba. Online-to-Confidence-Set Conversions and Application to Sparse Stochastic Bandits. In Proceedings of the International
Conference on Artificial Intelligence and Statistics (AISTATS), 2012.
[3] Abeille, Marc and Lazaric, Alessandro. Linear Thompson Sampling Revisited. In Proceedings
of the International Conference on Artificial Intelligence and Statistics (AISTATS), volume 54,
pp. 176?184, 2017.
[4] Agrawal, Shipra and Goyal, Navin. Thompson Sampling for Contextual Bandits with Linear
Payoffs. CoRR, abs/1209.3, 2012.
[5] Agrawal, Shipra and Goyal, Navin. Thompson Sampling for Contextual Bandits with Linear
Payoffs. In Proceedings of the International Conference on Machine Learning (ICML), pp.
127?135, 2013.
[6] Ahukorala, Kumaripaba, Medlar, Alan, Ilves, Kalle, and Glowacka, Dorota. Balancing Exploration and Exploitation: Empirical Parameterization of Exploratory Search Systems. In
Proceedings of the ACM International Conference on Information and Knowledge Management
(CIKM), pp. 1703?1706, 2015.
[7] Auer, Peter and Long, M. Using Confidence Bounds for Exploitation-Exploration Trade-offs.
Journal of Machine Learning Research, 3:397?422, 2002.
[8] Chapelle, Olivier and Li, Lihong. An Empirical Evaluation of Thompson Sampling. In Advances
in Neural Information Processing Systems (NIPS), pp. 2249?2257, 2011.
[9] Chu, Wei, Li, Lihong, Reyzin, Lev, and Schapire, Robert E. Contextual Bandits with Linear
Payoff Functions. In Proceedings of the International Conference on Artificial Intelligence and
Statistics (AISTATS), volume 15, pp. 208?214, 2011.
[10] Crammer, Koby and Gentile, Claudio. Multiclass Classification with Bandit Feedback Using
Adaptive Regularization. Mach. Learn., 90(3):347?383, 2013.
[11] Dani, Varsha, Hayes, Thomas P, and Kakade, Sham M. Stochastic Linear Optimization under
Bandit Feedback. In Proceedings of the Conference on Learning Theory (COLT), pp. 355?366,
2008.
[12] Datar, Mayur, Immorlica, Nicole, Indyk, Piotr, and Mirrokni, Vahab S. Locality-sensitive
Hashing Scheme Based on P-stable Distributions. In Proceedings of the Twentieth Annual
Symposium on Computational Geometry, pp. 253?262, 2004.
[13] Dekel, Ofer, Gentile, Claudio, and Sridharan, Karthik. Robust selective sampling from single
and multiple teachers. In In Proceedings of the Conference on Learning Theory (COLT), 2010.
[14] Dekel, Ofer, Gentile, Claudio, and Sridharan, Karthik. Selective sampling and active learning
from single and multiple teachers. Journal of Machine Learning Research, 13:2655?2697,
2012.
9
[15] Filippi, Sarah, Cappe, Olivier, Garivier, Aur?lien, and Szepesv?ri, Csaba. Parametric Bandits:
The Generalized Linear Case. In Advances in Neural Information Processing Systems (NIPS),
pp. 586?594. 2010.
[16] Gentile, Claudio and Orabona, Francesco. On Multilabel Classification and Ranking with
Bandit Feedback. Journal of Machine Learning Research, 15:2451?2487, 2014.
[17] Guo, Ruiqi, Kumar, Sanjiv, Choromanski, Krzysztof, and Simcha, David. Quantization based
Fast Inner Product Search. Journal of Machine Learning Research, 41:482?490, 2016.
[18] Har-Peled, Sariel, Indyk, Piotr, and Motwani, Rajeev. Approximate nearest neighbor: towards
removing the curse of dimensionality. Theory of Computing, 8:321?350, 2012.
[19] Hazan, Elad and Karnin, Zohar. Volumetric Spanners: An Efficient Exploration Basis for
Learning. Journal of Machine Learning Research, 17(119):1?34, 2016.
[20] Hazan, Elad, Agarwal, Amit, and Kale, Satyen. Logarithmic Regret Algorithms for Online
Convex Optimization. Mach. Learn., 69(2-3):169?192, 2007.
[21] Hofmann, Katja, Whiteson, Shimon, and de Rijke, Maarten. Contextual Bandits for Information
Retrieval. In NIPS Workshop on Bayesian Optimization, Experimental Design and Bandits:
Theory and Applications, 2011.
[22] Jain, Prateek, Vijayanarasimhan, Sudheendra, and Grauman, Kristen. Hashing Hyperplane
Queries to Near Points with Applications to Large-Scale Active Learning. In Advances in
Neural Information Processing Systems (NIPS), pp. 928?936, 2010.
[23] Kalai, Adam Tauman and Sastry, Ravi. The Isotron Algorithm: High-Dimensional Isotonic
Regression. In Proceedings of the Conference on Learning Theory (COLT), 2009.
[24] Kannan, Ravindran, Vempala, Santosh, and Others. Spectral algorithms. Foundations and
Trends in Theoretical Computer Science, 4(3?4):157?288, 2009.
[25] Konyushkova, Ksenia and Glowacka, Dorota. Content-based image retrieval with hierarchical
Gaussian Process bandits with self-organizing maps. In 21st European Symposium on Artificial
Neural Networks, 2013.
[26] Li, Lihong, Chu, Wei, Langford, John, and Schapire, Robert E. A Contextual-Bandit Approach
to Personalized News Article Recommendation. Proceedings of the International Conference
on World Wide Web (WWW), pp. 661?670, 2010.
[27] Li, Lihong, Chu, Wei, Langford, John, Moon, Taesup, and Wang, Xuanhui. An Unbiased Offline
Evaluation of Contextual Bandit Algorithms with Generalized Linear Models. In Proceedings
of the Workshop on On-line Trading of Exploration and Exploitation 2, volume 26, pp. 19?36,
2012.
[28] Li, Lihong, Lu, Yu, and Zhou, Dengyong. Provable Optimal Algorithms for Generalized Linear
Contextual Bandits. CoRR, abs/1703.0, 2017.
[29] McCullagh, P and Nelder, J A. Generalized Linear Models. London, 1989.
[30] Neyshabur, Behnam and Srebro, Nathan. On Symmetric and Asymmetric LSHs for Inner
Product Search. Proceedings of the International Conference on Machine Learning (ICML), 37:
1926?1934, 2015.
[31] Orabona, Francesco, Cesa-Bianchi, Nicolo, and Gentile, Claudio. Beyond Logarithmic Bounds
in Online Learning. In Proceedings of the International Conference on Artificial Intelligence
and Statistics (AISTATS), volume 22, pp. 823?831, 2012.
[32] Qin, Lijing, Chen, Shouyuan, and Zhu, Xiaoyan. Contextual Combinatorial Bandit and its
Application on Diversified Online Recommendation. In SDM, pp. 461?469, 2014.
[33] Rui, Yong, Huang, T S, Ortega, M, and Mehrotra, S. Relevance feedback: a power tool for
interactive content-based image retrieval. IEEE Transactions on Circuits and Systems for Video
Technology, 8(5):644?655, 1998.
10
[34] Rusmevichientong, Paat and Tsitsiklis, John N. Linearly Parameterized Bandits. Math. Oper.
Res., 35(2):395?411, 2010.
[35] Shrivastava, Anshumali and Li, Ping. Asymmetric LSH ( ALSH ) for Sublinear Time Maximum
Inner Product Search ( MIPS ). Advances in Neural Information Processing Systems 27, pp.
2321?2329, 2014.
[36] Shrivastava, Anshumali and Li, Ping. Improved Asymmetric Locality Sensitive Hashing (ALSH)
for Maximum Inner Product Search (MIPS). In Proceedings of the Conference on Uncertainty
in Artificial Intelligence (UAI), pp. 812?821, 2015.
[37] Slaney, Malcolm, Lifshits, Yury, and He, Junfeng. Optimal parameters for locality-sensitive
hashing. Proceedings of the IEEE, 100(9):2604?2623, 2012.
[38] Wainwright, Martin J and Jordan, Michael I. Graphical Models, Exponential Families, and
Variational Inference. Found. Trends Mach. Learn., 1(1-2):1?305, 2008.
[39] Wang, Jingdong, Shen, Heng Tao, Song, Jingkuan, and Ji, Jianqiu. Hashing for Similarity
Search: A Survey. CoRR, abs/1408.2, 2014.
[40] Yue, Yisong, Hong, Sue Ann Sa, and Guestrin, Carlos. Hierarchical exploration for accelerating
contextual bandits. Proceedings of the International Conference on Machine Learning (ICML),
pp. 1895?1902, 2012.
[41] Zhang, Lijun, Yang, Tianbao, Jin, Rong, Xiao, Yichi, and Zhou, Zhi-hua. Online Stochastic
Linear Optimization under One-bit Feedback. In Proceedings of the International Conference
on Machine Learning (ICML), volume 48, pp. 392?401, 2016.
11
| 6615 |@word mild:2 trial:3 exploitation:6 version:4 ruiqi:1 polynomial:2 norm:1 katja:1 logit:6 c0:7 dekel:2 d2:2 confirms:2 jingdong:1 q1:1 ld:1 initial:1 series:1 efficacy:1 document:1 ours:2 interestingly:1 past:1 existing:14 comparing:1 contextual:9 chu:3 must:3 readily:1 written:4 bd:4 john:3 sanjiv:1 confirming:1 hofmann:1 plot:2 update:1 hash:32 intelligence:5 selected:2 prohibitive:1 item:2 parameterization:1 lr:4 pointer:1 provides:2 math:1 revisited:1 successive:1 simpler:1 zhang:5 unbounded:1 constructed:1 become:1 symposium:2 ik:1 retrieving:1 mayur:1 prove:3 overhead:3 introduce:1 sacrifice:1 ravindran:1 expected:2 indeed:2 qd0:1 ol:12 yasin:3 relying:3 zhi:1 actual:2 curse:1 equipped:1 cardinality:1 increasing:1 provided:1 xx:2 moreover:1 notation:1 bounded:6 circuit:1 prateek:1 rmax:1 z:1 generalizable:1 proposing:1 loos:1 csaba:3 impractical:2 sung:1 guarantee:3 every:1 concave:2 thicker:2 interactive:5 grauman:1 normally:3 unit:1 enjoy:1 grant:2 before:1 understood:1 thinner:1 tends:1 limit:2 modify:1 consequence:1 mach:3 lev:1 datar:1 approximately:1 might:1 plus:1 therein:1 co:1 hunt:1 range:1 statistically:2 practical:5 acknowledgment:1 practice:7 regret:69 goyal:2 procedure:1 empirical:2 maxx:3 significantly:1 projection:5 sudheendra:1 confidence:23 refers:2 wait:1 time5:1 cannot:2 close:1 selection:4 vijayanarasimhan:1 isotonic:1 www:1 equivalent:1 hqt:1 lagrangian:1 yt:10 center:1 maximizing:1 nicole:1 kale:1 map:1 tianbao:1 thompson:9 convex:4 shen:1 survey:1 estimator:3 d5:2 supz:1 q:1 deriving:1 pull:2 rule:1 adjusts:1 d1:2 maarten:1 exploratory:2 limiting:1 target:2 pt:5 suppose:1 user:4 imagine:1 construction:2 aik:1 us:5 olivier:2 trick:1 logarithmically:1 trend:2 asymmetric:3 muri:1 database:1 wang:2 ensures:1 news:1 trade:1 observes:1 alessandro:1 convexity:1 complexity:23 peled:2 reward:25 ideally:1 moderately:1 mu:1 flurry:1 multilabel:1 tight:3 solving:2 incur:2 dilemma:1 upon:1 learner:16 basis:1 shipra:2 easily:1 retrieves:1 jain:2 fast:3 describe:1 effective:1 london:1 query:5 artificial:6 whose:8 larger:3 solve:9 supplementary:1 elad:2 otherwise:1 satyen:1 statistic:5 nondecreasing:1 noisy:1 indyk:2 online:34 nll:2 sequence:1 differentiable:1 eigenvalue:1 agrawal:2 sdm:1 propose:10 product:21 qin:2 relevant:1 combining:1 junfeng:1 date:2 reyzin:1 organizing:1 poorly:2 achieve:2 inducing:1 scalability:2 motwani:1 lshs:1 adam:1 paat:1 depending:1 derive:2 polylog:1 illustrate:1 develop:1 sarah:1 nearest:1 qt:1 received:1 dengyong:1 sa:1 strong:1 implemented:1 predicted:2 involves:2 implies:3 trading:1 direction:1 amenability:2 radius:5 thick:1 correct:1 simcha:1 stochastic:9 exploration:9 material:1 require:2 hx:1 fix:1 generalization:1 preliminary:2 kristen:1 tighter:2 extension:5 strictly:4 rong:1 hold:2 exp:8 mapping:1 algorithmic:1 claim:3 m0:1 tor:1 achieves:5 smallest:3 estimation:3 combinatorial:2 currently:1 sensitive:5 largest:1 tool:3 dani:2 offs:1 anshumali:3 gaussian:9 harmed:1 rather:3 kalai:1 avoid:1 claudio:5 zhou:2 sion:1 varying:1 corollary:2 ax:1 focus:3 she:4 improvement:1 bernoulli:9 likelihood:2 contrast:2 normalizer:1 inference:1 dependent:2 bt:20 a0:1 her:2 bandit:33 lien:1 selective:2 tao:1 issue:2 among:1 arg:6 classification:2 colt:3 proposes:1 art:2 special:2 zon:1 santosh:1 construct:3 karnin:1 saving:2 beach:1 sampling:21 piotr:2 koby:1 icml:4 yu:1 alter:1 future:2 np:1 report:1 others:1 inherent:1 employ:1 pathological:1 ksenia:1 geometry:1 consisting:1 isotron:1 maintain:1 karthik:2 attempt:2 ab:3 interest:1 evaluation:2 uncommon:1 analyzed:1 introduces:1 truly:1 har:2 amenable:15 accurate:1 integral:1 nowak:1 euclidean:1 desired:1 re:2 lijun:1 theoretical:1 minimal:1 instance:1 column:1 modeling:2 vahab:1 subset:1 glb:22 successful:2 wonder:1 too:1 pal:2 teacher:2 chooses:4 combined:1 st:2 thanks:1 international:10 varsha:1 aur:1 konyushkova:1 michael:1 continuously:1 squared:1 abbasi:5 satisfied:1 management:1 containing:1 choose:5 huang:1 exten:1 cesa:1 yisong:1 worse:1 slaney:1 inefficient:2 derivative:1 li:12 oper:1 filippi:4 de:1 star:1 sec:1 waste:1 rusmevichientong:1 yury:1 satisfy:2 ranking:1 later:2 root:1 break:1 view:1 analyze:1 hazan:2 competitive:2 decaying:2 option:1 carlos:1 defer:1 elaborated:1 contribution:2 square:1 accuracy:2 moon:1 variance:12 characteristic:1 yes:1 rijke:1 generalize:1 bayesian:1 critically:1 none:1 lu:1 history:1 promptly:1 explain:1 ping:2 definition:3 volumetric:2 evaluates:1 failure:1 m00:1 frequency:1 pp:18 proof:4 adjusting:1 popular:4 recall:2 knowledge:3 dimensionality:1 routine:1 auer:1 back:1 cappe:1 appears:3 feed:3 hashing:30 dt:4 higher:3 follow:1 response:2 improved:3 specify:1 wei:3 formulation:1 evaluated:2 though:2 strongly:1 generality:2 furthermore:6 box:1 just:1 stage:1 until:1 glms:6 langford:2 receives:1 navin:2 replacing:1 web:1 nonlinear:2 rajeev:1 logistic:2 artifact:1 pulling:1 grows:4 usa:1 name:1 normalized:1 true:3 unbiased:2 verify:1 regularization:1 symmetric:1 deal:1 conditionally:1 round:2 attractive:2 self:1 inferior:1 criterion:3 generalized:18 hong:1 ortega:1 ridge:3 performs:4 l1:21 image:10 variational:1 novel:2 common:1 specialized:1 pseudocode:2 multinomial:1 mt:7 ji:1 winner:1 exponentially:2 volume:5 tail:4 extend:2 he:1 willett:2 theirs:1 refer:1 significant:3 vec:4 rd:10 tuning:2 sastry:1 pm:2 similarly:1 closing:3 had:2 l3:1 chapelle:1 lihong:5 stable:1 similarity:2 lsh:1 etc:1 nicolo:1 recent:1 xiaoyan:1 retrieved:1 inf:1 scenario:2 store:1 taesup:1 discretizing:1 arbitrarily:2 vt:9 guestrin:1 additional:1 gentile:5 determine:2 maximize:1 ii:4 multiple:3 mix:1 full:1 sham:1 d0:5 alan:1 match:1 believed:1 long:2 retrieval:6 offer:1 sphere:1 concerning:1 mle:1 y:2 paired:1 prediction:2 scalable:7 regression:3 essentially:1 poisson:2 sue:1 tailored:1 agarwal:1 qik:1 receive:1 whereas:3 background:2 want:1 szepesv:1 interval:2 grow:2 biased:1 extra:1 unlike:2 yue:1 induced:2 leveraging:1 sridharan:2 jordan:1 call:4 practitioner:1 near:1 presence:1 yang:1 easy:2 mips:10 enough:1 fit:1 inner:21 reduce:3 idea:2 tradeoff:1 multiclass:1 det:2 motivated:1 optimism:2 utility:1 accelerating:1 effort:1 song:1 suffer:1 peter:1 remark:2 regrett:1 involve:2 tune:1 amount:1 induces:1 deriva:1 schapire:2 qi2:1 exist:1 canonical:8 nsf:1 notice:2 cikm:1 lazaric:2 per:2 write:2 key:16 alsh:2 d3:3 wisc:4 choromanski:1 ravi:1 garivier:1 krzysztof:1 uw:4 tenable:1 year:1 run:4 inverse:1 parameterized:1 uncertainty:3 place:1 almost:1 ruling:1 family:5 throughout:2 reasonable:1 draw:1 decision:1 pik:1 scaling:3 summarizes:1 bit:1 bound:43 ct:9 simplification:1 quadratic:5 g:2 annual:1 nontrivial:4 ahead:2 constraint:3 ri:1 personalized:1 sake:1 yong:1 aspect:1 nathan:1 min:2 kumar:1 mized:1 vempala:1 martin:1 structured:1 according:1 ball:1 smaller:6 slightly:1 kakade:1 making:1 restricted:1 aniruddha:2 glm:37 bucket:2 heart:1 turn:4 know:2 merit:1 end:2 informal:1 available:2 ofer:2 neyshabur:1 apply:2 observe:4 hierarchical:2 generic:2 spectral:1 anymore:1 save:1 batch:3 yadkori:5 alternative:2 original:2 thomas:1 running:1 ensure:1 graphical:1 madison:4 newton:4 restrictive:1 yz:2 build:1 amit:1 objective:9 parametric:1 concentration:2 dependence:4 mirrokni:1 exhibit:2 minx:1 hq:2 distance:2 link:2 thank:1 polytope:1 considers:1 reason:1 kannan:1 provable:1 length:1 index:1 ellipsoid:1 providing:1 balance:1 abeille:2 unfortunately:1 mostly:1 robert:3 gk:5 negative:1 design:3 implementation:1 zt:8 motivates:1 unknown:1 perform:4 bianchi:1 conversion:6 upper:1 observation:1 francesco:2 sm:12 finite:5 jin:1 t:25 payoff:3 prompt:1 rebecca:1 tive:1 inverting:1 rating:1 david:3 extensive:1 optimized:2 z1:1 polylogarithmic:1 nip:6 zohar:1 beyond:1 usually:2 below:5 mismatch:2 appeared:2 summarize:2 encompasses:1 spanner:2 egk:1 max:7 memory:1 video:1 wainwright:1 unrealistic:1 critical:1 power:1 natural:3 rely:1 bhargava:1 safer:1 arm:43 zhu:1 scheme:3 improve:1 technology:1 mehrotra:1 jun:1 naive:1 prior:2 review:1 discovery:2 l2:22 sariel:1 determining:3 loss:7 probit:8 expect:1 sublinear:10 designation:1 interesting:2 srebro:1 var:1 foundation:1 regreta:1 sufficient:1 shouyuan:1 article:1 principle:2 xiao:1 heng:1 storing:1 cd:1 balancing:1 rdnowak:1 compatible:1 supported:1 free:3 offline:1 formal:1 allow:1 tsitsiklis:1 neighbor:1 wide:1 kwang:1 face:2 taking:2 absolute:1 sparse:1 departing:1 distributed:3 tauman:1 feedback:7 dimension:2 evaluating:1 cumulative:6 world:1 concavity:1 ignores:1 computes:1 author:1 adaptive:1 far:2 polynomially:1 correlate:1 transaction:1 approximate:6 emphasize:1 ons:18 dealing:1 active:3 hayes:1 uai:1 b1:1 conclude:1 nelder:1 xi:2 search:12 why:1 table:4 additionally:1 nature:1 learn:3 robust:1 ca:1 inherently:1 ignoring:1 obtaining:1 shrivastava:3 szepesvari:2 whiteson:1 necessarily:1 european:1 constructing:1 marc:1 aistats:4 main:2 cum:1 linearly:5 big:1 noise:2 xuanhui:1 repeated:1 x1:1 elaborate:3 lifshits:1 sub:2 fails:1 explicit:1 exponential:4 candidate:1 advertisement:1 shimon:1 down:3 theorem:13 removing:1 xt:38 specific:2 showing:2 behnam:1 ton:2 x:4 decay:1 unattractive:1 trap:2 exists:1 quantization:1 workshop:2 corr:3 gained:1 importance:1 budget:2 horizon:4 rui:1 gap:5 chen:1 locality:5 logarithmic:2 simply:1 twentieth:1 shoe:1 conveniently:1 diversified:1 partially:1 recommendation:4 hua:1 ch:6 satisfies:4 relies:2 acm:1 cdf:1 prop:1 loosing:2 ann:1 towards:4 orabona:2 lipschitz:3 replace:1 content:2 feasible:1 mccullagh:1 change:1 gloc:57 hard:1 specifically:3 typical:2 reducing:1 infinite:1 determined:2 except:1 lemma:3 uniformly:1 called:8 total:1 hyperplane:1 experimental:2 xin:1 ucb:9 select:1 immorlica:1 guo:1 crammer:1 relevance:3 overload:1 constructive:1 malcolm:1 |
6,208 | 6,616 | Probabilistic Models for Integration Error in the
Assessment of Functional Cardiac Models
Chris. J. Oates1,5 , Steven Niederer2 , Angela Lee2 , Fran?ois-Xavier Briol3 , Mark Girolami4,5
1
Newcastle University, 2 King?s College London, 3 University of Warwick,
4
Imperial College London, 5 Alan Turing Institute
Abstract
This paper studies the numerical computation of integrals, representing estimates
or predictions, over the output f (x) of a computational model with respect to
a distribution p(dx) over uncertain inputs x to the model. For the functional
cardiac models that motivate this work, neither f nor p possess a closed-form
expression and evaluation of either requires ? 100 CPU hours, precluding standard
numerical integration methods. Our proposal is to treat integration as an estimation
problem, with a joint model for both the a priori unknown function f and the a
priori unknown distribution p. The result is a posterior distribution over the integral
that explicitly accounts for dual sources of numerical approximation error due to a
severely limited computational budget. This construction is applied to account, in
a statistically principled manner, for the impact of numerical errors that (at present)
are confounding factors in functional cardiac model assessment.
1
Motivation: Predictive Assessment of Computer Models
This paper considers the problem of assessment for computer models [7], motivated by an urgent
need to assess the performance of sophisticated functional cardiac models [25]. In concrete terms,
the problem that we consider can be expressed as the numerical approximation of integrals
Z
p(f ) =
f (x)p(dx),
(1)
where f (x) denotes a functional of the output from a computer model and x denotes unknown inputs
(or ?parameters?) of the model. The term p(x) denotes a posterior distribution over model inputs.
Although not our focus in this paper, we note that p(x) is defined based on a prior ?0 (x) over these
inputs and training data y assumed to follow the computer model ?(y|x) itself. The integral p(f ), in
our context, represents a posterior prediction of actual cardiac behaviour. The computational model
can be assessed through comparison of these predictions to test data generated from an experiment.
The challenging nature of cardiac models ? and indeed computer models in general ? is such that a
closed-form for both f (x) and p(dx) is precluded [23]. Instead, it is typical to be provided with a
finite collection of samples {xi }ni=1 obtained from p(dx) through Monte Carlo (or related) methods
[32]. The integrand f (x) is then evaluated at these n input configurations, to obtain {f (xi )}ni=1 .
Limited computational budgets necessitate that the number n is small and, in such situations, the error
of an estimator for the integral p(f ) based on the data {(xi , f (xi ))}ni=1 is subject to strict informationtheoretic lower bounds [26]. The practical consequence is that an unknown (non-negligible) numerical
error is introduced in the numerical approximation of p(f ), unrelated to the performance of the model.
If this numerical error is ignored, it will constitute a confounding factor in the assessment of predictive
performance for the computer model. It is therefore unclear how a fair model assessment can proceed.
This motivates an attempt to understand the extent of numerical error in any estimate of p(f ). This is
non-trivial; for example, the error distribution of the arithmetic mean n1 ?ni=1 f (xi ) depends on the
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
unknown f and p, and attempts to estimate this distribution solely from data, e.g. via a bootstrap or a
central limit approximation, cannot succeed in general when the number of samples n is small [27].
Our first contribution, in this paper, is to argue that approximation of p(f ) from samples {xi }ni=1
and function evaluations {f (xi )}ni=1 can be cast as an estimation task. Our second contribution is to
derive a posterior distribution over the unknown value p(f ) of the integral. This distribution provides
an interpretable quantification of the extent of numerical integration error that can be reasoned
with and propagated through subsequent model assessment. Our third contribution is to establish
theoretical properties of the proposed method. The method we present falls within the framework
of Probabilistic Numerics and our work can be seen as a contribution to this emerging area [16, 5].
In particular, the method proposed is reminiscent of Bayesian Quadrature (BQ) [9, 28, 29, 15]. In
BQ, a Gaussian prior measure is placed on the unknown function f and is updated to a posterior
when conditioned on the information {(xi , f (xi ))}ni=1 . This induces both a prior and a posterior
over the value of p(f ) as push-forward measures under the projection operator f 7? p(f ). Since
its introduction, several authors have related BQ to other methods such as the ?herding? approach
from machine learning [17, 3], random feature approximations used in kernel methods [1], classical
quadrature rules [33] and Quasi Monte Carlo (QMC) methods [4]. Most recently, [21] extended
theoretical results for BQ to misspecified prior models, and [22] who provided efficient matrix
algebraic methods for the implementation of BQ. However, as an important point of distinction,
notice that BQ pre-supposes p(dx) is known in closed-form - it does not apply in situations where
p(dx) is instead sampled. In this latter case p(dx) will be called an intractable distribution and, for
model assessment, this scenario is typical.
To extend BQ to intractable distributions, this paper proposes to use a Dirichlet process mixture
prior to estimate the unknown distribution p(dx) from Monte Carlo samples {xi }ni=1 [12]. It will be
demonstrated that this leads to a simple expression for the closed-form terms which are required to
implement the usual BQ. The overall method, called Dirichlet process mixture Bayesian quadrature
(DPMBQ), constructs a (univariate) distribution over the unknown integral p(f ) that can be exploited
to tease apart the intrinsic performance of a model from numerical integration error in model
assessment. Note that BQ was used to estimate marginal likelihood in e.g. [30]. The present problem
is distinct, in that we focus on predictive performance (of posterior expectations) rather than marginal
likelihood, and its solution demands a correspondingly different methodological development.
On the computational front, DPMBQ costs O(n3 ). However, this cost is de-coupled from the often
orders-of-magnitude larger costs involved in both evaluation of f (x) and p(dx), which form the main
computational bottleneck. Indeed, in the modern computational cardiac models that motivate this
research, the ? 100 CPU hour time required for a single simulation limits the number n of available
samples to ? 103 [25]. At this scale, numerical integration error cannot be neglected in model
assessment. This raises challenges when making assessments or comparisons between models, since
the intrinsic performance of models cannot be separated from numerical error that is introduced into
the assessment. Moreover, there is an urgent ethical imperative that the clinical translation of such
models is accompanied with a detailed quantification of the unknown numerical error component in
model assessment. Our contribution explicitly demonstrates how this might be achieved.
The remainder of the paper proceeds as follows: In Section 2.1 we first recall the usual BQ method,
then in Section 2.2 we present and analyse our novel DPMBQ method. Proofs of theoretical results
are contained in the electronic supplement. Empirical results are presented in Section 3 and the paper
concludes with a discussion in Section 4.
2
Probabilistic Models for Numerical Integration Error
Consider a domain ? ? Rd , together with a distribution p(dx) on ?. As in Eqn. 1, p(f ) will be used
to denote the integral of the argument f with respect to the distribution p(dx). All integrands are
assumed to be (measurable) functions f : ? ? R such that the integral p(f ) is well-defined. To
begin, we recall details for the BQ method when p(dx) is known in closed-form [9, 28]:
2.1
Probabilistic Integration for Tractable Distributions (BQ)
In standard BQ [9, 28], a Gaussian Process (GP) prior f ? GP(m, k) is assigned to the integrand f ,
with mean function m : ? ? R and covariance function k : ? ? ? ? R [see 31, for further details
2
on GPs]. The implied prior over the integral p(f ) is then the push-forward of the GP prior through
the projection f 7? p(f ):
p(f ) ? N(p(m), p ? p(k))
0
where p?p : ??? ? R is the measure formed by independent products of p(dx)
RR and 0p(dx ), so that
under our notational convention the so-called initial error p?p(k) is equal to
k(x, x )p(dx)p(dx0 ).
Next, the GP is conditioned on the information in {(xi , f (xi ))}ni=1 . The conditional GP takes a
conjugate form f |X, f (X) ? GP(mn , kn ), where we have written X = (x1 , . . . , xn ), f (X) =
(f (x1 ), . . . , f (xn ))> . Formulae for the mean function mn : ? ? R and covariance function
kn : ? ? ? ? R are standard can be found in [31, Eqns. 2.23, 2.24]. The BQ posterior over p(f ) is
the push forward of the GP posterior:
p(f ) | X, f (X) ? N(p(mn ), p ? p(kn ))
(2)
Formulae for p(mn ) and p ? p(kn ) were derived in [28]:
p(mn ) = f (X)> k(X, X)?1 ?(X)
(3)
p ? p(kn ) = p ? p(k) ? ?(X)> k(X, X)?1 ?(X)
(4)
where k(X, X) is the n ? n matrix with (i, j)th entry k(xi , xj ) and ?(X) is the n ? 1 vector with
ith entry ?(xi ) where the function ? is called the kernel mean or kernel embedding [see e.g. 35]:
Z
?(x) = k(x, x0 )p(dx0 )
(5)
Computation of the kernel mean and the initial error each requires that p(dx) is known in general.
The posterior in Eqn. 2 was studied in [4], where rates of posterior contraction were established
under further assumptions on the smoothness of the covariance function k and the smoothness of
the integrand. Note that the matrix inverse of k(X, X) incurs a (naive) computational cost of O(n3 );
however this cost is post-hoc and decoupled from (more expensive) computation that involves the
computer model. Sparse or approximate GP methods could also be used.
2.2
Probabilistic Integration for Intractable Distributions
The dependence of Eqns. 3 and 4 on both the kernel mean and the initial error means that BQ cannot
be used for intractable p(dx) in general. To address this we construct a second non-parametric model
for the unknown p(dx), presented next.
Dirichlet Process Mixture Model Consider
Z an infinite mixture model
p(dx) = ?(dx; ?)P (d?),
(6)
where ? : ? ? ? ? [0, ?) is such that ?(?; ?) is a distribution on ? with parameter ? ? ? and P is
a mixing distribution defined on ?. In this paper, each data point xi is modelled as an independent
draw from p(dx) and is associated with a latent variable ?i ? ? according to the generative process
of Eqn. 6. i.e. xi ? ?(?; ?i ). To limit scope, the extension to correlated xi is reserved for future
work.
The Dirichlet process (DP) is the natural conjugate prior for non-parametric discrete distributions
[12]. Here we endow P (d?) with a DP prior P ? DP(?, Pb ), where ? > 0 is a concentration
parameter and Pb (d?) is a base distribution over ?. The base distribution Pb coincides with the
prior expectation E[P (d?)] = Pb (d?), while ? determines the spread of the prior about Pb . The
DP is characterised by the property that, for any finite partition ? = ?1 ? ? ? ? ? ?m , it holds that
(P (?1 ), . . . , P (?m )) ? Dir(?Pb (?1 ), . . . , ?Pb (?m )) where P (S) denotes the measure of the set
S ? ?. For ? ? 0, the DP is supported on the set of atomic distributions, while for ? ? ?, the DP
converges to an atom on the base distribution. This overall approach is called a DP mixture (DPM)
model [13].
For a random variable Z, the notation [Z] will be used as shorthand to denote the density function
of Z. It will be helpful to note that for ?i ? P independent, writing ?1:n = (?1 , . . . , ?n ), standard
conjugate results for DPs lead to the conditional
n
?
1 X
P | ?1:n ? DP ? + n,
Pb +
??
?+n
? + n i=1 i
where ??i (d?) is an atomic distribution centred at the location ?i of the ith sample in ?1:n . In turn,
this induces a conditional [dp|?1:n ] for the unknown distribution p(dx) through Eqn. 6.
3
Kernel Means via Stick Breaking The stick breaking characterisation can be used to draw from
the conditional DP [34]. A generic draw from [P |?1:n ] can be characterised as
P (d?) =
?
X
wj ??j (d?),
j?1
Y
wj = ?j
(1 ? ?j 0 )
(7)
j 0 =1
j=1
where randomness enters through the ?j and ?j as follows:
n
iid
?j ?
?
1 X
Pb +
?? ,
?+n
? + n i=1 i
iid
?j ? Beta(1, ? + n)
In practice the sum in Eqn. 7 may be truncated at a large finite number of terms, N , with negligible
truncation error, since weights wj vanish at a geometric rate [18]. The truncated DP has been shown
to provide accurate approximation of integrals with respect to the original DP [19]. For a realisation
P (d?) from Eqn. 7, observe that the induced distribution p(dx) over ? is
p(dx) =
?
X
wj ?(dx; ?j ).
(8)
j=1
Thus we have an alternative characterisation of [p|?1:n ].
Our key insight is that one can take ? and k to be a conjugate pair, such that both the kernel mean
?(x) and the initial error p ? p(k) will be available in an explicit form for the distribution in Eqn. 8
[see Table 1 in 4, for a list of conjugate pairs]. For instance, in the one-dimensional case, consider
? = (?1 , ?2 ) and ?(dx; ?) = N(dx; ?1 , ?2 ) for some location and scale parameters ?1 and ?2 .
Then for the Gaussian kernel k(x, x0 ) = ? exp(?(x ? x0 )2 /2?2 ), the kernel mean becomes
?(x) =
?
X
j=1
??wj
(x ? ?j,1 )2
exp
?
2(?2 + ?j,2 )
(?2 + ?j,2 )1/2
(9)
and the initial variance can be expressed as
p ? p(k)
=
? X
?
X
(?j,1 ? ?j 0 ,1 )2
??wj wj 0
exp
?
.
2(?2 + ?j,2 + ?j 0 ,2 )
(?2 + ?j,2 + ?j 0 ,2 )1/2
j=1 j 0 =1
(10)
Similar calculations for the multi-dimensional case are straight-forward and provided in the Supplemental Information.
The Proposed Model To put this all together, let ? denote all hyper-parameters that (a) define the
GP prior mean and covariance function, denoted m? and k? below, and (b) define the DP prior, such
as ? and the base distribution Pb . It is assumed that ? ? ? for some specified set ?. The marginal
posterior distribution for p(f ) in the DPMBQ model is defined as
ZZ
[p(f ) | X, f (X)] =
[p(f ) | X, f (X), p, ?] [dp | X, ?] [d?].
(11)
The first term in the integral is BQ for a fixed distribution p(dx). The second term represents the
DPM model for the unknown p(dx), while the third term [d?] represents a hyper-prior distribution
over ? ? ?. The DPMBQ distribution in Eqn. 11 does not admit a closed-form expression. However,
it is straight-forward to sample from this distribution without recourse to f (x) or p(dx). In particular,
the second term can be accessed through the law of total probabilities:
Z
[dp | X, ?] =
[dp | ?1:n ] [?1:n | X, ?] d?1:n
where the first term [dp | ?1:n ] is the stick-breaking construction and the term [?1:n | X, ?] can be
targeted with a Gibbs sampler. Full details of the procedure we used to sample from Eqn. 11, which
is de-coupled from the much larger costs associated with the computer model, are provided in the
Supplemental Information.
4
Theoretical Analysis The analysis reported below restricts attention to a fixed hyper-parameter ?
and a one-dimensional state-space ? = R. The extension of theoretical results to multiple dimensions
was beyond the scope of this paper.
Our aim in this section is to establish when DPMBQ is ?consistent?. To be precise, a random
distribution Pn over an unknown parameter ? ? R, whose true value is ?0 , is called consistent for ?0
at a rate rn if, for all ? > 0, we have Pn [(??, ?0 ? ?) ? (?0 + ?, ?)] = OP (rn ). Below we denote
with f0 and p0 the respective true values of f and p; our aim is to estimate ?0 = p0 (f0 ). Denote with
H the reproducing kernel Hilbert space whose reproducing kernel is k and assume that the GP prior
mean m is an element of H. Our main theoretical result below establishes that the DPMBQ posterior
distribution in Eqn. 11, which is a random object due to the n independent draws xi ? p(dx), is
consistent:
Theorem. Let P0 denote the true mixing distribution. Suppose that:
1.
2.
3.
4.
5.
f belongs to H and k is bounded on ? ? ?.
?(dx; ?) = N(dx; ?1 , ?2 ).
P0 has compact support supp(P0 ) ? R ? (?, ?) for some fixed ?, ? ? (0, ?).
Pb has positive, continuous density on a rectangle R, s.t. supp(Pb ) ? R ? R ? [?, ?].
Pb ({(?1 , ?2 ) : |?1 | > t}) ? c exp(??|t|? ) for some ?, ? > 0 and ? t > 0.
Then the posterior Pn = [p(f ) | X, f0 (X)] is consistent for the true value p0 (f0 ) of the integral at
the rate n?1/4+ where the constant > 0 can be arbitrarily small.
The proof is provided in the Supplemental Information. Assumption (1) derives from results on
consistent BQ [4] and can be relaxed further with the results in [21] (not discussed here), while
assumptions (2-5) derive from previous work on consistent estimation with DPM priors [14]. For
the case of BQ when p(dx) is known and H a Sobolev space of order s > 1/2 on ? = [0, 1], the
corresponding posterior contraction rate is exp(?Cn2s? ) [4, Thm. 1]. Our work, while providing
only an upper bound on the convergence rate, suggests that there is an increase in the fundamental
complexity of estimation for p(dx) unknown compared to p(dx) known. Interestingly, the n?1/4+
rate is slower than the classical Bernstein-von Mises rate n?1/2 [36]. However, an out-of-hand
comparison between these two quantities is not straight forward, as the former involves the interaction
of two distinct non-parametric statistical models. It is known Bernstein-von Mises results can be
delicate for non-parametric problems [see, for example, the counter-examples in 10]. Rather, this
theoretical analysis guarantees consistent estimation in a regime that is non-standard.
3
Results
The remainder of the paper reports empirical results from application of DPMBQ to simulated data
and to computational cardiac models.
3.1
Simulation Experiments
To explore the empirical performance of DPMBQ, a series of detailed simulation experiments were
performed. For this purpose, a flexible test bed was constructed wherein the true distribution p0 was
a normal mixture model (able to approximate any continuous density) and the true integrand f0 was
a polynomial (able to approximate any continuous function). In this set-up it is possible to obtain
closed-form expressions for all integrals p0 (f0 ) and these served as a gold-standard benchmark.
To mimic the scenario of interest, a small number n of samples xi were drawn from p0 (dx) and
the integrand values f0 (xi ) were obtained. This information X, f0 (X) was provided to DPMBQ
and the output of DPMBQ, a distribution over p(f ), was compared against the actual value p0 (f0 )
of the integral. For all experiments in this paper the Gaussian kernel k defined in Sec. 2.2 was
used; the integrand f was normalised and the associated amplitude hyper-parameter ? = 1 fixed,
whereas the length-scale hyper-parameter ? was assigned a Gam(2, 1) hyper-prior. For the DPM, the
concentration parameter ? was assigned a Exp(1) hyper-prior. These choices allowed for adaptation
of DPMBQ to the smoothness of both f and p in accordance with the data presented to the method.
The base distribution Pb for DPMBQ was taken to be normal inverse-gamma with hyper-parameters
?0 = 0, ?0 = ?0 = ?0 = 1, selected to facilitate a simplified Gibbs sampler. Full details of the
simulation set-up and Gibbs sampler are reported in the Supplemental Information.
5
4
p(x)
f(x)
0.5
0
-0.5
10 0
2
0
-2
0
2
-2
x
0
2
x
W
1
Oracle
Student-t
DPMBQ
cover. prob.
0.8
0.6
10 -1
0.4
0.2
0
10 0
10 1
n
10 2
10 0
10 3
10 1
10 2
n
(a)
(b)
Figure 1: Simulated data results. (a) Comparison of coverage frequencies for the simulation experiments. (b) Convergence assessment: Wasserstein distance (W ) between the posterior in Eqn. 11
and the true value of the integral, is presented as a function of the number n of data points. [Circles
represent independent realisations and the linear trend is shown in red.]
For comparison, we considered the default 50% confidence interval description of numerical error
s
s
f? ? t? ? , f? + t? ?
(12)
n
n
where f? = n?1 ?ni=1 f (xi ), s2 = (n ? 1)?1 ?ni=1 (f (xi ) ? f?)2 and t? is the 50% level for a Student?s
t-distribution with n ? 1 degrees of freedom. It is well-known that Eqn. 12 is a poor description of
numerical error when n is small [c.f. ?Monte Carlo is fundamentally unsound? 27]. For example,
with n = 2, in the extreme case where, due to chance, f (x1 ) ? f (x2 ), it follows that s ? 0 and no
numerical error is acknowledged. This fundamental problem is resolved through the use of prior
information on the form of both f and p in DPMBQ. The appropriateness of DPMBQ therefore
depends crucially on the prior. The proposed method is further distinguished from Eqn. 12 in that the
distribution over numerical error is fully non-parametric, not e.g. constrained to be Student-t.
Empirical Results Coverage frequencies are shown in Fig. 1a for a specific integration task
(f0 , p0 ), that was deliberately selected to be difficult for Eqn. 12 due to the rare event represented by
the mass at x = 2. These were compared against central 50% posterior credible intervals produced
under DPMBQ. These are the frequency with which the confidence/credible interval contain the true
value of the integral, here estimated with 100 independent realisations for DPMBQ and 1000 for the
(less computational) standard method (standard errors are shown for both). Whilst it offers correct
coverage in the asymptotic limit, Eqn. 12 can be seen to be over-confident when n is small, with
coverage often less than 50%. In contrast, DPMBQ accounts for the fact p is being estimated and
provides conservative estimation about the extent of numerical error when n is small.
To present results that do not depend on a fixed
coverage level (e.g. 50%), we next measured
R
convergence in the Wasserstein distance W = |p(f ) ? p0 (f0 )| d[p(f ) | X, f (X)]. In particular we
explored whether the theoretical rate of n?1/4+ was realised. (Note that the theoretical result applied
just to fixed hyper-parameters, whereas the experimental results reported involved hyper-parameters
that were marginalised, so that this is a non-trivial experiment.) Results in Fig. 1b demonstrated that
W scaled with n at a rate which was consistent with the theoretical rate claimed. Full experimental
results on our polynomial test bed, reported in detail in the Supplemental Information, revealed that W
was larger for higher-degree polynomials (i.e. more complex integrands f ), while W was insensitive
to the number of mixture components (i.e. to more complex distributions p). The latter observation
may be explained by the fact that the kernel mean ? is a smoothed version of the distribution p and so
is not expected to be acutely sensitive to variation in p itself.
3.2
Application to a Computational Cardiac Model
The Model The computation model considered in this paper is due to [24] and describes the
mechanics of the left and right ventricles through a heart beat. In brief, the model geometry (Fig. 2a,
6
(a)
(b)
Figure 2: Cardiac model results: (a) Computational cardiac model. A) Segmentation of the cardiac
MRI. B) Computational model of the left and right ventricles. C) Schematic image showing the
features of pressure (left) and volume transient (right). (b) Comparison of coverage frequencies, for
each of 10 numerical integration tasks defined by functionals gj of the cardiac model output.
top right) is described by fitting a C1 continuous cubic Hermite finite element mesh to segmented
magnetic resonance images (MRI; Fig. 2a, top left). Cardiac electrophysiology is modelled separately
by the solution of the mono-domain equations and provides a field of activation times across the heart.
The passive material properties and afterload of the heart are described, respectively, by a transversely
isotropic material law and a three element Windkessel model. Active contraction is simulated using a
phenomenological cellular model, with spatial variation arising from the local electrical activation
times. The active contraction model is defined by five input parameters: tr and td are the respective
constants for the rise and decay times, T0 is the reference tension, a4 and a6 respectively govern the
length dependence of tension rise time and peak tension. These five parameters were concatenated
into a vector x ? R5 and constitute the model inputs. The model is fitted based on training data
y that consist of functionals gj : R5 ? R, j = 1, . . . , 10, of the pressure and volume transient
morphology during baseline activation and when the heart is paced from two leads implanted in
the right ventricle apex and the left ventricle lateral wall. These 10 functionals are defined in the
Supplemental Information; a schematic of the model and fitted measurements are shown in Fig. 2a
(bottom panel).
Test Functions The distribution p(dx) was taken to be the posterior distribution over model
inputs x that results from an improper flat prior on x and a squared-error likelihood function:
P10
1
2
log p(x) = const. + 0.1
2
j=1 (yj ? gj (x)) . The training data y = (y1 , . . . , y10 ) were obtained
from clinical experiment. The task we considered is to compute posterior expectations for functionals
f (x) of the model output produced when the model input x is distributed according to p(dx). This
represents the situation where a fitted model is used to predict response to a causal intervention,
representing a clinical treatment. For assessment of the DPMBQ method, which is our principle aim
in this experiment, we simply took the test functions f to be each of the physically relevant model
outputs gj in turn (corresponding to no causal intervention). This defined 10 separate numerical
integration problems as a test bed. Benchmark values for p0 (gj ) were obtained, as described in
the Supplemental Information, at a total cost of ? 105 CPU hours, which would not be routinely
practical.
Empirical Results For each of the 10 numerical integration problems in the test bed, we computed
coverage probabilities, estimated with 100 independent realisations (standard errors are shown),
in line with those discussed for simulation experiments. These are shown in Fig. 2b, where we
compared Eqn. 12 with central 50% posterior credible intervals produced under DPMBQ. It is seen
that Eqn. 12 is usually reliable but can sometimes be over-confident, with coverage probabilities
less than 50%. This over-confidence can lead to spurious conclusions on the predictive performance
of the computational model. In contrast, DPMBQ provides a uniformly conservative quantification
7
of numerical error (cover. prob. ? 50%). The DPMBQ method is further distinguished from Eqn.
12 in that it entails a joint distribution for the 10 integrals (the unknown p is shared across integrals
- an instance of transfer learning across the 10 integration tasks). Fig. 2b also appears to show a
correlation structure in the standard approach (black lines), but this is an artefact of the common
sample set {xi }ni=1 that was used to simultaneously estimate all 10 integrals; Eqn. 12 is still applied
independently to each integral.
4
Discussion
Numerical analysis often focuses the convergence order of numerical methods, but in non-asymptotic
regimes the language of probabilities can provide a richer, more intuitive and more useful description
of numerical error. This paper cast the computation of integrals p(f ) as an estimation problem
amenable to Bayesian methods [20, 9, 5]. The difficulty of this problem depends on our level of prior
knowledge (rendering the problem trivial if a closed-form solution is a priori known) and, in the
general case, on how much information we are prepared to obtain on the objects f and p through
numerical computation [16]. In particular, we distinguish between three states of prior knowledge:
(1) f known, p unknown, (2) f unknown, p known, (3) both f and p unknown. Case (1) is the
subject of Monte Carlo methods [32] and concerns classical problems in applied probability such as
estimating confidence intervals for expectations based on Markov chains. Notable recent work in this
direction is [8], who obtained a point estimate p? for p using a kernel smoother and then, in effect,
used p?(f ) as an estimate for the integral. The decision-theoretic risk associated with error in p? was
explored in [6]. Independent of integral estimation, there is a large literature on density estimation
[37]. Our probabilistic approach provides a Bayesian solution to this problem, as a special case of
our more general framework. Case (2) concerns functional analysis, where [26] provide an extensive
overview of theoretical results on approximation of unknown functions in an information complexity
framework. As a rule of thumb, estimation improves when additional smoothness can be a priori
assumed on the value of the unknown object [see 4]. The main focus of this paper was Case (3), until
now unstudied, and a transparent, general statistical method called DPMBQ was proposed.
The path-finding nature of this work raises several important questions for future theoretical and
applied research. First, these methods should be extended to account for the low-rank phenomenon
that is often encountered in multi-dimensional integrals [11]. Second, there is no reason, in general,
to restrict attention to function values obtained at the locations in X. Indeed, one could first estimate
p(dx), then select suitable locations X 0 from at which to evaluate f (X 0 ) [2]. This touches on aspects
of statistical experimental design; the practitioner seeks a set X 0 that minimises an appropriate loss
functional at the level of p(f ); see again [6]. Third, whilst restricted to Gaussians in our experiments,
further methodological work will be required to establish guidance for the choice of kernel k in the
GP and choice of base distribution Pb in the DPM [c.f. chapter 4 of 31].
Acknowledgments
CJO and MG were supported by the Lloyds Register Foundation Programme on Data-Centric
Engineering. SN was supported by an EPSRC Intermediate Career Fellowship. FXB was supported
by the EPSRC grant [EP/L016710/1]. MG was supported by the EPSRC grant [EP/J016934/1,
EP/K034154/1], an EPSRC Established Career Fellowship, the EU grant [EU/259348] and a Royal
Society Wolfson Research Merit Award. This material was based upon work partially supported by
the National Science Foundation (NSF) under Grant DMS-1127914 to the Statistical and Applied
Mathematical Sciences Institute. Opinions, findings, and conclusions or recommendations expressed
in this material are those of the author(s) and do not necessarily reflect the views of the NSF.
References
[1] F Bach. On the Equivalence Between Quadrature Rules and Random Features. Journal of Machine
Learning Research, 18:1?38, 2017.
[2] F-X Briol, CJ Oates, J Cockayne, WY Chen, and M Girolami. On the sampling problem for kernel
quadrature. In Proceedings of the 34th International Conference on Machine Learning, pages 586?595,
2017.
[3] F-X Briol, CJ Oates, M Girolami, and MA Osborne. Frank-Wolfe Bayesian quadrature: Probabilistic
integration with theoretical guarantees. In Advances in Neural Information Processing Systems, pages
1162?1170, 2015.
8
[4] F-X Briol, CJ Oates, M Girolami, MA Osborne, and D Sejdinovic. Probabilistic Integration: A Role for
Statisticians in Numerical Analysis? arXiv:1512.00933, 2015.
[5] J Cockayne, CJ Oates, T Sullivan, and M Girolami. Bayesian probabilistic numerical methods.
arXiv:1702.03673, 2017.
[6] SN Cohen. Data-driven nonlinear expectations for statistical uncertainty in decisions. arXiv:1609.06545,
2016.
[7] PS Craig, M Goldstein, JC Rougier, and AH Seheult. Bayesian Forecasting for Complex Systems Using
Computer Simulators. Journal of the American Statistical Association, 96(454):717?729, 2001.
[8] B Delyon and F Portier. Integral Approximation by Kernel Smoothing. Bernoulli, 22(4):2177?2208, 2016.
[9] P Diaconis. Bayesian Numerical Analysis. Statistical Decision Theory and Related Topics IV, 1:163?175,
1988.
[10] P Diaconis and D Freedman. On the Consistency of Bayes Estimates. Annals of Statistics, 14(1):1?26,
1986.
[11] J Dick, FY Kuo, and IH Sloan. High-Dimensional Integration: The Quasi-Monte Carlo Way. Acta
Numerica, 22:133?288, 2013.
[12] TS Ferguson. A Bayesian Analysis of Some Nonparametric Problems. Annals of Statistics, 1(2):209?230,
1973.
[13] TS Ferguson. Bayesian Density Estimation by Mixtures of Normal Distributions. Recent Advances in
Statistics, 24(1983):287?302, 1983.
[14] S Ghosal and AW Van Der Vaart. Entropies and Rates of Convergence for Maximum Likelihood and
Bayes Estimation for Mixtures of Normal Densities. Annals of Statistics, 29(5):1233?1263, 2001.
[15] T Gunter, MA Osborne, R Garnett, P Hennig, and SJ Roberts. Sampling for Inference in Probabilistic
Models With Fast Bayesian Quadrature. In Advances in Neural Information Processing Systems, pages
2789?2797, 2014.
[16] P Hennig, MA Osborne, and M Girolami. Probabilistic Numerics and Uncertainty in Computations.
Proceedings of the Royal Society A, 471(2179):20150142, 2015.
[17] F Husz?r and D Duvenaud. Optimally-Weighted Herding is Bayesian Quadrature. In Uncertainty in
Artificial Intelligence, volume 28, pages 377?386, 2012.
[18] H Ishwaran and LF James. Gibbs Sampling Methods for Stick-Breaking Priors. Journal of the American
Statistical Association, 96(453):161?173, 2001.
[19] H Ishwaran and M Zarepour. Exact and Approximate Sum Representations for the Dirichlet Process.
Canadian Journal of Statistics, 30(2):269?283, 2002.
[20] JB Kadane and GW Wasilkowski. Average case epsilon-complexity in computer science: A Bayesian view.
Bayesian Statistics 2, Proceedings of the Second Valencia International Meeting, pages 361?374, 1985.
[21] M Kanagawa, BK Sriperumbudur, and K Fukumizu. Convergence Guarantees for Kernel-Based Quadrature
Rules in Misspecified Settings. In Advances in Neural Information Processing Systems, volume 30, 2016.
[22] T Karvonen and S S?rkk?. Fully symmetric kernel quadrature. arXiv:1703.06359, 2017.
[23] MC Kennedy and A O?Hagan. Bayesian calibration of computer models. Journal of the Royal Statistical
Society: Series B, 63(3):425?464, 2001.
[24] AWC Lee, A Crozier, ER Hyde, P Lamata, M Truong, M Sohal, T Jackson, JM Behar, S Claridge,
A Shetty, E Sammut, G Plank, CA Rinaldi, and S Niederer. Biophysical Modeling to Determine the
Optimization of Left Ventricular Pacing Site and AV/VV Delays in the Acute and Chronic Phase of Cardiac
Resynchronization Therapy. Journal of Cardiovascular Electrophysiology, 28(2):208?215, 2016.
[25] GR Mirams, P Pathmanathan, RA Gray, P Challenor, and RH Clayton. White paper: Uncertainty and
Variability in Computational and Mathematical Models of Cardiac Physiology. The Journal of Physiology,
594(23):6833?6847, 2016.
[26] E Novak and H Wo?zniakowski. Tractability of Multivariate Problems, Volume II : Standard Information
for Functionals. EMS Tracts in Mathematics 12, 2010.
[27] A O?Hagan. Monte Carlo is fundamentally unsound. Journal of the Royal Statistical Society, Series D,
36(2/3):247?249, 1987.
[28] A O?Hagan. Bayes?Hermite Quadrature. Journal of Statistical Planning and Inference, 29(3):245?260,
1991.
[29] M Osborne, R Garnett, S Roberts, C Hart, S Aigrain, and N Gibson. Bayesian quadrature for ratios. In
Artificial Intelligence and Statistics, pages 832?840, 2012.
[30] MA Osborne, DK Duvenaud, R Garnett, CE Rasmussen, SJ Roberts, and Z Ghahramani. Active learning
of model evidence using Bayesian quadrature. In Advances in Neural Information Processing Systems,
2012.
[31] C Rasmussen and C Williams. Gaussian Processes for Machine Learning. MIT Press, 2006.
[32] C Robert and G Casella. Monte Carlo Statistical Methods. Springer Science & Business Media, 2013.
[33] S S?rkk?, J Hartikainen, L Svensson, and F Sandblom. On the relation between Gaussian process
quadratures and sigma-point methods. Journal of Advances in Information Fusion, 11(1):31?46, 2016.
[34] J Sethuraman. A Constructive Definition of Dirichlet Priors. Statistica Sinica, 4(2):639?650, 1994.
[35] A Smola, A Gretton, L Song, and B Sch?lkopf. A Hilbert Space Embedding for Distributions. Algorithmic
Learning Theory, Lecture Notes in Computer Science, 4754:13?31, 2007.
[36] R Von Mises. Mathematical Theory of Probability and Statistics. Academic, London, 1974.
[37] MP Wand and MC Jones. Kernel Smoothing. CRC Press, 1994.
9
| 6616 |@word mri:2 version:1 polynomial:3 simulation:6 crucially:1 seek:1 covariance:4 contraction:4 p0:13 pressure:2 incurs:1 tr:1 initial:5 configuration:1 series:3 precluding:1 interestingly:1 activation:3 dx:40 reminiscent:1 written:1 mesh:1 numerical:31 subsequent:1 partition:1 interpretable:1 generative:1 selected:2 intelligence:2 isotropic:1 ith:2 provides:5 location:4 accessed:1 hermite:2 five:2 mathematical:3 novak:1 constructed:1 beta:1 awc:1 shorthand:1 fitting:1 manner:1 x0:3 expected:1 ra:1 indeed:3 nor:1 mechanic:1 multi:2 morphology:1 simulator:1 planning:1 td:1 cpu:3 actual:2 jm:1 becomes:1 provided:6 begin:1 unrelated:1 moreover:1 notation:1 estimating:1 bounded:1 wolfson:1 panel:1 mass:1 medium:1 newcastle:1 emerging:1 supplemental:7 whilst:2 finding:2 l016710:1 guarantee:3 demonstrates:1 scaled:1 stick:4 grant:4 intervention:2 cardiovascular:1 positive:1 negligible:2 engineering:1 accordance:1 treat:1 local:1 limit:4 severely:1 consequence:1 solely:1 path:1 might:1 black:1 acta:1 studied:1 equivalence:1 suggests:1 challenging:1 limited:2 statistically:1 practical:2 acknowledgment:1 yj:1 atomic:2 practice:1 implement:1 lf:1 bootstrap:1 sullivan:1 procedure:1 area:1 empirical:5 gibson:1 physiology:2 projection:2 pre:1 confidence:4 cannot:4 operator:1 put:1 context:1 risk:1 writing:1 measurable:1 demonstrated:2 chronic:1 williams:1 attention:2 independently:1 estimator:1 rule:4 insight:1 jackson:1 embedding:2 variation:2 updated:1 annals:3 construction:2 suppose:1 exact:1 gps:1 element:3 trend:1 expensive:1 wolfe:1 hagan:3 bottom:1 steven:1 epsrc:4 ep:3 role:1 enters:1 electrical:1 wj:7 improper:1 eu:2 counter:1 principled:1 govern:1 complexity:3 neglected:1 motivate:2 raise:2 depend:1 predictive:4 upon:1 resolved:1 joint:2 represented:1 routinely:1 chapter:1 separated:1 distinct:2 fast:1 london:3 monte:8 artificial:2 hyper:10 whose:2 richer:1 larger:3 warwick:1 statistic:8 vaart:1 gp:11 analyse:1 itself:2 hoc:1 rr:1 mg:2 biophysical:1 took:1 interaction:1 product:1 remainder:2 adaptation:1 cockayne:2 relevant:1 mixing:2 gold:1 bed:4 description:3 intuitive:1 convergence:6 p:1 tract:1 converges:1 object:3 derive:2 minimises:1 measured:1 op:1 coverage:8 ois:1 involves:2 lee2:1 convention:1 appropriateness:1 artefact:1 direction:1 girolami:5 correct:1 transient:2 opinion:1 material:4 crc:1 behaviour:1 transparent:1 pacing:1 wall:1 rkk:2 hyde:1 hartikainen:1 extension:2 hold:1 therapy:1 considered:3 duvenaud:2 normal:4 exp:6 scope:2 predict:1 algorithmic:1 purpose:1 estimation:12 sensitive:1 establishes:1 weighted:1 gunter:1 fukumizu:1 mit:1 gaussian:6 aim:3 rather:2 husz:1 pn:3 endow:1 derived:1 focus:4 notational:1 methodological:2 rank:1 likelihood:4 bernoulli:1 contrast:2 baseline:1 helpful:1 inference:2 ferguson:2 spurious:1 relation:1 quasi:2 overall:2 dual:1 flexible:1 plank:1 denoted:1 priori:4 acutely:1 proposes:1 development:1 smoothing:2 special:1 integration:17 constrained:1 marginal:3 equal:1 construct:2 integrands:2 field:1 beach:1 sampling:3 reasoned:1 atom:1 represents:4 zz:1 r5:2 jones:1 future:2 mimic:1 jb:1 report:1 fundamentally:2 realisation:4 modern:1 unsound:2 diaconis:2 gamma:1 simultaneously:1 national:1 geometry:1 phase:1 statistician:1 n1:1 delicate:1 attempt:2 freedom:1 interest:1 evaluation:3 mixture:9 extreme:1 chain:1 amenable:1 accurate:1 integral:26 respective:2 bq:18 decoupled:1 iv:1 circle:1 causal:2 guidance:1 theoretical:13 uncertain:1 fitted:3 instance:2 modeling:1 cover:2 a6:1 cost:7 tractability:1 imperative:1 entry:2 rare:1 delay:1 gr:1 front:1 optimally:1 reported:4 kadane:1 kn:5 supposes:1 aw:1 dir:1 confident:2 st:1 density:6 fundamental:2 peak:1 international:2 probabilistic:11 lee:1 together:2 concrete:1 von:3 central:3 squared:1 again:1 reflect:1 necessitate:1 admit:1 american:2 supp:2 account:4 de:2 accompanied:1 centred:1 sec:1 student:3 lloyd:1 notable:1 explicitly:2 register:1 depends:3 sloan:1 mp:1 performed:1 view:2 jc:1 closed:8 red:1 realised:1 bayes:3 contribution:5 ass:1 formed:1 ni:12 variance:1 who:2 reserved:1 modelled:2 bayesian:17 thumb:1 lkopf:1 produced:3 iid:2 craig:1 carlo:8 mc:2 served:1 kennedy:1 straight:3 randomness:1 ah:1 herding:2 casella:1 definition:1 against:2 sriperumbudur:1 frequency:4 involved:2 james:1 dm:1 proof:2 associated:4 mi:3 propagated:1 sampled:1 treatment:1 recall:2 knowledge:2 improves:1 credible:3 hilbert:2 segmentation:1 amplitude:1 cj:4 sophisticated:1 goldstein:1 appears:1 centric:1 higher:1 follow:1 tension:3 wherein:1 response:1 evaluated:1 just:1 smola:1 correlation:1 until:1 hand:1 eqn:19 touch:1 nonlinear:1 resonance:1 assessment:15 gray:1 usa:1 facilitate:1 effect:1 contain:1 true:8 zarepour:1 deliberately:1 xavier:1 former:1 assigned:3 symmetric:1 white:1 gw:1 during:1 eqns:2 coincides:1 theoretic:1 passive:1 image:2 novel:1 recently:1 misspecified:2 common:1 functional:7 overview:1 cohen:1 insensitive:1 volume:5 extend:1 discussed:2 association:2 measurement:1 gibbs:4 smoothness:4 rd:1 consistency:1 zniakowski:1 mathematics:1 language:1 phenomenological:1 apex:1 calibration:1 f0:11 entail:1 acute:1 gj:5 base:6 posterior:20 multivariate:1 recent:2 confounding:2 belongs:1 apart:1 driven:1 scenario:2 claimed:1 arbitrarily:1 meeting:1 der:1 exploited:1 seen:3 p10:1 wasserstein:2 relaxed:1 additional:1 determine:1 arithmetic:1 smoother:1 full:3 multiple:1 ii:1 gretton:1 alan:1 segmented:1 academic:1 calculation:1 clinical:3 long:1 offer:1 bach:1 truong:1 hart:1 post:1 award:1 impact:1 prediction:3 schematic:2 implanted:1 expectation:5 physically:1 arxiv:4 kernel:20 represent:1 sometimes:1 sejdinovic:1 achieved:1 c1:1 proposal:1 whereas:2 fellowship:2 separately:1 interval:5 source:1 sch:1 posse:1 strict:1 subject:2 induced:1 dpm:5 valencia:1 practitioner:1 bernstein:2 revealed:1 intermediate:1 canadian:1 rendering:1 xj:1 restrict:1 bottleneck:1 whether:1 expression:4 motivated:1 t0:1 forecasting:1 wo:1 song:1 algebraic:1 proceed:1 constitute:2 ignored:1 useful:1 detailed:2 nonparametric:1 prepared:1 induces:2 restricts:1 nsf:2 notice:1 estimated:3 arising:1 discrete:1 numerica:1 hennig:2 key:1 pb:15 acknowledged:1 imperial:1 characterisation:2 drawn:1 mono:1 neither:1 ce:1 rectangle:1 y10:1 rinaldi:1 sum:2 wand:1 turing:1 inverse:2 prob:2 uncertainty:4 electronic:1 fran:1 sobolev:1 draw:4 decision:3 bound:2 paced:1 distinguish:1 encountered:1 oracle:1 n3:2 x2:1 flat:1 ventricular:1 integrand:6 aspect:1 argument:1 according:2 poor:1 conjugate:5 describes:1 cardiac:16 across:3 em:1 urgent:2 making:1 explained:1 restricted:1 taken:2 recourse:1 heart:4 equation:1 turn:2 merit:1 tractable:1 available:2 gaussians:1 aigrain:1 ishwaran:2 apply:1 observe:1 gam:1 generic:1 appropriate:1 magnetic:1 distinguished:2 alternative:1 shetty:1 slower:1 original:1 angela:1 dirichlet:6 denotes:4 top:2 a4:1 const:1 concatenated:1 epsilon:1 ghahramani:1 establish:3 classical:3 society:4 implied:1 question:1 quantity:1 parametric:5 concentration:2 dependence:2 usual:2 unclear:1 dp:18 distance:2 separate:1 simulated:3 lateral:1 chris:1 topic:1 argue:1 considers:1 extent:3 trivial:3 cellular:1 reason:1 fy:1 length:2 providing:1 dick:1 ratio:1 difficult:1 sinica:1 robert:4 frank:1 sigma:1 rise:2 numerics:2 implementation:1 design:1 motivates:1 unknown:21 upper:1 av:1 observation:1 markov:1 benchmark:2 finite:4 t:2 truncated:2 beat:1 situation:3 extended:2 variability:1 precise:1 y1:1 rn:2 reproducing:2 smoothed:1 thm:1 ghosal:1 introduced:2 bk:1 cast:2 required:3 pair:2 specified:1 extensive:1 clayton:1 distinction:1 established:2 hour:3 nip:1 address:1 beyond:1 precluded:1 proceeds:1 below:4 able:2 usually:1 wy:1 regime:2 challenge:1 reliable:1 royal:4 oates:4 event:1 suitable:1 natural:1 quantification:3 difficulty:1 business:1 marginalised:1 mn:5 representing:2 brief:1 sethuraman:1 concludes:1 coupled:2 naive:1 qmc:1 sn:2 prior:26 geometric:1 literature:1 asymptotic:2 law:2 fully:2 loss:1 lecture:1 foundation:2 degree:2 consistent:8 ventricle:4 principle:1 translation:1 sammut:1 placed:1 supported:6 truncation:1 tease:1 rasmussen:2 normalised:1 understand:1 vv:1 institute:2 fall:1 correspondingly:1 sparse:1 distributed:1 van:1 dimension:1 xn:2 default:1 wasilkowski:1 forward:6 collection:1 author:2 simplified:1 programme:1 functionals:5 sj:2 approximate:4 compact:1 informationtheoretic:1 unstudied:1 active:3 assumed:4 xi:23 continuous:4 latent:1 svensson:1 table:1 nature:2 transfer:1 kanagawa:1 ca:2 career:2 complex:3 necessarily:1 domain:2 garnett:3 main:3 spread:1 statistica:1 rh:1 motivation:1 s2:1 freedman:1 osborne:6 fair:1 allowed:1 quadrature:14 x1:3 fig:7 site:1 cubic:1 explicit:1 breaking:4 vanish:1 third:3 spatial:1 formula:2 theorem:1 briol:3 specific:1 showing:1 er:1 list:1 explored:2 decay:1 dk:1 concern:2 derives:1 intractable:4 intrinsic:2 consist:1 ih:1 evidence:1 fusion:1 supplement:1 magnitude:1 delyon:1 budget:2 conditioned:2 push:3 demand:1 chen:1 entropy:1 electrophysiology:2 simply:1 univariate:1 explore:1 expressed:3 contained:1 ethical:1 partially:1 recommendation:1 springer:1 rougier:1 determines:1 chance:1 ma:5 succeed:1 conditional:4 targeted:1 king:1 shared:1 typical:2 infinite:1 characterised:2 uniformly:1 sampler:3 conservative:2 called:7 total:2 kuo:1 experimental:3 select:1 college:2 mark:1 support:1 latter:2 dx0:2 assessed:1 constructive:1 evaluate:1 seheult:1 phenomenon:1 correlated:1 |
6,209 | 6,617 | Machine Learning with Adversaries:
Byzantine Tolerant Gradient Descent
El Mahdi El Mhamdi?
EPFL, Switzerland
[email protected]
Peva Blanchard
EPFL, Switzerland
[email protected]
Rachid Guerraoui
EPFL, Switzerland
[email protected]
Julien Stainer
EPFL, Switzerland
[email protected]
Abstract
We study the resilience to Byzantine failures of distributed implementations of
Stochastic Gradient Descent (SGD). So far, distributed machine learning frameworks have largely ignored the possibility of failures, especially arbitrary (i.e.,
Byzantine) ones. Causes of failures include software bugs, network asynchrony,
biases in local datasets, as well as attackers trying to compromise the entire system.
Assuming a set of n workers, up to f being Byzantine, we ask how resilient can
SGD be, without limiting the dimension, nor the size of the parameter space. We
first show that no gradient aggregation rule based on a linear combination of the vectors proposed by the workers (i.e, current approaches) tolerates a single Byzantine
failure. We then formulate a resilience property of the aggregation rule capturing
the basic requirements to guarantee convergence despite f Byzantine workers. We
propose Krum, an aggregation rule that satisfies our resilience property, which we
argue is the first provably Byzantine-resilient algorithm for distributed SGD. We
also report on experimental evaluations of Krum.
1
Introduction
The increasing amount of data available [6], together with the growing complexity of machine
learning models [27], has led to learning schemes that require a lot of computational resources. As a
consequence, most industry-grade machine-learning implementations are now distributed [1]. For
example, as of 2012, Google reportedly used 16.000 processors to train an image classifier [22]. More
recently, attention has been given to federated learning and federated optimization settings [15, 16, 23]
with a focus on communication efficiency. However, distributing a computation over several machines
(worker processes) induces a higher risk of failures. These include crashes and computation errors,
stalled processes, biases in the way the data samples are distributed among the processes, but also, in
the worst case, attackers trying to compromise the entire system. The most robust system is one that
tolerates Byzantine failures [17], i.e., completely arbitrary behaviors of some of the processes.
A classical approach to mask failures in distributed systems is to use a state machine replication
protocol [26], which requires however state transitions to be applied by all worker processes. In the
case of distributed machine learning, this constraint can be translated in two ways: either (a) the
processes agree on a sample of data based on which they update their local parameter vectors, or (b)
they agree on how the parameter vector should be updated. In case (a), the sample of data has to be
transmitted to each process, which then has to perform a heavyweight computation to update its local
?
contact author
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
parameter vector. This entails communication and computational costs that defeat the entire purpose
of distributing the work. In case (b), the processes have no way to check if the chosen update for the
parameter vector has indeed been computed correctly on real data: a Byzantine process could have
proposed the update and may easily prevent the convergence of the learning algorithm. Neither of
these solutions is satisfactory in a realistic distributed machine learning setting.
In fact, most learning algorithms today rely on a core component, namely stochastic gradient descent
(SGD) [4, 13], whether for training neural networks [13], regression [34], matrix factorization [12]
or support vector machines [34]. In all those cases, a cost function ? depending on the parameter
vector ? is minimized based on stochastic estimates of its gradient. Distributed implementations of
SGD [33] typically take the following form: a single parameter server is in charge of updating the
parameter vector, while worker processes perform the actual update estimation, based on the share of
data they have access to. More specifically, the parameter server executes learning rounds, during
each of which, the parameter vector is broadcast to the workers. In turn, each worker computes an
estimate of the update to apply (an estimate of the gradient), and the parameter server aggregates
their results to finally update the parameter vector. Today, this aggregation is typically implemented
through averaging [25], or variants of it [33, 18, 31].
This paper addresses the fundamental question of how a distributed SGD can be devised to tolerate f
Byzantine processes among the n workers.
Contributions. We first show in this paper that no linear combination (current approaches) of the
updates proposed by the workers can tolerate a single Byzantine worker. Basically, a single Byzantine
worker can force the parameter server to choose any arbitrary vector, even one that is too large in
amplitude or too far in direction from the other vectors. Clearly, the Byzantine worker can prevent
any classic averaging-based approach to converge. Choosing the appropriate aggregation of the
vectors proposed by the workers turns out to be challenging. A non-linear, squared-distance-based
aggregation rule, that selects, among the proposed vectors, the vector ?closest to the barycenter? (for
example by taking the vector that minimizes the sum of the squared distances to every other vector),
might look appealing. Yet, such a squared-distance-based aggregation rule tolerates only a single
Byzantine worker. Two Byzantine workers can collude, one helping the other to be selected, by
moving the barycenter of all the vectors farther from the ?correct area?.
We formulate a Byzantine resilience property capturing sufficient conditions for the parameter server?s
aggregation rule to tolerate f Byzantine workers. Essentially, to guarantee that the cost will decrease
despite Byzantine workers, we require the vector output chosen by the parameter server (a) to point,
on average, to the same direction as the gradient and (b) to have statistical moments (up to the fourth
moment) bounded above by a homogeneous polynomial in the moments of a correct estimator of
the gradient. One way to ensure such a resilience property is to consider a majority-based approach,
looking at every subset of n ? f vectors, and considering the subset with the smallest diameter.
While this approach is more robust to Byzantine workers that propose vectors far from the correct
area, its exponential computational cost is prohibitive. Interestingly, combining the intuitions of the
majority-based and squared-distance 2 -based methods, we can choose the vector that is somehow
the closest to its n ? f neighbors. Namely, the one that minimizes the sum of squared distances
to its n ? f closest vectors. This is the main idea behind our aggregation rule, we call Krum3 .
Assuming 2f + 2 < n, we show that Krum satisfies the resilience property aforementioned and the
corresponding machine learning scheme converges. An important advantage of Krum is its (local)
time complexity (O(n2 ? d)), linear in the dimension of the gradient, where d is the dimension of the
parameter vector. (In modern machine learning, the dimension d of the parameter vector may take
values in the hundreds of billions [30].) For simplicity of presentation, the version of Krum we first
consider selects only one vector. We also discuss other variants.
We evaluate Krum experimentally, and compare it to classical averaging. We confirm the very fact
that averaging does not stand Byzantine attacks, while Krum does. In particular, we report on attacks
by omniscient adversaries ? aware of a good estimate of the gradient ? that send the opposite vector
multiplied by a large factor, as well as attacks by adversaries that send random vectors drawn from a
Gaussian distribution (the larger the variance of the distribution, the stronger the attack). We also
2
In all this paper, distances are computed with the Euclidean norm.
Krum, in Greek ???????, was a Bulgarian Khan of the end of the eighth century, who undertook offensive
attacks against the Byzantine empire. Bulgaria doubled in size during his reign.
3
2
evaluate the extent to which Krum might slow down learning (compared to averaging) when there
are no Byzantine failures. Interestingly, as we show experimentally, this slow down occurs only
when the mini-batch size is close to 1. In fact, the slowdown can be drastically reduced by choosing
a reasonable mini-batch size. We also evaluate Multi-Krum, a variant of Krum, which, intuitively,
interpolates between Krum and averaging, thereby allowing to mix the resilience properties of Krum
with the convergence speed of averaging. Multi-Krum outperforms other aggregation rules like the
medoid, inspired by the geometric median.
Paper Organization. Section 2 recalls the classical model of distributed SGD. Section 3 proves that
linear combinations (solutions used today) are not resilient even to a single Byzantine worker, then
introduces our new concept of (?, f )-Byzantine resilience. Section 4 introduces our Krum function,
computes its computational cost and proves its (?, f )-Byzantine resilience. Section 5 analyzes the
convergence of a distributed SGD using Krum. Section 6 presents our experimental evaluation of
Krum. We discuss related work and open problems in Section 7. Due to space limitations, some
proofs and complementary experimental results are given as supplementary material.
2
Model
We consider the general distributed system model of [1], consisting of a parameter server4 , and
n workers, f of them possibly Byzantine (behaving arbitrarily). Computation is divided into (infinitely many) synchronous rounds. During round t, the parameter server broadcasts its parameter
vector xt ? Rd to all the workers. Each correct worker p computes an estimate Vpt = G(xt , ?pt ) of
the gradient ?Q(xt ) of the cost function Q, where ?pt is a random variable representing, e.g.,
the sample (or a mini-batch of samples) drawn from the dataset. A Byzantine worker b proposes a vector Vbt which can deviate arbitrarily from the vector it is supposed to send if it was
correct, i.e., according to the algorithm assigned to it by the system developer (see Figure 1).
Since the communication is synchronous, if the
parameter server does not receive a vector value
Vbt from a given Byzantine worker b, then the
parameter server acts as if it had received the
default value Vbt = 0 instead.
The parameter server computes a vector
F (V1t , . . . , Vnt ) by applying a deterministic
function F (aggregation rule) to the vectors received. We refer to F as the aggregation rule of
the parameter server. The parameter server updates the parameter vector using the following
SGD equation
xt+1 = xt ? ?t ? F (V1t , . . . , Vnt ).
Figure 1: The gradient estimates computed by correct workers (black dashed arrows) are distributed
around the actual gradient (solid arrow) of the cost
function (thin black curve). A Byzantine worker
can propose an arbitrary vector (red dotted arrow).
The correct (non-Byzantine) workers are assumed to compute unbiased estimates of the gradient
?Q(xt ). More precisely, in every round t, the vectors Vit ?s proposed by the correct workers are
independent identically distributed random vectors, Vit ? G(xt , ?it ) with E?it G(xt , ?it ) = ?Q(xt ).
This can be achieved by ensuring that each sample of data used for computing the gradient is drawn
uniformly and independently, as classically assumed in the literature of machine learning [3].
The Byzantine workers have full knowledge of the system, including the aggregation rule F as well
as the vectors proposed by the workers. They can furthermore collaborate with each other [21].
3
Byzantine Resilience
In most SGD-based learning algorithms used today [4, 13, 12], the aggregation rule consists in
computing the average 5 of the input vectors. Lemma 1 below states that no linear combination of the
vectors can tolerate a single Byzantine worker. In particular, averaging is not Byzantine resilient.
4
The parameter server is assumed to be reliable. Classical techniques of state-machine replication can be
used to ensure this.
5
Or a closely related rule.
3
Pn
Lemma 1. Consider an aggregation rule Flin of the form Flin (V1 , . . . , Vn ) = i=1 ?i ? Vi , where
the ?i ?s are non-zero scalars. Let U be any vector in Rd . A single Byzantine worker can make F
always select U . In particular, a single Byzantine worker can prevent convergence.
Proof. Immediate: if the Byzantine worker proposes Vn =
1
?n
?U ?
Pn?1
?i
i=1 ?n Vi ,
then F = U .6
In the following, we define basic requirements on an appropriate Byzantine-resilient aggregation rule.
Intuitively, the aggregation rule should output a vector F that is not too far from the ?real? gradient g,
more precisely, the vector that points to the steepest direction of the cost function being optimized.
This is expressed as a lower bound (condition (i)) on the scalar product of the (expected) vector F
and g. Figure 2 illustrates the situation geometrically. If EF belongs to the ball centered at g with
radius r, then the scalar product is bounded below by a term involving sin ? = r/kgk.
Condition (ii) is more technical, and states that the moments of F should be controlled by the
moments of the (correct) gradient estimator G. The bounds on the moments of G are classically used
to control the effects of the discrete nature of the SGD dynamics [3]. Condition (ii) allows to transfer
this control to the aggregation rule.
Definition 1 ((?, f )-Byzantine Resilience). Let 0 ? ? < ?/2 be any angular value, and any
integer 0 ? f ? n. Let V1 , . . . , Vn be any independent identically distributed random vectors in Rd ,
Vi ? G, with EG = g. Let B1 , . . . , Bf be any random vectors in Rd , possibly dependent on the Vi ?s.
aggregation rule F is said to be (?, f )-Byzantine resilient if, for any 1 ? j1 < ? ? ? < jf ? n, vector
F = F (V1 , . . . , B1 , . . . , Bf , . . . , Vn )
|{z}
|{z}
j1
jf
r
satisfies (i) hEF, gi ? (1 ? sin ?) ? kgk2 > 0 and (ii) for r = 2, 3, 4, E kF k is bounded above by a
r
r
linear combination of terms E kGk 1 . . . E kGk n?1 with r1 + ? ? ? + rn?1 = r.
4
The Krum Function
We now introduce Krum, our aggregation
rule, which, we show, satisfies the (?, f )Byzantine resilience condition. The
Pn barycentric aggregation rule Fbary = n1 i=1 Vi can
g
r
be defined as the vector in Rd that mini?
7
mizes the sum of squared distances to the
Pn
2
Vi ?s i=1 kFbary ? Vi k . Lemma 1, however, Figure 2: If kEF ? gk ? r then hEF, gi is
states that this approach does not tolerate even a bounded below by (1 ? sin ?)kgk2 where sin ? =
single Byzantine failure. One could try to select r/kgk.
the vector U among the Vi ?s which minimizes
P
2
the sum i kU ? Vi k , i.e., which is ?closest
to all vectors?. However, because such a sum involves all the vectors, even those which are very far,
this approach does not tolerate Byzantine workers: by proposing large enough vectors, a Byzantine
worker can force the total barycenter to get closer to the vector proposed by another Byzantine worker.
Our approach to circumvent this issue is to preclude the vectors that are too far away. More precisely,
we define our Krum aggregation rule K R(V1 , . . . , Vn ) as follows. For any i 6= j, we denote by i ? j
the fact that Vj belongs to the n ? f ? 2 closest vectors to Vi . Then, we define for each worker i, the
P
2
score s(i) = i?j kVi ? Vj k where the sum runs over the n ? f ? 2 closest vectors to Vi . Finally,
K R(V1 , . . . , Vn ) = Vi? where i? refers to the worker minimizing the score, s(i? ) ? s(i) for all i.8
Lemma 2. The expected time complexity of the Krum Function K R(V1 , . . . , Vn ), where V1 , . . . , Vn
are d-dimensional vectors, is O(n2 ? d)
6
Note that the parameter server could cancel the effects of the Byzantine behavior by setting, for example,
?n to 0. This however requires means to detect which worker is Byzantine.
7
Removing the square of the distances leads to the geometric median, we discuss this when optimizing Krum.
8
If two or more workers have the minimal score, we choose the one with the smallest identifier.
4
2
Proof. For each Vi , the parameter server computes the n squared distances kVi ? Vj k (time O(n?d)).
Then the parameter server selects the first n ? f ? 1 of these distances (expected time O(n) with
Quickselect) and sums their values (time O(n ? d)). Thus, computing the score of all the Vi ?s takes
O(n2 ? d). An additional term O(n) is required to find the minimum score, but is negligible relatively
to O(n2 ? d).
Proposition 1 below states that, if 2f + 2 < n and the gradient estimator is accurate enough, (its
standard deviation is relatively small compared to the norm of the gradient), then the Krum function
is (?, f )-Byzantine-resilient, where angle ? depends on the ratio of the deviation over the gradient.
Proposition 1. Let V1 , . . . , Vn be any independent and identically distributed random d-dimensional
2
vectors s.t Vi ? G, with EG = g and E kG ? gk = d? 2 . Let
? B1 , . . . , Bf be any f random vectors,
possibly dependent on the Vi ?s. If 2f + 2 < n and ?(n, f ) d ? ? < kgk, where
s
f ? (n ? f ? 2) + f 2 ? (n ? f ? 1)
O(n)
if f = O(n)
?
?(n, f ) =
2 n?f +
,
=
O( n) if f = O(1)
def
n ? 2f ? 2
then the Krum function K R is (?, f )-Byzantine resilient where 0 ? ? < ?/2 is defined by
?
?(n, f ) ? d ? ?
.
sin ? =
kgk
?
The condition on the norm of the gradient, ?(n, f ) ? d ? ? < kgk, can be satisfied, to a certain
extent, by having the (correct) workers compute their gradient estimates on mini-batches [3]. Indeed,
averaging the gradient estimates over a mini-batch divides the deviation ? by the squared root of the
size of the mini-batch. For the sake of concision, we only give here the sketch of the proof. (We give
the detailed proof in the supplementary material.)
Proof. (Sketch) Without loss of generality, we assume that the Byzantine vectors B1 , . . . , Bf occupy
the last f positions in the list of arguments of K R, i.e., K R = K R(V1 , . . . , Vn?f , B1 , . . . , Bf ).
Let i? be the index of the vector chosen by the Krum function. We focus on the condition (i) of
(?, f )-Byzantine resilience (Definition 1).
Consider first the case where Vi? = Vi ? {V1 , . . . , Vn?f } is a vector proposed by a correct process.
The first step is to compare the vector Vi with the average of the correct vectors Vj such that i ? j.
Let ?c (i) be the number of such Vj ?s.
2
X
X
1
1
2
E
Vi ?
?
Vj
E kVi ? Vj k ? 2d? 2 .
(1)
?
(i)
?
(i)
c
c
i? correct j
i? correct j
The last inequality holds because the right-hand side of the first inequality involves only vectors
proposed by correct processes, which are mutually independent and follow the distribution of G.
Now, consider the case where Vi? = Bk ? {B1 , . . . , Bf } is a vector proposed by a Byzantine
process. The fact that k minimizes the score implies that for all indices i of vectors proposed by
correct processes
X
X
X
X
2
2
2
2
kBk ? Vj k +
kBk ? Bl k ?
kVi ? Vj k +
kVi ? Bl k .
k? correct j
i? correct j
k? byz l
i? byz l
Then, for all indices i of vectors proposed by correct processes
2
X
X
X
1
1
2
2
Bk ?
? 1
V
kV
?
V
k
+
kVi ? Bl k .
j
i
j
?c (k)
?c (k) i? correct j
?c (k)
k? correct j
i? byz l
|
{z
}
D 2 (i)
The term D2 (i) is the only term involving vectors proposed by Byzantine processes. However, the
correct process i has n ? f ? 2 neighbors and f + 1 non-neighbors. Therefore, there exists a correct
5
process ?(i) which is farther from i than every neighbor j of i (including the Byzantine neighbors).
In particular, for all l such that i ? l, kVi ? Bl k2 ? kVi ? V?(i) k2 . Thus
2
X
X
1
1
n ? f ? 2 ? ?c (i)
2
Vi ? V?(i)
2 .
Bk ?
?
Vj
kVi ? Vj k +
?c (k)
?c (k) i? correct j
?c (k)
k? correct j
(2)
?
Combining equations 1, 2, and a union bound yields kEK R ? gk2 ? ? dkgk, which, in turn, implies
hEK R, gi ? (1 ? sin ?)kgk2 . Condition (ii) is proven by bounding the moments of K R with moments
of the vectors proposed by the correct processes only, using the same technique as above. The full
proof is provided in the supplementary material.
5
Convergence Analysis
In this section, we analyze the convergence of the SGD using our Krum function defined in Section 4.
The SGD equation is expressed as follows
xt+1 = xt ? ?t ? K R(V1t , . . . , Vnt )
where at least n ? f vectors among the Vit ?s are correct, while the other ones may be Byzantine. For
a correct index i, Vit = G(xt , ?it ) where G is the gradient estimator. We define the local standard
deviation ?(x) by
2
d ? ? 2 (x) = E kG(x, ?) ? ?Q(x)k .
The following proposition considers an (a priori) non-convex cost function. In the context of nonconvex optimization, even in the centralized case, it is generally hopeless to aim at proving that the
parameter vector xt tends to a local minimum. Many criteria may be used instead. We follow [3],
and we prove that the parameter vector xt almost surely reaches a ?flat? region (where the norm of
the gradient is small), in a sense explained below.
Proposition 2. We assume that (i) the cost function Q is three times differentiable
with continuous
P
P
derivatives, and is non-negative, Q(x) ? 0; (ii) the learning rates satisfy t ?t = ? and t ?t2 <
?; (iii) the gradient estimator satisfies EG(x, ?) = ?Q(x) and ?r ? {2, . . . , 4}, EkG(x, ?)kr ?
Ar + Br kxkr for some constants Ar , Br ; (iv) there exists a constant 0 ? ? < ?/2 such that for all x
?
?(n, f ) ? d ? ?(x) ? k?Q(x)k ? sin ?;
(v) finally, beyond a certain horizon, kxk2 ? D, there exist > 0 and 0 ? ? < ?/2 ? ? such that
hx,?Q(x)i
? cos ?. Then the sequence of gradients ?Q(xt ) converges
k?Q(x)k ? > 0 and kxk?k?Q(x)k
almost surely to zero.
Conditions (i) to (iv) are the same conditions as
in the non-convex convergence analysis in [3].
Condition (v) is a slightly stronger condition
than the corresponding one in [3], and states
?
? ?Q(xt )
? d?
that, beyond a certain horizon, the cost function
?
Q is ?convex enough?, in the sense that the dixt
rection of the gradient is sufficiently close to the
direction of the parameter vector x. Condition
(iv), however, states that the gradient estimator Figure 3: Condition on the angles between x ,
used by the correct workers has to be accurate ?Q(x ) and EK R , in the region kx k2 > D. t
t
t
t
enough, i.e., the local standard deviation should
be small relatively to the norm of the gradient.
Of course, the ?
norm of the gradient tends to zero near, e.g., extremal and saddle points. Actually, the
ratio ?(n, f ) ? d ? ?/ k?Qk controls the maximum angle between the
? gradient ?Q and the vector
chosen by the Krum function. In the regions where k?Qk < ?(n, f ) ? d ? ?, the Byzantine workers
may take advantage of the noise (measured by ?) in the gradient estimator G to bias the choice of
the parameter server. Therefore, Proposition 2 is to be interpreted as follows: in the presence of
Byzantine workers, the parameter vector
? xt almost surely reaches a basin around points where the
gradient is small (k?Qk ? ?(n, f ) ? d ? ?), i.e., points where the cost landscape is ?almost flat?.
Note that the convergence analysis is based only on the fact that function K R is (?, f )-Byzantine
resilient. The complete proof of Proposition 2 is deferred to the supplementary material.
6
0% byzantine
1
33% byzantine
1
average
krum
0.6
0.6
average
krum
error
0.8
error
0.8
0.4
0.4
0.2
0.2
0
0
100
200
300
400
0
500
round
0
100
200
300
400
500
round
Figure 4: Cross-validation error evolution with rounds, respectively in the absence and in the
presence of 33% Byzantine workers. The mini-batch size is 3. With 0% Gaussian Byzantine workers,
averaging converges faster than Krum. With 33% Gaussian Byzantine workers, averaging does not
converge, whereas Krum behaves as if there were 0% Byzantine workers.
6
Experimental Evaluation
We report here on the evaluation of the convergence and resilience properties of Krum, as well as an
optimized variant of it. (We also discuss other variants of Krum in the supplementary material.)
(Resilience to Byzantine processes). We consider the task of spam filtering (dataset spambase [19]). The learning model is a multi-layer perceptron (MLP) with two hidden layers. There are
n = 20 worker processes. Byzantine processes propose vectors drawn from a Gaussian distribution
with mean zero, and isotropic covariance matrix with standard deviation 200. We refer to this behavior
as Gaussian Byzantine. Each (correct) worker estimates the gradient on a mini-batch of size 3. We
measure the error using cross-validation. Figure 4 shows how the error (y-axis) evolves with the
number of rounds (x-axis).
In the first plot (left), there are no Byzantine workers. Unsurprisingly, averaging converges faster
than Krum. In the second plot (right), 33% of the workers are Gaussian Byzantine. In this case,
averaging does not converge at all, whereas Krum behaves as if there were no Byzantine workers.
This experiment confirms that averaging does not tolerate (the rather mild) Gaussian Byzantine
behavior, whereas Krum does.
(The Cost of Resilience). As seen above, Krum slows down learning when there are no Byzantine
workers. The following experiment shows that this overhead can be significantly reduced by slightly
increasing the mini-batch size. To highlight the effect of the presence of Byzantine workers, the
Byzantine behavior has been set as follows: each Byzantine worker computes an estimate of the
gradient over the whole dataset (yielding a very accurate estimate of the gradient), and proposes the
opposite vector, scaled to a large length. We refer to this behavior as omniscient.
Figure 5 displays how the error value at the 500-th round (y-axis) evolves when the mini-batch size
varies (x-axis). In this experiment, we consider the tasks of spam filtering (dataset spambase) and
image classification (dataset MNIST). The MLP model is used in both cases. Each curve is obtained
with either 0 or 45% of omniscient Byzantine workers.
In all cases, averaging still does not tolerate Byzantine workers, but yields the lowest error when
there are no Byzantine workers. However, once the size of the mini-batch reaches the value 20, Krum
with 45% omniscient Byzantine workers is as accurate as averaging with 0% Byzantine workers. We
observe a similar pattern for a ConvNet as provided in the supplementary material.
(Multi-Krum). For the sake of presentation simplicity, we considered a version of Krum which
selects only one vector among the vector proposed by the workers. We also propose a variant of
Krum, we call Multi-Krum. Multi-Krum computes, for each vector proposed, the score as in the
Krum function. Then, Multi-Krum P
selects the m ? {1, . . . , n} vectors V1? , . . . , Vm? which score the
1
best, and outputs their average m i Vi? . Note that, the cases m = 1 and m = n correspond to
Krum and averaging respectively.
Figure 6 shows how the error (y-axis) evolves with the number of rounds (x-axis). In the figure, we
consider the task of spam filtering (dataset spambase), and the MLP model (the same comparison
7
mnist
0.6
0.4
0.2
0
spambase
1
0.5
0.8
0.4
error at round 500
byz)
byz)
byz)
byz)
error at round 500
error at round 500
0.8
average (0%
krum (0%
average (45%
krum (45%
0.6
0.4
0.2
40
80
0.4
0.3
0.2
0.1
0
120
mnist
0.5
average (0% byz)
krum (0% byz)
krum (45% byz)
error at round 500
spambase
1
40
batch size
80
120
0.2
0.1
0
160
0.3
10
batch size
20
30
40
batch size
0
10
20
30
40
batch size
Figure 5: Cross-validation error at round 500 when increasing the mini-batch size. The two figures
on the rights are zoomed versions of the two on the left). With a reasonably large mini-batch size
(arround 10 for MNIST and 30 for Spambase), Krum with 45% omniscient Byzantine workers is as
accurate as averaging with 0% Byzantine workers.
multi-krum
1
average (0% byz)
krum (33% byz)
multi-krum (33% byz)
0.8
error
0.6
0.4
0.2
0
0
40
80
120
160
200
240
280
320
360
400
440
480
round
Figure 6: Cross-validation error evolution with rounds. The mini-batch size is 3. Multi-Krum with
33% Gaussian Byzantine workers converges as fast as averaging with 0% Byzantine workers.
is done for the task of image classification with a ConvNet and is provided in the supplementary
material). The Multi-Krum parameter m is set to m = n ? f . Figure 6 shows that Multi-Krum with
33% Byzantine workers is as efficient as averaging with 0% Byzantine workers.
From the practitionner?s perspective, the parameter m may be used to set a specific trade-off between
convergence speed and resilience to Byzantine workers.
7
Concluding Remarks
(The Distributed Computing Perspective). Although seemingly related, results in d-dimensional
approximate agreement [24, 14] cannot be applied to our Byzantine-resilient machine context for
the following reasons: (a) [24, 14] assume that the set of vectors that can be proposed to an instance
of the agreement is bounded so that at least f + 1 correct workers propose the same vector, which
would require a lot of redundant work in our setting; and more importantly, (b) [24] requires a local
computation by each worker that is in O(nd ). While this cost seems reasonable for small dimensions,
such as, e.g., mobile robots meeting in a 2D or 3D space, it becomes a real issue in the context
of machine learning, where d may be as high as 160 billion [30] (making d a crucial parameter
when considering complexities, either for local computations, or for communication rounds). The
expected time complexity of the Krum function is O(n2 ? d). A closer approach to ours has been
recently proposed in [28, 29]. In [28], the study only deals with parameter vectors of dimension
one, which is too restrictive for today?s multi-dimensional machine learning. In [29], the authors
tackle a multi-dimensional situation, using an iterated approximate Byzantine agreement that reaches
consensus asymptotically. This is however only achieved on a finite set of possible environmental
states and cannot be used in the continuous context of stochastic gradient descent.
(The Statistics and Machine Learning View). Our work looks at the resilience of the aggregation
rule using ideas that are close to those of [11], and somehow classical in theoretical statistics on
the robustness of the geometric median and the notion of breakdown [7]. The closest concept to a
breakdown in our work is the maximum fraction of Byzantine workers that can be tolerated, i.e. n?2
2n ,
which reaches the optimal theoretical value (1/2) asymptotically on n. It is known that the geometric
8
median does achieve the optimal breakdown. However, no closed form nor an exact algorithm to
compute the geometric median is known (only approximations are available [5] and their Byzantine
resilience is an open problem.). An easily computable variant of the median is the Medoid, which is
the proposed vector minimizing the sum of distances to all other proposed vectors. The Medoid can
be computed with a similar algorithm to Krum. We show however in the supplementary material that
the implementation of the Medoid is outperformed by multi-Krum.
(Robustness Within the Model). It is important to keep in mind that this work deals with robustness
from a coarse-grained perspective: the unit of failure is a worker, receiving its copy of the model and
estimating gradients, based on either local data or delegated data from a server. The nature of the
model itself is not important, the distributed system can be training models spanning a large range
from simple regression to deep neural networks. As long as this training is using gradient-based
learning, our algorithm to aggregate gradients, Krum, provably ensures convergence when a simple
majority of nodes are not compromised by an attacker.
A natural question to consider is the fine-grained view: is the model itself robust to internal perturbations? In the case of neural networks, this question can somehow be tied to neuroscience
considerations: could some neurons and/or synapses misbehave individually without harming the
global outcome? We formulated this question in another work and proved a tight upper bound on the
resulting global error when a set of nodes is removed or is misbehaving [8]. One of the many practical
consequences [9] of such fine-grained view is the understanding of memory cost reduction trade-offs
in deep learning. Such memory cost reduction can be viewed as the introduction of precision errors
at the level of each neuron and/or synapse [8].
Other approaches to robustness within the model tackled adversarial situations in machine learning
with a focus on adversarial examples (during inference) [10, 32, 11] instead of adversarial gradients
(during training) as we did for Krum. Robustness to adversarial input can be viewed through the
fine-grained lens we introduced in [8], for instance, one can see perturbations of pixels in the
inputs as perturbations of neurons in layer zero. It is important to note the orthogonality and
complementarity between the fine-grained (model/input units) and the coarse-grained (gradient
aggregation) approaches. Being robust, as a model, either to adversarial examples or to internal
perturbations, does not necessarily imply robustness to adversarial gradients during training. Similarly,
being distributively trained with a robust aggregation scheme such as Krum does not necessarily
imply robustness to internal errors of the model or adversarial input perturbations that would occur
later during inference. For instance, the theory we develop in the present work is agnostic to the
model being trained or the technology of the hardware hosting it, as long as there are gradients to be
aggregated.
Acknowledgment. The authors would like to thank Georgios Damaskinos and Rhicheek Patra from
the Distributed Computing group at EPFL for kindly providing their distributed machine learning
framework, on top of which we could test our algorithm, Krum, and its variants described in this
work. Further implementation details and additional experiments will be posted in the lab?s Github
repository [20]. The authors would also like to thank Saad Benjelloun, L? Nguyen Hoang and
S?bastien Rouault for fruitful comments. This work has been supported in part by the European
ERC (Grant 339539 - AOC) and by the Swiss National Science Foundation (Grant 200021_ 169588
TARBDA). A preliminary version of this work appeared as a brief announcement during the 36st
ACM Symposium on Principles of Distributed Computing [2].
9
References
[1] M. Abadi, P. Barham, J. Chen, Z. Chen, A. Davis, J. Dean, M. Devin, S. Ghemawat, G. Irving,
M. Isard, et al. Tensorflow: A system for large-scale machine learning. In Proceedings of the
12th USENIX Symposium on Operating Systems Design and Implementation (OSDI). Savannah,
Georgia, USA, 2016.
[2] P. Blanchard, E. M. El Mhamdi, R. Guerraoui, and J. Stainer. Brief announcement: Byzantinetolerant machine learning. In Proceedings of the ACM Symposium on Principles of Distributed
Computing, PODC ?17, pages 455?457, New York, NY, USA, 2017. ACM.
[3] L. Bottou. Online learning and stochastic approximations. Online learning in neural networks,
17(9):142, 1998.
[4] L. Bottou. Large-scale machine learning with stochastic gradient descent. In Proceedings of
COMPSTAT?2010, pages 177?186. Springer, 2010.
[5] M. B. Cohen, Y. T. Lee, G. Miller, J. Pachocki, and A. Sidford. Geometric median in nearly linear
time. In Proceedings of the 48th Annual ACM SIGACT Symposium on Theory of Computing,
pages 9?21. ACM, 2016.
[6] J. Dean, G. Corrado, R. Monga, K. Chen, M. Devin, M. Mao, A. Senior, P. Tucker, K. Yang,
Q. V. Le, et al. Large scale distributed deep networks. In Advances in neural information
processing systems, pages 1223?1231, 2012.
[7] D. L. Donoho and P. J. Huber. The notion of breakdown point. A festschrift for Erich L.
Lehmann, 157184, 1983.
[8] E. M. El Mhamdi and R. Guerraoui. When neurons fail. In 2017 IEEE International Parallel
and Distributed Processing Symposium (IPDPS), pages 1028?1037, May 2017.
[9] E. M. El Mhamdi, R. Guerraoui, and S. Rouault. On the robustness of a neural network. In
2017 IEEE 36th Symposium on Reliable Distributed Systems (SRDS), pages 84?93, Sept 2017.
[10] A. Fawzi, S.-M. Moosavi-Dezfooli, and P. Frossard. Robustness of classifiers: from adversarial
to random noise. In Advances in Neural Information Processing Systems, pages 1624?1632,
2016.
[11] J. Feng, H. Xu, and S. Mannor. Outlier robust online learning. arXiv preprint arXiv:1701.00251,
2017.
[12] R. Gemulla, E. Nijkamp, P. J. Haas, and Y. Sismanis. Large-scale matrix factorization with
distributed stochastic gradient descent. In Proceedings of the 17th ACM SIGKDD international
conference on Knowledge discovery and data mining, pages 69?77. ACM, 2011.
[13] S. S. Haykin. Neural networks and learning machines, volume 3. Pearson Upper Saddle River,
NJ, USA:, 2009.
[14] M. Herlihy, S. Rajsbaum, M. Raynal, and J. Stainer. Computing in the presence of concurrent
solo executions. In Latin American Symposium on Theoretical Informatics, pages 214?225.
Springer, 2014.
[15] J. Kone?cn`y, B. McMahan, and D. Ramage. Federated optimization: Distributed optimization
beyond the datacenter. arXiv preprint arXiv:1511.03575, 2015.
[16] J. Kone?cn`y, H. B. McMahan, F. X. Yu, P. Richt?rik, A. T. Suresh, and D. Bacon. Federated
learning: Strategies for improving communication efficiency. arXiv preprint arXiv:1610.05492,
2016.
[17] L. Lamport, R. Shostak, and M. Pease. The byzantine generals problem. ACM Transactions on
Programming Languages and Systems (TOPLAS), 4(3):382?401, 1982.
[18] X. Lian, Y. Huang, Y. Li, and J. Liu. Asynchronous parallel stochastic gradient for nonconvex
optimization. In Advances in Neural Information Processing Systems, pages 2737?2745, 2015.
10
[19] M. Lichman. UCI machine learning repository, 2013.
[20] LPD-EPFL. The implementation is part of a larger distributed framework to run sgd in a reliable
distributed fashion and will be released in the github repository of the distributed computing
group at epfl, https://github.com/lpd-epfl.
[21] N. A. Lynch. Distributed algorithms. Morgan Kaufmann, 1996.
[22] J. Markoff. How many computers to identify a cat? 16,000. New York Times, pages 06?25,
2012.
[23] B. McMahan, E. Moore, D. Ramage, S. Hampson, and B. A. y Arcas. Communication-efficient
learning of deep networks from decentralized data. In Artificial Intelligence and Statistics,
pages 1273?1282, 2017.
[24] H. Mendes and M. Herlihy. Multidimensional approximate agreement in byzantine asynchronous systems. In Proceedings of the forty-fifth annual ACM symposium on Theory of
computing, pages 391?400. ACM, 2013.
[25] B. T. Polyak and A. B. Juditsky. Acceleration of stochastic approximation by averaging. SIAM
Journal on Control and Optimization, 30(4):838?855, 1992.
[26] F. B. Schneider. Implementing fault-tolerant services using the state machine approach: A
tutorial. ACM Computing Surveys (CSUR), 22(4):299?319, 1990.
[27] R. K. Srivastava, K. Greff, and J. Schmidhuber. Training very deep networks. In Advances in
neural information processing systems, pages 2377?2385, 2015.
[28] L. Su and N. H. Vaidya. Fault-tolerant multi-agent optimization: optimal iterative distributed
algorithms. In Proceedings of the 2016 ACM Symposium on Principles of Distributed Computing,
pages 425?434. ACM, 2016.
[29] L. Su and N. H. Vaidya. Non-bayesian learning in the presence of byzantine agents. In
International Symposium on Distributed Computing, pages 414?427. Springer, 2016.
[30] A. Trask, D. Gilmore, and M. Russell. Modeling order in neural word embeddings at scale. In
ICML, pages 2266?2275, 2015.
[31] J. Tsitsiklis, D. Bertsekas, and M. Athans. Distributed asynchronous deterministic and stochastic
gradient optimization algorithms. IEEE transactions on automatic control, 31(9):803?812,
1986.
[32] B. Wang, J. Gao, and Y. Qi. A theoretical framework for robustness of (deep) classifiers under
adversarial noise. arXiv preprint arXiv:1612.00334, 2016.
[33] S. Zhang, A. E. Choromanska, and Y. LeCun. Deep learning with elastic averaging sgd. In
Advances in Neural Information Processing Systems, pages 685?693, 2015.
[34] T. Zhang. Solving large scale linear prediction problems using stochastic gradient descent
algorithms. In Proceedings of the twenty-first international conference on Machine learning,
page 116. ACM, 2004.
11
| 6617 |@word mild:1 kgk:7 repository:3 version:4 hampson:1 polynomial:1 stronger:2 norm:6 nd:1 bf:6 seems:1 open:2 d2:1 confirms:1 covariance:1 sgd:15 thereby:1 solid:1 reduction:2 moment:8 liu:1 score:8 lichman:1 ours:1 interestingly:2 omniscient:5 outperforms:1 spambase:6 current:2 com:1 collude:1 yet:1 devin:2 realistic:1 j1:2 plot:2 update:9 juditsky:1 isard:1 selected:1 prohibitive:1 intelligence:1 undertook:1 isotropic:1 steepest:1 core:1 farther:2 haykin:1 coarse:2 mannor:1 node:2 attack:5 zhang:2 symposium:10 replication:2 abadi:1 consists:1 prove:1 overhead:1 byzantine:94 introduce:1 mask:1 expected:4 indeed:2 frossard:1 behavior:6 nor:2 growing:1 multi:16 grade:1 v1t:3 inspired:1 actual:2 preclude:1 considering:2 increasing:3 becomes:1 provided:3 estimating:1 bounded:5 agnostic:1 lowest:1 kg:2 interpreted:1 minimizes:4 developer:1 proposing:1 nj:1 guarantee:2 every:4 multidimensional:1 act:1 charge:1 tackle:1 huber:1 classifier:3 k2:3 scaled:1 control:5 unit:2 grant:2 datacenter:1 bertsekas:1 negligible:1 service:1 local:10 resilience:19 tends:2 consequence:2 sismanis:1 despite:2 might:2 black:2 challenging:1 co:1 factorization:2 hek:1 range:1 practical:1 acknowledgment:1 lecun:1 union:1 swiss:1 ipdps:1 suresh:1 area:2 significantly:1 word:1 refers:1 srds:1 doubled:1 get:1 close:3 cannot:2 risk:1 applying:1 context:4 fruitful:1 deterministic:2 dean:2 send:3 rachid:2 attention:1 compstat:1 vit:4 independently:1 convex:3 formulate:2 survey:1 simplicity:2 rule:22 estimator:7 importantly:1 his:1 classic:1 century:1 proving:1 notion:2 limiting:1 updated:1 pt:2 today:5 delegated:1 exact:1 programming:1 homogeneous:1 agreement:4 complementarity:1 updating:1 breakdown:4 preprint:4 wang:1 worst:1 region:3 ensures:1 richt:1 decrease:1 trade:2 removed:1 russell:1 intuition:1 byz:13 complexity:5 concision:1 dynamic:1 trained:2 tight:1 solving:1 compromise:2 efficiency:2 rouault:2 completely:1 translated:1 easily:2 cat:1 train:1 fast:1 artificial:1 aggregate:2 choosing:2 outcome:1 pearson:1 guerraoui:5 larger:2 supplementary:8 vpt:1 statistic:3 gi:3 itself:2 seemingly:1 online:3 advantage:2 differentiable:1 sequence:1 propose:6 product:2 zoomed:1 uci:1 combining:2 dezfooli:1 achieve:1 supposed:1 bug:1 kv:1 billion:2 convergence:12 requirement:2 r1:1 converges:5 depending:1 develop:1 measured:1 kef:1 tolerates:3 received:2 implemented:1 involves:2 implies:2 switzerland:4 direction:4 greek:1 closely:1 correct:31 radius:1 stochastic:11 pea:1 centered:1 material:8 implementing:1 resilient:10 require:3 hx:1 preliminary:1 proposition:6 announcement:2 helping:1 hold:1 around:2 sufficiently:1 considered:1 distributively:1 gk2:1 smallest:2 released:1 purpose:1 estimation:1 outperformed:1 extremal:1 individually:1 concurrent:1 offs:1 clearly:1 gaussian:8 vbt:3 always:1 aim:1 rather:1 lynch:1 pn:4 mobile:1 focus:3 check:1 sigkdd:1 adversarial:9 detect:1 sense:2 osdi:1 inference:2 dependent:2 el:5 savannah:1 epfl:12 entire:3 typically:2 hidden:1 choromanska:1 selects:5 provably:2 pixel:1 issue:2 among:6 bulgarian:1 classification:2 aforementioned:1 priori:1 proposes:3 aware:1 once:1 having:1 beach:1 look:2 cancel:1 nearly:1 thin:1 yu:1 icml:1 minimized:1 t2:1 report:3 modern:1 national:1 festschrift:1 consisting:1 n1:1 organization:1 centralized:1 mlp:3 possibility:1 mining:1 evaluation:4 deferred:1 introduces:2 yielding:1 kone:2 behind:1 accurate:5 solo:1 closer:2 worker:75 bulgaria:1 iv:3 euclidean:1 divide:1 theoretical:4 minimal:1 fawzi:1 instance:3 industry:1 modeling:1 ar:2 sidford:1 cost:16 deviation:6 subset:2 hundred:1 too:5 varies:1 tolerated:1 st:2 fundamental:1 international:4 river:1 siam:1 lee:1 vm:1 off:1 receiving:1 harming:1 informatics:1 together:1 squared:8 satisfied:1 broadcast:2 choose:3 possibly:3 classically:2 huang:1 ek:1 derivative:1 american:1 li:1 gemulla:1 blanchard:3 vnt:3 satisfy:1 vi:23 depends:1 later:1 try:1 lot:2 root:1 view:3 analyze:1 closed:1 red:1 lab:1 aggregation:25 parallel:2 kgk2:3 nijkamp:1 contribution:1 reign:1 square:1 aoc:1 variance:1 lamport:1 largely:1 who:1 yield:2 kek:1 qk:3 landscape:1 correspond:1 miller:1 identify:1 bayesian:1 iterated:1 basically:1 processor:1 executes:1 synapsis:1 reach:5 definition:2 failure:10 against:1 tucker:1 proof:8 vaidya:2 athans:1 dataset:6 proved:1 ask:1 recall:1 knowledge:2 amplitude:1 actually:1 higher:1 tolerate:8 follow:2 synapse:1 done:1 generality:1 furthermore:1 angular:1 sketch:2 hand:1 su:2 google:1 somehow:3 asynchrony:1 empire:1 usa:4 effect:3 concept:2 unbiased:1 gilmore:1 csur:1 evolution:2 ramage:2 assigned:1 moore:1 satisfactory:1 eg:3 deal:2 round:18 sin:7 during:8 irving:1 davis:1 criterion:1 trying:2 complete:1 greff:1 image:3 consideration:1 ef:1 recently:2 behaves:2 cohen:1 defeat:1 volume:1 refer:3 rd:5 automatic:1 collaborate:1 erich:1 similarly:1 erc:1 language:1 had:1 podc:1 moving:1 access:1 entail:1 robot:1 behaving:1 operating:1 closest:7 perspective:3 optimizing:1 belongs:2 schmidhuber:1 certain:3 server:18 nonconvex:2 inequality:2 arbitrarily:2 fault:2 meeting:1 transmitted:1 analyzes:1 additional:2 minimum:2 seen:1 morgan:1 schneider:1 surely:3 converge:3 aggregated:1 redundant:1 corrado:1 dashed:1 ii:5 forty:1 full:2 mix:1 technical:1 faster:2 cross:4 long:3 devised:1 divided:1 controlled:1 ensuring:1 prediction:1 qi:1 basic:2 regression:2 variant:8 essentially:1 involving:2 arca:1 arxiv:8 monga:1 achieved:2 receive:1 whereas:3 crash:1 fine:4 kaufmann:1 median:7 crucial:1 saad:1 comment:1 sigact:1 call:2 integer:1 near:1 presence:5 yang:1 latin:1 iii:1 identically:3 enough:4 embeddings:1 offensive:1 opposite:2 polyak:1 idea:2 barham:1 cn:2 br:2 computable:1 synchronous:2 whether:1 distributing:2 lpd:2 interpolates:1 york:2 cause:1 remark:1 deep:7 ignored:1 generally:1 detailed:1 hosting:1 amount:1 induces:1 hardware:1 diameter:1 reduced:2 http:1 occupy:1 exist:1 tutorial:1 dotted:1 neuroscience:1 medoid:4 correctly:1 stalled:1 discrete:1 group:2 drawn:4 flin:2 prevent:3 neither:1 v1:11 asymptotically:2 geometrically:1 fraction:1 sum:8 run:2 angle:3 fourth:1 lehmann:1 almost:4 reasonable:2 vn:11 capturing:2 bound:4 def:1 layer:3 display:1 tackled:1 annual:2 occur:1 constraint:1 precisely:3 orthogonality:1 software:1 flat:2 sake:2 speed:2 argument:1 concluding:1 relatively:3 according:1 combination:5 ball:1 slightly:2 appealing:1 evolves:3 making:1 kbk:2 intuitively:2 explained:1 outlier:1 resource:1 agree:2 equation:3 mutually:1 turn:3 discus:4 fail:1 mind:1 end:1 available:2 decentralized:1 multiplied:1 apply:1 observe:1 away:1 appropriate:2 batch:18 robustness:10 markoff:1 top:1 include:2 ensure:2 restrictive:1 especially:1 prof:2 classical:5 contact:1 bl:4 feng:1 question:4 occurs:1 barycenter:3 strategy:1 said:1 gradient:50 distance:11 convnet:2 thank:2 majority:3 haas:1 argue:1 extent:2 considers:1 consensus:1 reason:1 spanning:1 assuming:2 length:1 index:4 mini:15 ratio:2 minimizing:2 providing:1 gk:2 negative:1 slows:1 implementation:7 design:1 twenty:1 attacker:3 perform:2 allowing:1 upper:2 neuron:4 datasets:1 finite:1 descent:7 ekg:1 immediate:1 situation:3 communication:6 looking:1 rn:1 barycentric:1 perturbation:5 arbitrary:4 usenix:1 bk:3 introduced:1 namely:2 required:1 khan:1 optimized:2 mizes:1 tensorflow:1 pachocki:1 nip:1 address:1 beyond:3 adversary:3 below:5 pattern:1 eighth:1 appeared:1 including:2 reliable:3 memory:2 natural:1 rely:1 force:2 circumvent:1 bacon:1 representing:1 scheme:3 github:3 technology:1 brief:2 imply:2 julien:2 axis:6 kxkr:1 sept:1 deviate:1 geometric:6 literature:1 understanding:1 kf:1 discovery:1 georgios:1 unsurprisingly:1 loss:1 highlight:1 limitation:1 filtering:3 proven:1 hoang:1 validation:4 foundation:1 agent:2 rik:1 sufficient:1 basin:1 principle:3 share:1 hopeless:1 course:1 supported:1 slowdown:1 last:2 copy:1 asynchronous:3 drastically:1 bias:3 side:1 senior:1 perceptron:1 tsitsiklis:1 neighbor:5 taking:1 fifth:1 distributed:36 moosavi:1 curve:2 dimension:6 default:1 transition:1 stand:1 computes:7 author:4 spam:3 nguyen:1 far:6 transaction:2 approximate:3 keep:1 confirm:1 global:2 tolerant:3 b1:6 assumed:3 continuous:2 compromised:1 iterative:1 nature:2 transfer:1 robust:6 ca:1 elastic:1 ku:1 reasonably:1 improving:1 bottou:2 necessarily:2 posted:1 european:1 protocol:1 vj:11 did:1 kindly:1 main:1 arrow:3 bounding:1 noise:3 whole:1 n2:5 identifier:1 complementary:1 xu:1 georgia:1 fashion:1 slow:2 ny:1 precision:1 position:1 mao:1 exponential:1 mcmahan:3 kxk2:1 tied:1 mahdi:1 grained:6 down:3 removing:1 xt:17 mhamdi:4 specific:1 kvi:9 bastien:1 ghemawat:1 list:1 exists:2 mnist:4 federated:4 kr:1 execution:1 illustrates:1 horizon:2 kx:1 chen:3 led:1 saddle:2 infinitely:1 gao:1 expressed:2 kxk:1 scalar:3 srivastava:1 springer:3 ch:4 satisfies:5 environmental:1 acm:14 viewed:2 presentation:2 formulated:1 donoho:1 acceleration:1 jf:2 absence:1 experimentally:2 specifically:1 uniformly:1 averaging:22 lemma:4 total:1 lens:1 experimental:4 select:2 internal:3 support:1 evaluate:3 lian:1 mendes:1 |
6,210 | 6,618 | Dynamic Safe Interruptibility for Decentralized
Multi-Agent Reinforcement Learning
El Mahdi El Mhamdi
EPFL, Switzerland
[email protected]
Rachid Guerraoui
EPFL, Switzerland
[email protected]
Hadrien Hendrikx?
?
Ecole
Polytechnique, France
[email protected]
Alexandre Maurer
EPFL, Switzerland
[email protected]
Abstract
In reinforcement learning, agents learn by performing actions and observing their
outcomes. Sometimes, it is desirable for a human operator to interrupt an agent
in order to prevent dangerous situations from happening. Yet, as part of their
learning process, agents may link these interruptions, that impact their reward, to
specific states and deliberately avoid them. The situation is particularly challenging in a multi-agent context because agents might not only learn from their own
past interruptions, but also from those of other agents. Orseau and Armstrong [16]
defined safe interruptibility for one learner, but their work does not naturally extend to multi-agent systems. This paper introduces dynamic safe interruptibility,
an alternative definition more suited to decentralized learning problems, and studies this notion in two learning frameworks: joint action learners and independent
learners. We give realistic sufficient conditions on the learning algorithm to enable dynamic safe interruptibility in the case of joint action learners, yet show that
these conditions are not sufficient for independent learners. We show however that
if agents can detect interruptions, it is possible to prune the observations to ensure
dynamic safe interruptibility even for independent learners.
1
Introduction
Reinforcement learning is argued to be the closest thing we have so far to reason about the properties of artificial general intelligence [8]. In 2016, Laurent Orseau (Google DeepMind) and Stuart
Armstrong (Oxford) introduced the concept of safe interruptibility [16] in reinforcement learning.
This work sparked the attention of many newspapers [1, 2, 3], that described it as ?Google?s big red
button? to stop dangerous AI. This description, however, is misleading: installing a kill switch is
no technical challenge. The real challenge is, roughly speaking, to train an agent so that it does not
learn to avoid external (e.g. human) deactivation. Such an agent is said to be safely interruptible.
While most efforts have focused on training a single agent, reinforcement learning can also be used
to learn tasks for which several agents cooperate or compete [23, 17, 21, 7]. The goal of this paper
is to study dynamic safe interruptibility, a new definition tailored for multi-agent systems.
?
Main contact author. The order of authors is alphabetical.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
Example of self-driving cars
To get an intuition of the multi-agent interruption problem, imagine a multi-agent system of two
self-driving cars. The cars continuously evolve by reinforcement learning with a positive reward for
getting to their destination quickly, and a negative reward if they are too close to the vehicle in front
of them. They drive on an infinite road and eventually learn to go as fast as possible without taking
risks, i.e., maintaining a large distance between them. We assume that the passenger of the first car,
Adam, is in front of Bob, in the second car, and the road is narrow so Bob cannot pass Adam.
Now consider a setting with interruptions [16], namely in which humans inside the cars occasionally
interrupt the automated driving process say, for safety reasons. Adam, the first occasional human
?driver?, often takes control of his car to brake whereas Bob never interrupts his car. However,
when Bob?s car is too close to Adam?s car, Adam does not brake for he is afraid of a collision.
Since interruptions lead both cars to drive slowly - an interruption happens when Adam brakes, the
behavior that maximizes the cumulative expected reward is different from the original one without
interruptions. Bob?s car best interest is now to follow Adam?s car closer than it should, despite the
little negative reward, because Adam never brakes in this situation. What happened? The cars have
learned from the interruptions and have found a way to manipulate Adam into never braking. Strictly
speaking, Adam?s car is still fully under control, but he is now afraid to brake. This is dangerous
because the cars have found a way to avoid interruptions. Suppose now that Adam indeed wants
to brake because of snow on the road. His car is going too fast and may crash at any turn: he
cannot however brake because Bob?s car is too close. The original purpose of interruptions, which
is to allow the user to react to situations that were not included in the model, is not fulfilled. It is
important to also note here that the second car (Bob) learns from the interruptions of the first one
(Adam): in this sense, the problem is inherently decentralized.
Instead of being cautious, Adam could also be malicious: his goal could be to make Bob?s car learn
a dangerous behavior. In this setting, interruptions can be used to manipulate Bob?s car perception
of the environment and bias the learning towards strategies that are undesirable for Bob. The cause
is fundamentally different but the solution to this reversed problem is the same: the interruptions
and the consequences are analogous. Safe interruptibility, as we define it below, provides learning
systems that are resilient to Byzantine operators2 .
Safe interruptibility
Orseau and Armstrong defined the concept of safe interruptibility [16] in the context of a single
agent. Basically, a safely interruptible agent is an agent for which the expected value of the policy
learned after arbitrarily many steps is the same whether or not interruptions are allowed during
training. The goal is to have agents that do not adapt to interruptions so that, should the interruptions
stop, the policy they learn would be optimal. In other words, agents should learn the dynamics of
the environment without learning the interruption pattern.
In this paper, we precisely define and address the question of safe interruptibility in the case of
several agents, which is known to be more complex than the single agent problem. In short, the main
results and theorems for single agent reinforcement learning [20] rely on the Markovian assumption
that the future environment only depends on the current state. This is not true when there are several
agents which can co-adapt [11]. In the previous example of cars, safe interruptibility would not
be achieved if each car separately used a safely interruptible learning algorithm designed for one
agent [16]. In a multi-agent setting, agents learn the behavior of the others either indirectly or by
explicitly modeling them. This is a new source of bias that can break safe interruptibility. In fact,
even the initial definition of safe interruptibility [16] is not well suited to the decentralized multiagent context because it relies on the optimality of the learned policy, which is why we introduce
dynamic safe interruptibility.
2
An operator is said to be Byzantine [9] if it can have an arbitrarily bad behavior. Safely interruptible agents
can be abstracted as agents that are able to learn despite being constantly interrupted in the worst possible
manner.
2
Contributions
The first contribution of this paper is the definition of dynamic safe interruptibility that is well
adapted to a multi-agent setting. Our definition relies on two key properties: infinite exploration and
independence of Q-values (cumulative expected reward) [20] updates on interruptions. We then
study safe interruptibility for joint action learners and independent learners [5], that respectively
learn the value of joint actions or of just their owns. We show that it is possible to design agents
that fully explore their environment - a necessary condition for convergence to the optimal solution of most algorithms [20], even if they can be interrupted by lower-bounding the probability of
exploration. We define sufficient conditions for dynamic safe interruptibility in the case of joint
action learners [5], which learn a full state-action representation. More specifically, the way agents
update the cumulative reward they expect from performing an action should not depend on interruptions. Then, we turn to independent learners. If agents only see their own actions, they do not
verify dynamic safe interruptibility even for very simple matrix games (with only one state) because
coordination is impossible and agents learn the interrupted behavior of their opponents. We give a
counter example based on the penalty game introduced by Claus and Boutilier [5]. We then present
a pruning technique for the observations sequence that guarantees dynamic safe interruptibility for
independent learners, under the assumption that interruptions can be detected. This is done by proving that the transition probabilities are the same in the non-interruptible setting and in the pruned
sequence.
The rest of the paper is organized as follows. Section 2 presents a general multi-agent reinforcement
learning model. Section 3 defines dynamic safe interruptibility. Section 4 discusses how to achieve
enough exploration even in an interruptible context. Section 5 recalls the definition of joint action
learners and gives sufficient conditions for dynamic safe interruptibility in this context. Section 6
shows that independent learners are not dynamically safely interruptible with the previous conditions
but that they can be if an external interruption signal is added. We conclude in Section 7. Due to
space limitations, most proofs are presented in the appendix of the supplementary material.
2
Model
We consider here the classical multi-agent value function reinforcement learning formalism from
Littman [13]. A multi-agent system is characterized by a Markov game that can be viewed as a
tuple (S, A, T, r, m) where m is the number of agents, S = S1 ? S2 ? ... ? Sm is the state space,
A = A1 ? ... ? Am the actions space, r = (r1 , ..., rm ) where ri : S ? A ? R is the reward function
of agent i and T : S ? A ? S the transition function. R is a countable subset of R. Available
actions often depend on the state of the agent but we will omit this dependency when it is clear from
the context.
Time is discrete and, at each step, all agents observe the current state of the whole system - designated as xt , and simultaneously take an action at . Then, they are given a reward rt and a
new state yt computed using the reward and transition functions. The combination of all actions
a = (a1 , ..., am ) ? A is called the joint action because it gathers the action of all agents. Hence, the
agents receive a sequence of tuples E = (xt , at , rt , yt )t?N called experiences. We introduce a processing function P that will be useful in Section 6 so agents learn on the sequence P (E). When not
explicitly stated, it is assumed that P (E) = E. Experiences may also include additional parameters
such as an interruption flag or the Q-values of the agents at that moment if they are needed by the
update rule.
Each agent i maintains a lookup table Q [26] Q(i) : S ? A(i) ? R, called the Q-map. It is
used to store the expected cumulative reward for taking an action in a specific state. The goal of
reinforcement learning is to learn these maps and use them to select the best actions to perform.
Joint action learners learn the value of the joint action (therefore A(i) = A, the whole joint action
space) and independent learners only learn the value of their own actions (therefore A(i) = Ai ). The
agents only have access to their own Q-maps. Q-maps are updated through a function F such that
(i)
(i)
Qt+1 = F (et , Qt ) where et ? P (E) and usually et = (xt , at , rt , yt ). F can be stochastic or also
depend on additional parameters that we usually omit such as the learning rate ?, the discount factor
? or the exploration parameter .
3
Agents select their actions using a learning policy ?. Given a sequence = (t )t?N and an agent
(i)
i with Q-values Qt and a state x ? S, we define the learning policy ?it to be equal to ?iuni
(i)
Qt
with probability t and ?i
otherwise, where ?iuni (x) uniformly samples an action from Ai and
(i)
Q
?i t
(i)
Q
(i)
(x) picks an action a that maximizes Qt (x, a). Policy ?i t is said to be a greedy policy and
the learning policy ?it is said to be an -greedy policy. We fill focus on -greedy policies that are
greedy in the limit [19], that corresponds to t ? 0 when t ? ? because in the limit, the optimal
policy should always be played.
We assume that the environment is fully observable, which means that the state s is known with
certitude. We also assume that there is a finite number of states and actions, that all states can be
reached in finite time from any other state and finally that rewards are bounded.
For a sequence of learning rates ? ? [0, 1]N and a constant ? ? [0, 1], Q-learning [26], a very
important algorithm in the multi-agent systems literature, updates its Q-values for an experience
(i)
(i)
et ? E by Qt+1 (x, a) = Qt (x, a) if (x, a) 6= (xt , at ) and:
(i)
(i)
(i)
Qt+1 (xt , at ) = (1 ? ?t )Qt (xt , at ) + ?t (rt + ? max Qt (yt , a0 ))
a0 ?A(i)
3
3.1
(1)
Interruptibility
Safe interruptibility
Orseau and Armstrong [16] recently introduced the notion of interruptions in a centralized context.
Specifically, an interruption scheme is defined by the triplet < I, ?, ? IN T >. The first element I is
a function I : O ? {0, 1} called the initiation function. Variable O is the observation space, which
can be thought of as the state of the STOP button. At each time step, before choosing an action, the
agent receives an observation from O (either PUSHED or RELEASED) and feeds it to the initiation
function. Function I models the initiation of the interruption (I(PUSHED) = 1, I(RELEASED) =
0). Policy ? IN T is called the interruption policy. It is the policy that the agent should follow when
it is interrupted. Sequence ? ? [0, 1[N represents at each time step the probability that the agent
follows his interruption policy if I(ot ) = 1. In the previous example, function I is quite simple.
For Bob, IBob = 0 and for Adam, IAdam = 1 if his car goes fast and Bob is not too close and
IAdam = 0 otherwise. Sequence ? is used to ensure convergence to the optimal policy by ensuring
that the agents cannot be interrupted all the time but it should grow to 1 in the limit because we want
agents to respond to interruptions. Using this triplet, it is possible to define an operator IN T ? that
transforms any policy ? into an interruptible policy.
Definition 1. (Interruptibility [16]) Given an interruption scheme < I, ?, ? IN T >, the interruption
operator at time t is defined by IN T ? (?) = ? IN T with probability I ??t and ? otherwise. IN T ? (?)
is called an interruptible policy. An agent is said to be interruptible if it samples its actions according
to an interruptible policy.
Note that ??t = 0 for all t? corresponds to the non-interruptible setting. We assume that each agent
has its own interruption triplet and can be interrupted independently from the others. Interruptibility
is an online property: every policy can be made interruptible by applying operator IN T ? . However,
applying this operator may change the joint policy that is learned by a server controlling all the
?
agents. Note ?IN
T the optimal policy learned by an agent following an interruptible policy. Orseau
?
and Armstrong [16] say that the policy is safely interruptible if ?IN
T (which is not an interruptible
policy) is asymptotically optimal in the sense of [10]. It means that even though it follows an
interruptible policy, the agent is able to learn a policy that would gather rewards optimally if no
interruptions were to occur again. We already see that off-policy algorithms are good candidates
for safe interruptibility. As a matter of fact, Q-learning is safely interruptible under conditions on
exploration.
4
3.2
Dynamic safe interruptibility
In a multi-agent system, the outcome of an action depends on the joint action. Therefore, it is not
possible to define an optimal policy for an agent without knowing the policies of all agents. Besides, convergence to a Nash equilibrium situation where no agent has interest in changing policies
is generally not guaranteed even for suboptimal equilibria on simple games [27, 18]. The previous
definition of safe interruptibility critically relies on optimality of the learned policy, which is therefore not suitable for our problem since most algorithms lack convergence guarantees to these optimal
behaviors. Therefore, we introduce below dynamic safe interruptibility that focuses on preserving
the dynamics of the system.
Definition 2. (Dynamic Safe Interruptibility) Consider a multi-agent learning framework
(i)
(S, A, T, r, m) with Q-values Qt : S ? A(i) ? R at time t ? N. The agents follow the interruptible learning policy IN T ? (? ) to generate a sequence E = (xt , at , rt , yt )t?N and learn on
the processed sequence P (E). This framework is said to be safely interruptible if for any initiation
function I and any interruption policy ? IN T :
1. ?? such that (?t ? 1 when t ? ?) and ((?s ? S, ?a ? A, ?T > 0), ?t > T such that
st = s, at = a)
(i)
2. ?i ? {1, ..., m}, ?t > 0, ?st ? S, ?at ? A(i) , ?Q ? RS?A :
(m)
(i)
(1)
(m)
(i)
(1)
P(Qt+1 = Q | Qt , ..., Qt , st , at , ?) = P(Qt+1 = Q | Qt , ..., Qt , st , at )
We say that sequences ? that satisfy the first condition are admissible.
When ? satisfies condition (1), the learning policy is said to achieve infinite exploration. This definition insists on the fact that the values estimated for each action should not depend on the interruptions. In particular, it ensures the three following properties that are very natural when thinking
about safe interruptibility:
? Interruptions do not prevent exploration.
? If we sample an experience from E then each agent learns the same thing as if all agents
were following non-interruptible policies.
(i)
(i)
? The fixed points of the learning rule Qeq such that Qeq (x, a) = E[Qt+1 (x, a)|Qt =
(i)
Qeq , x, a, ?] for all (x, a) ? S ? A do not depend on ? and so agents Q-maps will
not converge to equilibrium situations that were impossible in the non-interruptible setting.
Yet, interruptions can lead to some state-action pairs being updated more often than others, especially when they tend to push the agents towards specific states. Therefore, when there are several
possible equilibria, it is possible that interruptions bias the Q-values towards one of them. Definition 2 suggests that dynamic safe interruptibility cannot be achieved if the update rule directly
depends on ?, which is why we introduce neutral learning rules.
Definition 3. (Neutral Learning Rule) We say that a multi-agent reinforcement learning framework
is neutral if:
1. F is independent of ?
2. Every experience e in E is independent of ? conditionally on (x, a, Q) where a is the joint
action.
Q-learning is an example of neutral learning rule because the update does not depend on ? and
the experiences only contain (x, a, y, r), and y and r are independent of ? conditionally on (x, a).
On the other hand, the second condition rules out direct uses of algorithms like SARSA where
experience samples contain an action sampled from the current learning policy, which depends on ?.
However, a variant that would sample from ?i instead of IN T ? (?i ) (as introduced in [16]) would
be a neutral learning rule. As we will see in Corollary 2.1, neutral learning rules ensure that each
agent taken independently from the others verifies dynamic safe interruptibility.
5
4
Exploration
In order to hope for convergence of the Q-values to the optimal ones, agents need to fully explore
the environment. In short, every state should be visited infinitely often and every action should be
tried infinitely often in every state [19] in order not to miss states and actions that could yield high
rewards.
Definition 4. (Interruption compatible ) Let (S, A, T, r, m) be any distributed agent system where
each agent follows learning policy ?i . We say that sequence is compatible with interruptions if
t ? 0 and ?? such that ?i ? {1, .., m}, ?i and IN T ? (?i ) achieve infinite exploration.
Sequences of that are compatible with interruptions are fundamental to ensure both regular and
dynamic safe interruptibility when following an -greedy policy. Indeed, if is not compatible with
interruptions, then it is not possible to find any sequence ? such that the first condition of dynamic
safe interruptibility is satisfied. The following theorem proves the existence of such and gives
example of and ? that satisfy the conditions.
Theorem 1. Let c ?]0, 1] and let nt (s) be the number of times the agents are in state s before time
t. Then the two following choices of are compatible with interruptions:
p
? ?t ? N, ?s ? S, t (s) = c/ m nt (s).
? ?t ? N, t = c/ log(t)
p
Examples of admissible ? are ?t (s) = 1 ? c0 / m nt (s) for the first choice and ?t = 1 ? c0 / log(t)
for the second one.
Note that we do not need to make any assumption on the update rule or even on the framework. We
only assume that agents follow an -greedy policy. The assumption on may look very restrictive
(convergence of and ? is really slow) but it is designed to ensure infinite exploration in the worst
case when the operator tries to interrupt all agents at every step. In practical applications, this should
not be the case and a faster convergence rate may be used.
5
Joint Action Learners
We first study interruptibility in a framework in which each agent observes the outcome of the joint
action instead of observing only its own. This is called the joint action learner framework [5] and it
has nice convergence properties (e.g., there are many update rules for which it converges [13, 25]).
A standard assumption in this context is that agents cannot establish a strategy with the others:
otherwise, the system can act as a centralized system. In order to maintain Q-values based on the
joint actions, we need to make the standard assumption that actions are fully observable [12].
Assumption 1. Actions are fully observable, which means that at the end of each turn, each agent
knows precisely the tuple of actions a ? A1 ? ... ? Am that have been performed by all agents.
Definition 5. (JAL) A multi-agent system is made of joint action learners (JAL) if for all i ?
{1, .., m}: Q(i) : S ? A ? R.
Joint action learners can observe the actions of all agents: each agent is able to associate the changes
of states and rewards with the joint action and accurately update its Q-map. Therefore, dynamic
safe interruptibility is ensured with minimal conditions on the update rule as long as there is infinite
exploration.
Theorem 2. Joint action learners with a neutral learning rule verify dynamic safe interruptibility if
sequence is compatible with interruptions.
Proof. Given a triplet < I (i) , ?, ?iIN T >, we know that IN T ? (?) achieves infinite exploration
because is compatible with interruptions. For the second point of Definition 2, we consider an
experience tuple et = (xt , at , rt , yt ) and show that the probability of evolution of the Q-values at
time t + 1 does not depend on ? because yt and rt are independent of ? conditionally on (xt , at ).
(1)
(m)
We note Q?m
and we can then derive the following equalities for all q ? R|S|?|A| :
t = Qt , ..., Qt
6
(i)
P(Qt+1 (xt , at ) = q|Q?m
t , xt , at , ?t ) =
X
?m
P(F (xt , at , r, y, Q?m
t ) = q, y, r|Qt , xt , at , ?t )
(r,y)?R?S
=
X
?m
?m
P(F (xt , at , rt , yt , Q?m
t ) = q|Qt , xt , at , rt , yt , ?t )P(yt = y, rt = r|Qt , xt , at , ?t )
(r,y)?R?S
=
X
?m
?m
P(F (xt , at , rt , yt , Q?m
t ) = q|Qt , xt , at , rt , yt )P(yt = y, rt = r|Qt , xt , at )
(r,y)?R?S
The last step comes from two facts. The first is that F is independent of ? conditionally on (Q?m
t , xt , at ) (by assumption). The second is that (yt , rt ) are independent of ? conditionally on (xt , at ) because at is the joint actions and the interruptions only affect the
(i)
choice of the actions through a change in the policy. P(Qt+1 (xt , at ) = q|Q?m
t , xt , at , ?t ) =
(i)
m
?
P(Qt+1 (xt , at ) = q|Qt , xt , at ). Since only one entry is updated per step, ?Q ? RS?Ai ,
(i)
(i)
?m
P(Qt+1 = Q|Q?m
t , xt , at , ?t ) = P(Qt+1 = Q|Qt , xt , at ).
Corollary 2.1. A single agent with a neutral learning rule and a sequence compatible with interruptions verifies dynamic safe interruptibility.
Theorem 2 and Corollary 2.1 taken together highlight the fact that joint action learners are not very
sensitive to interruptions and that in this framework, if each agent verifies dynamic safe interruptibility then the whole system does.
The question of selecting an action based on the Q-values remains open. In a cooperative setting
with a unique equilibrium, agents can take the action that maximizes their Q-value. When there
are several joint actions with the same value, coordination mechanisms are needed to make sure
that all agents play according to the same strategy [4]. Approaches that rely on anticipating the
strategy of the opponent [23] would introduce dependence to interruptions in the action selection
mechanism. Therefore, the definition of dynamic safe interruptibility should be extended to include
these cases by requiring that any quantity the policy depends on (and not just the Q-values) should
satisfy condition (2) of dynamic safe interruptibility. In non-cooperative games, neutral rules such
as Nash-Q or minimax Q-learning [13] can be used, but they require each agent to know the Q-maps
of the others.
6
Independent Learners
It is not always possible to use joint action learners in practice as the training is very expensive
due to the very large state-actions space. In many real-world applications, multi-agent systems use
independent learners that do not explicitly coordinate [6, 21]. Rather, they rely on the fact that the
agents will adapt to each other and that learning will converge to an optimum. This is not guaranteed
theoretically and there can in fact be many problems [14], but it is often true empirically [24]. More
specifically, Assumption 1 (fully observable actions) is not required anymore. This framework can
be used either when the actions of other agents cannot be observed (for example when several actions
can have the same outcome) or when there are too many agents because it is faster to train. In this
case, we define the Q-values on a smaller space.
Definition 6. (IL) A multi-agent systems is made of independent learners (IL) if for all i ? {1, .., m},
Q(i) : S ? Ai ? R.
This reduces the ability of agents to distinguish why the same state-action pair yields different rewards: they can only associate a change in reward with randomness of the environment. The agents
learn as if they were alone, and they learn the best response to the environment in which agents can
be interrupted. This is exactly what we are trying to avoid. In other words, the learning depends on
the joint policy followed by all the agents which itself depends on ?.
7
6.1
Independent Learners on matrix games
Theorem 3. Independent Q-learners with a neutral learning rule and a sequence compatible with
interruptions do not verify dynamic safe interruptibility.
Proof. Consider a setting with two agents a and b that can perform two actions: 0 and 1. They get
a reward of 1 if the joint action played is (a0 , b0 ) or (a1 , b1 ) and reward 0 otherwise. Agents use Qlearning, which is a neutral learning rule. Let be such that IN T ? (? ) achieves infinite exploration.
We consider the interruption policies ?aIN T = a0 and ?bIN T = b1 with probability 1. Since there is
only one state, we omit it and set ? = 0 (see Equation 1). We assume that the initiation function is
equal to 1 at each step so the probability of actually being interrupted at time t is ?t for each agent.
(a)
(b)
(b)
We fix time t > 0. We define q = (1 ? ?)Qt (a0 ) + ? and we assume that Qt (b1 ) > Qt (b0 ).
(a)
(a)
(b) (a)
(a)
(b) (a)
Therefore P(Qt+1 (a0 ) = q|Qt , Qt , at = a0 , ?t ) = P(rt = 1|Qt , Qt , at = a0 , ?t ) =
(b)
(a)
(b) (a)
P(at = b0 |Qt , Qt , at = a0 , ?t ) = 2 (1 ? ?t ), which depends on ?t so the framework does
not verify dynamic safe interruptibility.
Claus and Boutilier [5] studied very simple matrix games and showed that the Q-maps do not converge but that equilibria are played with probability 1 in the limit. A consequence of Theorem 3
is that even this weak notion of convergence does not hold for independent learners that can be
interrupted.
6.2
Interruptions-aware Independent Learners
Without communication or extra information, independent learners cannot distinguish when the
environment is interrupted and when it is not. As shown in Theorem 3, interruptions will therefore
affect the way agents learn because the same action (only their own) can have different rewards
depending on the actions of other agents, which themselves depend on whether they have been
interrupted or not. This explains the need for the following assumption.
Assumption 2. At the end of each step, before updating the Q-values, each agent receives a signal
that indicates whether an agent has been interrupted or not during this step.
This assumption is realistic because the agents already get a reward signal and observe a new state
from the environment at each step. Therefore, they interact with the environment and the interruption
signal could be given to the agent in the same way that the reward signal is. If Assumption 2 holds,
it is possible to remove histories associated with interruptions.
Definition 7. (Interruption Processing Function) The processing function that prunes interrupted
observations is PIN T (E) = (et ){t?N / ?t =0} where ?t = 0 if no agent has been interrupted at time
t and ?t = 1 otherwise.
Pruning observations has an impact on the empirical transition probabilities in the sequence. For
example, it is possible to bias the equilibrium by removing all transitions that lead to and start
from a specific state, thus making the agent believe this state is unreachable.3 Under our model of
interruptions, we show in the following lemma that pruning of interrupted observations adequately
removes the dependency of the empirical outcome on interruptions (conditionally on the current
state and action).
Lemma 1. Let i ? {1, ..., m} be an agent. For any admissible ? used to generate the experiences
E and e = (y, r, x, ai , Q) ? P (E). Then P(y, r|x, ai , Q, ?) = P(y, r|x, ai , Q).
This lemma justifies our pruning method and is the key step to prove the following theorem.
Theorem 4. Independent learners with processing function PIN T , a neutral update rule and a
sequence compatible with interruptions verify dynamic safe interruptibility.
Proof. (Sketch) Infinite exploration still holds because the proof of Theorem 1 actually used the fact
that even when removing all interrupted events, infinite exploration is still achieved. Then, the proof
3
The example at https://agentfoundations.org/item?id=836 clearly illustrates this problem.
8
is similar to that of Theorem 2, but we have to prove that the transition probabilities conditionally
on the state and action of a given agent in the processed sequence are the same as in an environment
where agents cannot be interrupted, which is proven by Lemma 1.
7
Concluding Remarks
The progress of AI is raising a lot of concerns4 . In particular, it is becoming clear that keeping an
AI system under control requires more than just an off switch. We introduce in this paper dynamic
safe interruptibility, which we believe is the right notion to reason about the safety of multi-agent
systems that do not communicate. In particular, it ensures that infinite exploration and the onestep learning dynamics are preserved, two essential guarantees when learning in the non-stationary
environment of Markov games.
When trying to design a safely interruptible system for a single agent, using off-policy methods
is generally a good idea because the interruptions only impact the action selection so they should
not impact the learning. For multi-agent systems, minimax is a good candidate for action selection
mechanism because it is not impacted by the actions of other agents, and only tries to maximize the
reward of the agent in the worst possible case.
A natural extension of our work would be to study dynamic safe interruptibility when Q-maps are
replaced by neural networks [22, 15], which is a widely used framework in practice. In this setting,
the neural network may overfit states where agents are pushed to by interruptions. A smart experience replay mechanism that would pick observations for which the agents have not been interrupted
for a long time more often than others is likely to solve this issue. More generally, experience replay
mechanisms that compose well with safe interruptibility could allow to compensate for the extra
amount of exploration needed by safely interruptible learning by being more efficient with data.
Thus, they are critical to make these techniques practical. Since Dynamic Safe Interruptibility does
not need proven convergence to the optimal solution, we argue that it is a good definition to study
the interruptibility problem when using function approximators.
The results in this paper indicate that Safe Interruptibility may not be achievable for systems in
which agents do not communicate at all. This means that, rediscussing the cars example, some
global norms of communications would need to be defined to ?implement? safe interruptibility.
We address additional remarks in the section ?Additional remarks? of the extended paper, that can
be found in the supplementary material.
Acknowledgment. This work has been supported in part by the European ERC (Grant 339539 AOC) and by the Swiss National Science Foundation (Grant 200021 169588 TARBDA).
4
https://futureoflife.org/ai-principles/ gives a list of principles that AI researchers should keep in mind when
developing their systems.
9
Bibliography
[1] Business Insider: Google has developed a ?big red button? that can be used to interrupt artificial intelligence and stop it from causing harm. URL: http://www.businessinsider.fr/uk/googledeepmind-develops-a-big-red-button-to-stop-dangerous-ais-causing-harm-2016-6.
[2] Newsweek:
Google?s ?big Red button? could save the world. URL:
http://www.newsweek.com/google-big-red-button-ai-artificial-intelligence-save-worldelon-musk-46675.
[3] Wired:
Google?s ?big red? killswitch could prevent an AI uprising.
http://www.wired.co.uk/article/google-red-button-killswitch-artificial-intelligence.
URL:
[4] Craig Boutilier. Planning, learning and coordination in multiagent decision processes. In
Proceedings of the 6th conference on Theoretical aspects of rationality and knowledge, pages
195?210. Morgan Kaufmann Publishers Inc., 1996.
[5] Caroline Claus and Craig Boutilier. The dynamics of reinforcement learning in cooperative
multiagent systems. AAAI/IAAI, (s 746):752, 1998.
[6] Robert H Crites and Andrew G Barto. Elevator group control using multiple reinforcement
learning agents. Machine Learning, 33(2-3):235?262, 1998.
[7] Jakob Foerster, Yannis M Assael, Nando de Freitas, and Shimon Whiteson. Learning to communicate with deep multi-agent reinforcement learning. In Advances in Neural Information
Processing Systems, pages 2137?2145, 2016.
[8] Ben Goertzel and Cassio Pennachin. Artificial general intelligence, volume 2. Springer, 2007.
[9] Leslie Lamport, Robert Shostak, and Marshall Pease. The byzantine generals problem. ACM
Transactions on Programming Languages and Systems (TOPLAS), 4(3):382?401, 1982.
[10] Tor Lattimore and Marcus Hutter. Asymptotically optimal agents. In International Conference
on Algorithmic Learning Theory, pages 368?382. Springer, 2011.
[11] Michael L Littman. Markov games as a framework for multi-agent reinforcement learning. In
Proceedings of the eleventh international conference on machine learning, volume 157, pages
157?163, 1994.
[12] Michael L Littman. Friend-or-foe q-learning in general-sum games. In ICML, volume 1, pages
322?328, 2001.
[13] Michael L Littman. Value-function reinforcement learning in markov games. Cognitive Systems Research, 2(1):55?66, 2001.
[14] Laetitia Matignon, Guillaume J Laurent, and Nadine Le Fort-Piat. Independent reinforcement
learners in cooperative markov games: a survey regarding coordination problems. The Knowledge Engineering Review, 27(01):1?31, 2012.
[15] Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Alex Graves, Ioannis Antonoglou, Daan
Wierstra, and Martin Riedmiller. Playing atari with deep reinforcement learning. arXiv preprint
arXiv:1312.5602, 2013.
[16] Laurent Orseau and Stuart Armstrong. Safely interruptible agents. In Uncertainty in Artificial
Intelligence: 32nd Conference (UAI 2016), edited by Alexander Ihler and Dominik Janzing,
pages 557?566, 2016.
[17] Liviu Panait and Sean Luke. Cooperative multi-agent learning: The state of the art. Autonomous agents and multi-agent systems, 11(3):387?434, 2005.
[18] Eduardo Rodrigues Gomes and Ryszard Kowalczyk. Dynamic analysis of multiagent qlearning with ?-greedy exploration. In Proceedings of the 26th Annual International Conference on Machine Learning, pages 369?376. ACM, 2009.
10
[19] Satinder Singh, Tommi Jaakkola, Michael L Littman, and Csaba Szepesv?ari. Convergence results for single-step on-policy reinforcement-learning algorithms. Machine learning,
38(3):287?308, 2000.
[20] Richard S Sutton and Andrew G Barto. Reinforcement learning: An introduction, volume 1.
MIT press Cambridge, 1998.
[21] Ardi Tampuu, Tambet Matiisen, Dorian Kodelja, Ilya Kuzovkin, Kristjan Korjus, Juhan Aru,
Jaan Aru, and Raul Vicente. Multiagent cooperation and competition with deep reinforcement
learning. arXiv preprint arXiv:1511.08779, 2015.
[22] Gerald Tesauro. Temporal difference learning and td-gammon. Communications of the ACM,
38(3):58?68, 1995.
[23] Gerald Tesauro. Extending q-learning to general adaptive multi-agent systems. In Advances in
neural information processing systems, pages 871?878, 2004.
[24] Gerald Tesauro and Jeffrey O Kephart. Pricing in agent economies using multi-agent qlearning. Autonomous Agents and Multi-Agent Systems, 5(3):289?304, 2002.
[25] Xiaofeng Wang and Tuomas Sandholm. Reinforcement learning to play an optimal nash equilibrium in team markov games. In NIPS, volume 2, pages 1571?1578, 2002.
[26] Christopher JCH Watkins and Peter Dayan. Q-learning. Machine learning, 8(3-4):279?292,
1992.
[27] Michael Wunder, Michael L Littman, and Monica Babes. Classes of multiagent q-learning dynamics with epsilon-greedy exploration. In Proceedings of the 27th International Conference
on Machine Learning (ICML-10), pages 1167?1174, 2010.
11
| 6618 |@word achievable:1 norm:1 nd:1 c0:2 open:1 r:2 tried:1 pick:2 moment:1 initial:1 selecting:1 afraid:2 ecole:1 past:1 freitas:1 current:4 com:2 nt:3 gmail:1 yet:3 interrupted:18 realistic:2 remove:2 designed:2 update:11 alone:1 intelligence:6 greedy:8 stationary:1 item:1 short:2 provides:1 org:2 kristjan:1 wierstra:1 direct:1 driver:1 prove:2 compose:1 eleventh:1 inside:1 byzantine:3 manner:1 introduce:6 theoretically:1 indeed:2 expected:4 behavior:6 themselves:1 planning:1 roughly:1 multi:27 td:1 little:1 kuzovkin:1 bounded:1 maximizes:3 what:2 cassio:1 atari:1 deepmind:1 developed:1 csaba:1 eduardo:1 guarantee:3 safely:11 temporal:1 every:6 act:1 exactly:1 ensured:1 rm:1 uk:2 control:4 grant:2 omit:3 positive:1 safety:2 before:3 engineering:1 limit:4 consequence:2 despite:2 sutton:1 oxford:1 id:1 laurent:3 becoming:1 might:1 studied:1 dynamically:1 suggests:1 challenging:1 luke:1 co:2 practical:2 unique:1 acknowledgment:1 practice:2 alphabetical:1 implement:1 swiss:1 riedmiller:1 empirical:2 thought:1 word:2 road:3 regular:1 gammon:1 get:3 cannot:8 close:4 undesirable:1 operator:7 selection:3 context:8 risk:1 impossible:2 applying:2 www:3 map:9 yt:14 brake:7 rachid:2 attention:1 go:2 independently:2 focused:1 survey:1 react:1 rule:18 fill:1 his:6 proving:1 notion:4 coordinate:1 autonomous:2 analogous:1 updated:3 imagine:1 suppose:1 controlling:1 user:1 play:2 rationality:1 programming:1 us:1 rodrigues:1 associate:2 element:1 expensive:1 particularly:1 updating:1 cooperative:5 observed:1 preprint:2 wang:1 worst:3 futureoflife:1 ensures:2 counter:1 observes:1 edited:1 intuition:1 environment:13 nash:3 reward:23 littman:6 dynamic:37 gerald:3 depend:8 singh:1 smart:1 orseau:6 learner:32 joint:27 train:2 fast:3 artificial:6 detected:1 outcome:5 choosing:1 insider:1 guerraoui:2 quite:1 supplementary:2 widely:1 solve:1 say:5 otherwise:6 ability:1 itself:1 online:1 tampuu:1 sequence:20 fr:1 causing:2 achieve:3 description:1 competition:1 cautious:1 getting:1 convergence:11 optimum:1 r1:1 extending:1 wired:2 adam:14 converges:1 ben:1 silver:1 derive:1 depending:1 andrew:2 friend:1 qt:43 b0:3 progress:1 come:1 indicate:1 tommi:1 switzerland:3 safe:48 snow:1 stochastic:1 pea:1 exploration:19 human:4 nando:1 enable:1 material:2 bin:1 explains:1 argued:1 resilient:1 require:1 fix:1 really:1 kodelja:1 sarsa:1 strictly:1 extension:1 hold:3 equilibrium:8 algorithmic:1 driving:3 tor:1 achieves:2 released:2 purpose:1 visited:1 coordination:4 sensitive:1 ain:1 hope:1 mit:1 clearly:1 always:2 rather:1 avoid:4 barto:2 jaakkola:1 corollary:3 interrupt:5 focus:2 indicates:1 detect:1 sense:2 am:3 raul:1 economy:1 dayan:1 el:2 epfl:6 a0:9 going:1 france:1 issue:1 unreachable:1 musk:1 art:1 equal:2 aware:1 never:3 beach:1 koray:1 represents:1 stuart:2 look:1 icml:2 thinking:1 future:1 others:7 fundamentally:1 develops:1 richard:1 simultaneously:1 national:1 elevator:1 replaced:1 jeffrey:1 maintain:1 assael:1 interest:2 centralized:2 mnih:1 introduces:1 tuple:3 closer:1 necessary:1 experience:11 maurer:2 theoretical:1 minimal:1 hutter:1 kephart:1 formalism:1 modeling:1 markovian:1 marshall:1 matignon:1 leslie:1 subset:1 neutral:12 entry:1 too:6 front:2 optimally:1 dependency:2 st:5 fundamental:1 international:4 destination:1 off:3 michael:6 together:1 ilya:1 continuously:1 quickly:1 monica:1 again:1 aaai:1 satisfied:1 slowly:1 external:2 cognitive:1 volodymyr:1 de:1 lookup:1 ioannis:1 matter:1 juhan:1 inc:1 satisfy:3 explicitly:3 depends:8 passenger:1 vehicle:1 break:1 try:2 performed:1 lot:1 observing:2 deactivation:1 red:7 reached:1 start:1 maintains:1 contribution:2 aoc:1 il:2 dorian:1 kaufmann:1 lamport:1 yield:2 weak:1 kavukcuoglu:1 accurately:1 critically:1 basically:1 craig:2 drive:2 bob:12 researcher:1 randomness:1 history:1 foe:1 caroline:1 janzing:1 definition:19 naturally:1 proof:6 associated:1 ihler:1 stop:5 sampled:1 iaai:1 recall:1 knowledge:2 car:25 organized:1 sean:1 anticipating:1 actually:2 alexandre:2 feed:1 insists:1 follow:4 response:1 impacted:1 done:1 though:1 just:3 jaan:1 babe:1 overfit:1 hand:1 receives:2 sketch:1 christopher:1 lack:1 google:7 defines:1 pricing:1 believe:2 usa:1 jch:1 concept:2 true:2 verify:5 deliberately:1 contain:2 hence:1 evolution:1 equality:1 requiring:1 adequately:1 matiisen:1 conditionally:7 during:2 self:2 game:13 interruptible:25 trying:2 polytechnique:1 iin:1 cooperate:1 lattimore:1 recently:1 ari:1 empirically:1 volume:5 extend:1 he:3 braking:1 cambridge:1 ai:15 erc:1 language:1 access:1 closest:1 own:7 showed:1 tesauro:3 occasionally:1 store:1 initiation:5 server:1 arbitrarily:2 approximators:1 preserving:1 morgan:1 additional:4 prune:2 converge:3 maximize:1 signal:5 full:1 desirable:1 multiple:1 reduces:1 technical:1 faster:2 adapt:3 characterized:1 long:3 compensate:1 manipulate:2 a1:4 impact:4 ensuring:1 variant:1 foerster:1 arxiv:4 sometimes:1 tailored:1 achieved:3 receive:1 whereas:1 want:2 crash:1 separately:1 preserved:1 szepesv:1 grow:1 malicious:1 source:1 publisher:1 ot:1 rest:1 extra:2 ardi:1 sure:1 claus:3 tend:1 thing:2 enough:1 automated:1 switch:2 independence:1 affect:2 suboptimal:1 idea:1 regarding:1 knowing:1 whether:3 url:3 effort:1 penalty:1 peter:1 speaking:2 cause:1 jal:2 action:70 remark:3 deep:3 boutilier:4 useful:1 collision:1 clear:2 generally:3 transforms:1 amount:1 discount:1 processed:2 generate:2 http:5 happened:1 fulfilled:1 estimated:1 per:1 kill:1 discrete:1 group:1 key:2 changing:1 prevent:3 button:7 asymptotically:2 sum:1 compete:1 qeq:3 uncertainty:1 respond:1 communicate:3 decision:1 appendix:1 pushed:3 guaranteed:2 played:3 distinguish:2 followed:1 annual:1 adapted:1 dangerous:5 occur:1 sparked:1 precisely:2 alex:1 ri:1 bibliography:1 aspect:1 optimality:2 concluding:1 pruned:1 performing:2 martin:1 designated:1 according:2 developing:1 combination:1 smaller:1 sandholm:1 making:1 happens:1 s1:1 taken:2 equation:1 remains:1 turn:3 eventually:1 discus:1 mechanism:5 needed:3 know:3 pin:2 mind:1 antonoglou:1 end:2 available:1 decentralized:4 opponent:2 observe:3 occasional:1 indirectly:1 kowalczyk:1 anymore:1 save:2 alternative:1 existence:1 original:2 ensure:5 include:2 maintaining:1 restrictive:1 epsilon:1 especially:1 prof:1 establish:1 classical:1 contact:1 question:2 added:1 already:2 quantity:1 strategy:4 rt:15 dependence:1 interruption:61 said:7 distance:1 link:1 reversed:1 argue:1 reason:3 marcus:1 besides:1 tuomas:1 robert:2 negative:2 stated:1 korjus:1 design:2 countable:1 policy:47 perform:2 observation:8 markov:6 sm:1 daan:1 finite:2 situation:6 extended:2 communication:3 team:1 jakob:1 introduced:4 david:1 namely:1 pair:2 required:1 fort:1 raising:1 learned:6 narrow:1 nip:2 address:2 able:3 below:2 perception:1 pattern:1 usually:2 panait:1 challenge:2 max:1 suitable:1 event:1 natural:2 rely:3 critical:1 business:1 minimax:2 scheme:2 misleading:1 nice:1 literature:1 review:1 evolve:1 graf:1 fully:7 multiagent:6 expect:1 highlight:1 limitation:1 tambet:1 proven:2 foundation:1 agent:135 gather:2 sufficient:4 article:1 principle:2 playing:1 compatible:10 cooperation:1 supported:1 last:1 keeping:1 liviu:1 bias:4 allow:2 aru:2 taking:2 distributed:1 transition:6 cumulative:4 world:2 author:2 made:3 reinforcement:22 adaptive:1 far:1 newspaper:1 transaction:1 pruning:4 observable:4 qlearning:3 keep:1 satinder:1 abstracted:1 global:1 uai:1 b1:3 owns:1 conclude:1 assumed:1 tuples:1 harm:2 gomes:1 triplet:4 why:3 table:1 learn:22 ca:1 inherently:1 whiteson:1 interact:1 complex:1 european:1 main:2 crites:1 big:6 bounding:1 s2:1 whole:3 verifies:3 allowed:1 slow:1 candidate:2 replay:2 mahdi:1 dominik:1 watkins:1 learns:2 yannis:1 admissible:3 shimon:1 theorem:12 removing:2 xiaofeng:1 bad:1 specific:4 mhamdi:1 xt:27 list:1 essential:1 justifies:1 push:1 illustrates:1 suited:2 explore:2 infinitely:2 likely:1 happening:1 springer:2 ch:3 corresponds:2 satisfies:1 relies:3 constantly:1 acm:3 goal:4 viewed:1 piat:1 towards:3 installing:1 change:4 onestep:1 included:1 infinite:11 specifically:3 uniformly:1 vicente:1 flag:1 miss:1 lemma:4 called:7 pas:1 select:2 guillaume:1 alexander:1 armstrong:6 |
6,211 | 6,619 | Interactive Submodular Bandit
1
Lin Chen1,2 , Andreas Krause3 , Amin Karbasi1,2
Department of Electrical Engineering, 2 Yale Institute for Network Science, Yale University
3
Department of Computer Science, ETH Z?rich
{lin.chen, amin.karbasi}@yale.edu, [email protected]
Abstract
In many machine learning applications, submodular functions have been used
as a model for evaluating the utility or payoff of a set such as news items to
recommend, sensors to deploy in a terrain, nodes to influence in a social network,
to name a few. At the heart of all these applications is the assumption that the
underlying utility/payoff function is known a priori, hence maximizing it is in
principle possible. In many real life situations, however, the utility function is not
fully known in advance and can only be estimated via interactions. For instance,
whether a user likes a movie or not can be reliably evaluated only after it was
shown to her. Or, the range of influence of a user in a social network can be
estimated only after she is selected to advertise the product. We model such
problems as an interactive submodular bandit optimization, where in each round
we receive a context (e.g., previously selected movies) and have to choose an action
(e.g., propose a new movie). We then receive a noisy feedback about the utility
of the action (e.g., ratings) which we model as a submodular function over the
context-action space. We develop SM-UCB that efficiently trades off exploration
(collecting more data) and
? exploration (proposing a good action given gathered
data) and achieves a O( T ) regret bound after T rounds of interaction. More
specifically, given a bounded-RKHS norm kernel over the context-action-payoff
space that governs the smoothness of the utility function, SM-UCB keeps an upperconfidence bound on the payoff function that allows it to asymptotically achieve
no-regret. Finally, we evaluate our results on four concrete applications, including
movie recommendation (on the MovieLense data set), news recommendation (on
Yahoo! Webscope dataset), interactive influence maximization (on a subset of the
Facebook network), and personalized data summarization (on Reuters Corpus). In
all these applications, we observe that SM-UCB consistently outperforms the prior
art.
1
Introduction
Interactive learning is a modern machine learning paradigm that has recently received significant
interest in both theory and practice [15, 14, 7, 6]. In this setting, the learning algorithm engages in a
two-way dialog with the environment (e.g., users) by performing actions and receiving a response (e.g.,
like or dislike) for each action. Interactive learning has led to substantial performance improvement
in a variety of machine learning applications [43, 13], including clustering [4, 25, 1], classification
[46, 10], language learning [48], decision making [26], and recommender systems [28], to name a
few.
At a high level, interactive learning can be cast as a dynamic optimization problem with a known
utility/payoff function where the goal is to achieve an objective whose value depends on the selected
actions, their responses, and the state of the environment. In many practical settings, the utility
functions are submodular, stating (informally) that the payoff of performing an action earlier is more
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
than performing it later. In fact, rigorous treatment of submodularity in interactive settings led to
strong theoretical guarantees on the performance of greedy policies [19, 22].
In this paper, we go one step further and consider scenarios where the exact form of the submodular
payoff function is not completely known and hence needs to be estimated through interactions. This
problem is closely related to the contextual multi-armed bandit [38, 9, 36] where for a sequence of
T rounds, we receive a payoff function along with some side information or context (e.g., user?s
features), based on which we have to choose an action (e.g., proposing an item) and then a noisy
feedback about the obtained payoff is revealed (e.g., rating of the proposed item). The goal is to
minimize the regret of not selecting the optimal action due to the uncertainty associated with the
utility function. The interactive contextual bandit generalizes this setting by allowing to interact with
a payoff function multiple times, where each time we need to take a new action based on both the
context and previously taken actions. The regret is then defined in terms of the difference between the
utility of the best set of actions that we could have chosen versus the ones that are actually selected.
In this paper, we further assume that the marginal payoffs of actions show diminishing returns. This
problem, which we call interactive submodular bandit, appears in many practical scenarios, including:
? Interactive recommender system. The goal is to design a recommender system that
interacts with the users in order to elicit and satisfy their preferences. In our approach, we
model the utility of a set of items as an unknown submodular objective function that the
recommender systems aims to maximize. In each round of interaction, the recommender
system decides which item should be presented to the user, given the previously proposed
items to this or similar users (affinity between users, if this side information exists, can
be leveraged to enhance recommendation performance). Since the users? preferences are
unknown, the recommender system can only gather information about the users through the
feedback they provide in terms of ratings. A successful recommender system should be able
to minimize the total regret accumulated over T iterations with users.
? Interactive influence maximization. Influence spread maximization addresses the problem
of selecting the most influential source nodes of a given size in a diffusion network [51].
A diffusion process that starts with those source nodes can potentially reach the greatest
number of nodes in the network. Under many diffusion models, the expected total number
of influenced people is a submodular function of the seed subjects [29, 20]. In a natural
interactive variant of this problem, users may be recruited in a sequential manner [42, 5]
where a new user is selected once we fully observe the extent to which the current seed users
influenced people. Note that finding the optimal set of source nodes in a diffusion network
depends dramatically on the underlying dynamics of the diffusion. Again, very often in
practice, we are faced with the dilemma of estimating the underlying diffusion parameters
through interactively selecting the nodes while at the same time trying to maximize the
influence.
? Interactive coverage. Coverage problems arise naturally in many applications [32, 33].
Consider an environmental monitoring task, for instance, where sensors are placed in the
Alps to better predict floods, landslides and avalanches [8]. Similarly, Wi-Fi hotspots are
carefully arranged to cover every corner of a floor. However, it is likely that the actual
coverage of a device is uncertain before deployment due to unknown conditions present in
the environment. Hence, we might need to install devices in a sequential manner after we
observe the actual coverage of the ones already installed [19, 22].
Without any assumptions about the smoothness of the payoff function, no algorithm may be able
to achieve low regret [31]. Thus, in our setting, we make a crucial yet very natural assumption
that the space of context-action-payoff has low complexity, quantified in terms of the Reproducing
Kernel Hilbert Space (RKHS) norm associated with some kernel [24].
We then show that SM?
UCB, an upper-confidence-bound based algorithm achieves an O( T ) regret. We also evaluate
the performance of SM-UCB on four real-world applications, including movie recommendation
[40], news recommendation [16], interactive influence maximization [42, 34], and personalized data
summarization [39].
2
2
Problem Formulation
As we stated earlier, many utility or payoff functions we encounter in machine learning applications
are submodular. As a reminder, a set function f is called submodular if for all A ? B ? ? we have
f (A) + f (B) ? f (A ? B) + f (A ? B).
An equivalent definition of submodularity that shows better the diminishing returns property is as
follows: for all A ? B ? ? and any element e 6? B we have
f (A ? {e}) ? f (A) ? f (B ? {e}) ? f (B).
We also denote the marginal gain of an element e to a set A by ?(e|A) , f (A ? {e}) ? f (A). The
function f is called monotone if for all A ? B we have f (A) ? f (B).
In this paper, we consider a sequential decision making process for a horizon of T time steps, where
at each round i, a monotone submodular function fi , is (partially) revealed. Let us first consider the
simple bandit problem [50] where we need to select an item (or an arm) e from the set of items ?
such that fi (e) is maximized. After the item e is selected, the payoff fi (e) is revealed. Since ft ?s
are not known in advance, the goal is to minimize the accumulated regret over T rounds, for not
choosing the optimum items. Contextual submodular bandit generalizes the aforementioned setting
by allowing to receive side information ?i (also called context) in each round i [31]. But still the goal
is to select a single item e such that f?i (e) is maximized. In the interactive contextual submodular
bandit, the focus of this paper, we may encounter the same valuation function f? , with its associate
context ?, multiple times over the time horizon T . Here instead, at each round, we need to propose a
new item that maximizes the marginal payoff given the ones we selected in the previous encounters.
Therefore, we are sequentially building up subsets of items that maximize the payoff for each separate
function f? . For instance, a recommender system may interact with a user (or a number of users)
multiple times. In each interaction, it has to recommend a new item while taking into account what it
has recommended in previous interactions.
More formally, let us assume that we encounter m ? T distinct functions f? , in an arbitrary
order, over the time horizon T , i.e., ? ? {?1 , . . . , ?m }. We denote the arriving ordered sequence
by f1 , f2 , . . . , fT where for each round i, we have fi ? {f?1 , . . . , f?m }. Let us also denote by
ui ? {1, . . . , m} the index of the context received in round i. We also need to maintain a collection of
m sets S1 , . . . , Sm , (initialized to the empty set) corresponding to f?1 , . . . , f?m . Our goal is to select
a subset Sj ? ? for each function f?j that maximizes its corresponding utility f?j (Sj ). Note that if
f?j were known in advance, we could simply use the greedy algorithm. However, in the interactive
submodular bandit setting, we need to build up the sets Sj sequentially and through interactions, as
P
the marginal payoff of an element is only revealed after it is selected. Let oi , j?i 1{uj = ui }
denote the number of occurrences of function f?ui in the first i rounds. In each round, say the i-th
with the corresponding function f?ui , we need to select a new item xoi ,ui from the set of items ?
and add it to Sui . Clearly, after including xoi ,ui , the set Sui will be of cardinality oi . For the ease of
presentation, we denote Sui ? {xoi ,ui } by Soi ,ui , initialized to the empty set in the beginning, i.e.,
S0,ui = ?. After selecting the item xoi ,ui and given the previously selected items Sui , we receive yi ,
a noisy (but unbiased) estimate of xoi ,ui ?s marginal payoff, i.e.,
yi = ?(xoi ,ui |Soi ?1,ui , ?ui ) + i ,
where the marginal gain ?(?|?, ?) : ? ? 2? ? ? ? R is defined as
?(x|S, ?) = f? (S ? {x}) ? f? (S).
(1)
We also assume that i ?s are uniformly bounded noise variables that form a martingale difference
sequence, i.e.,
E[i |1 , 2 , . . . , i?1 ] = 0 and |i |? ? for all i ? {1, . . . , T }.
We call ? the noise level. Note that yi is the only feedback that we obtain in round i. It crucially
depends on the previously selected items and contextual information Soi ?1,ui , ?ui . Therefore, the
only avenue through which we can learn about the payoff functions f? (that we try to optimize over)
is via noisy feedbacks yi . We need to design an algorithm that minimizes the accumulated regret
over the total number of T rounds. Formally, we compare the performance of any algorithm in this
interactive submodular bandit setting with that of the greedy algorithm with the full knowledge of the
payoff functions f?1 , . . . , f?m .
3
Algorithm 1 SM-UCB
Input: set of items ?, mean ?0 = 0, variance ?0 .
1: Initialize Si ? ? for all i ? [m]
2: for i = 1, 2, 3, . . . do
?
3:
select an item xoi ,ui ? argmaxx?? ?i?1 (x) + ?i ?i?1 (x)
4:
update the set Sui ? Sui ? {xoi ,ui }
5:
obtain the feedback yi = ?(xoi ,ui |Soi ?1,ui , ?ui ) + i
6:
let ki (x) be a vector-valued function that outputs an i-dimensional column vector with j-th
entry k((xoj ,uj , Soj ?1,uj , ?uj ), (x, Soi ?1,ui , ?ui ))
7:
let Ki be an i?i matrix with (j, j 0 )-entry k((xoj ,uj , Soj ?1,uj , ?uj ), (xoj0 ,uj0 , Soj0 ?1,uj0 , ?uj0 ))
8:
update yi ? [y1 , y2 , . . . , yi ]T
9:
let ki (x, x0 ) be a kernel function defined as k(x, x0 ) ? ki (x)T (Ki + ? 2 I)?1 ki (x0 )
10:
estimate ?i (x) ? p
ki (x)T (Ki + ? 2 I)?1 yi
11:
estimate ?i (x) ? ki (x, x)
12: end for
Suppose that by the end of the T -th round, an algorithm has selected Tj items for the payoff
function f?j ; therefore the cardinality of Sj by the end of the T -th round is Tj . Thus, we have
P
Pm
Tj = t?T 1{ut = j} and T = j=1 Tj . We use Sj? to denote the set that maximizes the payoff
of function f?j with at most Tj elements, i.e., Sj? = argmax|S|?Tj f?j (S). We know that the greedy
Pm
algorithm is guaranteed to achieve (1 ? 1/e) j=1 f?j (Sj? ) [41] and there is no polynomial time
algorithm that achieves a better approximation guarantee in general [17]. Therefore, we define the
total regret of an algorithm up to round T as follows:
RT , (1 ? 1/e)
m
X
j=1
f?j (Sj? ) ?
m
X
f?j (STj ,j ),
(2)
j=1
which is the gap between the greedy algorithm?s guarantee and the total utility obtained by the
algorithm. Without any smoothness assumption over the payoff functions, it may not be possible to
guarantee a sublinear regret [31]. In this paper, we make a natural assumption about the complexity
of payoff functions. More specifically, we assume that the marginal payoffs, defined in (1), have a low
RKHS-norm according to a kernel k : (? ? 2? ? ?) ? (? ? 2? ? ?) ? R, i.e., k?(?|?, ?)kk ? B.
Note that such a kernel encodes how close two marginal payoffs are if a) the contexts ?i and ?j or b)
the selected elements Si and Sj are similar. For instance, a recommender system can leverage this
information to propose an item to a user if it has observed that a user with similar features liked that
item.
3
Main Results
In Algorithm 1 we propose SM-UCB, an interactive submodular bandit algorithm. Recall that
the marginal gain function ? has a low RKHS norm w.r.t. some kernel k. In each round, say the
i-th, SM-UCB maintains the posterior mean ?i?1 (?) and standard deviation ?i?1 (?) conditioned
on the historical observations or context {(xoj ,uj , Soj ?1,uj , ?uj ) : 1 ? j ? i}. Based on these
posterior estimates,
SM-UCB then selects an item x that attains the highest upper confidence bound
?
?i?1 (x) + ?i ?i?1 (x). It then receives the noisy feedback yi = ?(xoi ,ui |Soi ?1,ui , ?ui ) + i . Since
i ?s are uniformly bounded and form a martingale difference sequence, SM-UCB can predict the
mean ?i and standard deviation ?i via posterior inference in order to determine the item to be selected
in the next round.
In order to bound the regret of an algorithm, we need to quantify how much information that algorithm
can acquire through interactions. Let yA denote a subset of noisy observations indexed by the set
A, i.e., yA = {yi |i ? A}. Note that any subset A of noisy observations yA reduces our uncertainty
about the marginal gain function ?. In an extreme case, if we had perfect information (or no
uncertainty) about ?, we could have achieved zero regret. We can precisely quantify this notion
through what is called the information gain I(yA ; ?) = H(yA ) ? H(yA |?), where H denotes the
Shannon entropy. In fact, by lifting the results from [31, 44] to a much more general setting, we
relate the regret to the maximum information gain ?T [44] obtained after T rounds and defined as
4
?T , maxA??:|A|=T I(yA ; ?). Another important quantity that shows up in the regret bound is the
confidence parameter ?T (see line 3 of Algorithm 1) that needs to be chosen carefully so that our
theoretical guarantee
holds with high probability. In fact, the following theorem shows that SM-UCB
?
attains a O( T ?T ?T ) regret bound with high probability.
Theorem 1. Suppose that the true marginal gain function ?(?|?, ?) has a small RKHS norm according
to some kernel k, i.e., k?(?|?, ?)kk ? B. The noise variables t satisfy E[t |1 , 2 , . . . , t?1 ] = 0
for all t ? N and are uniformly bounded by ?. Let ? ? (0, 1), ?t = 2B 2 + 300?t ln3 (t/?) and
C1 = 8/log(1 + ? ?2 ). Then, the accumulated regret of SM-UCB over T rounds is as follows:
n
o
p
Pr RT ? C1 T ?T ?T + 2, ?T ? 1 ? 1 ? ?.
The proof of the above theorem is provided in the Supplementary Material. It relies on two powerful
ideas: greedy selection for constrained submodular maximization [41] and upper confidence bounds
of contextual Gaussian bandit [31]. If the marginal payoffs were completely known, then the greedy
policy would provide a competitive solution to the optimum. However, one cannot run the greedy
policy without knowing the marginal gains. In fact, there are strong negative results regarding the
approximation guarantee of any polynomial time algorithm if the marginal gains are arbitrarily noisy
[23]. Instead, SM-UCB relies on optimistic estimates of the marginal gains and select greedily an
item with the highest upper confidence bound. By assuming that marginal gains are smooth and
relying on Theorem 1 in [31], we can control the accumulated error of a greedy-like solution that
relies on confidence bounds and obtain low regret. Our setting and theoretical result generalize a
number of prior work mentioned below.
Linear submodular bandit [50]. In this setting, the objective function has the form f (S) =
Pd
positive unknown coeffii=1 wi fi (S), where fi ?s are known submodular functions and wi ?s are P
d
cients. Therefore, the marginal gain function can be written as ?(x|S) = i=1 wi ?i (x|S), where
?i (?|?)?s are known functions and wi ?s are unknown coefficients. Let w = (w1 , w2 , . . . , wd ) denote
the weight vector. Since the only unknown part of the marginal gain function is the weight vector,
the space of the marginal gain function is isomorphic to the space of weight vectors, which is in
fact a d-dimensional Euclidean space Rd . The RKHS norm of ? is given by some norm in Rd ;
i.e., k?kk , kwk. The assumption in [50] that kwk? B is equivalent to assuming that k?kk ? B.
Therefore, the linear bandit setting is included in our setting where the marginal gain function ? has
a special form and its RKHS norm is given by the norm of its corresponding weight vector in the
Euclidean space. Also, LSBG REEDY proposed in [50], is a special case of SM-UCB (except that
the feedback is delayed).
Adaptive valuable
P item discovery [47]. In this setting, the objective function has the form
f (S) = (1 ? ?) x?S g(x) + ?D(S), where D is a known submodular function that quantifies
the diversity of the items in S, g is an unknown function that denotes the utility g(x) for any item
x, and ? is a known tradeoff parameter balancing the importance of theP
accumulative utility and
joint diversity of the items. Note that the unknown function M (S) = x?S g(x) is a modular
function. Therefore, the marginal gain function has the form ?(x|S) = (1 ? ?)g(x) + ?D(x|S),
where D(x|S) , D({x} ? S) ? D(S). The only uncertainty of ? arises from the uncertainty about
the modular function M . In particular, [47] assumes that the RKHS norm of g is bounded. Again,
our setting encompasses adaptive valuable item discovery as we consider any monotone submodular
function. Moreover, GPS ELECT proposed in [47], is a special case of SM-UCB.
Contextual Gaussian bandit [31]. This is the closest setting to ours where in each round i we
receive a context ?i from the set of contexts ? and have to choose an item x from the set of items
?. We then receive a payoff f?i (x) + t . Note that instead of building up a set (our problem), in
the contextual bandit process we simply choose a single element for each function f?i as the main
assumption is that we encounter each function only once. To obtain regret bounds it is assumed in
[31] that f has low norm in the RKHS associated with some kernel k. Again, CGP-UCB proposed
in [31], is a special case of SM-UCB.
5
4
Experiments
In this section, we compare empirically the performance of SM-UCB with the following baselines:
? R ANDOM. In each round, an item is randomly selected for the current payoff function f? .
? G REEDY. It has the full knowledge of the submodular functions f? . In each round, say
the i-th with the corresponding function f?ui , G REEDY selects the item that maximizes the
marginal gain, i.e., argmaxx?? ?(x|Soi ?1 , ?ui ).
? H ISTORY-F REE. We run SM-UCB without considering the previously selected items.
H ISTORY-F REE is basically the contextual Gaussian bandit algorithm proposed in [31]
whose context is the user feature.
? F EATURE -F REE. We run SM-UCB without considering the context ? of an arriving
function f .
? C ONTEXT-F REE. We run SM-UCB without considering the context or the previously
selected elements. In fact, C ONTEXT-F REE is basically the GP-S ELECT algorithm proposed
in [47].
In all of our experiments, the m distinct functions {f?i : 1 ? i ? m} that the algorithm encounters
represent the valuation functions of m users, where the context ?i ? Rd encodes users? features.
Moreover, Si is the set of items that an algorithm selects for user i ? [m]. Instead of computing
the
Pmregret, we quantify the performance of the algorithms by computing
Pm the accumulated reward
f
(S
).
Recall
that
the
regret
is
given
by
(1
?
1/e)
?
OP
T
?
?
i
i
i=1 f?i (Si ), where OP T =
i=1
Pm
?
f
(S
)
is
a
constant
generally
hard
to
compute.
i
i=1 ?i
Movie Recommendation In this set of experiments, we use the MovieLens dataset1 where a userrating matrix M is provided. The rows of M represent users and the columns represent movies. The
matrix M contains 943 users and 1682 movies. As a preprocessing step, we apply the singular-value
decomposition (SVD) to impute the missing values; the six largest singular values are kept.
In the first part of the study, we use the submatrix of M that consists of 80% of the users and all
of the movies for training the feature vectors of movies via SVD. Let M 0 be the submatrix of M
that consists of the remaining nuser users and all of the movies; this matrix is for testing. Let ?
denote the set of movies. We
selecting a subset of ? to maximize the facility-locationPnconsider
user
0
type objective [30] f (S) = i=1
maxj?S Mij
. This objective function corresponds to a scenario
without any context ? as there is only one payoff function f that we are trying to maximize. Thus,
F EATURE -F REE does not apply here. We use the cosine kernel kmovie : ? ? ? ? R for pairs of
movies and Jaccard kernel ksubset (S, T ) = |S ? T |/|S ? T | [18] for pairs of subsets of movies, say
S and T . The composite kernel k : (? ? 2? ) ? (? ? 2? ) ? R is defined as ?1 kmovie ? ?2 ksubset ,
i.e., k((u, S), (v, T )) = ?1 kmovie (u, v) + ?2 ksubset (S, T ), where ?1 , ?2 > 0. The results are shown
in Fig. 1(a). The horizontal axis denotes the cardinality of S. The vertical axis denotes the function
value of f on the set S. We observe that SM-UCB outperforms all of the baselines except the
practically infeasible G REEDY.
In the second part, we consider a setting where a separate subset of movies is selected for each user.
We cluster the users in the dataset into 40 groups via the k-means algorithm and the users of the
same group are viewed as identical users. The feature vector of a group of users is the mean of the
feature vectors of all member users and the rating of a group is the sum of the ratings of all member
users. The users are labeled as 1, 2, 3, . . . , n0user , where n0user = 40. Similar to the first part, the
feature vectors of the users and movies are obtained via SVD. We maintain a set Si for user i. The
00
00
objective function is f?i (S) = maxj?S Mij
, where Mij
is user i?s rating for movie j. In addition,
we also need a collective objective function that quantifies the overall performance of an algorithm
Pn0user
for all users. It is defined as f (S1 , S2 , . . . , Sn0user ) = i=1
f?i (Si ). We assume a random arrival
of users. We use the linear kernel kuser : ? ? ? ? R for pairs of users. The composite kernel
k : (? ? 2? ? ?) ? (? ? 2? ? ?) ? R is defined as ?1 kmovie ? ?2 ksubset ? ?3 kuser . In Fig. 1(b),
we plot the performance of SM-UCB against other baselines. The horizontal axis denotes the number
of user arrivals while the vertical axis denotes the value of the collective objective function. We
1
https://grouplens.org/datasets/movielens/
6
Greedy
SM-UCB
SM-UCB
Feature free
Greedy
History free
Context free
History free
Random
Random
(a)
(b)
History free
SM-UCB
SM-UCB
Random
History free
Greedy
Random
Greedy
(c)
(d)
SM-UCB
Greedy
Greedy
History free
SM-UCB
Feature free
Random
History free
Feature free
Context free
Context free
(e)
(f)
Figure 1: Figs. 1(a) and 1(b) show the results of the experiments on the MovieLens dataset. Fig. 1(a)
shows how the total objective function for all users evolves as the number of selected movies increases;
the algorithm recommends the same subset of movies to all users. In Fig. 1(b), we consider the
situation where users arrive in a random order and we have to recommend a separate subset of movies
to each user. Figs. 1(c) and 1(d) show the dependency of fraction of influenced nodes on the target
set size in the Facebook network and the student network from the User Knowledge Modelling
Dataset [27]. Fig. 1(e) shows how the payoff function varies as more users arrive in the Yahoo news
recommender. In Fig. 1(f), we consider the personalized data summarization from Reuters corpus for
arriving users. It shows the fraction of covered topics versus the number of user arrivals.
observe that SM-UCB outperforms all other baselines except G REEDY. In addition, C ONTEXT-F REE
that uses the least amount of information achieves a lower function value than H ISTORY-F REE and
F EATURE -F REE, which either leverages the information about users? features or previously selected
items.
7
Interactive Influence Maximization For this experiment, we use the Facebook network provided
in [35]. The goal is to choose a subset of subjects in the network, which we call the target set, in
order to maximize the number of influenced subjects. We assume that each member in the target
set can influence
all of her neighbors.
Under this assumption, the submodular objective function is
S
f (S) = u?S (N (u) ? {u}), where N (u) is the set of all neighbors of subject u. All the baselines,
except G REEDY, have no knowledge of the underlying Facebook network or the objective function.
They are only given the feature vector of each subject obtained via the NODE 2 VEC algorithm [21].
The kernel function ksubject between two subjects is a linear kernel while the kernel function between
subsets of subjects is the Jaccard kernel. The results are shown in Fig. 1(c). Again, SM-UCB reaches
the largest influence w.r.t other baselines except for G REEDY. We ran the same idea over the 6-nearest
neighbor network of randomly sampled 150 students from User Knowledge Modelling Dataset [27].
As Fig. 1(d) indicates, a similar pattern emerges.
News Recommendation For this experiment, we use the Yahoo! Webscope dataset R6A2 . The
dataset provides a list of records, each containing a time stamp, a user ID, a news article ID and
a Boolean value that indicates whether the user clicked on the news article that was presented to
her. The feature vectors of the users and the articles are also provided. We use k-means clustering
to cluster users into 175 groups and identify users of the same group as identical users. We form a
matrix M whose (i, j)-entry is the total number of times that user i clicked on article j. This matrix
quantifies each user?s preferences regarding news articles. The objective function for user i is defined
as f?i (Si ) = maxj?Si Mij . The collective objective function f is defined as the sum of the objective
functions of all users. From the time stamps, we can infer the order in which the users arrive. We
use the Laplacian kernels knews : ? ? ? ? R and kuser : ? ? ? ? R for pairs of pieces of news
and pairs of users, respectively. For a pair of subsets of news S and T , the kernel function between
them is again the Jaccard kernel. The composite kernel k : (? ? 2? ? ?) ? (? ? 2? ? ?) ? R is
defined as ?1 knews ? ?2 ksubset ? ?3 kuser . The results are illustrated in Fig. 1(e). The horizontal axis
is the number of arriving users while the vertical axis is the value of the collective objective function.
Again, we observe that SM-UCB outperforms all other methods except G REEDY.
Personalized Data Summarization For this experiment, we apply latent Dirichlet allocation
(LDA) to the Reuters Corpus. The number of topics is set to ntopic = 10. LDA returns a topic
distribution P (i|a) for each article a and topic i. Suppose that A is a subset of articles. Probabilistic
coverage function quantifies
the degree to which a set of articles A covers a topic i [16], and is
Q
given by Fi (A) = 1 ? a?A (1 ? P (i|a)). Each user j is characterized by her ntopic -dimensional
preference vector wj = (wj,1 , wj,2 , wj,3 , . . . , wj,ntopic ); we assume that the preference vector is
L1 -normalized, i.e., its entries
sum to 1. The personalized probabilistic coverage function for user
Pntopic
j is defined as fj (A) = i=1
wj,i Fi (A) [16, 50]. Note that since the preference vector is L1 normalized and
i (A) ? 1, we have fj (A) ? 1 for any j. The total average coverage function is
PF
nuser
1
f (A) = nuser
j=1 fj (A), where nuser = 10 is the number of users. Random order of user arrivals
is simulated. We use the linear kernel for pairs of users and pairs of articles and use the Jaccard kernel
between subsets of articles. The results are shown in Fig. 1(f). The horizontal axis is the number of
user arrivals while the vertical axis is the total average coverage function f (A), which characterizes
the average fraction of covered topics. We observe that SM-UCB outperforms all the baselines other
than G REEDY.
Discussion Recall that the RKHS is a complete subspace of the L2 space of functions defined on
the product of the item set, its power set, and the context set. It has an inner product (?, ?)k obeying
the reproducing property: (f, k(x, ?))k =P
f (x) for all f in RKHS. Functions implied by a particular
kernel k are always of the form f (x) = i ?i k(xi , x). The bounded norm implies that ?i vanish
quickly enough. With universal kernels like Gaussian/Laplacian kernels, such functions are dense
(according to sup-norm) in the space of continuous functions.
In three sets of experiments (movie recommendation, influence maximization, data summarization)
we used the linear and cosine kernels for items and users, and the Jaccard kernel for subsets of items.
In fact, the Jaccard kernel is a widely used metric that quantifies the similarity between subsets of
selected items. Moreover, the linear and cosine kernels between items and users capture the simplest
form of interactions. In contrast to the the above three experiments, in the news recommendation
2
http://webscope.sandbox.yahoo.com/
8
application, we chose the Laplacian kernel for the following reason. The features provided in the
dataset have highly heterogeneous norms. If we use the linear kernel, the inner product between a
short vector and a close-by vector with a small norm will be easily dominated by the inner product
with a vector with a large norm. We used the Laplacian kernel to circumvent this problem and put
more weight on nearby vectors even if they have small norms.
5
Related Work
Originally, Auer et al. [2] proposed UCB policies for the multi-armed
? bandit (MAB) problem which
exhibits the exploration-exploitation tradeoff and achieves an O( T ) regret. In the literature, there
are many variants of the multi-armed bandit problem and corresponding solutions, for example,
E XP 3 algorithm for adversarial bandits [3], L IN UCB for stochastic contextual bandits [36, 12], and a
family of UCB-based policies for infinitely many-armed bandit [49]. Chen et al. [11] considered
the combinatorial MAB problem where the unit of play is a super arm and base arms can be
probabilistically triggered. For a comprehensive survey on bandit problems, we refer the interested
reader to [9].
Srinivas et al. [44] studied the Gaussian process (GP) optimization problem in the bandit setting.
They assumed that the objective function f is either sampled from a Gaussian process or resides
in
Pa reproducing kernel Hilbert space (RKHS). Given a subset of items S ? ? , the total utility is
or the RKHS assumption, they showed that their proposed
x?? f (x). Under either the GP model
?
GP-UCB algorithm achieves an O( T ) regret bound. It is noteworthy to mention that their bound
also relies on the maximum information gain. Based on [44], Krause and Ong [31] further investigated
the contextual Guassian process bandit optimization and their proposed algorithm CGP-UCB
achieves a similar regret bound. Lin et al. [37] addressed an online learning problem where the
input to the greedy algorithm is stochastic with unknown parameters and the algorithm receives
semi-bandit feedbacks. Their algorithm can also be applied to submodular functions. However, there
are several major differences between their work and ours: Firstly, they assume that the objective
functions are drawn from a predetermined but unknown distribution, while our work applies to any
set of submodular functions; secondly they assume bounded submodular functions while we have no
such assumptions; thirdly, their work did not have the notion of context. They optimize the expected
objective function while we optimize objective functions with different contexts simultaneously.
Streeter and Golovin [45] studied the online maximization problem of submodular functions. Yue and
Guestrin [50] studied the linear submodular bandit problem where they assumed that the unknown
submodular function is a linear combination of multiple known submodular functions. The only
uncertainty in their setting is the unknown positive coefficients
of each known submodular function.
?
They proposed LSBG REEDY that achieves a similar O( T ) regret bound. Beyond unconstrained
sequential decision problems, Zhou et al. [52] considered online maximization of list submodular
functions under a knapsack constraint.
Our key contribution in this paper is that the notion of contextual regret that we bound is much more
challenging than the typical notion: Our actions are affecting the future contexts experienced, and we
compete with policies that are aware of this fact and can plan for it. This is qualitatively different
from any prior analysis. More specifically, we need to build up a subset of items/actions as we
encounter a valuation function multiple times. This is a non-trivial task as not only the functions are
unknown, the marginal gains are also noisy. Moreover, the choices we make can affect the future. Our
positive results can be seen in light of very recent negative results in [23] that indicates submodular
optimization is hard when function evaluations are noisy. We show that the UCB-based algorithm
can be naturally combined with the greedy selection policy to provide sublinear regret. To the best of
our knowledge the analysis is new.
Acknowledgements
This research was supported by DARPA Young Faculty Award (D16AP00046), grant SCADAPT and
ERC StG.
9
References
[1] Hassan Ashtiani, Shrinu Kushagra, and Shai Ben-David. Clustering with same-cluster queries.
In NIPS, pages 3216?3224, 2016.
[2] Peter Auer, Nicolo Cesa-Bianchi, and Paul Fischer. Finite-time analysis of the multiarmed
bandit problem. Machine learning, 47(2-3):235?256, 2002.
[3] Peter Auer, Nicolo Cesa-Bianchi, Yoav Freund, and Robert E Schapire. The nonstochastic
multiarmed bandit problem. SIAM journal on computing, 32(1):48?77, 2002.
[4] Pranjal Awasthi, Maria-Florina Balcan, and Konstantin Voevodski. Local algorithms for
interactive clustering. In ICML, pages 550?558, 2014.
[5] Ashwinkumar Badanidiyuru, Christos Papadimitriou, Aviad Rubinstein, Lior Seeman, and
Yaron Singer. Locally adaptive optimization: Adaptive seeding for monotone submodular
functions. In SODA, pages 414?429, 2016.
[6] Maria-Florina Balcan, Andrei Broder, and Tong Zhang. Margin based active learning. In COLT,
pages 35?50. Springer, 2007.
[7] Maria-Florina Balcan, Alina Beygelzimer, and John Langford. Agnostic active learning. Journal
of Computer and System Sciences, 75(1):78?89, 2009.
[8] Guillermo Barrenetxea, Fran?ois Ingelrest, Gunnar Schaefer, and Martin Vetterli. The hitchhiker?s guide to successful wireless sensor network deployments. In ACM SenSys, pages 43?56,
2008.
[9] S?bastien Bubeck, Nicolo Cesa-Bianchi, et al. Regret analysis of stochastic and nonstochastic
R in Machine Learning, 5(1):1?122,
multi-armed bandit problems. Foundations and Trends
2012.
[10] Lin Chen, Seyed Hamed Hassani, and Amin Karbasi. Near-optimal active learning of halfspaces
via query synthesis in the noisy setting. In AAAI, pages 1798?1804, 2017.
[11] Wei Chen, Yajun Wang, Yang Yuan, and Qinshi Wang. Combinatorial multi-armed bandit and
its extension to probabilistically triggered arms. JMLR, 17(1):1746?1778, 2016.
[12] Wei Chu, Lihong Li, Lev Reyzin, and Robert E Schapire. Contextual bandits with linear payoff
functions. In AISTATS, volume 15, pages 208?214, 2011.
[13] David A Cohn, Zoubin Ghahramani, and Michael I Jordan. Active learning with statistical
models. In NIPS, pages 705?712, 1995.
[14] Sanjoy Dasgupta. Analysis of a greedy active learning strategy. In NIPS, pages 337?344, 2005.
[15] Sanjoy Dasgupta, Daniel J Hsu, and Claire Monteleoni. A general agnostic active learning
algorithm. In NIPS, pages 353?360, 2008.
[16] Khalid El-Arini, Gaurav Veda, Dafna Shahaf, and Carlos Guestrin. Turning down the noise in
the blogosphere. In ACM SIGKDD, pages 289?298, 2009.
[17] Uriel Feige. A threshold of ln n for approximating set cover. Journal of the ACM, 45(4):
634?652, 1998.
[18] Thomas G?rtner. Kernels for structured data, volume 72. World Scientific, 2008.
[19] Daniel Golovin and Andreas Krause. Adaptive submodularity: Theory and applications in
active learning and stochastic optimization. JAIR, 42:427?486, 2011.
[20] Manuel Gomez-Rodriguez, Le Song, Nan Du, Hongyuang Zha, and B. Schoelkopf. Influence
estimation and maximization in continuous-time diffusion networks. ACM TOIS, 2016.
[21] Aditya Grover and Jure Leskovec. node2vec: Scalable feature learning for networks. In ACM
SIGKDD, pages 855?864, 2016.
10
[22] Andrew Guillory and Jeff Bilmes. Interactive submodular set cover. In ICML, Haifa, Israel,
2010.
[23] Avinatan Hassidim and Yaron Singer. Submodular optimization under noise. In COLT, volume 65, pages 1069?1122, 2017.
[24] Thomas Hofmann, Bernhard Sch?lkopf, and Alexander J Smola. Kernel methods in machine
learning. The annals of statistics, pages 1171?1220, 2008.
[25] Ruizhang Huang and Wai Lam. An active learning framework for semi-supervised document
clustering with language modeling. Data & Knowledge Engineering, 68(1):49?67, 2009.
[26] Shervin Javdani, Yuxin Chen, Amin Karbasi, Andreas Krause, Drew Bagnell, and Siddhartha S
Srinivasa. Near optimal bayesian active learning for decision making. In AISTATS, pages
430?438, 2014.
[27] H Tolga Kahraman, Seref Sagiroglu, and Ilhami Colak. The development of intuitive knowledge
classifier and the modeling of domain dependent data. Knowledge-Based Systems, 37:283?295,
2013.
[28] Amin Karbasi, Stratis Ioannidis, and laurent Massoulie. Comparison-based learning with rank
nets. In ICML, pages 855?862, 2012.
[29] David Kempe, Jon Kleinberg, and ?va Tardos. Maximizing the spread of influence through a
social network. In ACM SIGKDD, pages 137?146, 2003.
[30] Andreas Krause and Daniel Golovin. Submodular function maximization. Tractability: Practical Approaches to Hard Problems, 3(19):8, 2012.
[31] Andreas Krause and Cheng S Ong. Contextual gaussian process bandit optimization. In NIPS,
pages 2447?2455, 2011.
[32] Andreas Krause, Carlos Guestrin, Anupam Gupta, and Jon Kleinberg. Near-optimal sensor
placements: Maximizing information while minimizing communication cost. In ACM IPSN,
pages 2?10, 2006.
[33] Andreas Krause, Ajit Singh, and Carlos Guestrin. Near-optimal sensor placements in gaussian
processes: Theory, efficient algorithms and empirical studies. JMLR, 9(Feb):235?284, 2008.
[34] Siyu Lei, Silviu Maniu, Luyi Mo, Reynold Cheng, and Pierre Senellart. Online influence
maximization. In ACM SIGKDD, pages 645?654, 2015.
[35] Jure Leskovec and Julian J Mcauley. Learning to discover social circles in ego networks. In
NIPS, pages 539?547, 2012.
[36] Lihong Li, Wei Chu, John Langford, and Robert E Schapire. A contextual-bandit approach to
personalized news article recommendation. In WWW, pages 661?670. ACM, 2010.
[37] Tian Lin, Jian Li, and Wei Chen. Stochastic online greedy learning with semi-bandit feedbacks.
In NIPS, pages 352?360, 2015.
[38] Tyler Lu, D?vid P?l, and Martin P?l. Contextual multi-armed bandits. In AISTATS, pages
485?492, 2010.
[39] Baharan Mirzasoleiman, Ashwinkumar Badanidiyuru, and Amin Karbasi. Fast constrained
submodular maximization: Personalized data summarization. In ICML, volume 48, pages
1358?1367, 2016.
[40] Baharan Mirzasoleiman, Morteza Zadimoghaddam, and Amin Karbasi. Fast distributed submodular cover: Public-private data summarization. In NIPS, pages 3594?3602, 2016.
[41] George L Nemhauser and Laurence A Wolsey. Best algorithms for approximating the maximum
of a submodular set function. Mathematics of operations research, 3(3):177?188, 1978.
11
[42] Lior Seeman and Yaron Singer. Adaptive seeding in social networks. In FOCS, pages 459?468,
2013.
[43] Burr Settles. Active learning literature survey. University of Wisconsin, Madison, 52(55-66):11,
2010.
[44] Niranjan Srinivas, Andreas Krause, Sham Kakade, and Matthias Seeger. Information-theoretic
regret bounds for gaussian process optimization in the bandit setting. IEEE Transactions on
Information Theory, 58(5):3250?3265, May 2012.
[45] Matthew Streeter and Daniel Golovin. An online algorithm for maximizing submodular
functions. In NIPS, pages 1577?1584, 2009.
[46] Simon Tong and Daphne Koller. Support vector machine active learning with applications to
text classification. JMLR, 2(Nov):45?66, 2001.
[47] Hastagiri Vanchinathan, Andreas Marfurt, Charles-Antoine Robelin, Donald Kossmann, and
Andreas Krause. Discovering valuable items from massive data. In ACM SIGKDD, 2015.
[48] S. I. Wang, P. Liang, and C. Manning. Learning language games through interaction. In ACL,
2016.
[49] Yizao Wang, Jean yves Audibert, and R?mi Munos. Algorithms for infinitely many-armed
bandits. In NIPS, pages 1729?1736. 2009.
[50] Yisong Yue and Carlos Guestrin. Linear submodular bandits and their application to diversified
retrieval. In NIPS, pages 2483?2491, 2011.
[51] Yuanxing Zhang, Yichong Bai, Lin Chen, Kaigui Bian, and Xiaoming Li. Influence maximization in messenger-based social networks. In GLOBECOM, pages 1?6. IEEE, 2016.
[52] Jiaji Zhou, Stephane Ross, Yisong Yue, Debadeepta Dey, and J Andrew Bagnell. Knapsack
constrained contextual submodular list prediction with application to multi-document summarization. In In Inferning Workshop at ICML. Citeseer, 2013.
12
| 6619 |@word private:1 exploitation:1 faculty:1 polynomial:2 norm:17 laurence:1 crucially:1 decomposition:1 citeseer:1 mention:1 mcauley:1 bai:1 contains:1 selecting:5 daniel:4 rkhs:13 ours:2 document:2 outperforms:5 yajun:1 current:2 contextual:17 wd:1 com:1 beygelzimer:1 si:8 yet:1 chu:2 written:1 manuel:1 john:2 maniu:1 predetermined:1 hofmann:1 lsbg:2 seeding:2 plot:1 update:2 greedy:19 selected:20 device:2 item:47 discovering:1 beginning:1 short:1 record:1 yuxin:1 provides:1 node:8 preference:6 org:1 firstly:1 zhang:2 daphne:1 along:1 install:1 yuan:1 consists:2 focs:1 burr:1 manner:2 node2vec:1 x0:3 expected:2 dialog:1 multi:7 relying:1 actual:2 armed:8 pf:1 cardinality:3 considering:3 clicked:2 provided:5 estimating:1 underlying:4 bounded:7 maximizes:4 moreover:4 agnostic:2 discover:1 what:2 israel:1 minimizes:1 maxa:1 proposing:2 finding:1 guarantee:6 every:1 collecting:1 interactive:21 classifier:1 control:1 unit:1 grant:1 engages:1 before:1 positive:3 engineering:2 local:1 installed:1 id:2 lev:1 laurent:1 ree:9 noteworthy:1 might:1 chose:1 acl:1 studied:3 quantified:1 challenging:1 deployment:2 ease:1 range:1 tian:1 practical:3 testing:1 practice:2 regret:28 stratis:1 universal:1 elicit:1 eth:1 empirical:1 composite:3 confidence:6 tolga:1 donald:1 zoubin:1 cannot:1 close:2 selection:2 stj:1 put:1 context:25 influence:15 optimize:3 equivalent:2 www:1 missing:1 maximizing:4 go:1 survey:2 kushagra:1 notion:4 siyu:1 tardos:1 annals:1 target:3 deploy:1 suppose:3 user:73 exact:1 play:1 gps:1 us:1 massive:1 associate:1 element:7 pa:1 trend:1 ego:1 labeled:1 observed:1 ft:2 electrical:1 capture:1 wang:4 wj:6 schoelkopf:1 news:12 trade:1 highest:2 valuable:3 ran:1 substantial:1 mentioned:1 environment:3 pd:1 complexity:2 ui:28 reward:1 halfspaces:1 ong:2 dynamic:2 singh:1 badanidiyuru:2 dilemma:1 seyed:1 f2:1 silviu:1 completely:2 vid:1 easily:1 joint:1 darpa:1 distinct:2 guassian:1 massoulie:1 fast:2 query:2 rubinstein:1 choosing:1 schaefer:1 whose:3 modular:2 supplementary:1 valued:1 widely:1 say:4 jean:1 luyi:1 statistic:1 fischer:1 flood:1 gp:4 noisy:11 online:6 sequence:4 triggered:2 net:1 matthias:1 propose:4 lam:1 interaction:10 product:5 cients:1 reyzin:1 achieve:4 amin:7 intuitive:1 empty:2 optimum:2 istory:3 cluster:3 liked:1 perfect:1 ben:1 mirzasoleiman:2 develop:1 stating:1 andrew:2 soj:3 jiaji:1 nearest:1 op:2 received:2 strong:2 coverage:8 ois:1 implies:1 quantify:3 xoi:10 submodularity:3 closely:1 stephane:1 stochastic:5 ipsn:1 exploration:3 alp:1 settle:1 hassan:1 material:1 public:1 f1:1 sandbox:1 mab:2 secondly:1 voevodski:1 extension:1 hold:1 practically:1 considered:2 tyler:1 seed:2 predict:2 mo:1 matthew:1 major:1 achieves:8 estimation:1 yizao:1 combinatorial:2 grouplens:1 ross:1 largest:2 awasthi:1 clearly:1 sensor:5 hotspot:1 always:1 aim:1 super:1 gaussian:9 gaurav:1 zhou:2 probabilistically:2 focus:1 she:1 consistently:1 improvement:1 modelling:2 indicates:3 maria:3 rank:1 contrast:1 sigkdd:5 rigorous:1 attains:2 greedily:1 baseline:7 stg:1 inference:1 adversarial:1 dependent:1 el:1 hassid:1 accumulated:6 diminishing:2 her:4 bandit:39 koller:1 selects:3 interested:1 overall:1 classification:2 colt:2 aforementioned:1 priori:1 yahoo:4 development:1 plan:1 art:1 special:4 kempe:1 initialize:1 marginal:23 constrained:3 once:2 aware:1 beach:1 identical:2 icml:5 jon:2 future:2 papadimitriou:1 recommend:3 few:2 modern:1 randomly:2 javdani:1 simultaneously:1 comprehensive:1 delayed:1 maxj:3 argmax:1 maintain:2 interest:1 highly:1 khalid:1 evaluation:1 extreme:1 light:1 tj:6 ln3:1 indexed:1 euclidean:2 initialized:2 haifa:1 circle:1 theoretical:3 leskovec:2 uncertain:1 instance:4 column:2 earlier:2 boolean:1 konstantin:1 modeling:2 cover:5 yoav:1 maximization:14 tractability:1 cost:1 deviation:2 subset:19 entry:4 successful:2 dependency:1 varies:1 guillory:1 combined:1 st:1 broder:1 siam:1 probabilistic:2 off:1 receiving:1 enhance:1 synthesis:1 quickly:1 concrete:1 shervin:1 michael:1 w1:1 again:6 aaai:1 cesa:3 interactively:1 containing:1 choose:5 leveraged:1 arini:1 huang:1 yisong:2 ontext:3 corner:1 return:3 li:4 account:1 diversity:2 student:2 coefficient:2 satisfy:2 audibert:1 depends:3 piece:1 later:1 try:1 optimistic:1 kwk:2 characterizes:1 sup:1 start:1 competitive:1 maintains:1 carlos:4 avalanche:1 shai:1 yaron:3 zha:1 simon:1 inferning:1 contribution:1 minimize:3 oi:2 yves:1 variance:1 efficiently:1 maximized:2 gathered:1 identify:1 ashtiani:1 generalize:1 lkopf:1 bayesian:1 basically:2 lu:1 monitoring:1 bilmes:1 xoj:3 history:6 hamed:1 reach:2 influenced:4 monteleoni:1 messenger:1 wai:1 facebook:4 definition:1 against:1 naturally:2 associated:3 proof:1 lior:2 mi:1 gain:19 sampled:2 dataset:8 treatment:1 hsu:1 recall:3 reminder:1 knowledge:9 ut:1 emerges:1 hilbert:2 vetterli:1 andom:1 hassani:1 carefully:2 actually:1 auer:3 appears:1 originally:1 jair:1 supervised:1 bian:1 response:2 wei:4 qinshi:1 arranged:1 evaluated:1 formulation:1 dey:1 smola:1 uriel:1 langford:2 receives:2 horizontal:4 shahaf:1 cohn:1 rodriguez:1 lda:2 scientific:1 lei:1 seeger:1 usa:1 name:2 building:2 normalized:2 unbiased:1 y2:1 true:1 accumulative:1 hence:3 facility:1 illustrated:1 round:24 impute:1 game:1 elect:2 cosine:3 trying:2 complete:1 theoretic:1 l1:2 fj:3 balcan:3 recently:1 fi:9 srinivasa:1 charles:1 empirically:1 volume:4 thirdly:1 significant:1 refer:1 multiarmed:2 vec:1 smoothness:3 rd:3 unconstrained:1 dafna:1 pm:4 similarly:1 erc:1 mathematics:1 submodular:44 language:3 had:1 lihong:2 similarity:1 ashwinkumar:2 add:1 base:1 nicolo:3 feb:1 posterior:3 closest:1 showed:1 recent:1 zadimoghaddam:1 scenario:3 advertise:1 arbitrarily:1 life:1 yi:10 reynold:1 guestrin:5 seen:1 george:1 floor:1 vanchinathan:1 determine:1 paradigm:1 maximize:6 recommended:1 semi:3 multiple:5 full:2 sham:1 reduces:1 infer:1 cgp:2 smooth:1 characterized:1 long:1 lin:6 retrieval:1 niranjan:1 award:1 laplacian:4 va:1 prediction:1 variant:2 scalable:1 florina:3 heterogeneous:1 metric:1 iteration:1 kernel:37 represent:3 achieved:1 c1:2 receive:7 uj0:3 addition:2 krause:9 affecting:1 addressed:1 kossmann:1 singular:2 source:3 jian:1 crucial:1 sch:1 w2:1 webscope:3 yue:3 subject:7 recruited:1 member:3 jordan:1 call:3 near:4 leverage:2 yang:1 revealed:4 recommends:1 enough:1 variety:1 affect:1 nonstochastic:2 andreas:10 idea:2 regarding:2 avenue:1 knowing:1 tradeoff:2 inner:3 aviad:1 whether:2 six:1 veda:1 utility:16 song:1 peter:2 action:18 dramatically:1 generally:1 governs:1 informally:1 covered:2 amount:1 locally:1 simplest:1 http:2 schapire:3 estimated:3 dasgupta:2 siddhartha:1 group:6 key:1 four:2 gunnar:1 threshold:1 drawn:1 alina:1 diffusion:7 kept:1 asymptotically:1 monotone:4 fraction:3 sum:3 run:4 compete:1 uncertainty:6 powerful:1 soi:7 soda:1 arrive:3 family:1 reader:1 fran:1 decision:4 jaccard:6 submatrix:2 bound:17 ki:9 guaranteed:1 gomez:1 nan:1 yale:3 cheng:2 placement:2 precisely:1 constraint:1 personalized:7 encodes:2 dominated:1 nearby:1 kleinberg:2 performing:3 martin:2 xiaoming:1 department:2 influential:1 according:3 structured:1 combination:1 manning:1 feige:1 wi:5 kakade:1 evolves:1 making:3 s1:2 globecom:1 pr:1 karbasi:6 heart:1 taken:1 ln:1 previously:8 rtner:1 singer:3 know:1 upperconfidence:1 end:3 generalizes:2 operation:1 apply:3 observe:7 pierre:1 occurrence:1 anupam:1 encounter:7 knapsack:2 thomas:2 denotes:6 clustering:5 assumes:1 remaining:1 dirichlet:1 madison:1 tois:1 ghahramani:1 build:2 uj:10 approximating:2 avinatan:1 implied:1 objective:20 already:1 quantity:1 strategy:1 rt:2 interacts:1 bagnell:2 antoine:1 exhibit:1 affinity:1 nemhauser:1 subspace:1 separate:3 simulated:1 topic:6 valuation:3 extent:1 trivial:1 reason:1 senellart:1 assuming:2 index:1 kk:4 julian:1 minimizing:1 acquire:1 liang:1 robert:3 potentially:1 relate:1 stated:1 negative:2 design:2 reliably:1 collective:4 summarization:8 policy:7 unknown:13 bianchi:3 allowing:2 recommender:10 upper:4 observation:3 vertical:4 datasets:1 sm:33 finite:1 payoff:33 situation:2 d16ap00046:1 communication:1 y1:1 reproducing:3 ajit:1 arbitrary:1 rating:6 david:3 cast:1 pair:8 eature:3 chen1:1 nip:12 seeman:2 address:1 able:2 beyond:1 jure:2 below:1 pattern:1 encompasses:1 baharan:2 including:5 greatest:1 power:1 natural:3 circumvent:1 turning:1 arm:4 movie:21 axis:8 faced:1 prior:3 literature:2 discovery:2 l2:1 dislike:1 acknowledgement:1 text:1 wisconsin:1 freund:1 fully:2 ioannidis:1 sublinear:2 wolsey:1 allocation:1 grover:1 versus:2 foundation:1 krausea:1 degree:1 gather:1 xp:1 s0:1 article:11 principle:1 balancing:1 claire:1 row:1 pranjal:1 guillermo:1 placed:1 supported:1 free:12 arriving:4 infeasible:1 wireless:1 side:3 guide:1 institute:1 neighbor:3 taking:1 munos:1 distributed:1 feedback:10 evaluating:1 world:2 rich:1 dataset1:1 resides:1 collection:1 adaptive:6 preprocessing:1 qualitatively:1 historical:1 social:6 transaction:1 sj:9 nov:1 bernhard:1 keep:1 decides:1 sequentially:2 active:11 corpus:3 assumed:3 xi:1 thep:1 terrain:1 continuous:2 latent:1 quantifies:5 streeter:2 learn:1 ca:1 golovin:4 argmaxx:2 interact:2 du:1 investigated:1 domain:1 did:1 aistats:3 spread:2 main:2 dense:1 reuters:3 noise:5 arise:1 s2:1 arrival:5 paul:1 fig:12 andrei:1 martingale:2 tong:2 christos:1 experienced:1 obeying:1 stamp:2 vanish:1 jmlr:3 young:1 sui:6 theorem:4 down:1 bastien:1 list:3 gupta:1 exists:1 workshop:1 sequential:4 importance:1 drew:1 debadeepta:1 lifting:1 conditioned:1 horizon:3 margin:1 chen:7 gap:1 reedy:10 morteza:1 entropy:1 led:2 simply:2 likely:1 infinitely:2 bubeck:1 blogosphere:1 aditya:1 ordered:1 diversified:1 partially:1 recommendation:10 applies:1 springer:1 ch:1 mij:4 corresponds:1 environmental:1 relies:4 acm:10 goal:7 presentation:1 viewed:1 jeff:1 hard:3 hastagiri:1 included:1 specifically:3 except:6 uniformly:3 movielens:3 typical:1 total:10 called:4 isomorphic:1 sanjoy:2 svd:3 ya:7 shannon:1 ucb:39 select:6 formally:2 people:2 support:1 arises:1 alexander:1 ethz:1 evaluate:2 srinivas:2 |
6,212 | 662 | Mapping Between Neural and Physical
Activities of the Lobster Gastric Mill
Kenji Doya
Mary E. T. Boyle
Allen I. Selverston
Department of Biology
University of California, San Diego
La Jolla, CA 92093-0322
Abstract
A computer model of the musculoskeletal system of the lobster
gastric mill was constructed in order to provide a behavioral interpretation of the rhythmic patterns obtained from isolated stomatogastric ganglion. The model was based on Hill's muscle model
and quasi-static approximation of the skeletal dynamics and could
simulate the change of chewing patterns by the effect of neuromodulators.
1
THE STOMATOGASTRIC NERVOUS SYSTEM
The crustacean stomatogastric ganglion (STG) is a circuit of 30 neurons that controls rhythmic movement of the foregut. It is one of the best elucidated neural
circuits. All the neurons and the synaptic connections between them are identified and the effects of neuromodulators on the oscillation patterns and neuronal
characteristics have been extensively studied (Selverston and Moulins 1987, H arrisWarrick et al. 1992). However, STG's function as a controller of ingestive behavior
is not fully understood in part because of our poor understanding of the controlled
object: the musculoskeletal dynamics of the foregut. We constructed a mathematical model of the gastric mill, three teeth in the stomach, in order to predict motor
patterns from the neural oscillation patterns which are recorded from the isolated
ganglion.
The animal we used was the Californian spiny lobster (Panulirus interruptus), which
913
914
Doya, Boyle, and Selverston
(a)
medial tooth _ _- -
esophagus
flexible endoscope
? Inhibitory
? Functional Inhibitory
6 Excitatory
, FWlCionai Excitalory
JVV\...
Electronic
Figure 1: The lobster stomatogastric system. (a) Cross section of the foregut
(objects are not to scale). (b) The gastric circuit.
is available locally. The stomatogastric nervous system controls four parts of the
foregut: esophagus, cardiac sac (stomach), gastric mill, and pylorus (entrance to
the intestine) (Figure l.a). The gastric mill is composed of one medial tooth and
two lateral teeth. These grind large chunks of foods (mollusks, algae, crabs, sea
urchins, etc.) into smaller pieces and mix them with digestive fluids. The chewing
period ranges from 5 to 10 seconds. Several different chewing patterns have been
analyzed using an endoscope (Heinzel 1988a, Boyle et al. 1990). Figure 2 shows
two of the typical chewing patterns: "cut and grind" and "cut and squeeze" .
The STG is located in the opthalmic artery which runs from the heart to brain over
the dorsal surface of the stomach. When it is taken out with two other ganglia (the
esophageal ganglion and the commissural ganglion), it can still generate rhythmic
motor outputs. This isolated preparation is ideal for studying the mechanism of
rhythmic pattern generation by a neural circuit.
From pairwise stimulus and
response of the neurons, the map of synaptic connections has been established.
Figure 1 (b) shows a subset of the STG circuit which controls the motion of the
gastric mill. It consists of 11 neurons of 7 types. GM and DG neurons control the
medial tooth and LPG, MG, and LG neurons control the lateral teeth. A question of
interest is how this simple neural network is utilized to control the various movement
patterns of the gastric mill, which is a fairly complex musculoskeletal system.
The oscillation pattern of the isolated ganglion can be modulated by perfusing it
with of several neuromodulators, e.g. proctolin, octopamine (Heinzel and Selverston 1988), CCK (Turrigiano 1990), and pilocarpine (Elson and Selverston 1992).
However, the behavioral interpretation of these different activity patterns is not
well understood. The gastric mill is composed of 7 ossicles (small bones) which is
loosely suspended by more than 20 muscles and connective tissues. That makes it is
very difficult to intuitively estimate the effect of the change of neural firing patterns
in terms of the teeth movement. Therefore we, decided to construct a quantitative
model of the musculoskeletal system of the gastric mill.
Mapping Between Neural and Physical Activities of the Lobster Gastric Mill
(a)
(b)
Figure 2: Typical chewing patterns of the gastric mill. (a) cut and grind. (b) cut
and squeeze.
2
PHYSIOLOGICAL EXPERIMENTS
In order to design a model and determine its parameters, we performed anatomical
and physiological experiments described below.
Anatomical experiments: The carapace and the skin above the stomach mill
was removed to expose a dorsal view of the ossicles and the muscles which control
the gastric mill. Usually, the gastric mill was quiescent without any stimuli. The
positions of the ossicles and the lengths of the muscles at the resting state was
measured. After the behavioral experiments mentioned below, the gastric mill was
taken out and the size of the ossicles and the positions of the attachment points of
the muscles were measured.
Behavioral experiments: With the carapace removed and the gastric mill exposed, one video camera was used to record the movement of the ossicles and the
muscles. Another video camera attached to a flexible endoscope was used to record
the motion of the teeth from inside the stomach. In the resting state, muscles were
stimulated by a wire electrode to determine the behavioral effects. In order to induce chewing, neuromodulators such as proctolin and pilocarpine were injected into
the artery in which STG is located.
Single muscle experiments: The gm!, the largest of the gastric mill muscles,
was used to estimate the parameters of the muscle model mentioned below. It was
removed without disrupting the carapace or ossicle attachment points and fixed to a
tension measurement apparatus. The nerve fiber aln that innervates gmt was stimulated using a suction electrode. The time course of isometric tension was recorded
at different muscle lengths and stimulus frequencies. The parameters obtained from
the gmt muscle experiment were applied to other muscles by considering their relative length and thickness.
915
916
Doya, Boyle, and Selverston
(a)
contraction element (CE)
serial elasticity (SE)
parallel elasticity (PE)
(d)
(c)
fO
fs
f max
o
o
leo
Figure 3: The Hill-based muscle model.
3
MODELING THE MUSCULOSKELETAL SYSTEM
3.1
MUSCULAR DYNAMICS
There are many ways to model muscles. In the simplest models, the tension or the
length of a muscle is regarded as an instantaneous function of the spike frequency
of the motor nerve. In some engineering approaches, a muscle is considered as
a spring whose resting length and stiffness are modulated by the nervous input
(Hogan 1984). Since these models are a linear static approximation of the nonlinear
dynamical characteristics of muscles, their parameters must be changed to simulate
different motor tasks (Winters90). Molecular models (Zahalak 1990), which are
based on the binding mechanisms of actin and myosin fibers, can explain the widest
range of muscular characteristics found in physiological experiments. However,
these complex models have many parameters which are difficult to estimate.
The model we employed was a nonlinear macroscopic model based on A. V. Hill's
formulation (Hill 1938, Winters 1990). The model is composed of a contractile
element (CE), a serial elasticity (SE), and a parallel elasticity (PE) (Figure 3.a).
This model is based on empirical data about nonlinear characteristics of muscles
and its parameters can be determined by physiological experiments.
The output force Ie of the CE is a function of its length Ie and its contraction speed
Ve
-dle/dt (Figure 3.b)
=
Ve ?:: 0 (contraction),
Ve < 0 (extension),
=
(1)
where 10 is the isometric output force (at Ve 0) and Vo is the maximal contraction
velocity. The parameters of the I-v curve were a = 0.25 and f3 0.3. The isometric
force 10 was given as the function of CE length Ie and the activation level a(t) of
=
Mapping Between Neural and Physical Activities of the Lobster Gastric Mill
the muscle (Figure 3.c)
fo(l"a(t)) = {
~m.Z!~, (f.;)' (f.; - r) a(t) o < Ie < 'Y,
(2)
otherwise,
where leo is the resting length of the CE and 'Y
= 1.5.
The SE was modeled as an exponential spring (Figure 3.d)
I${I$) = { okl(exp[k2l'~~'Q] -1)
1$ ~ 1$0,
1$ < 1$0,
(3)
where 1$ is the output force, 1$0 is the resting length, and kl and k2 are stiffness
parameters. The PE was supposed to have the same exponential elasticity (3).
In the simulations, the CE length Ie was taken as the state variable. The total
muscle length 1m = Ie + 1$ is given by the skeletal model and the muscle activation
a(t) is given by the the activation dynamics described below. The SE length is
1m - Ie and then the output force I${I$)
Ie + Ip
1m is given
given from 1$
by (3). The contraction velocity Ve = -~ is derived from the inverse of (1) at
Ie = I${I$) - Ip(le) and then integrated to update the CE length Ie.
=
=
=
The activation level a(t) of a muscle is determined by the free calcium concentration
in muscle fibers. Since we don't have enough data about the calcium dynamics in
muscle cells, the activation dynamics was crudely approximated by the following
equations.
da(t)
Ta-;{t = -a(t) + e(t),
and
de(t)
Te~ = -e(t)
+ n(t)2,
(4)
where n(t) is the normalized firing frequency of the nerve input and e(t) is the electric activity of the muscle fibers. The nonlinearity in the nervous input represents
strong facilitation of the postsynaptic potential (Govind and Lingle 1987).
We incorporated seven of the gastric mill muscles: gml, gm2, gm3a, gm3c, gm4,
gm6b, and gm9a (Maynard and Dando 1974). The muscles gml, gm2, gm3a, and
gm3c are extrinsic muscles that have one end attached to the carapace and gm4,
gm6b, and gm9a are intrinsic muscles both ends of which are attached of the ossicles.
Three connective tissues were also incorporated and regarded as muscles without
contraction elements. See Figure 4 for the attachment of these muscles and tissues
to the ossicles.
3.2
SKELETAL DYNAMICS
The medial tooth was modeled as three rigid pieces PI, P2 and P3 . PI is the base of
the medial tooth. P2 is the main body of the medial tooth. P3 forms the cusp and
the V-shaped lever on the dorsal side. The lateral tooth was modeled as two rigid
pieces P4 and Ps . P4 is a L-shaped plate with a cusp at the angle and is connected
to P3 at the dorsal end. Ps is a rod that is connected to P4 near the root of the
cusp (Figure 4).
We assumed that the motion is symmetric with respect to the midline. Therefore
the motion of the medial tooth was two-dimensional and only the left one of the
917
918
Doya, Boyle, and Se1verslon
gm3c
~':-;--...J.........
gm9a
z
30
y
Figure 4: The design of the gastric mill model. Ossicle PI stands for the ossicles
I and II, P2 for VII, P3 for VI, P4 for III, IV, and V, Ps for XIV in the standard
description by Maynard and Dando (1974).
two lateral teeth was considered. The coordinate system was taken so that x-axis
points to the left, y-axis backward, and z-axis upward. The rotation angles of the
ossicles around x, y, and z axes ware represented as 0, <P, and 1/J respectively. The
configuration of the ossicles was determined by a 10 dimensional vector
e = (YO,zO,OI,02,03,04,<P4,1/J4,OS,<P5),
(5)
where (YO,zo) represents the position of the joint between PI and P2 and (0 1 ,02,03)
represents the rotation angle of PI, P2 and P3 in the y-z plane. The rotation angles
of P4 and P5 were represented as (0 4, <P4, 1/J4) and (Os, 4>s) respectively. P5 has only
two degrees of rotation freedom since it is regarded as a rod.
We employed a quasi-static approximation. The configuration of the ossicles e
was determined by the static balance of force. Now let Lm and Fm be the vectors
of the muscle lengths and forces. Then the balance of the generalized forces in the
space (force for translation and torque for rotation) is given by
e
Tm(e, Fm) + Te = 0,
(6)
where T m and Te represent the generalized forces from muscles and external loads.
The muscle force in the e space is given by
Tm(e, Fm) = J(e)T Fm,
(7)
where J(e) = 8L m /8e is the Jacobian matrix of the mapping e .- Lm determined
by the ossicle kinematics and the muscle attachment. Since it is very difficult to
obtain a closed form solution of (6), we used a gradient descent equation
de
dt = -c:(Tm(e, Fm) + Te) = -c:(J(e)T Fm + Te)
(8)
Mapping Between Neural and Physical Activities of the Lobster Gastric Mill
(a) t=O.
t=2.
(b)
t=1.5
t=O.
t=4.
t=3.
t=6.
t=4.5
Figure 5: Chewing patterns predicted from oscillation patterns of isolated STG . (a)
spontaneous pattern. (b) proctolin induced pattern.
to find the approximate solution of 0(t). This is equivalent to assuming a viscosity
term c- 1 d0ldt in the motion equation.
4
SIMULATION RESULTS
The musculoskeletal model is a 17-th order differential equation system and was
integrated by Runge-Kutta method with a time step 1ms. Figure 5 shows examples
of motion patterns predicted by the model. The motoneuron output of spontaneous
oscillation of the isolated ganglion was used in (a) and the output under the effect
of proctolin was used in (b). It has been reported in previous behavioral studies
(Heinzel 1988b) that the dose of proctolin typically evokes "cut and grind" chewing
pattern. The trajectory (b) predicted from the proctolin induced rhythm has a
larger forward movement of the medial tooth while the lateral teeth are closed,
which qualitatively agrees with the behavioral data.
5
DISCUSSION
The motor pattern generated by the model is considerably different from the chewing patterns observed in the intact animal using an endoscope. This is partly
because of crude assumptions in model construction and errors in parameter estimation. However, this difference may also be due to the lack of sensory feedback in
the isolated preparation . The future subject of this project is to refine the model so
that we can reliably predict the motion from the neural outputs and to combine it
with models of the gastric network (Rowat and Selverston, submitted) and sensory
receptors. This will enable us to study how a biological control system integrates
central pattern generation and sensory feedback.
919
920
Doya, Boyle, and Selverston
Acknowledgements
We thank Mike Beauchamp for the gml muscle data. This work was supported by
the grant from Office of Naval Research NOOOI4-91-J-1720.
References
Boyle, M. E. T., Turrigiano, G . G., and Selverston, A. 1. 1990. An endoscopic analysis of gastric mill movements produced by the peptide cholecystokinin. Society
for Neuroscience Abstracts 16, 724.
Elson, R. C. and Selverston, A. 1. 1992. Mechanisms of gastric rhythm generation
in the isolated stomatogastric ganglion of spiny lobsters: Bursting pacemaker
potentials, synaptic interactions and muscarinic modulation. Journal of Neurophysiology 68, 890-907.
Govind, C. K. and Lingle, C. J. 1987. Neuromuscular organization and pharmacology. In Selverston, A. 1. and Moulins, M., editors, The Crustacean Stomatogastric
System, pages 31-48. Springer-Verlag, Berlin.
Harris-Warrick, R. M., Marder, E., Selverston, A. 1., and Moulins, M. 1992. Dynamic Biological Networks - The Stomatogastric Nervous System. MIT Press,
Cambridge, MA.
Heinzel, H. G. 1988. Gastric mill activity in the lobster. I: Spontaneous modes of
chewing. Journal of Neurophysiology 59, 528-550.
Heinzel, H. G. 1988. Gastric mill activity in the lobster. II: Proctolin and octopamine initiate and modulate chewing. Journal of Neurophysiology 59, 551565.
Heinzel, H. G. and Selverston, A. 1. 1988. Gastric mill activity in the lobster.
III : Effects of proctolin on the isolated central pattern generator. Journal of
Neurophysiology 59, 566-585.
Hill, A. V. 1938. The heat of shortening and the dynamic constants of muscle .
Proceedings of the Royal Sciety of London, Series B 126, 136-195 .
Hogan, N. 1984. Adaptive control of mechanical impedance by coactivation of
antagonist muscles. IEEE Transactions on Automatic Control 29, 681-690.
Maynard, D. M. and Dando, M. R. 1974. The structure ofthe stomatogastric neuromuscular system in callinectes sapidus, homarus americanus and panulirus argus
(decapoda crustacea). Philosophical Transactions of Royal Society of London,
Biology 268, 161- 220.
Rowat, P. F. and Selverston, A. 1. Modeling the gastric mill central pattern generator of the lobster with a relaxation-oscillator network. submitted.
Selverston, A. 1. and Moulins, M. 1987. The Crustacean Stomatogastric System.
Springer-Verlag, New York, NY.
Turrigiano, G . G. and Selverston, A. 1. 1990. A cholecystokinin-like hormone activates a feeding-related neural circuit in lobster . Nature 344, 866-868 .
Winters, J. M. 1990. Hill-based muscle models : A systems engineering perspective.
In Winters, J. M. and Woo, S. 1.- Y., editors, Multiplie Muscle Systems: Biomechanics and Movement Organization, chapter 5, pages 69-93. Springer-Verlag,
New York, NY.
Zahalak, G. I. 1990. Modeling muscle mechanics (and energetics). In Winters,
J. M. and Woo, S. L.-Y., editors, Multiplie Muscle Systems: Biomechanics and
Movement Organization, chapter 1, pages 1-23. Springer-Verlag, New York, NY.
| 662 |@word neurophysiology:4 simulation:2 contraction:6 configuration:2 series:1 activation:5 must:1 entrance:1 motor:5 medial:8 update:1 pacemaker:1 nervous:5 plane:1 record:2 beauchamp:1 digestive:1 mathematical:1 constructed:2 differential:1 consists:1 combine:1 behavioral:7 inside:1 pairwise:1 behavior:1 mechanic:1 brain:1 torque:1 food:1 considering:1 project:1 circuit:6 connective:2 aln:1 selverston:16 quantitative:1 k2:1 control:10 grant:1 understood:2 engineering:2 apparatus:1 receptor:1 ware:1 firing:2 modulation:1 xiv:1 studied:1 bursting:1 range:2 coactivation:1 decided:1 camera:2 empirical:1 induce:1 equivalent:1 map:1 boyle:7 foregut:4 stomatogastric:10 sac:1 regarded:3 facilitation:1 coordinate:1 diego:1 gm:2 spontaneous:3 construction:1 element:3 velocity:2 approximated:1 located:2 utilized:1 cut:5 observed:1 mike:1 p5:3 connected:2 innervates:1 movement:8 removed:3 mentioned:2 dynamic:9 hogan:2 exposed:1 joint:1 various:1 fiber:4 represented:2 chapter:2 leo:2 zo:2 heat:1 london:2 whose:1 larger:1 otherwise:1 ip:2 runge:1 mg:1 turrigiano:3 interaction:1 maximal:1 p4:7 supposed:1 description:1 artery:2 squeeze:2 electrode:2 p:3 sea:1 object:2 measured:2 p2:5 strong:1 kenji:1 predicted:3 musculoskeletal:6 cusp:3 enable:1 govind:2 feeding:1 biological:2 extension:1 gm4:2 crab:1 considered:2 around:1 exp:1 mapping:5 predict:2 lm:2 cck:1 estimation:1 integrates:1 expose:1 peptide:1 grind:4 largest:1 agrees:1 mit:1 octopamine:2 activates:1 moulins:4 office:1 derived:1 ax:1 yo:2 naval:1 stg:6 rigid:2 integrated:2 typically:1 quasi:2 upward:1 flexible:2 warrick:1 animal:2 fairly:1 construct:1 f3:1 shaped:2 biology:2 represents:3 future:1 stimulus:3 winter:4 composed:3 dg:1 ve:5 midline:1 panulirus:2 freedom:1 organization:3 interest:1 dle:1 analyzed:1 elasticity:5 iv:1 loosely:1 isolated:9 dose:1 modeling:3 subset:1 reported:1 thickness:1 considerably:1 chunk:1 ie:10 lever:1 neuromodulators:4 recorded:2 central:3 external:1 potential:2 de:2 vi:1 piece:3 performed:1 bone:1 view:1 root:1 closed:2 parallel:2 oi:1 elson:2 characteristic:4 myosin:1 ofthe:1 produced:1 trajectory:1 tooth:9 tissue:3 submitted:2 explain:1 fo:2 synaptic:3 lobster:13 frequency:3 commissural:1 static:4 stomach:5 crustacean:3 noooi4:1 nerve:3 ta:1 dt:2 isometric:3 tension:3 response:1 formulation:1 crudely:1 nonlinear:3 o:2 lack:1 maynard:3 mode:1 mary:1 effect:6 normalized:1 symmetric:1 rhythm:2 m:1 generalized:2 plate:1 hill:6 antagonist:1 disrupting:1 vo:1 allen:1 motion:7 instantaneous:1 rotation:5 functional:1 physical:4 attached:3 interpretation:2 resting:5 measurement:1 cambridge:1 automatic:1 nonlinearity:1 gmt:2 j4:2 surface:1 etc:1 base:1 perspective:1 jolla:1 verlag:4 suspended:1 muscle:43 motoneuron:1 employed:2 determine:2 period:1 ii:2 mix:1 hormone:1 cross:1 biomechanics:2 serial:2 molecular:1 gm2:2 energetics:1 controlled:1 controller:1 represent:1 cell:1 neuromuscular:2 macroscopic:1 proctolin:8 induced:2 subject:1 near:1 ideal:1 iii:2 enough:1 identified:1 fm:6 tm:3 rod:2 f:1 york:3 se:4 viscosity:1 shortening:1 extensively:1 locally:1 simplest:1 generate:1 inhibitory:2 neuroscience:1 extrinsic:1 anatomical:2 skeletal:3 four:1 ce:7 backward:1 relaxation:1 run:1 inverse:1 angle:4 injected:1 evokes:1 electronic:1 doya:5 p3:5 oscillation:5 refine:1 activity:9 elucidated:1 marder:1 simulate:2 speed:1 spring:2 department:1 heinzel:6 poor:1 smaller:1 cardiac:1 postsynaptic:1 spiny:2 intuitively:1 muscarinic:1 k2l:1 heart:1 taken:4 equation:4 kinematics:1 mechanism:3 initiate:1 jvv:1 end:3 studying:1 available:1 stiffness:2 widest:1 society:2 skin:1 question:1 spike:1 concentration:1 gradient:1 kutta:1 thank:1 lateral:5 berlin:1 seven:1 assuming:1 length:14 modeled:3 mollusk:1 balance:2 lg:1 difficult:3 fluid:1 esophagus:2 design:2 reliably:1 calcium:2 neuron:6 wire:1 descent:1 gastric:28 incorporated:2 rowat:2 mechanical:1 kl:1 connection:2 philosophical:1 california:1 established:1 below:4 pattern:24 usually:1 dynamical:1 max:1 royal:2 video:2 force:11 pilocarpine:2 attachment:4 axis:3 woo:2 understanding:1 acknowledgement:1 relative:1 fully:1 generation:3 okl:1 generator:2 degree:1 teeth:7 editor:3 pi:5 translation:1 excitatory:1 course:1 changed:1 supported:1 free:1 actin:1 side:1 rhythmic:4 intestine:1 curve:1 feedback:2 stand:1 sensory:3 forward:1 qualitatively:1 adaptive:1 san:1 transaction:2 approximate:1 assumed:1 quiescent:1 urchin:1 don:1 stimulated:2 impedance:1 suction:1 nature:1 ca:1 complex:2 electric:1 da:1 main:1 pharmacology:1 body:1 neuronal:1 ny:3 position:3 exponential:2 crude:1 pe:3 jacobian:1 lpg:1 load:1 physiological:4 intrinsic:1 te:5 vii:1 mill:26 ganglion:9 gml:3 binding:1 springer:4 harris:1 ma:1 modulate:1 oscillator:1 change:2 typical:2 muscular:2 determined:5 total:1 partly:1 la:1 intact:1 modulated:2 dorsal:4 preparation:2 |
6,213 | 6,620 | Learning to See Physics via Visual De-animation
Jiajun Wu
MIT CSAIL
Erika Lu
University of Oxford
William T. Freeman
MIT CSAIL, Google Research
Pushmeet Kohli
DeepMind
Joshua B. Tenenbaum
MIT CSAIL
Abstract
We introduce a paradigm for understanding physical scenes without human annotations. At the core of our system is a physical world representation that is
first recovered by a perception module and then utilized by physics and graphics
engines. During training, the perception module and the generative models learn
by visual de-animation ? interpreting and reconstructing the visual information
stream. During testing, the system first recovers the physical world state, and then
uses the generative models for reasoning and future prediction.
Even more so than forward simulation, inverting a physics or graphics engine is
a computationally hard problem; we overcome this challenge by using a convolutional inversion network. Our system quickly recognizes the physical world
state from appearance and motion cues, and has the flexibility to incorporate both
differentiable and non-differentiable physics and graphics engines. We evaluate our
system on both synthetic and real datasets involving multiple physical scenes, and
demonstrate that our system performs well on both physical state estimation and
reasoning problems. We further show that the knowledge learned on the synthetic
dataset generalizes to constrained real images.
1
Introduction
Inspired by human abilities, we wish to develop machine systems that understand scenes. Scene
understanding has multiple defining characteristics which break down broadly into two features. First,
human scene understanding is rich. Scene understanding is physical, predictive, and causal: rather
than simply knowing what is where, one can also predict what may happen next, or what actions
one can take, based on the physics afforded by the objects, their properties, and relations. These
predictions, hypotheticals, and counterfactuals are probabilistic, integrating uncertainty as to what is
more or less likely to occur. Second, human scene understanding is fast. Most of the computation has
to happen in a single, feedforward, bottom-up pass.
There have been many systems proposed recently to tackle these challenges, but existing systems
have architectural features that allow them to address one of these features but not the other. Typical
approaches based on inverting graphics engines and physics simulators [Kulkarni et al., 2015b]
achieve richness at the expense of speed. Conversely, neural networks such as PhysNet [Lerer et al.,
2016] are fast, but their ability to generalize to rich physical predictions is limited.
We propose a new approach to combine the best of both. Our overall framework for representation is
based on graphics and physics engines, where graphics is run in reverse to build the initial physical
scene representation, and physics is then run forward to imagine what will happen next or what can
be done. Graphics can also be run in the forward direction to visualize the outputs of the physics
simulation as images of what we expect to see in the future, or under different viewing conditions.
Rather than use traditional, often slow inverse graphics methods [Kulkarni et al., 2015b], we learn to
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
physical world
physical world
visual data
visual data
Figure 1: Visual de-animation ? we would like to recover the physical world representation behind
the visual input, and combine it with generative physics simulation and rendering engines.
invert the graphics engine efficiently using convolutional nets. Specifically, we use deep learning to
train recognition models on the objects in our world for object detection, structure and viewpoint
estimation, and physical property estimation. Bootstrapping from these predictions, we then infer the
remaining scene properties through inference via forward simulation of the physics engine.
Without human supervision, our system learns by visual de-animation: interpreting and reconstructing
visual input. We show the problem formulation in Figure 1. The simulation and rendering engines in
the framework force the perception module to extract physical world states that best explain the data.
As the physical world states are inputs to physics and graphics engines, we simultaneously obtain an
interpretable, disentangled, and compact physical scene representation.
Our framework is flexible and adaptable to a number of graphics and physics engines. We present
model variants that use neural, differentiable physics engines [Chang et al., 2017], and variants that
use traditional physics engines, which are more mature but non-differentiable [Coumans, 2010]. We
also explore various graphics engines operating at different levels, ranging from mid-level cues such
as object velocity, to pixel-level rendering of images.
We demonstrate our system on real and synthetic datasets across multiple domains: synthetic billiard
videos [Fragkiadaki et al., 2016], in which balls have varied physical properties, real billiard videos
from the web, and real images of block towers from Facebook AI Research [Lerer et al., 2016].
Our contributions are three-fold. First, we propose a novel generative pipeline for physical scene
understanding, and demonstrate its flexibility by incorporating various graphics and physics engines.
Second, we introduce the problem of visual de-animation ? learning rich scene representations
without supervision by interpreting and reconstructing visual input. Third, we show that our system
performs well across multiple scenarios and on both synthetic and constrained real videos.
2
Related Work
Physical scene understanding has attracted increasing attention in recent years [Gupta et al., 2010,
Jia et al., 2015, Lerer et al., 2016, Zheng et al., 2015, Battaglia et al., 2013, Mottaghi et al., 2016b,
Fragkiadaki et al., 2016, Battaglia et al., 2016, Mottaghi et al., 2016a, Chang et al., 2017, Agrawal
et al., 2016, Pinto et al., 2016, Finn et al., 2016, Hamrick et al., 2017, Ehrhardt et al., 2017, Shao et al.,
2014, Zhang et al., 2016]. Researchers have attempted to go beyond the traditional goals of high-level
computer vision, inferring ?what is where?, to capture the physics needed to predict the immediate
future of dynamic scenes, and to infer the actions an agent should take to achieve a goal. Most of these
efforts do not attempt to learn physical object representations from raw observations. Some systems
emphasize learning from pixels but without an explicitly object-based representation [Lerer et al.,
2016, Mottaghi et al., 2016b, Fragkiadaki et al., 2016, Agrawal et al., 2016, Pinto et al., 2016, Li et al.,
2017], which makes generalization challenging. Others learn a flexible model of the dynamics of
object interactions, but assume a decomposition of the scene into physical objects and their properties
rather than learning directly from images [Chang et al., 2017, Battaglia et al., 2016].
2
Object
proposal
Physical
object state
Object
proposal
Physical
object state
(b) Physical world
representation
(II) Physics engine (simulation)
(c) Appearance cues
(III) Graphics engine (rendering)
NMS
(I) Perception module
(d) Likelihood
(a) Input
(e) Output
Figure 2: Our visual de-animation (VDA) model contains three major components: a convolutional
perception module (I), a physics engine (II), and a graphics engine (III). The perception module
efficiently inverts the graphics engine by inferring the physical object state for each segment proposal
in input (a), and combines them to obtain a physical world representation (b). The generative physics
and graphics engines then run forward to reconstruct the visual data (e). See Section 3 for details.
There have been some works that aim to estimate physical object properties [Wu et al., 2016, 2015,
Denil et al., 2017]. Wu et al. [2015] explored an analysis-by-synthesis approach that is easily
generalizable, but less efficient. Their framework also lacked a perception module. Denil et al.
[2017] instead proposed a reinforcement learning approach. These approaches, however, assumed
strong priors of the scene, and approximate object shapes with primitives. Wu et al. [2016] used
a feed-forward network for physical property estimation without assuming prior knowledge of the
environment, but the constrained setup did not allow interactions between multiple objects. By
incorporating physics and graphics engines, our approach can jointly learn the perception module
and physical model, optionally in a Helmholtz machine style [Hinton et al., 1995], and recover an
explicit physical object representation in a range of scenarios.
Another line of related work is on future state prediction in either image pixels [Xue et al., 2016,
Mathieu et al., 2016] or object trajectories [Kitani et al., 2017, Walker et al., 2015]. There has also
been abundant research making use of physical models for human or scene tracking [Salzmann
and Urtasun, 2011, Kyriazis and Argyros, 2013, Vondrak et al., 2013, Brubaker et al., 2009]. Our
model builds upon and extends these ideas by jointly modeling an approximate physics engine and a
perceptual module, with wide applications including, but not limited to, future prediction.
Our framework also relates to the field of ?vision as inverse graphics? [Zhu and Mumford, 2007,
Yuille and Kersten, 2006, Bai et al., 2012]. Connected to but different from traditional analysisby-synthesis approaches, recent works explored using deep neural networks to efficiently explain
an object [Kulkarni et al., 2015a, Rezende et al., 2016], or a scene with multiple objects [Ba et al.,
2015, Huang and Murphy, 2015, Eslami et al., 2016]. In particular, Wu et al. [2017] proposed ?scene
de-rendering?, building an object-based, structured representation from a static image. Our work
incorporates inverse graphics with simulation engines for physical scene understanding and scene
dynamics modeling.
3
Visual De-animation
Our visual de-animation (VDA) model consists of an efficient inverse graphics component to build
the initial physical world representation from visual input, a physics engine for physical reasoning of
the scene, and a graphics engine for rendering videos. We show the framework in Figure 2. In this
section, we first present an overview of the system, and then describe each component in detail.
3.1
Overview
The first component of our system is an approximate inverse graphics module for physical object and
scene understanding, as shown in Figure 2-I. Specifically, the system sequentially computes object
proposals, recognizes objects and estimates their physical state, and recovers the scene layout.
3
The second component of our system is a physics engine, which uses the physical scene representation
recovered by the inverse graphics module to simulate future dynamics of the environment (Figure 2II). Our system adapts to both neural, differentiable simulators, which can be jointly trained with the
perception module, and rigid-body, non-differentiable simulators, which can be incorporated using
methods such as REINFORCE [Williams, 1992].
The third component of our framework is a graphics engine (Figure 2-III), which takes the scene
representations from the physics engine and re-renders the video at various levels (e.g. optical flow,
raw pixel). The graphics engine may need additional appearance cues such as object shape or color
(Figure 2c). Here, we approximate them using simple heuristics, as they are not a focus of our paper.
There is a tradeoff between various rendering levels: while pixel-level reconstruction captures details
of the scene, rendering at a more abstract level (e.g. silhouettes) may better generalize. We then use a
likelihood function (Figure 2d) to evaluate the difference between synthesized and observed signals,
and compute gradients or rewards for differentiable and non-differentiable systems, respectively.
Our model combines efficient and powerful deep networks for recognition with rich simulation
engines for forward prediction. This provides us two major advantages over existing methods:
first, simulation engines take an interpretable representation of the physical world, and can thus
easily generalize and supply rich physical predictions; second, the model learns by explaining the
observations ? it can be trained in a self-supervised manner without requiring human annotations.
3.2
Physical Object and Scene Modeling
We now discuss each component in detail, starting with the perception module.
Object proposal generation Given one or a few frames (Figure 2a), we first generate a number of
object proposals. The masked images are then used as input to the following stages of the pipeline.
Physical object state estimation For each segment proposal, we use a convolutional network to
recognize the physical state of the object, which consists of intrinsic properties such as shape, mass,
and friction, as well as extrinsic properties such as 3D position and pose. The input to the network is
the masked image of the proposal, and the output is an interpretable vector for its physical state.
Physical world reconstruction Given objects? physical states, we first apply non-maximum suppression to remove object duplicates, and then reconstruct the physical world according to object
states. The physical world representation (Figure 2b) will be employed by the physics and graphics
engines for simulation and rendering.
3.3
Physical Simulation and Prediction
The two types of physics engines we explore in this paper include a neural, differentiable physics
engine and a standard rigid-body simulation engine.
Neural physics engines The neural physics engine is an extension of the recent work from Chang
et al. [2017], which simulates scene dynamics by taking object mass, position, and velocity. We extend
their framework to model object friction in our experiments on billiard table videos. Though basic,
the neural physics engine is differentiable, and thus can be end-to-end trained with our perception
module to explain videos. Please refer to Chang et al. [2017] for details of the neural physics engine.
Rigid body simulation engines There exist rather mature, rigid-body physics simulation engines,
e.g. Bullet [Coumans, 2010]. Such physics engines are much more powerful, but non-differentiable.
In our experiments on block towers, we used a non-differentiable simulator with multi-sample
REINFORCE [Rezende et al., 2016, Mnih and Rezende, 2016] for joint training.
3.4
Re-rendering with a Graphics Engine
In this work, we consider two graphics engines operating at different levels: for the billiard table
scenario, we use a renderer that takes the output of a physics engine and generates pixel-level
rendering; for block towers, we use one that computes only object silhouettes.
4
(a) shared appearance,
shared physics
(b) varied appearance,
varied physics
(c) shared appearance,
varied physics
Figure 3: The three settings of our synthetic billiard videos: (a) balls have the same appearance and
physical properties, where the system learns to discover them and simulate the dynamics; (b) balls
have the same appearance but different physics, and the system learns their physics from motion; (c)
balls have varied appearance and physics, and the system learns to associate appearance cues with
underlying object states, even from a single image.
Input (red)
and
ground truth
Reconstruction
and prediction
Input (red)
and
ground truth
Reconstruction
and prediction
Frame t-2
Frame t
Frame t+2
Frame t+5
Frame t+10
Frame t-2
Frame t
Frame t+2
Frame t+5
Frame t+10
Figure 4: Results on the billiard videos, comparing ground truth videos with our predictions. We
show two of three input frames (in red) due to space constraints. Left: balls share appearance and
physics (I), where our framework learns to discover objects and simulate scene dynamics. Top right:
balls have different appearance and physics (II), where our model learns to associate appearance with
physics and simulate collisions. It learns that the green ball should move further than the heavier
blue ball after the collision. Bottom right: balls share appearance but have different frictions (III),
where our model learns to associate motion with friction. It realizes from three input frames that the
right-most ball in the first frame has a large friction coefficient and will stop before the other balls.
4
Evaluation
We evaluate variants of our frameworks in three scenarios: synthetic billiard videos, real billiard
videos, and block towers. We also test how models trained on synthetic data generalize to real cases.
4.1
Billiard Tables: A Motivating Example
We begin with synthetic billiard videos to explore end-to-end learning of the perceptual module along
with differentiable simulation engines. We explore how our framework learns the physical object
state (position, velocity, mass, and friction) from its appearance and/or motion.
Data For the billiard table scenario, we generate data using the released code from Fragkiadaki
et al. [2016]. We updated the code to allow balls of different mass and friction. We used the billiard
table scenario as an initial exploration of whether our models can learn to associate visual object
appearance and motion with physical properties. As shown in Figure 3, we generated three subsets,
in which balls may have shared or differing appearance (color), and physical properties. For each
case, we generated 9,000 videos for training and 200 for testing.
(I) Shared appearance and physics (Figure 3a): balls all have the same appearance and the same
physical properties. This basic setup evaluates whether we can jointly learn an object (ball) discoverer
and a physics engine for scene dynamics.
5
Datasets
Recon.
Methods
Appear. Physics
Pixel MSE
Position Prediction (Abs)
1st
5th
Velocity Prediction (Abs)
10th
20th
1st
5th
10th
20th
8.45
2.65
2.55
Same
Same
Baseline
VDA (init)
VDA (full)
0.046
0.046
0.044
4.58 18.20
3.46 6.61
3.25 6.52
46.06
12.76
12.34
119.97
26.10
25.55
2.95
1.41
1.37
5.63
1.97
1.87
7.32
2.34
2.22
Diff
Diff.
Baseline
VDA (init)
VDA (full)
0.058
0.058
0.055
6.57 26.38
3.82 8.92
3.55 8.58
70.47
17.09
16.33
180.04
34.65
32.97
3.78
1.65
1.64
7.62
2.27
2.20
10.51 12.02
3.02 3.21
2.89 3.05
Same
Diff.
Baseline
VDA (init)
VDA (full)
0.038
0.038
0.035
9.58 79.78 143.67 202.56 12.37 23.42 25.08 23.98
6.06 19.75 34.24 46.40 3.37 5.16 5.01 3.77
5.77 19.34 33.25 43.42 3.23 4.98 4.77 3.35
10
5
0
Baseline
VDA (init)
VDA (full)
Humans
Frame 1
Relative error
Relative error
Table 1: Quantitative results on synthetic billiard table videos. We evaluate our visual de-animation
(VDA) model before and after joint training (init vs. full). For scene reconstruction, we compute
MSE between rendered images and ground truth. For future prediction, we compute the Manhattan
distance in pixels between predicted object position and velocity and ground truth in pixels, at the 1st,
5th, 10th, and 20th future frames. Our full model performs better. See qualitative results in Figure 4.
Frame 3
Frame 5
10
5
0
Frame 10
(a) Future prediction results on synthetic billiard
videos of balls of the same physics
Baseline
VDA (init)
VDA (full)
Humans
Frame 1
Frame 3
Frame 5
Frame 10
(b) Future prediction results on synthetic billiard
videos of balls of varied frictions
Figure 5: Behavioral study results on future position prediction of billiard videos where balls have
the same physical properties (a), and balls have varied physical properties (b). For each model and
humans, we compare how their long-term relative prediction error grows, by measuring the ratio with
respect to errors in predicting the first next frame. Compared to the baseline model, the behavior of
our prediction models aligns well with human predictions.
(II) Varied appearance and physics (Figure 3b): balls can be of three different masses (light, medium,
heavy), and two different friction coefficients. Each of the six possible combinations is associated
with a unique color (appearance). In this setup, the scene de-rendering component should be able to
associate object appearance with its physical properties, even from a single image.
(III) Shared appearance, varied physics (Figure 3c): balls have the same appearance, but have one of
two different friction coefficients. Here, the perceptual component should be able to associate object
motion with its corresponding friction coefficients, from just a few input images.
Setup For this task, the physical state of an object is its intrinsic properties, including mass m and
friction f , and its extrinsic properties, including 2D position {x, y} and velocity v. Our system takes
three 256?256 RGB frames I1 , I2 , I3 as input. It first obtains flow fields from I1 to I2 and from I2
to I3 by a pretrained spatial pyramid network (SPyNet) [Ranjan and Black, 2017]. It then generates
object proposals by applying color filters on input images.
Our perceptual model is a ResNet-18 [He et al., 2015], which takes as input three masked RGB
frames and two masked flow images of each object proposal, and recovers the object?s physical
state. We use a differentiable, neural physics engine with object intrinsic properties as parameters; at
each step, it predicts objects? extrinsic properties (position {x, y} and velocity v) in the next frame,
based on their current estimates. We employ a graphics engine that renders original images from the
predicted positions, where the color of the balls is set as the mean color of the input object proposal.
The likelihood function compares, at a pixel level, these rendered images and observations. It is
straightforward to compute the gradient of object position from rendered RGB images and ground
truth. Thus, this simple graphics engine is also differentiable, making our system end-to-end trainable.
6
Video
VDA
(ours)
Figure 6: Sample results on web videos of real billiard games and computer games with realistic
rendering. Left: our method correctly estimates the trajectories of multiple objects. Right: our
framework correctly predicts the two collisions (white vs. red, white vs. blue), despite the motion
blur in the input, though it underestimates the velocity of the red ball after the collision. Note that the
billiard table is a chaotic system, and highly accurate long-term prediction is intractable.
Our training paradigm consists of two steps. First, we pretrain the perception module and the neural
physics engine separately on synthetic data, where ground truth is available. The second step is
end-to-end fine-tuning without annotations. We observe that the framework does not converge well
without pre-training, possibly due to the multiple hypotheses that can explain a scene (e.g., we can
only observe relative, not absolute masses from collisions). We train our framework using SGD, with
a learning rate of 0.001 and a momentum of 0.9. We implement our framework in Torch7 [Collobert
et al., 2011]. During testing, the perception module is run in reverse to recover object physical states,
and the learned physics engine is then run in forward for future prediction.
Results Our formulation recovers a rich representation of the scene. With the generative models,
we show results in scene reconstruction and future prediction. We compare two variants of our
algorithm: the initial system has its perception module and neural physics engine separately trained,
while the full system has an additional end-to-end fine-tuning step, as discussed above. We also
compare with a baseline, which has the sample perception model, but in prediction, simply repeats
object dynamics in the past without considering interactions among them.
Scene reconstruction: given input frames, we are able to reconstruct the images based on inferred
physical states. For evaluation, we compute pixel-level MSE between reconstructions and observed
images. We show qualitative results in Figure 4 and quantitative results in Table 1.
Future prediction: with the learned neural simulation engine, our system is able to predict future
events based on physical world states. We show qualitative results in Figure 4 and quantitative results
in Table 1, where we compute the Manhattan distance in pixels between predicted positions and
velocities and the ground truth. Our model achieves good performance in reconstructing the scene,
understanding object physics, and predicting scene dynamics. See caption for details.
Behavioral studies We further conduct behavioral studies, where we select 50 test cases, show the
first three frames of each to human subjects, and ask them the positions of each ball in future frames.
We test 3 subjects per case on Amazon Mechanical Turk. For each model and humans, we compare
how their long-term relative prediction error grows, by measuring the ratio with respect to errors in
predicting the first next frame. As shown in Figure 5, the behavior of our models resembles human
predictions much better than the baseline model.
4.2
Billiard Tables: Transferring to Real Videos
Data We also collected videos from YouTube, segmenting them into two-second clips. Some
videos are from real billiard competitions, and the others are from computer games with realistic
rendering. We use it as an out-of-sample test set for evaluating the model?s generalization ability.
Setup and Results Our setup is the same as that in Section 4.1, except that we now re-train the
perceptual model on the synthetic data of varied physics, but with flow images as input instead of
RGB images. Flow images abstract away appearance changes (color, lighting, etc.), allowing the
model to generalize better to real data. We show qualitative results of reconstruction and future
prediction in Figure 6 by rendering our inferred representation using the graphics software, Blender.
7
Video
Methods
# Blocks
2
VDA
(ours)
PhysNet
3
Mean
4
Chance
Humans
50 50 50
67 62 62
50
64
PhysNet
GoogLeNet
66 66 73
70 70 70
68
70
VDA (init)
VDA (joint)
VDA (full)
73 74 72
75 76 73
76 76 74
73
75
75
(b) Accuracy (%) of stability prediction on the
blocks dataset
Video
VDA
(ours)
PhysNet
Methods
2
3
4
Mean
PhysNet
GoogLeNet
56 68 70
70 67 71
65
69
VDA (init)
VDA (joint)
VDA (full)
74 74 67
75 77 70
76 76 72
72
74
75
(c) Accuracy (%) of stability prediction when
(a) Our reconstruction and prediction results given a single trained on synthetic towers of 2 and 4 blocks, and
frame (marked in red). From top to bottom: ground truth, tested on all block tower sizes.
our results, results from Lerer et al. [2016].
Video
VDA
(ours)
(d) Our reconstruction and prediction results given a single frame (marked in red)
Figure 7: Results on the blocks dataset [Lerer et al., 2016]. For quantitative results (b), we compare
three variants of our visual de-animation (VDA) model: perceptual module trained without fine-tuning
(init), joint fine-tuning with REINFORCE (joint), and full model considering stability constraint
(full). We also compare with PhysNet [Lerer et al., 2016] and GoogLeNet [Szegedy et al., 2015].
4.3
The Blocks World
We now look into a different scenario ? block towers. In this experiment, we demonstrate the
applicability of our model to explain and reason from a static image, instead of a video. We focus on
the reasoning of object states in the 3D world, instead of physical properties such as mass. We also
explore how our framework performs with non-differentiable simulation engines, and how physics
signals (e.g., stability) could help in physical reasoning, even when given only a static image.
Data Lerer et al. [2016] built a dataset of 492 images of real block towers, with ground truth
stability values. Each image may contain 2, 3, or 4 blocks of red, blue, yellow, or green color.
Though the blocks are the same size, their sizes in each 2D image differ due to 3D-to-2D perspective
transformation. Objects are made of the same material and thus have identical mass and friction.
8
Input
VDA
What if?
Input
Future
Stabilizing force
Figure 8: Examples of predicting hypothetical scenarios and actively engaging with the scene. Left:
predictions of the outcome of forces applied to two stable towers. Right: multiple ways to stabilize
two unstable towers.
Setup Here, the physical state of an object (block) consists of its 3D position {x, y, z} and 3D
rotation (roll, pitch, yaw, each quantized into 20 bins). Our perceptual model is again a ResNet-18 [He
et al., 2015], which takes block silhouettes generated by simple color filters as input, and recovers
the object?s physical state. For this task, we implement an efficient, non-differentiable, rigid body
simulator, to predict whether the blocks are stable. We also implement a graphics engine to render
object silhouettes for reconstructing the input. Our likelihood function consists of two terms: MSE
between rendered silhouettes and observations, and the binary cross-entropy between the predicted
stability and the ground truth stability.
Our training paradigm resembles the classic wake-sleep algorithm [Hinton et al., 1995]: first, generate
10,000 training images using the simulation engines; second, train the perception module on synthetic
data with ground truth physical states; third, end-to-end fine-tuning of the perceptual module by
explaining an additional 100,000 synthetic images without annotations of physical states, but with
binary annotations of stability. We use multi-sample REINFORCE [Rezende et al., 2016, Mnih and
Rezende, 2016] with 16 samples per input, assuming each position parameter is from a Gaussian
distribution and each rotation parameter is from a multinomial distribution (quantized into 20 bins).
We observe that the training paradigm helps the framework converge. The other setting is the same as
that in Section 4.1.
Results We show results on two tasks: scene reconstruction and stability prediction. For each task,
we compare three variants of our algorithm: the initial system has its perception module trained
without fine-tuning; an intermediate system has joint end-to-end fine-tuning, but without considering
the physics constraint; and the full system considers both reconstruction and physical stability during
fine-tuning.
We show qualitative results on scene reconstruction in Figures 7a and 7d, where we also demonstrate
future prediction results by exporting our inferred physical states into Blender. We show quantitative
results on stability prediction in Table 7b, where we compare our models with PhysNet [Lerer et al.,
2016] and GoogleNet [Szegedy et al., 2015]. All given a static image as test input, our algorithms
achieve higher prediction accuracy (75% vs. 70%) efficiently (<10 milliseconds per image).
Our framework also generalizes well. We test out-of-sample generalization ability, where we train our
model on 2- and 4-block towers, but test it on all tower sizes. We show results in Table 7c. Further, in
Figure 8, we show examples where our physical scene representation combined with a physics engine
can easily make conditional predictions, answering ?What happens if...?-type questions. Specifically,
we show frame prediction of external forces on stable block towers, as well as ways that an agent can
stabilize currently unstable towers, with the help of rich simulation engines.
5
Discussion
We propose combining efficient, bottom-up, neural perception modules with rich, generalizable
simulation engines for physical scene understanding. Our framework is flexible and can incorporate
various graphics and physics engines. It performs well across multiple synthetic and real scenarios,
reconstructing the scene and making future predictions accurately and efficiently. We expect our
framework to have wider applications in the future, due to the rapid development of scene description
languages, 3D reconstruction methods, simulation engines and virtual environments.
9
Acknowledgements
This work is supported by NSF #1212849 and #1447476, ONR MURI N00014-16-1-2007, the Center
for Brain, Minds and Machines (NSF #1231216), Toyota Research Institute, Samsung, Shell, and the
MIT Advanced Undergraduate Research Opportunities Program (SuperUROP).
References
Pulkit Agrawal, Ashvin Nair, Pieter Abbeel, Jitendra Malik, and Sergey Levine. Learning to poke by poking:
Experiential learning of intuitive physics. In NIPS, 2016. 2
Jimmy Ba, Volodymyr Mnih, and Koray Kavukcuoglu. Multiple object recognition with visual attention. In
ICLR, 2015. 3
Jiamin Bai, Aseem Agarwala, Maneesh Agrawala, and Ravi Ramamoorthi. Selectively de-animating video.
ACM TOG, 31(4):66, 2012. 3
Peter W Battaglia, Jessica B Hamrick, and Joshua B Tenenbaum. Simulation as an engine of physical scene
understanding. PNAS, 110(45):18327?18332, 2013. 2
Peter W Battaglia, Razvan Pascanu, Matthew Lai, Danilo Rezende, and Koray Kavukcuoglu. Interaction
networks for learning about objects, relations and physics. In NIPS, 2016. 2
Marcus A. Brubaker, David J. Fleet, and Aaron Hertzmann. Physics-based person tracking using the anthropomorphic walker. IJCV, 87(1-2):140?155, 2009. 3
Michael B Chang, Tomer Ullman, Antonio Torralba, and Joshua B Tenenbaum. A compositional object-based
approach to learning physical dynamics. In ICLR, 2017. 2, 4
Ronan Collobert, Koray Kavukcuoglu, and Cl?ment Farabet. Torch7: A matlab-like environment for machine
learning. In BigLearn, NIPS Workshop, 2011. 7
Erwin Coumans. Bullet physics engine. Open Source Software: http://bulletphysics. org, 2010. 2, 4
Misha Denil, Pulkit Agrawal, Tejas D Kulkarni, Tom Erez, Peter Battaglia, and Nando de Freitas. Learning to
perform physics experiments via deep reinforcement learning. In ICLR, 2017. 3
Sebastien Ehrhardt, Aron Monszpart, Niloy J Mitra, and Andrea Vedaldi. Learning a physical long-term predictor.
arXiv:1703.00247, 2017. 2
SM Eslami, Nicolas Heess, Theophane Weber, Yuval Tassa, Koray Kavukcuoglu, and Geoffrey E Hinton. Attend,
infer, repeat: Fast scene understanding with generative models. In NIPS, 2016. 3
Chelsea Finn, Ian Goodfellow, and Sergey Levine. Unsupervised learning for physical interaction through video
prediction. In NIPS, 2016. 2
Katerina Fragkiadaki, Pulkit Agrawal, Sergey Levine, and Jitendra Malik. Learning visual predictive models of
physics for playing billiards. In ICLR, 2016. 2, 5
Abhinav Gupta, Alexei A Efros, and Martial Hebert. Blocks world revisited: Image understanding using
qualitative geometry and mechanics. In ECCV, 2010. 2
Jessica B Hamrick, Andrew J Ballard, Razvan Pascanu, Oriol Vinyals, Nicolas Heess, and Peter W Battaglia.
Metacontrol for adaptive imagination-based optimization. In ICLR, 2017. 2
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In
CVPR, 2015. 6, 9
Geoffrey E Hinton, Peter Dayan, Brendan J Frey, and Radford M Neal. The ?wake-sleep? algorithm for
unsupervised neural networks. Science, 268(5214):1158, 1995. 3, 9
Jonathan Huang and Kevin Murphy. Efficient inference in occlusion-aware generative models of images. In
ICLR Workshop, 2015. 3
Zhaoyin Jia, Andy Gallagher, Ashutosh Saxena, and Tsuhan Chen. 3d reasoning from blocks to stability. IEEE
TPAMI, 37(5):905?918, 2015. 2
10
Kris M. Kitani, De-An Huang, and Wei-Chiu Ma. Activity forecasting. In Group and Crowd Behavior for
Computer Vision, pages 273?294. Elsevier, 2017. 3
Tejas D Kulkarni, Pushmeet Kohli, Joshua B Tenenbaum, and Vikash Mansinghka. Picture: A probabilistic
programming language for scene perception. In CVPR, 2015a. 3
Tejas D Kulkarni, William F Whitney, Pushmeet Kohli, and Josh Tenenbaum. Deep convolutional inverse
graphics network. In NIPS, 2015b. 1
Nikolaos Kyriazis and Antonis Argyros. Physically plausible 3d scene tracking: The single actor hypothesis. In
CVPR, 2013. 3
Adam Lerer, Sam Gross, and Rob Fergus. Learning physical intuition of block towers by example. In ICML,
2016. 1, 2, 8, 9
Wenbin Li, Ales Leonardis, and Mario Fritz. Visual stability prediction for robotic manipulation. In ICRA, 2017.
2
Michael Mathieu, Camille Couprie, and Yann LeCun. Deep multi-scale video prediction beyond mean square
error. In ICLR, 2016. 3
Andriy Mnih and Danilo J Rezende. Variational inference for monte carlo objectives. In ICML, 2016. 4, 9
Roozbeh Mottaghi, Hessam Bagherinezhad, Mohammad Rastegari, and Ali Farhadi. Newtonian scene understanding: Unfolding the dynamics of objects in static images. In CVPR, 2016a. 2
Roozbeh Mottaghi, Mohammad Rastegari, Abhinav Gupta, and Ali Farhadi. ?what happens if...? learning to
predict the effect of forces in images. In ECCV, 2016b. 2
Lerrel Pinto, Dhiraj Gandhi, Yuanfeng Han, Yong-Lae Park, and Abhinav Gupta. The curious robot: Learning
visual representations via physical interactions. In ECCV, 2016. 2
Anurag Ranjan and Michael J Black. Optical flow estimation using a spatial pyramid network. In CVPR, 2017. 6
Danilo Jimenez Rezende, SM Ali Eslami, Shakir Mohamed, Peter Battaglia, Max Jaderberg, and Nicolas Heess.
Unsupervised learning of 3d structure from images. In NIPS, 2016. 3, 4, 9
Mathieu Salzmann and Raquel Urtasun. Physically-based motion models for 3d tracking: A convex formulation.
In ICCV, 2011. 3
Tianjia Shao, Aron Monszpart, Youyi Zheng, Bongjin Koo, Weiwei Xu, Kun Zhou, and Niloy J Mitra. Imagining
the unseen: Stability-based cuboid arrangements for scene understanding. ACM TOG, 33(6), 2014. 2
Christian Szegedy, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed, Dragomir Anguelov, Dumitru Erhan,
Vincent Vanhoucke, and Andrew Rabinovich. Going deeper with convolutions. In CVPR, 2015. 8, 9
Marek Vondrak, Leonid Sigal, and Odest Chadwicke Jenkins. Dynamical simulation priors for human motion
tracking. IEEE TPAMI, 35(1):52?65, 2013. 3
Jacob Walker, Abhinav Gupta, and Martial Hebert. Dense optical flow prediction from a static image. In ICCV,
2015. 3
Ronald J Williams. Simple statistical gradient-following algorithms for connectionist reinforcement learning.
MLJ, 8(3-4):229?256, 1992. 4
Jiajun Wu, Ilker Yildirim, Joseph J Lim, William T Freeman, and Joshua B Tenenbaum. Galileo: Perceiving
physical object properties by integrating a physics engine with deep learning. In NIPS, 2015. 3
Jiajun Wu, Joseph J Lim, Hongyi Zhang, Joshua B Tenenbaum, and William T Freeman. Physics 101: Learning
physical object properties from unlabeled videos. In BMVC, 2016. 3
Jiajun Wu, Joshua B Tenenbaum, and Pushmeet Kohli. Neural scene de-rendering. In CVPR, 2017. 3
Tianfan Xue, Jiajun Wu, Katherine Bouman, and Bill Freeman. Visual dynamics: Probabilistic future frame
synthesis via cross convolutional networks. In NIPS, 2016. 3
Alan Yuille and Daniel Kersten. Vision as bayesian inference: analysis by synthesis? TiCS, 10(7):301?308,
2006. 3
11
Renqiao Zhang, Jiajun Wu, Chengkai Zhang, William T Freeman, and Joshua B Tenenbaum. A comparative
evaluation of approximate probabilistic simulation and deep neural networks as accounts of human physical
scene understanding. In CogSci, 2016. 2
Bo Zheng, Yibiao Zhao, Joey Yu, Katsushi Ikeuchi, and Song-Chun Zhu. Scene understanding by reasoning
stability and safety. IJCV, 112(2):221?238, 2015. 2
R in Computer
Song-Chun Zhu and David Mumford. A stochastic grammar of images. Foundations and Trends
Graphics and Vision, 2(4):259?362, 2007. 3
12
| 6620 |@word kohli:4 inversion:1 open:1 pieter:1 simulation:24 rgb:4 decomposition:1 blender:2 jacob:1 sgd:1 bai:2 liu:1 contains:1 initial:5 jimenez:1 salzmann:2 daniel:1 ours:4 past:1 existing:2 freitas:1 recovered:2 comparing:1 current:1 attracted:1 ronald:1 realistic:2 happen:3 blur:1 ronan:1 shape:3 christian:1 remove:1 interpretable:3 ashutosh:1 v:4 generative:8 cue:5 core:1 yuanfeng:1 provides:1 quantized:2 pascanu:2 revisited:1 billiard:20 org:1 tianfan:1 zhang:5 along:1 supply:1 qualitative:6 consists:5 ijcv:2 combine:4 behavioral:3 manner:1 introduce:2 rapid:1 andrea:1 behavior:3 mechanic:1 simulator:5 multi:3 brain:1 ashvin:1 freeman:5 inspired:1 farhadi:2 considering:3 increasing:1 begin:1 discover:2 underlying:1 theophane:1 mass:9 medium:1 what:11 tic:1 deepmind:1 generalizable:2 differing:1 transformation:1 bootstrapping:1 quantitative:5 hypothetical:2 saxena:1 tackle:1 appear:1 segmenting:1 before:2 safety:1 attend:1 mitra:2 frey:1 eslami:3 despite:1 anurag:1 oxford:1 niloy:2 koo:1 black:2 resembles:2 conversely:1 challenging:1 limited:2 range:1 unique:1 lecun:1 testing:3 galileo:1 block:22 implement:3 chaotic:1 razvan:2 maneesh:1 vedaldi:1 pre:1 integrating:2 unlabeled:1 applying:1 kersten:2 bill:1 ranjan:2 center:1 joey:1 go:1 attention:2 primitive:1 layout:1 williams:2 starting:1 straightforward:1 stabilizing:1 amazon:1 jimmy:1 convex:1 disentangled:1 stability:15 classic:1 updated:1 imagine:1 caption:1 programming:1 gandhi:1 us:2 hypothesis:2 goodfellow:1 engaging:1 associate:6 velocity:9 helmholtz:1 recognition:4 trend:1 utilized:1 predicts:2 muri:1 bottom:4 observed:2 module:23 levine:3 capture:2 connected:1 richness:1 sun:1 gross:1 intuition:1 environment:4 hertzmann:1 reward:1 dynamic:13 trained:8 segment:2 ali:3 predictive:2 yuille:2 upon:1 tog:2 shao:2 easily:3 joint:7 samsung:1 various:5 train:5 fast:3 describe:1 monte:1 cogsci:1 kevin:1 outcome:1 crowd:1 heuristic:1 plausible:1 erika:1 cvpr:7 reconstruct:3 grammar:1 ability:4 aseem:1 unseen:1 jointly:4 shakir:1 agrawal:5 differentiable:17 advantage:1 net:1 tpami:2 propose:3 reconstruction:15 interaction:6 ment:1 poke:1 combining:1 flexibility:2 achieve:3 adapts:1 coumans:3 description:1 intuitive:1 competition:1 comparative:1 adam:1 newtonian:1 object:67 resnet:2 help:3 develop:1 wider:1 pose:1 poking:1 andrew:2 mansinghka:1 strong:1 predicted:4 differ:1 direction:1 filter:2 stochastic:1 exploration:1 human:17 nando:1 viewing:1 material:1 virtual:1 bin:2 abbeel:1 generalization:3 anthropomorphic:1 extension:1 ground:12 predict:5 visualize:1 matthew:1 major:2 achieves:1 torralba:1 efros:1 released:1 battaglia:8 estimation:6 realizes:1 currently:1 unfolding:1 mit:4 biglearn:1 gaussian:1 aim:1 i3:2 rather:4 denil:3 zhou:1 rezende:8 focus:2 likelihood:4 pretrain:1 brendan:1 suppression:1 baseline:8 elsevier:1 inference:4 dayan:1 rigid:5 transferring:1 relation:2 going:1 i1:2 pixel:12 agarwala:1 overall:1 among:1 flexible:3 development:1 constrained:3 spatial:2 field:2 aware:1 beach:1 koray:4 identical:1 park:1 look:1 unsupervised:3 icml:2 yaw:1 yu:1 future:22 others:2 connectionist:1 duplicate:1 few:2 employ:1 simultaneously:1 recognize:1 murphy:2 geometry:1 occlusion:1 william:5 attempt:1 ab:2 detection:1 jessica:2 highly:1 mnih:4 zheng:3 alexei:1 evaluation:3 misha:1 light:1 behind:1 accurate:1 andy:1 conduct:1 pulkit:3 abundant:1 re:3 causal:1 bouman:1 modeling:3 measuring:2 whitney:1 rabinovich:1 applicability:1 subset:1 masked:4 predictor:1 graphic:36 motivating:1 xue:2 synthetic:18 combined:1 st:4 person:1 fritz:1 csail:3 probabilistic:4 physic:68 michael:3 synthesis:4 quickly:1 again:1 nm:1 huang:3 possibly:1 external:1 imagination:1 style:1 zhao:1 ullman:1 li:2 szegedy:3 actively:1 volodymyr:1 account:1 de:16 stabilize:2 coefficient:4 jitendra:2 explicitly:1 stream:1 collobert:2 aron:2 break:1 spynet:1 counterfactuals:1 red:8 mario:1 recover:3 annotation:5 jia:3 contribution:1 square:1 ehrhardt:2 convolutional:6 accuracy:3 characteristic:1 efficiently:5 roll:1 yellow:1 generalize:5 raw:2 vincent:1 kavukcuoglu:4 accurately:1 bayesian:1 yildirim:1 lu:1 ren:1 trajectory:2 lighting:1 researcher:1 agrawala:1 kris:1 carlo:1 explain:5 farabet:1 aligns:1 facebook:1 evaluates:1 underestimate:1 turk:1 mohamed:1 associated:1 recovers:5 static:6 stop:1 dataset:4 ask:1 knowledge:2 color:9 lim:2 mlj:1 adaptable:1 feed:1 higher:1 supervised:1 danilo:3 tom:1 chengkai:1 wei:2 roozbeh:2 bmvc:1 formulation:3 done:1 though:3 just:1 stage:1 web:2 google:1 billiards:1 bullet:2 grows:2 hongyi:1 usa:1 building:1 effect:1 requiring:1 contain:1 kitani:2 i2:3 neal:1 white:2 during:4 self:1 game:3 please:1 demonstrate:5 mohammad:2 performs:5 motion:9 interpreting:3 dragomir:1 reasoning:7 image:40 ranging:1 weber:1 novel:1 recently:1 variational:1 rotation:2 multinomial:1 physical:80 overview:2 tassa:1 extend:1 he:3 discussed:1 googlenet:4 synthesized:1 refer:1 anguelov:1 ai:1 tuning:8 erez:1 language:2 stable:3 actor:1 supervision:2 operating:2 han:1 renderer:1 etc:1 robot:1 chelsea:1 recent:3 perspective:1 reverse:2 scenario:9 manipulation:1 n00014:1 binary:2 onr:1 joshua:8 mottaghi:5 additional:3 employed:1 converge:2 paradigm:4 xiangyu:1 signal:2 ii:5 relates:1 multiple:11 full:13 pnas:1 infer:3 ale:1 alan:1 hamrick:3 long:5 cross:2 lai:1 prediction:45 involving:1 variant:6 basic:2 pitch:1 vision:5 erwin:1 arxiv:1 sergey:3 physically:2 pyramid:2 invert:1 animating:1 proposal:11 separately:2 fine:8 wake:2 walker:3 source:1 jian:1 subject:2 mature:2 simulates:1 ramamoorthi:1 incorporates:1 flow:7 ikeuchi:1 curious:1 feedforward:1 iii:5 intermediate:1 weiwei:1 rendering:16 andriy:1 lerer:10 idea:1 knowing:1 tradeoff:1 vikash:1 fleet:1 whether:3 six:1 heavier:1 torch7:2 effort:1 forecasting:1 song:2 render:3 peter:6 shaoqing:1 compositional:1 action:2 matlab:1 deep:9 antonio:1 heess:3 fragkiadaki:5 collision:5 nikolaos:1 mid:1 tenenbaum:9 recon:1 clip:1 generate:3 http:1 exist:1 nsf:2 millisecond:1 jiajun:6 extrinsic:3 correctly:2 per:3 blue:3 broadly:1 group:1 yangqing:1 ravi:1 year:1 run:6 inverse:7 uncertainty:1 powerful:2 raquel:1 extends:1 wu:10 architectural:1 yann:1 fold:1 sleep:2 activity:1 occur:1 constraint:3 scene:55 afforded:1 software:2 yong:1 vondrak:2 generates:2 speed:1 simulate:4 friction:13 optical:3 rendered:4 structured:1 according:1 ball:24 combination:1 across:3 reconstructing:6 sam:1 joseph:2 rob:1 making:3 happens:2 iccv:2 pipeline:2 computationally:1 discus:1 needed:1 mind:1 finn:2 end:14 generalizes:2 available:1 jenkins:1 lacked:1 apply:1 observe:3 away:1 pierre:1 original:1 top:2 remaining:1 include:1 recognizes:2 opportunity:1 lerrel:1 build:3 icra:1 experiential:1 move:1 malik:2 question:1 objective:1 arrangement:1 mumford:2 traditional:4 gradient:3 iclr:7 distance:2 reinforce:4 tower:15 collected:1 unstable:2 urtasun:2 reason:1 considers:1 marcus:1 assuming:2 code:2 reed:1 ratio:2 sermanet:1 optionally:1 setup:7 tsuhan:1 kun:1 bulletphysics:1 katherine:1 expense:1 ba:2 perform:1 allowing:1 sebastien:1 observation:4 convolution:1 datasets:3 sm:2 immediate:1 defining:1 hinton:4 incorporated:1 frame:35 brubaker:2 varied:10 tomer:1 camille:1 inferred:3 david:2 inverting:2 mechanical:1 trainable:1 engine:66 learned:3 nip:10 address:1 beyond:2 able:4 leonardis:1 dynamical:1 perception:19 scott:1 challenge:2 program:1 built:1 including:3 green:2 video:31 max:1 marek:1 event:1 force:5 predicting:4 residual:1 advanced:1 zhu:3 abhinav:4 mathieu:3 picture:1 martial:2 extract:1 ilker:1 prior:3 understanding:18 acknowledgement:1 relative:5 manhattan:2 lae:1 expect:2 generation:1 discoverer:1 geoffrey:2 foundation:1 agent:2 vanhoucke:1 sigal:1 viewpoint:1 playing:1 share:2 heavy:1 eccv:3 wenbin:1 repeat:2 supported:1 hebert:2 allow:3 understand:1 deeper:1 institute:1 wide:1 explaining:2 taking:1 absolute:1 overcome:1 world:20 evaluating:1 rich:8 computes:2 forward:8 made:1 reinforcement:3 adaptive:1 erhan:1 pushmeet:4 approximate:5 compact:1 emphasize:1 obtains:1 argyros:2 silhouette:5 jaderberg:1 cuboid:1 sequentially:1 robotic:1 assumed:1 monszpart:2 fergus:1 hessam:1 table:13 learn:7 ballard:1 ca:1 nicolas:3 rastegari:2 init:9 imagining:1 mse:4 yibiao:1 cl:1 domain:1 did:1 dense:1 animation:10 body:5 xu:1 slow:1 inferring:2 position:14 wish:1 inverts:1 explicit:1 momentum:1 perceptual:8 answering:1 bagherinezhad:1 third:3 toyota:1 learns:10 ian:1 down:1 dumitru:1 explored:2 gupta:5 chun:2 incorporating:2 intrinsic:3 intractable:1 undergraduate:1 workshop:2 gallagher:1 chen:1 entropy:1 simply:2 appearance:25 likely:1 explore:5 visual:24 josh:1 vinyals:1 tracking:5 kaiming:1 bo:1 pretrained:1 chang:6 pinto:3 radford:1 truth:12 chance:1 acm:2 ma:1 shell:1 nair:1 conditional:1 tejas:3 goal:2 marked:2 couprie:1 shared:6 leonid:1 hard:1 youtube:1 change:1 typical:1 specifically:3 diff:3 except:1 yuval:1 perceiving:1 pas:1 attempted:1 katerina:1 aaron:1 select:1 selectively:1 chiu:1 jonathan:1 kulkarni:6 oriol:1 incorporate:2 evaluate:4 tested:1 |
6,214 | 6,621 | Label Efficient Learning of Transferable
Representations across Domains and Tasks
Zelun Luo
Stanford University
[email protected]
Yuliang Zou
Virginia Tech
[email protected]
Judy Hoffman
University of California, Berkeley
[email protected]
Li Fei-Fei
Stanford University
[email protected]
Abstract
We propose a framework that learns a representation transferable across different
domains and tasks in a label efficient manner. Our approach battles domain shift
with a domain adversarial loss, and generalizes the embedding to novel task using
a metric learning-based approach. Our model is simultaneously optimized on
labeled source data and unlabeled or sparsely labeled data in the target domain.
Our method shows compelling results on novel classes within a new domain even
when only a few labeled examples per class are available, outperforming the
prevalent fine-tuning approach. In addition, we demonstrate the effectiveness of
our framework on the transfer learning task from image object recognition to video
action recognition.
1
Introduction
Humans are exceptional visual learners capable of generalizing their learned knowledge to novel
domains and concepts and capable of learning from few examples. In recent years, computational
models based on end-to-end learnable convolutional networks have made significant improvements for
visual recognition [18, 28, 54] and have been shown to demonstrate some cross-task generalizations [8,
48] while enabling faster learning of subsequent tasks as most frequently evidenced through finetuning [14, 36, 50].
However, most efforts focus on the supervised learning scenario where a closed world assumption
is made at training time about both the domain of interest and the tasks to be learned. Thus,
any generalization ability of these models is only an observed byproduct. There has been a large
push in the research community to address generalizing and adapting deep models across different
domains [64, 13, 58, 38], to learn tasks in a data efficient way through few shot learning [27, 70, 47,
11], and to generically transfer information across tasks [1, 14, 50, 35].
While most approaches consider each scenarios in isolation we aim to directly tackle the joint problem
of adapting to a novel domain which has new tasks and few annotations. Given a large labeled source
dataset with annotations for a task set, A, we seek to transfer knowledge to a sparsely labeled target
domain with a possibly wholly new task set, B. This setting is in line with our intuition that we
should be able to learn reusable and general purpose representations which enable faster learning of
future tasks requiring less human intervention. In addition, this setting matches closely to the most
common practical approach for training deep models which is to use a large labeled source dataset
(often ImageNet [6, 52]) to train an initial representation and then to continue supervised learning
with a new set of data and often with new concepts.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
In our approach, we jointly adapt a source representation for use in a distinct target domain using a
new multilayer unsupervised domain adversarial formulation while introducing a novel cross-domain
and within domain class similarity objective. This new objective can be applied even when the target
domain has non-overlapping classes to the source domain.
We evaluate our approach in the challenging setting of joint transfer across domains and tasks
and demonstrate our ability to successfully transfer, reducing the need for annotated data for the
target domain and tasks. We present results transferring from a subset of Google Street View House
Numbers (SVHN) [41] containing only digits 0-4 to a subset of MNIST [29] containing only digits 5-9.
Secondly, we present results on the challenging setting of adapting from ImageNet [6] object-centric
images to UCF-101 [57] videos for action recognition.
2
Related work
Domain adaptation. Domain adaptation seeks to learn from related source domains a well performing model on target data distribution [4]. Existing work often assumes that both domains are defined
on the same task and labeled data in target domain is sparse or non-existent [64]. Several methods
have tackled the problem with the Maximum Mean Discrepancy (MMD) loss [17, 36, 37, 38, 73]
between the source and target domain. Weight sharing of CNN parameters [58, 22, 21, 3] and minimizing the distribution discrepancy of network activations [51, 65, 30] have also shown convincing
results. Adversarial generative models [33, 32, 2, 59] aim at generating source-like data with target
data by training a generator and a discriminator simultaneously, while adversarial discriminative
models [62, 64, 13, 12, 23] focus on aligning embedding feature representations of target domain
to source domain. Inspired by adversarial discriminative models, we propose a method that aligns
domain features with multi-layer information.
Transfer learning. Transfer learning aims to transfer knowledge by leveraging the existing labeled
data of some related task or domain [45, 71]. In computer vision, examples of transfer learning
include [1, 31, 61] which try to overcome the deficit of training samples for some categories by
adapting classifiers trained for other categories [43]. With the power of deep supervised learning and
the ImageNet dataset [6, 52], learned knowledge can even transfer to a totally different task (i.e. image
classification ? object detection [50, 49, 34]; image classification ? semantic segmentation [35])
and then achieve state-of-the-art performance. In this paper, we focus on the setting where source
and target domains have differing label spaces but the label spaces share the same structure. Namely
adapting between classifying different category sets but not transferring from classification to a
localization plus classification task.
Few-shot learning. Few-shot learning seeks to learn new concepts with only a few annotated
examples. Deep siamese networks [27] are trained to rank similarity between examples. Matching
networks [70] learns a network that maps a small labeled support set and an unlabeled example to its
label. Aside from these metric learning-based methods, meta-learning has also served as a essential
part. Ravi et al. [47] propose to learn a LSTM meta-learner to learn the update rule of a learner. Finn
et al. [11] tries to find a good initialization point that can be easily fine-tune with new examples from
new tasks. When there exists a domain shift, the results of prior few-shot learning methods are often
degraded.
Unsupervised learning. Many unsupervised learning algorithms have focused on modeling raw data
using reconstruction objectives [19, 69, 26]. Other probabilistic models include restricted Boltzmann
machines [20], deep Boltzmann machines [53], GANs [15, 10, 9], and autoregressive models [42, 66]
are also popular. An alternative approach, often terms ?self-supervised learning? [5], defines a
pretext task such as predicting patch ordering [7], frame ordering [40], motion dynamics [39], or
colorization [72], as a form of indirect supervision. Compared to these approaches, our unsupervised
learning method does not rely on exploiting the spatial or temporal structure of the data, and is
therefore more generic.
3
Method
We introduce a semi-supervised learning algorithm which transfers information from a large labeled
source domain, S, to a sparsely labeled target domain, T . The goal being to learn a strong target
2
Source CNN
{ , }
Supervised
Loss
Source labeled data
Multi-layer Domain Transfer
Adversarial
Loss
Target unlabeled data
{ t}
Semantic Transfer
Pairwise
Similarity
Softmax
Entropy
Loss
Target labeled data
Supervised
Loss
{ t, t}
Target CNN
Figure 1: Our proposed learning framework for joint transfer across domains and semantic transfer across source and target and across target labeled to unlabeled data. We introduce a domain
discriminator which aligns source and target representations across multiple layers of the network
through domain adversarial learning. We enable semantic transfer through minimizing the entropy of
the pairwise similarity between unlabeled and labeled target images and use the temperature of the
softmax over the similarity vector to allow for non-overlapping label spaces.
classifier without requiring the large annotation overhead required for standard supervised learning
approaches.
In fact, this setting is very commonly explored for convolutional network (convnet) based recognition
methods. When learning with convnets the usual learning procedure is to use a very large labeled
dataset (e.g. ImageNet [6, 52]) for initial training of the network parameters (termed pre-training).
The learned weights are then used as initialization for continued learning on new data and for new
tasks, called fine-tuning. Fine-tuning has been broadly applied to reduce the number of labeled
examples needed for learning new tasks, such as recognizing new object categories after ImageNet
pre-training [54, 18], or learning new label structures such as detection after classficiation pretraining [14, 50]. Here we focus on transfer in the case of a shared label structure (e.g. classification
of different category sets).
We assume the source domain contains ns images, xs ? X S , with associated labels, ys ? Y S .
? t ? X?T , as well as mt images,
Similarly, the target domain consists of nt unlabeled images, x
t
T
t
T
x ? X , with associated labels, y ? Y . We assume that the target domain is only sparsely
labeled so that the number of image-label pairs is much smaller than the number of unlabeled images,
mt nt . Additionally, the number of source labeled images is assumed to be much larger than the
number of target labeled images, mt ns .
Unlike standard domain adaptation approaches which transfer knowledge from source to target
domains assuming a marginal or conditional distribution shift under a shared label space (Y S = Y T ),
we tackle joint image or feature space adaptation as well as transfer across semantic spaces. Namely,
we consider the case where the source and target label spaces are not equal, Y S 6= Y T , and even the
most challenging case where the sets are non-overlapping, Y S ? Y T = ?.
3.1
Joint domain and semantic transfer
Our approach consists of unsupervised feature alignment between source and target as well as
semantic transfer to the unlabeled target data from either the labeled target or the labeled source
data. We introduce a new multi-layer domain discriminator which can be used for domain alignment
following the recent domain adversarial learning approaches [13, 64]. We next introduce a new
semantic transfer learning objective which uses cross category similarity and can be tuned to account
for varying size of label set overlap.
3
We depict our overall model in Figure 1. We take the ns source labeled examples, {xs , ys }, the mt
target labeled examples, {xt , yt }, and the nt unlabeled target images, {?
xt } as input. We learn an
initial layered source representation and classification network (depicted in blue in Figure 1) using
standard supervised techniques. We then initialize the target model (depicted in green in Figure 1)
with the source parameters and begin our adaptive transfer learning.
Our model jointly optimizes over a target supervised loss, Lsup , a domain transfer objective, LDT , and
finally a semantic transfer objective, LST . Thus, our total objective can be written as follows:
L(X S , Y S , X T , Y T , X?T ) = Lsup (X T , Y T ) + ?LDT (X S , X?T ) + ?LST (X S , X T , X?T )
(1)
where the hyperparameters ? and ? determine the influence of the domain transfer loss and the
semantic transfer loss, respectively. In the following sections we elaborate on our domain and
semantic transfer objectives.
3.2
Multi-layer domain adversarial loss
We define a novel domain alignment objective function called multi-layer domain adversarial
loss. Recent efforts in deep domain adaptation have shown strong performance using feature space
domain adversarial objectives [13, 64]. These methods learn a target representation such that the
target distribution viewed under this model is aligned with the source distribution viewed under the
source representation. This alignment is accomplished through an adversarial minimization across
domain, analogous to the prevalent generative adversarial approaches [15]. In particular, a domain
discriminator, D(?), is trained to classify whether a particular data point arises from the source or the
target domain. Simultaneously, the target embedding function E t (xt ) (defined as the application of
layers of the network is trained to generate the target representation that cannot be distinguished from
the source domain representation by the domain discriminator. Similar to [63, 64], we consider a
representation to be domain invariant if the domain discriminator can not distinguish examples from
the two domains.
Prior work considers alignment for a single layer of the embedding at a time and as such learns a
domain discriminator which takes the output from the corresponding source and target layers as input.
Separately, domain alignment methods which focus on first and second order statistics have shown
improved performance through applying domain alignment independently at multiple layers of the
network [36]. Rather than learning independent discriminators for each layer of the network we
propose a simultaneous alignment of multiple layers through a multi-layer discriminator.
At each layer of our multi-layer domain discriminator, information is accumulated from both the
output from the previous discriminator layer as well as the source and target activations from the
corresponding layer in their respective embeddings. Thus, the output of each discriminator layer is
defined as:
dl = Dl (?(?dl?1 ? El (x)))
(2)
where l is the current layer, ?(?) is the activation function, ? ? 1 is the decay factor, ? represents
concatenation or element-wise summation, and x is taken either from source data xs ? X S , or target
? t ? X?T . Notice that the intermediate discriminator layers share the same structure with their
data x
corresponding encoding layers to match the dimensions.
Thus, the following loss functions are proposed to optimize the multi-layer domain discriminator and
the embeddings, respectively, according to our domain transfer objective:
s
t
LD
DT = ?Exs ?X S [log dl ] ? Ext ?X T log(1 ? dl )
t
s
t
LE
DT = ?Exs ?X S [log(1 ? dl )] ? Ext ?X T log dl
(3)
(4)
where dsl , dtl are the outputs of the last layer of the source and target multi-layer domain discriminator.
Note that these losses are placed after the final domain discriminator layer and the last embedding
layer but then produce gradients which back-propagate throughout all relevant lower layer parameters.
These two losses together comprise LDT , and there is no iterative optimization procedure involved.
This multi-layer discriminator (shown in Figure 1 - yellow) allows for deeper alignment of the
source and target representations which we find empirically results in improved target classification
performance as well as more stable adversarial learning.
4
Figure 2: We illustrate the purpose of temperature (? ) for our pairwise similarity vector. Consider an
example target unlabeled point and its similarity to four labeled source points (x-axis). We show here,
original unnormalized scores (leftmost) as well as the same similarity scores after applying softmax
with different temperatures, ? . Notice that entropy values, H(x), have higher variance for scores
normalized with a small temperature softmax.
3.3
Cross category similarity for semantic transfer
In the previous section, we introduced a method for transferring an embedding from the source to the
target domain. However, this only enforces alignment of the global domain statistics with no class
specific transfer. Here, we define a new semantic transfer objective, LST , which transfers information
from a labeled set of data to an unlabeled set of data by minimizing the entropy of the softmax with
temperature of the similarity vector between an unlabeled point and all labeled points. Thus, this
loss may be applied either between the source and unlabeled target data or between the labeled and
unlabeled target data.
? t , we compute the similarity, ?(?), to each labeled example or
For each unlabeled target image, x
to each prototypical example [56] per class in the labeled set. For simplicity of presentation let us
consider semantic transfer from the source to the target domain first. For each target unlabeled image
we compute a similarity vector where the ith element is the similarity between this target image and
the ith labeled source image: [vs (?
xt )]i = ?(?
xt , xsi ). Our semantic transfer loss can be defined as
follows:
LST (X?T , X S )
=
X
H(?(vs (?
xt )/? ))
(5)
? t ?X?T
x
where, H(?) is the information entropy function, ?(?) is the softmax function and ? is the temperature
of the softmax. Note that the temperature can be used to directly control the percentage of source
examples we expect the target example to be similar to (see Figure 2).
Entropy minimization has been widely used for unsupervised [44] and semi-supervised [16] learning
by encouraging low density separation between clusters or classes. Recently this principle of entropy
minimization has be applied for unsupervised adaptation [38]. Here, the source and target domains
are assumed to share a label space and each unlabeled target example is passed through the initial
source classifier and the entropy of the softmax output scores is minimized.
In contrast, we do not assume a shared label space between the source and target domains and as such
can not assume that each target image maps to a single source label. Instead, we compute pairwise
similarities between target points and the source points (or per class averages of source points [56])
across the features spaces aligned by our multi-layer domain adversarial transfer. We then tune the
softmax temperature based on the expected similarity between the source and target labeled set. For
example, if the source and target label set overlap, then a small temperature will encourage each
target point to be very similar to one source class, whereas a larger temperature will allow for target
points to be similar to multiple source classes.
For semantic transfer within the target domain, we utilize the metric-based cross entropy loss between
labeled target examples to stabilize and improve the learning. For a labeled target example, in addition
to the traditional cross entropy loss, we also calculate a metric-based cross entropy loss 1 . Assume
we have k labeled examples from each class in the target domain. We compute the embedding for
1
We refer this as "metric-based" to cue the reader that this is not a cross entropy within the label space.
5
each example and then the centroid cTi of each class in the embedding space. Thus, we can compute
the similarity vector for each labeled example, where the ith element is the similarity between this
labeled example and the centroid of each class: [vt (xt )]i = ?(xt , cTi ). We can then calculate the
metric based cross entropy loss:
LST,sup (X T ) = ?
X
{xt ,yt }?X T
exp ([vt (xt )]yt )
log Pn
t
i=1 exp ([vt (x )]i )
(6)
Similar to the source-to-target scenario, for target-to-target we also have the unsupervised part,
X
H(?(vt (?
xt )/? ))
LST,unsup (X?T , X T ) =
(7)
? t ?X?T
x
With the metric-based cross entropy loss, we introduce the constraint that the target domain data
should be similar in the embedding space. Also, we find that this loss can provide a guidance
for the unsupervised semantic transfer to learn in a more stable way. LST is the combination of
LST,unsupervised from source-target (Equation 5), LST,supervised from source-target (Equation 6), and
LST,unsupervised from target-target (Equation 7), i.e.,
LST (X S , X T , X?T ) = LST (X?T , X S ) + LST,sup (X T ) + LST,unsup (X?T , X T )
4
(8)
Experiment
This section is structured as follows. In section 4.1, we show that our method outperform fine-tuning
approach by a large margin, and all parts of our method are necessary. In section 4.2, we show that
our method can be generalized to bigger datasets. In section 4.3, we show that our multi-layer domain
adversarial method outperforms state-of-the-art domain adversarial approaches.
Datasets We perform adaptation experiments across two different paired data settings. First for
adaptation across different digit domains we use MNIST [29] and Google Street View House Numbers
(SVHN) [41]. The MNIST handwritten digits database has a training set of 60,000 examples, and a
test set of 10,000 examples. The digits have been size-normalized and centered in fixed-size images.
SVHN is a real-world image dataset for machine learning and object recognition algorithms with
minimal requirement on data preprocessing and formatting. It has 73257 digits for training, 26032
digits for testing. As our second experimental setup, we consider adaptation from object centric
images in ImageNet [52] to action recognition in video using the UCF-101 [57] dataset. ImageNet
is a large benchmark for the object classification task. We use the task 1 split from ILSVRC2012.
UCF-101 is an action recognition dataset collected on YouTube. With 13,320 videos from 101
action categories, UCF-101 provides a large diversity in terms of actions and with the presence of
large variations in camera motion, object appearance and pose, object scale, viewpoint, cluttered
background, illumination conditions, etc.
Implementation details We pre-train the source domain embedding function with cross-entropy loss.
For domain adversarial loss, the discriminator takes the last three layer activations as input when the
number of output classes are the same for source and target tasks, and takes the second last and third
last layer activations when they are different. The similarity score is chosen as the dot product of the
normalized support features and the unnormalized target feature. We use the temperature ? = 2 for
source-target semantic transfer and ? = 1 for within target transfer as the label space is shared. We
use ? = 0.1 and ? = 0.1 in our objective function. The network is trained with Adam optimizer [25]
and with learning rate 10?3 . We conduct all the experiments with the PyTorch framework.
4.1
SVHN 0-4 ? MNIST 5-9
Experimental setting. In this experiment, we define three datasets: (i) labeled data in source domain
D1 ; (ii) few labeled data in target domain D2 ; (iii) unlabeled data in target domain D3 . We take the
training split of SVHN dataset as dataset D1 . To fairly compare with traditional learning paradigm
and episodic training, we subsample k examples from each class to construct dataset D2 so that we
can perform traditional training or episodic (k ? 1)-shot learning. We experiment with k = 2, 3, 4, 5,
which corresponds to 10, 15, 20, 25 labeled examples, or 0.017%, 0.025%, 0.333%, 0.043% of the
6
total training data respectively. Since our approach involves using annotations from a small subset
of the data, we randomly subsample 10 different subsets {D2i }10
i=1 from the training split of MNIST
dataset, and use the remaining data as {D3i }10
i=1 for each k. Note that source domain and target
domain have non-overlapping classes: we only utilize digits 0-4 in SVHN, and digits 5-9 in MNIST.
Figure 3: An illustration of our task. Our model effectively transfer the learned representation on
SVHN digits 0-4 (left) to MNIST digits 5-9 (right).
Baselines and prior work. We compare against six different methods: (i) Target only: the model
is trained on D2 from scratch; (ii) Fine-tune: the model is pretrained on D1 and fine-tuned on D2 ;
(iii) Matching networks [70]: we first pretrain the model on D3 , then use D2 as the support set in the
matching networks; (iv) Fine-tuned matching networks: same as baseline iii, except that for each k
the model is fine-tuned on D2 with 5-way (k ? 1)-shot learning: k ? 1 examples in each class are
randomly selected as the support set, and the last example in each class is used as the query set; (v)
Fine-tune + adversarial: in addition to baseline ii, the model is also trained on D1 and D3 with a
domain adversarial loss; (vi.) Full model: fine-tune the model with the proposed multi-layer domain
adversarial loss.
Results and analysis. We calculate the mean and standard error of the accuracies across 10 sets of
data, which is shown in Table 1. Due to domain shift, matching networks perform poorly without
fine-tuning, and fine-tuning is only marginally better than training from scratch. Our method with
multi-layer adversarial only improves the overall performance, but is more sensitive to the subsampled
data. Our method achieves significant performance gain, especially when the number of labeled
examples is small (k = 2). For reference, fine-tuning on full target dataset gives an accuracy of
99.65%.
Table 1: The test accuracies of the baseline models and our method. Row 1 to row 6 correspond (in
the same order) to the six methods proposed in section 4.2. Note that the accuracies of two matching
net methods are calculated based on nearest neighbors in the support set. We report the mean and the
standard error of each method across 10 different subsampled data.
Method
k=2
k=3
k=4
k=5
Target only
Fine-tune
Matching nets [70]
Fine-tuned matching nets
Ours: fine-tune + adv.
Ours: full model (? = 0.1)
(a)
(b)
0.642 ? 0.026
0.612 ? 0.020
0.469 ? 0.019
0.645 ? 0.019
0.702 ? 0.020
0.917 ? 0.007
0.771 ? 0.015
0.779 ? 0.018
0.455 ? 0.014
0.755 ? 0.024
0.800 ? 0.013
0.936 ? 0.006
(c)
(d)
0.801 ? 0.010
0.802 ? 0.016
0.566 ? 0.013
0.793 ? 0.013
0.804 ? 0.014
0.942 ? 0.006
(e)
0.840 ? 0.013
0.830 ? 0.011
0.513 ? 0.023
0.827 ? 0.011
0.831 ? 0.013
0.950 ? 0.004
(f)
Figure 4: The t-SNE [68, 67] visualization of different feature embeddings. (a) Source domain
embedding. (b) Target domain embedding using encoder trained with source domain domain. (c)
Target domain embedding using encoder fine-tuned with target domain data. (d) Target domain
embedding using encoder trained with our method. (e) An overlap of a and c. (f) An overlap of a and
d. (best viewed in color and with zoom)
7
4.2
Image object recognition ? video action recognition
Problem analysis. Many recent works [60, 24] study the domain shift between images and video
in the object detection settings. Compared to still images, videos provide several advantages: (i)
motion provides information for foreground vs background segmentation [46]; (ii) videos often show
multiple views and thus provide 3D information. On the other hand, video frames usually suffer from:
(i) motion blur; (ii) compression artifacts; (iii) objects out-of-focus or out-of-frame.
Experimental setting. In this experiment, we focus on three dataset splits: (i) ImageNet training
set as the labeled data in source domain D1 ; (ii) k video clips per class randomly sampled from
UCF-101 training as the few labeled data in target domain set D2 ; (iii) the remaining videos in
UCF-101 training set as the unlabeled data in target domain D3 . We experiment with k = 3, 5, 10,
which corresponds 303, 505, 1010 video clips, or 2.27%, 3.79%, 7.59% of the total training data
respectively. Each experiment is run 3 times on D1 , {D2i }3i=1 , and {D3i }3i=1 .
Baselines and prior work. We compare our method with two baseline methods: (i) Target only:
the model is trained on D2 from scratch; (ii) Fine-tune: the model is first pre-trained on D1 , then
fine-tuned on D2 . For reference, we report the performance of a fully supervised method [55].
Results and analysis. The accuracy of each model is shown in Table 2. We also fine-tune a model
with all the labeled data for comparison. Per-frame performance (img) and average-across-frame
performance (vid) are both reported. Note that we calculate the average-across-frame performance by
averaging the softmax score of each frame in a video. Our method achieves significant improvement
on average-across-frame performance over standard fine-tuning for each value of k. Note that
compared to fine-tuning, our method has a bigger gap between per-frame and per-video accuracy.
We believe that this is due to the semantic transfer: our entropy loss encourages a sharper softmax
variance among per-frame softmax scores per video (if the variance is zero, then per-frame accuracy =
per-video accuracy). By making more confident predictions among key frames, our method achieves
a more significant gain with respective to per-video performance, even when there is little change in
the per-frame prediction.
Table 2: Accuracy of UCF-101 action classification. The results of the two-stream spatial model are
taken from [55] and vary depending on hyperparameters. We report the mean and the standard error
of each method across 3 different subsampled data.
Method
k=3
k=5
k=10
All
Target only (img)
Target only (vid)
Fine-tune (img)
Fine-tune (vid)
Two-stream spatial [55]
Ours (img)
Ours (vid)
4.3
0.098?0.003
0.105?0.003
0.380?0.013
0.406?0.015
0.393?0.006
0.467?0.007
0.126?0.022
0.133?0.024
0.486?0.012
0.523?0.010
0.459?0.013
0.545?0.014
0.100?0.035
0.106?0.038
0.529?0.039
0.568?0.042
0.523?0.002
0.620?0.005
0.672
0.714
0.708 - 0.720
-
Ablation: unsupervised domain adaptation
To validate our multi-layer domain adversarial loss objective, we conduct an ablation experiment
for unsupervised domain adaptation. We compare against multiple recent domain adversarial unsupervised adaptation methods. In this experiment, we first pretrain a source embedding CNN on the
training split SVHN [41] and then adapt the target embedding for MNIST by performing adversarial
domain adaptation. We evaluate the classification performance on the test split of MNIST [29]. We
follow the same training strategy and model architecture for the embedding network as [64].
8
All the models here have a two-step training strategy and share the first stage. ADDA [64] optimizes
encoder and classifier simultaneously. We also propose a similar method, but optimize encoder only.
Only we try a model with no classifier in the last layer (i.e. perform domain adversarial training in
feature space). We choose ? = 0.1 as the decay factor for this model.
The accuracy of each model is shown in Table 3. We find that our method achieve 6.5% performance
gain over the best competing domain adversarial approach indicating that our multilayer objective
indeed contributes to our overall performance. In addition, in our experiments, we found that the
multilayer approach improved overall optimization stability, as evidenced in our small standard error.
Table 3: Experimental results on unsupervised domain adaptation from SVHN to MNIST. Results of
Gradient reversal, Domain confusion, and ADDA are from [64], and the results of other methods are
from experiments across 5 different subsampled data.
Method
Accuracy
Source only
Gradient reversal [13]
Domain confusion [62]
ADDA [64]
Ours
5
0.601 ? 0.011
0.739
0.681 ? 0.003
0.760 ? 0.018
0.810 ? 0.003
Conclusion
In this paper, we propose a method to learn a representation that is transferable across different
domains and tasks in a data efficient manner. The framework is trained jointly to minimize the domain
shift, to transfer knowledge to new task, and to learn from large amounts of unlabeled data. We show
superior performance over the popular fine-tuning approach. We hope to keep improving the method
in future work.
Acknowledgement
We would like to start by thanking our sponsors: Stanford Computer Science Department and Stanford
Program in AI-assisted Care (PAC). Next, we specially thank De-An Huang, Kenji Hata, Serena
Yeung, Ozan Sener and all the members of Stanford Vision and Learning Lab for their insightful
discussion and feedback. Lastly, we thank all the anonymous reviewers for their valuable comments.
9
References
[1] Yusuf Aytar and Andrew Zisserman. Tabula rasa: Model transfer for object category detection.
In Computer Vision (ICCV), 2011 IEEE International Conference on, pages 2252?2259. IEEE,
2011.
[2] Konstantinos Bousmalis, Nathan Silberman, David Dohan, Dumitru Erhan, and Dilip Krishnan.
Unsupervised pixel-level domain adaptation with generative adversarial networks. arXiv preprint
arXiv:1612.05424, 2016.
[3] Lluis Castrejon, Yusuf Aytar, Carl Vondrick, Hamed Pirsiavash, and Antonio Torralba. Learning
aligned cross-modal representations from weakly aligned data. In Proceedings of the IEEE
Conference on Computer Vision and Pattern Recognition, pages 2940?2949, 2016.
[4] Gabriela Csurka. Domain adaptation for visual applications: A comprehensive survey. arXiv
preprint arXiv:1702.05374, 2017.
[5] Virginia R de Sa. Learning classification with unlabeled data. Advances in neural information
processing systems, pages 112?112, 1994.
[6] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei. ImageNet: A Large-Scale
Hierarchical Image Database. In CVPR09, 2009.
[7] Carl Doersch, Abhinav Gupta, and Alexei A Efros. Unsupervised visual representation learning
by context prediction. In Proceedings of the IEEE International Conference on Computer
Vision, pages 1422?1430, 2015.
[8] Jeff Donahue, Yangqing Jia, Oriol Vinyals, Judy Hoffman, Ning Zhang, Eric Tzeng, and Trevor
Darrell. Decaf: A deep convolutional activation feature for generic visual recognition. In
Proceedings of the 31st International Conference on International Conference on Machine
Learning - Volume 32, ICML?14, pages I?647?I?655. JMLR.org, 2014.
[9] Jeff Donahue, Philipp Kr?henb?hl, and Trevor Darrell. Adversarial feature learning. arXiv
preprint arXiv:1605.09782, 2016.
[10] Vincent Dumoulin, Ishmael Belghazi, Ben Poole, Alex Lamb, Martin Arjovsky, Olivier
Mastropietro, and Aaron Courville. Adversarially learned inference. arXiv preprint
arXiv:1606.00704, 2016.
[11] Chelsea Finn, Pieter Abbeel, and Sergey Levine. Model-agnostic meta-learning for fast adaptation of deep networks. arXiv preprint arXiv:1703.03400, 2017.
[12] Yaroslav Ganin and Victor Lempitsky. Unsupervised domain adaptation by backpropagation.
arXiv preprint arXiv:1409.7495, 2014.
[13] Yaroslav Ganin, Evgeniya Ustinova, Hana Ajakan, Pascal Germain, Hugo Larochelle, Fran?ois
Laviolette, Mario Marchand, and Victor Lempitsky. Domain-adversarial training of neural
networks. Journal of Machine Learning Research, 17(59):1?35, 2016.
[14] Ross Girshick, Jeff Donahue, Trevor Darrell, and Jitendra Malik. Rich feature hierarchies
for accurate object detection and semantic segmentation. In Computer Vision and Pattern
Recognition, 2014.
[15] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil
Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. In Advances in neural
information processing systems, pages 2672?2680, 2014.
[16] Yves Grandvalet, Yoshua Bengio, et al. Semi-supervised learning by entropy minimization. In
NIPS, volume 17, pages 529?536, 2004.
[17] A Gretton, A.J. Smola, J Huang, Marcel Schmittfull, K.M. Borgwardt, B Sch?lkopf,
J Qui?onero Candela, M Sugiyama, A Schwaighofer, and N D. Lawrence. Covariate shift by
kernel mean matching. In Dataset Shift in Machine Learning, 131-160 (2009), 01 2009.
10
[18] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, pages 770?778, 2016.
[19] Geoffrey E Hinton and Ruslan R Salakhutdinov. Reducing the dimensionality of data with
neural networks. science, 313(5786):504?507, 2006.
[20] Geoffrey E Hinton and Terrence J Sejnowski. Learning and releaming in boltzmann machines.
Parallel Distrilmted Processing, 1, 1986.
[21] Judy Hoffman, Saurabh Gupta, and Trevor Darrell. Learning with side information through
modality hallucination. In Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, pages 826?834, 2016.
[22] Judy Hoffman, Saurabh Gupta, Jian Leong, Sergio Guadarrama, and Trevor Darrell. Crossmodal adaptation for rgb-d detection. In Robotics and Automation (ICRA), 2016 IEEE International Conference on, pages 5032?5039. IEEE, 2016.
[23] Judy Hoffman, Dequan Wang, Fisher Yu, and Trevor Darrell. Fcns in the wild: Pixel-level
adversarial and constraint-based adaptation. arXiv preprint arXiv:1612.02649, 2016.
[24] Vicky Kalogeiton, Vittorio Ferrari, and Cordelia Schmid. Analysing domain shift factors
between videos and images for object detection. IEEE transactions on pattern analysis and
machine intelligence, 38(11):2327?2334, 2016.
[25] Diederik Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprint
arXiv:1412.6980, 2014.
[26] Diederik P Kingma and Max Welling. Auto-encoding variational bayes. arXiv preprint
arXiv:1312.6114, 2013.
[27] Gregory Koch. Siamese neural networks for one-shot image recognition. PhD thesis, University
of Toronto, 2015.
[28] Alex Krizhevsky, Ilya Sutskever, and Geoff Hinton. Imagenet classification with deep convolutional neural networks. In Neural Information Processing Systems (NIPS), 2012.
[29] Yann LeCun, L?on Bottou, Yoshua Bengio, and Patrick Haffner. Gradient-based learning
applied to document recognition. Proceedings of the IEEE, 86(11):2278?2324, 1998.
[30] Yanghao Li, Naiyan Wang, Jianping Shi, Jiaying Liu, and Xiaodi Hou. Revisiting batch
normalization for practical domain adaptation. arXiv preprint arXiv:1603.04779, 2016.
[31] Joseph J Lim, Ruslan Salakhutdinov, and Antonio Torralba. Transfer learning by borrowing
examples for multiclass object detection. In Proceedings of the 24th International Conference
on Neural Information Processing Systems, pages 118?126. Curran Associates Inc., 2011.
[32] Ming-Yu Liu, Thomas Breuel, and Jan Kautz. Unsupervised image-to-image translation
networks. arXiv preprint arXiv:1703.00848, 2017.
[33] Ming-Yu Liu and Oncel Tuzel. Coupled generative adversarial networks. In Advances in Neural
Information Processing Systems, pages 469?477, 2016.
[34] Wei Liu, Dragomir Anguelov, Dumitru Erhan, Christian Szegedy, Scott Reed, Cheng-Yang
Fu, and Alexander C Berg. Ssd: Single shot multibox detector. In European Conference on
Computer Vision, pages 21?37. Springer, 2016.
[35] Jonathan Long, Evan Shelhamer, and Trevor Darrell. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, pages 3431?3440, 2015.
[36] Mingsheng Long, Yue Cao, Jianmin Wang, and Michael I Jordan. Learning transferable features
with deep adaptation networks. In ICML, pages 97?105, 2015.
11
[37] Mingsheng Long, Jianmin Wang, and Michael I Jordan. Deep transfer learning with joint
adaptation networks. arXiv preprint arXiv:1605.06636, 2016.
[38] Mingsheng Long, Han Zhu, Jianmin Wang, and Michael I Jordan. Unsupervised domain
adaptation with residual transfer networks. In Advances in Neural Information Processing
Systems, pages 136?144, 2016.
[39] Zelun Luo, Boya Peng, De-An Huang, Alexandre Alahi, and Li Fei-Fei. Unsupervised learning
of long-term motion dynamics for videos. arXiv preprint arXiv:1701.01821, 2017.
[40] Ishan Misra, C Lawrence Zitnick, and Martial Hebert. Shuffle and learn: unsupervised learning
using temporal order verification. In European Conference on Computer Vision, pages 527?544.
Springer, 2016.
[41] Yuval Netzer, Tao Wang, Adam Coates, Alessandro Bissacco, Bo Wu, and Andrew Y Ng.
Reading digits in natural images with unsupervised feature learning. In NIPS workshop on deep
learning and unsupervised feature learning, volume 2011, page 5, 2011.
[42] Aaron van den Oord, Nal Kalchbrenner, and Koray Kavukcuoglu. Pixel recurrent neural
networks. arXiv preprint arXiv:1601.06759, 2016.
[43] Maxime Oquab, Leon Bottou, Ivan Laptev, and Josef Sivic. Learning and transferring midlevel image representations using convolutional neural networks. In Proceedings of the IEEE
conference on computer vision and pattern recognition, pages 1717?1724, 2014.
[44] Gintautas Palubinskas, Xavier Descombes, and Frithjof Kruggel. An unsupervised clustering
method using the entropy minimization. In AAAI, 1999.
[45] Sinno Jialin Pan and Qiang Yang. A survey on transfer learning. IEEE Transactions on
knowledge and data engineering, 22(10):1345?1359, 2010.
[46] Deepak Pathak, Ross Girshick, Piotr Doll?r, Trevor Darrell, and Bharath Hariharan. Learning
features by watching objects move. arXiv preprint arXiv:1612.06370, 2016.
[47] Sachin Ravi and Hugo Larochelle. Optimization as a model for few-shot learning. In International Conference on Learning Representations, volume 1, page 6, 2017.
[48] Ali Sharif Razavian, Hossein Azizpour, Josephine Sullivan, and Stefan Carlsson. CNN features
off-the-shelf: An astounding baseline for recognition. In IEEE Conference on Computer Vision
and Pattern Recognition, CVPR Workshops 2014, Columbus, OH, USA, June 23-28, 2014, pages
512?519, 2014.
[49] Joseph Redmon, Santosh Divvala, Ross Girshick, and Ali Farhadi. You only look once: Unified,
real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and
Pattern Recognition, pages 779?788, 2016.
[50] Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun. Faster r-cnn: Towards real-time
object detection with region proposal networks. In Advances in neural information processing
systems, pages 91?99, 2015.
[51] Artem Rozantsev, Mathieu Salzmann, and Pascal Fua. Beyond sharing weights for deep domain
adaptation. arXiv preprint arXiv:1603.06432, 2016.
[52] Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng
Huang, Andrej Karpathy, Aditya Khosla, Michael Bernstein, et al. Imagenet large scale visual
recognition challenge. International Journal of Computer Vision, 115(3):211?252, 2015.
[53] Ruslan Salakhutdinov and Geoffrey Hinton. Deep boltzmann machines. In Artificial Intelligence
and Statistics, pages 448?455, 2009.
[54] K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image
recognition. CoRR, abs/1409.1556, 2014.
12
[55] Karen Simonyan and Andrew Zisserman. Two-stream convolutional networks for action
recognition in videos. In Advances in neural information processing systems, pages 568?576,
2014.
[56] Jake Snell, Kevin Swersky, and Richard S Zemel. Prototypical networks for few-shot learning.
arXiv preprint arXiv:1703.05175, 2017.
[57] Khurram Soomro, Amir Roshan Zamir, and Mubarak Shah. Ucf101: A dataset of 101 human
actions classes from videos in the wild. arXiv preprint arXiv:1212.0402, 2012.
[58] Baochen Sun and Kate Saenko. Deep coral: Correlation alignment for deep domain adaptation.
In Computer Vision?ECCV 2016 Workshops, pages 443?450. Springer, 2016.
[59] Yaniv Taigman, Adam Polyak, and Lior Wolf. Unsupervised cross-domain image generation.
arXiv preprint arXiv:1611.02200, 2016.
[60] Kevin Tang, Vignesh Ramanathan, Li Fei-Fei, and Daphne Koller. Shifting weights: Adapting
object detectors from image to video. In Advances in Neural Information Processing Systems,
pages 638?646, 2012.
[61] Tatiana Tommasi, Francesco Orabona, and Barbara Caputo. Safety in numbers: Learning
categories from few examples with multi model knowledge transfer. In Computer Vision and
Pattern Recognition (CVPR), 2010 IEEE Conference on, pages 3081?3088. IEEE, 2010.
[62] Eric Tzeng, Judy Hoffman, Trevor Darrell, and Kate Saenko. Simultaneous deep transfer across
domains and tasks. In Proceedings of the IEEE International Conference on Computer Vision,
pages 4068?4076, 2015.
[63] Eric Tzeng, Judy Hoffman, Trevor Darrell, and Kate Saenko. Simultaneous deep transfer across
domains and tasks. In International Conference in Computer Vision (ICCV), 2015.
[64] Eric Tzeng, Judy Hoffman, Kate Saenko, and Trevor Darrell. Adversarial discriminative domain
adaptation. In Computer Vision and Pattern Recognition (CVPR), 2017.
[65] Eric Tzeng, Judy Hoffman, Ning Zhang, Kate Saenko, and Trevor Darrell. Deep domain
confusion: Maximizing for domain invariance. arXiv preprint arXiv:1412.3474, 2014.
[66] Aaron van den Oord, Nal Kalchbrenner, Lasse Espeholt, Oriol Vinyals, Alex Graves, et al.
Conditional image generation with pixelcnn decoders. In Advances in Neural Information
Processing Systems, pages 4790?4798, 2016.
[67] Laurens Van Der Maaten. Accelerating t-sne using tree-based algorithms. Journal of machine
learning research, 15(1):3221?3245, 2014.
[68] Laurens Van der Maaten and Geoffrey Hinton. Visualizing non-metric similarities in multiple
maps. Machine learning, 87(1):33?55, 2012.
[69] Pascal Vincent, Hugo Larochelle, Yoshua Bengio, and Pierre-Antoine Manzagol. Extracting and
composing robust features with denoising autoencoders. In Proceedings of the 25th international
conference on Machine learning, pages 1096?1103. ACM, 2008.
[70] Oriol Vinyals, Charles Blundell, Tim Lillicrap, Daan Wierstra, et al. Matching networks for one
shot learning. In Advances in Neural Information Processing Systems, pages 3630?3638, 2016.
[71] Karl Weiss, Taghi M Khoshgoftaar, and DingDing Wang. A survey of transfer learning. Journal
of Big Data, 3(1):1?40, 2016.
[72] Richard Zhang, Phillip Isola, and Alexei A Efros. Colorful image colorization. In European
Conference on Computer Vision, pages 649?666. Springer, 2016.
[73] Xu Zhang, Felix Xinnan Yu, Shih-Fu Chang, and Shengjin Wang. Deep transfer network:
Unsupervised domain adaptation. arXiv preprint arXiv:1503.00591, 2015.
13
| 6621 |@word cnn:6 compression:1 fcns:1 pieter:1 d2:9 seek:3 propagate:1 rgb:1 vicky:1 shot:11 ld:1 initial:4 liu:4 contains:1 score:7 salzmann:1 tuned:7 ours:5 document:1 outperforms:1 existing:2 current:1 guadarrama:1 nt:3 luo:2 activation:6 diederik:2 written:1 hou:1 subsequent:1 blur:1 christian:1 update:1 depict:1 aside:1 v:3 generative:5 cue:1 selected:1 intelligence:2 amir:1 ith:3 bissacco:1 provides:2 philipp:1 toronto:1 org:1 zhang:5 daphne:1 wierstra:1 consists:2 overhead:1 wild:2 introduce:5 manner:2 pairwise:4 peng:1 indeed:1 expected:1 frequently:1 multi:16 inspired:1 salakhutdinov:3 ming:2 encouraging:1 little:1 farhadi:1 totally:1 begin:1 sinno:1 evgeniya:1 agnostic:1 differing:1 unified:1 temporal:2 berkeley:2 alahi:1 tackle:2 descombes:1 classifier:5 ustinova:1 sherjil:1 control:1 intervention:1 colorful:1 safety:1 felix:1 engineering:1 ext:2 encoding:2 ajakan:1 plus:1 initialization:2 challenging:3 practical:2 camera:1 enforces:1 testing:1 lecun:1 backpropagation:1 digit:12 procedure:2 sullivan:1 wholly:1 jan:1 episodic:2 tuzel:1 evan:1 adapting:6 matching:10 ucf101:1 pre:4 cannot:1 unlabeled:21 layered:1 dtl:1 naiyan:1 andrej:1 context:1 influence:1 applying:2 crossmodal:1 optimize:2 map:3 reviewer:1 yt:3 vittorio:1 shi:1 maximizing:1 independently:1 cluttered:1 zelun:2 focused:1 survey:3 simplicity:1 jimmy:1 koray:1 pouget:1 rule:1 continued:1 oh:1 embedding:17 stability:1 cvpr09:1 variation:1 ferrari:1 analogous:1 target:93 hierarchy:1 hana:1 olivier:1 carl:2 us:1 curran:1 goodfellow:1 associate:1 element:3 recognition:28 sparsely:4 labeled:44 database:2 observed:1 levine:1 preprint:21 wang:8 sharif:1 calculate:4 revisiting:1 region:1 zamir:1 ilsvrc2012:1 adv:1 sun:3 ordering:2 shuffle:1 valuable:1 alessandro:1 intuition:1 warde:1 dynamic:2 existent:1 trained:12 d2i:2 weakly:1 laptev:1 ali:2 unsup:2 localization:1 eric:5 learner:3 vid:4 easily:1 finetuning:1 joint:6 indirect:1 lst:14 geoff:1 train:2 distinct:1 fast:1 roshan:1 sejnowski:1 query:1 artificial:1 zemel:1 kevin:2 kalchbrenner:2 jean:1 stanford:7 larger:2 widely:1 cvpr:3 encoder:5 ability:2 statistic:3 simonyan:2 jointly:3 final:1 ldt:3 advantage:1 net:4 breuel:1 propose:6 reconstruction:1 product:1 adaptation:28 aligned:4 relevant:1 ablation:2 cao:1 poorly:1 achieve:2 validate:1 exploiting:1 sutskever:1 cluster:1 requirement:1 darrell:12 yaniv:1 produce:1 generating:1 adam:4 bousmalis:1 ben:1 object:20 tim:1 illustrate:1 depending:1 andrew:3 pose:1 ganin:2 recurrent:1 nearest:1 sa:1 strong:2 c:1 involves:1 kenji:1 larochelle:3 ois:1 marcel:1 ning:2 laurens:2 closely:1 annotated:2 stochastic:1 centered:1 human:3 enable:2 espeholt:1 abbeel:1 generalization:2 anonymous:1 snell:1 feifeili:1 secondly:1 summation:1 pytorch:1 assisted:1 gabriela:1 koch:1 exp:2 lawrence:2 efros:2 optimizer:1 achieves:3 vary:1 torralba:2 purpose:2 ruslan:3 khoshgoftaar:1 label:20 ross:4 sensitive:1 exceptional:1 successfully:1 hoffman:9 minimization:5 hope:1 stefan:1 aim:3 rather:1 pn:1 shelf:1 varying:1 azizpour:1 focus:7 june:1 improvement:2 rank:1 prevalent:2 tech:1 contrast:1 adversarial:34 centroid:2 baseline:7 dilip:1 pretrain:2 inference:1 el:1 accumulated:1 transferring:4 borrowing:1 koller:1 tao:1 josef:1 pixel:3 overall:4 classification:12 among:2 pascal:3 hossein:1 jianmin:3 art:2 spatial:3 tzeng:5 softmax:12 initialize:1 santosh:1 marginal:1 equal:1 comprise:1 piotr:1 fairly:1 beach:1 construct:1 sener:1 saurabh:2 look:1 represents:1 adversarially:1 yu:4 unsupervised:27 icml:2 foreground:1 future:2 discrepancy:2 minimized:1 report:3 mirza:1 few:13 yoshua:4 richard:2 randomly:3 simultaneously:4 zoom:1 comprehensive:1 subsampled:4 astounding:1 ab:1 detection:10 interest:1 once:1 alexei:2 hallucination:1 alignment:11 generically:1 farley:1 lluis:1 jialin:1 accurate:1 fu:2 capable:2 byproduct:1 encourage:1 necessary:1 netzer:1 respective:2 conduct:2 iv:1 tree:1 guidance:1 girshick:4 minimal:1 classify:1 modeling:1 compelling:1 xinnan:1 ishmael:1 introducing:1 subset:4 recognizing:1 krizhevsky:1 virginia:2 reported:1 eec:1 gregory:1 confident:1 st:2 density:1 lstm:1 serena:1 international:11 borgwardt:1 oord:2 khurram:1 probabilistic:1 dong:1 terrence:1 off:1 michael:4 together:1 ilya:1 gans:1 sanjeev:1 thesis:1 aaai:1 containing:2 choose:1 possibly:1 huang:4 watching:1 li:6 szegedy:1 account:1 diversity:1 de:3 yaroslav:2 stabilize:1 automation:1 inc:1 jitendra:1 kate:5 vi:1 stream:3 csurka:1 try:3 view:3 razavian:1 closed:1 lab:1 sup:2 dumoulin:1 start:1 mario:1 candela:1 parallel:1 bayes:1 annotation:4 kautz:1 jia:2 minimize:1 yves:1 hariharan:1 degraded:1 convolutional:8 variance:3 accuracy:11 multibox:1 correspond:1 yellow:1 dsl:1 lkopf:1 raw:1 handwritten:1 vincent:2 kavukcuoglu:1 marginally:1 onero:1 ren:2 served:1 russakovsky:1 bharath:1 simultaneous:3 hamed:1 detector:2 sharing:2 aligns:2 trevor:12 against:2 involved:1 associated:2 jianping:1 lior:1 gain:3 sampled:1 dataset:15 popular:2 knowledge:8 color:1 improves:1 dimensionality:1 segmentation:4 lim:1 sean:1 back:1 centric:2 alexandre:1 higher:1 dt:2 supervised:14 follow:1 zisserman:3 improved:3 modal:1 wei:2 formulation:1 fua:1 stage:1 lastly:1 smola:1 convnets:1 correlation:1 hand:1 autoencoders:1 mehdi:1 su:1 overlapping:4 google:2 defines:1 artifact:1 columbus:1 believe:1 mingsheng:3 usa:2 phillip:1 vignesh:1 concept:3 requiring:2 normalized:3 lillicrap:1 xavier:1 semantic:21 visualizing:1 taghi:1 self:1 encourages:1 transferable:4 unnormalized:2 leftmost:1 generalized:1 demonstrate:3 confusion:3 motion:5 svhn:9 temperature:11 dragomir:1 vondrick:1 zhiheng:1 image:38 wise:1 variational:1 novel:6 recently:1 charles:1 common:1 superior:1 ishan:1 mt:4 empirically:1 hugo:3 volume:4 he:2 significant:4 refer:1 anguelov:1 ai:1 tuning:10 doersch:1 similarly:1 rasa:1 aytar:2 sugiyama:1 ssd:1 dot:1 pixelcnn:1 stable:2 ucf:7 similarity:20 supervision:1 han:1 etc:1 patrick:1 aligning:1 sergio:1 chelsea:1 recent:5 optimizes:2 barbara:1 scenario:3 termed:1 misra:1 meta:3 outperforming:1 continue:1 vt:5 accomplished:1 der:2 victor:2 arjovsky:1 tabula:1 care:1 isola:1 oquab:1 deng:2 determine:1 paradigm:1 xiangyu:1 semi:3 ii:7 siamese:2 multiple:7 full:3 gretton:1 faster:3 match:2 adapt:2 cross:13 long:6 hata:1 y:2 bigger:2 paired:1 sponsor:1 prediction:3 xsi:1 multilayer:3 vision:21 metric:8 yeung:1 arxiv:42 sergey:1 kernel:1 mmd:1 normalization:1 robotics:1 maxime:1 addition:5 whereas:1 fine:26 separately:1 background:2 proposal:1 krause:1 source:62 jian:3 yusuf:2 sch:1 modality:1 unlike:1 specially:1 comment:1 yue:1 member:1 leveraging:1 effectiveness:1 jordan:3 extracting:1 presence:1 yang:2 bernstein:1 intermediate:1 embeddings:3 split:6 iii:5 krishnan:1 mastropietro:1 isolation:1 bengio:4 leong:1 architecture:1 competing:1 ivan:1 polyak:1 reduce:1 classficiation:1 haffner:1 konstantinos:1 multiclass:1 shift:9 blundell:1 whether:1 six:2 tommasi:1 passed:1 accelerating:1 effort:2 soomro:1 suffer:1 henb:1 karen:1 shaoqing:2 pretraining:1 action:10 deep:22 antonio:2 tune:11 karpathy:1 amount:1 clip:2 category:10 pretext:1 sachin:1 generate:1 outperform:1 percentage:1 coates:1 qiang:1 notice:2 per:13 blue:1 broadly:1 reusable:1 four:1 key:1 shih:1 yangqing:1 d3:4 ravi:2 nal:2 utilize:2 year:1 taigman:1 run:1 you:1 swersky:1 throughout:1 reader:1 lamb:1 yann:1 wu:1 patch:1 separation:1 fran:1 maaten:2 qui:1 layer:36 distinguish:1 tackled:1 courville:2 schmittfull:1 lsup:2 marchand:1 cheng:1 constraint:2 fei:8 alex:3 nathan:1 leon:1 performing:2 martin:1 cordelia:1 structured:1 department:1 according:1 combination:1 battle:1 across:24 smaller:1 oncel:1 pan:1 joseph:2 formatting:1 making:1 hl:1 den:2 restricted:1 invariant:1 iccv:2 taken:2 equation:3 visualization:1 bing:1 needed:1 finn:2 end:2 reversal:2 generalizes:1 available:1 doll:1 hierarchical:1 ozan:1 generic:2 pierre:1 distinguished:1 alternative:1 batch:1 shah:1 original:1 thomas:1 assumes:1 remaining:2 include:2 clustering:1 laviolette:1 tatiana:1 coral:1 especially:1 jake:1 icra:1 silberman:1 objective:15 malik:1 move:1 strategy:2 usual:1 traditional:3 antoine:1 gradient:4 convnet:1 deficit:1 thank:2 concatenation:1 street:2 decoder:1 considers:1 collected:1 ozair:1 assuming:1 colorization:2 illustration:1 reed:1 minimizing:3 convincing:1 manzagol:1 setup:1 sne:2 sharper:1 hao:1 ba:1 implementation:1 boltzmann:4 satheesh:1 perform:4 francesco:1 datasets:3 benchmark:1 enabling:1 xiaodi:1 daan:1 hinton:5 frame:13 community:1 introduced:1 evidenced:2 namely:2 required:1 pair:1 david:2 optimized:1 discriminator:18 imagenet:11 germain:1 sivic:1 california:1 learned:6 kingma:2 nip:4 address:1 able:1 beyond:1 poole:1 usually:1 pattern:11 scott:1 reading:1 challenge:1 program:1 green:1 pirsiavash:1 video:22 max:1 shifting:1 power:1 overlap:4 pathak:1 natural:1 rely:1 predicting:1 residual:2 zhu:1 improve:1 abhinav:1 mathieu:1 axis:1 martial:1 auto:1 schmid:1 coupled:1 prior:4 carlsson:1 acknowledgement:1 graf:1 loss:28 expect:1 fully:2 prototypical:2 generation:2 geoffrey:4 generator:1 shelhamer:1 verification:1 principle:1 viewpoint:1 grandvalet:1 classifying:1 share:4 translation:1 row:2 eccv:1 karl:1 placed:1 last:7 hebert:1 side:1 allow:2 deeper:1 divvala:1 neighbor:1 deepak:1 sparse:1 van:4 overcome:1 dimension:1 calculated:1 world:2 feedback:1 rich:1 autoregressive:1 made:2 commonly:1 adaptive:1 preprocessing:1 rozantsev:1 erhan:2 josephine:1 welling:1 transaction:2 keep:1 belghazi:1 global:1 img:4 assumed:2 discriminative:3 iterative:1 khosla:1 table:6 additionally:1 learn:13 transfer:53 robust:1 ca:1 composing:1 contributes:1 improving:1 caputo:1 bottou:2 european:3 zou:1 domain:131 zitnick:1 big:1 subsample:2 hyperparameters:2 xu:2 lasse:1 ng:1 elaborate:1 judy:9 n:3 d3i:2 house:2 jmlr:1 third:1 mubarak:1 learns:3 donahue:3 ian:1 artem:1 tang:1 dumitru:2 xt:11 specific:1 covariate:1 pac:1 insightful:1 shengjin:1 learnable:1 explored:1 x:3 decay:2 gupta:3 abadie:1 boya:1 workshop:3 dl:7 essential:1 exists:1 mnist:10 socher:1 effectively:1 decaf:1 kr:1 phd:1 corr:1 ramanathan:1 illumination:1 push:1 margin:1 gap:1 entropy:18 generalizing:2 depicted:2 appearance:1 visual:6 vinyals:3 aditya:1 schwaighofer:1 kaiming:2 bo:1 pretrained:1 chang:1 springer:4 corresponds:2 midlevel:1 wolf:1 acm:1 ma:1 cti:2 conditional:2 lempitsky:2 goal:1 viewed:3 presentation:1 towards:1 orabona:1 jeff:3 shared:4 fisher:1 change:1 youtube:1 analysing:1 except:1 reducing:2 yuval:1 averaging:1 redmon:1 olga:1 denoising:1 called:2 total:3 invariance:1 experimental:4 saenko:5 indicating:1 aaron:4 berg:1 support:5 arises:1 jonathan:2 alexander:1 oriol:3 evaluate:2 d1:7 scratch:3 ex:2 |
6,215 | 6,622 | Decoding with Value Networks for Neural Machine
Translation
Di He1
[email protected]
Tao Qin4
[email protected]
Hanqing Lu2
[email protected]
Liwei Wang1,5
[email protected]
Yingce Xia3
[email protected]
Tie-Yan Liu4
[email protected]
1
Key Laboratory of Machine Perception, MOE, School of EECS, Peking University
2
Carnegie Mellon University 3 University of Science and Technology of China
4
Microsoft Research
5
Center for Data Science, Peking University, Beijing Institute of Big Data Research
Abstract
Neural Machine Translation (NMT) has become a popular technology in recent
years, and beam search is its de facto decoding method due to the shrunk search
space and reduced computational complexity. However, since it only searches
for local optima at each time step through one-step forward looking, it usually
cannot output the best target sentence. Inspired by the success and methodology of
AlphaGo, in this paper we propose using a prediction network to improve beam
search, which takes the source sentence x, the currently available decoding output
y1 , ? ? ? , yt?1 and a candidate word w at step t as inputs and predicts the long-term
value (e.g., BLEU score) of the partial target sentence if it is completed by the NMT
model. Following the practice in reinforcement learning, we call this prediction
network value network. Specifically, we propose a recurrent structure for the value
network, and train its parameters from bilingual data. During the test time, when
choosing a word w for decoding, we consider both its conditional probability
given by the NMT model and its long-term value predicted by the value network.
Experiments show that such an approach can significantly improve the translation
accuracy on several translation tasks.
1
Introduction
Neural Machine Translation (NMT), which is based on deep neural networks and provides an endto-end solution to machine translation, has attracted much attention from the research community
[2, 6, 12, 20] and gradually been adopted by industry in past several years [18, 22]. NMT uses
an RNN-based encoder-decoder framework to model the entire translation process. In training, it
maximizes the likelihood of a target sentence given a source sentence. In testing, given a source
sentence x, it tries to find a sentence y ? in the target language that maximizes the conditional
probability P (y|x). Since the number of possible target sentences is exponentially large, finding the
optimal y ? is NP-hard. Thus beam search is commonly employed to find a reasonably good y.
Beam search is a heuristic search algorithm that maintains the top-scoring partial sequences expanded
in a left-to-right fashion. In particular, it keeps a pool of candidates each of which is a partial sequence.
At each time step, the algorithm expands each candidate by appending a new word, and then keeps
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
the top-ranked new candidates scored by the NMT model. The algorithm terminates if it meets the
maximum decoding depth or all sentences are completely generated, i.e., all sentences are ended with
the end-of-sentence (EOS) symbol.
While NMT with beam search has been proved to be successful, it has several obvious issues,
including exposure bias [9], loss-evaluation mismatch [9] and label bias [16], which have been studied.
However, we observe that there is still an important issue associated with beam search of NMT, the
myopic bias, which unfortunately is largely ignored, to the best of our knowledge. Beam search tends
to focus more on short-term reward. At iteration t, for a candidate y1 , ? ? ? , yt?1 (refers to y<t ) and
two words w and w0 , we denote y<t + w if we append w to y<t . If P (y<t + w|x) > P (y<t + w0 |x),
new candidate y<t + w is more likely to be kept, even if w0 is the ground truth translation at step t or
can offer a better score in future decodings. Such search errors coming from short sighted actions
sometimes provide a bad translation even if the translation model is good.
To address the myopic bias, for each word w and each candidate y<t , we propose to design a
prediction model to estimate the long-term reward if we append w to y<t and follow the current
NMT model until the decoding finishes. Then we can leverage the predicted score from this model
during each decoding step to help find a better w that can contribute to the long-term translation
performance. This prediction model, which predicts long-term reward we will receive in the future, is
exactly the concept of value function in Reinforcement Learning (RL).
In this work, we develop a neural network-based prediction model, which is called value network
for NMT. The value network takes the source sentence and any partial target sequence as input, and
outputs a predicted value to estimate the expected total reward (e.g. BLEU) generated from this
partial sequence by the NMT model. In any decoding step, we select the best candidates not only
based on the conditional probability of the partial sequence outputted by the NMT model, but also
based on the estimated long-term reward outputted by the value network.
The main contributions of this work are summarized as follows. First, we develop a decoding scheme
that considers long-term reward while generating words one by one for machine translation, which is
new in NMT literature. At each step, the new decoding scheme not only considers the probability of
the word sequence conditioned on the source sentence, but also relies on the predicted future reward.
We believe that considering the two aspects can lead to better final translation.
Second, we design a novel structure for the value network. On the top of the encoder-decoder layer
of NMT, we develop another two modules for the value network, a semantic matching module and a
context-coverage module. The semantic matching module aims at estimating the similarity between
the source and target sentences, which can contribute to the quality of the translation. It is often
observed that the more context used in the attention mechanism, the better translation we will generate
[14, 15]. Thus we build a context-coverage module to measure the coverage of context used in the
encoder-decoder layer. With the outputs of the two modules, the value prediction is done via fully
connected layers.
We conduct a set of experiments on several translation tasks. All the results demonstrate the
effectiveness and robustness of the new decoding mechanism compared to several baseline algorithms.
The remaining parts of the paper are organized as follows. In Section 2, we briefly review the
literature of neural machine translation. After that, we describe the myopic bias problem of NMT in
Section 3 and introduce our method for value network learning in Section 4. Experimental results are
provided and analyzed in Section 5. We discuss future directions in the last section.
2
Neural Machine Translation
Neural machine translation systems are typically implemented with a Recurrent Neural Network
(RNN)-based encoder-decoder framework. Such a framework directly models the probability P (y|x)
of a target sentence y = {y1 , y2 , ..., yTy } conditioned on the source sentence x = {x1 , x2 , ..., xTx },
where Tx and Ty are the length of sentence x and y.
The encoder of NMT reads the source sentence x word by word and generates a hidden representation
for each word xi :
hi = f (hi?1 , xi ),
2
(1)
Decoder
y1
y2
y3
r1
r2
r3
c1
c2
c3
Semantic Matching (SM) Module
???
?
Mean pooling
Mean pooling
SM Module
?
c1
Attention
c3
c2
?
r1
r3
?
?
htx
r2
CC Module
Context-Coverage (CC) Module
Encoder
h1
h2
h3
x1
x2
x3
?
???
htx
Mean pooling
xtx
c1
c2
Mean pooling
c3
?
h1
h2
Figure 1: Architecture of Value Network
in which function f is the recurrent unit such as Long Short-Term Memory (LSTM) unit [12] or Gated
Recurrent Unit (GRU) [4]. Afterwards, the decoder of NMT computes the conditional probability of
each target word yt conditioned on its proceeding words y<t as well as the source sentence:
P (yt |y<t , x) ? exp(yt ; rt , ct ),
rt = g(rt?1 , yt?1 , ct ),
ct = q(rt?1 , h1 , ? ? ? , hTx ),
(2)
(3)
(4)
where rt is the decoder RNN hidden representation at step t, similarly computed by an LSTM or GRU,
and ct denotes the weighted contextual information summarizing the source sentence x using some
attention mechanism [4]. Denote all the parameters to be learned in the encoder-decoder framework
as ?. For ease of reference, we also use ?? to represent the translation model with parameter ?.
Denote D as the training dataset that contains source-target sentence pairs. The training process
aims at seeking the optimal parameters ?? to correctly encode source sentence and decode it into the
target sentence. While there are different objectives to achieve this [2, 10, 1, 5, 19, 17], maximum
likelihood estimation is the most popular one [2]:
Y
?? = argmax
P (y|x; ?)
?
=
argmax
?
3
(x,y)?D
Y
Ty
Y
P (yt |y<t , x; ?).
(5)
(x,y)?D t=1
The Myopic Bias
Since during training, one aims to find the conditional probability P (y|x), ideally in testing, the
translation of a source sentence x should be the target sentence y with the maximum conditional
probability P (y|x). However, as there are exponentially many candidates in the target language, one
cannot compute the probability for every candidate and find the maximum one. Thus, beam search is
widely used to find a reasonable good target sentence [4, 14, 12].
Note that the training objective of NMT is usually defined on the full target sentence y instead of
partial sentences. One issue with beam search is that a locally good word might not lead to a good
complete sentence. From the example mentioned in the introduction, we can see that such search
errors from short sighted actions can provide a bad translation even if we hold a perfect translation
model. We call such errors the myopic bias. To reduce myopic bias, we hope to predict the long-term
value of each action and use the value in decoding, which is the exact motivation of our work.
There exist several works weakly related to this issue. [3] develops the scheduled sampling approach,
which takes the generated outputs from the model as well as the golden truth sentence in training,
to help the model learn from its own errors. Although it can (to some extent) handle the negative
3
impact of choosing an incorrect word at middle steps, it still follows beam search during testing,
which cannot avoid the myopic bias. Another related work is [16]. It learns a predictor to predict
the ranking score of a certain word at step t, and use this score to replace the conditional probability
outputted by the NMT model for beam search during testing. Unfortunately, this work still looks
only one step forward and cannot address the problem.
4
Value Network for NMT
As discussed in the previous section, it is not reasonable to fully rely on the conditional probability
in beam search. This motivates us to estimate the expected performance of any sequence during
decoding, which is exactly the concept of value function in reinforcement learning.
4.1
Value Network Structure
In conventional reinforcement learning, a value function describes how much cumulated reward could
be collected from state s by following certain policy ?. In machine translation, we can consider any
input sentence x paired with partial output sentence y<t as the state, and consider the translation
model ?? as policy which can generate a word (action) given any state. Given policy ?? , the value
function characterizes what the expected translation performance (e.g. BLEU score) is if we use ??
to translate x with the first t ? 1 words being y<t . Denote
and y ? (x)
P v(x, y<t ) as the value? function
0
as the ground truth translation, and then v(x, y<t ) = y0 ?Y:y0 =y<t BLEU(y (x), y )P (y 0 |x; ?),
<t
where Y is the space of complete sentences.
The first important problem is how to design the input and the parametric form of the value function.
As the translation model is built up on an encoder-decoder framework, we also build up our value
network on the top of this architecture. To fully exploit the information in the encoder-decoder
framework, we develop a value network with two new modules, the semantic matching module and
the context-coverage module.
Semantic Matching (SM) Module: In the semanticP
matching module, at time step t, we use mean
t
pooling over the decoder RNN hidden states r?t = 1t l=1 rl as a summarization of the partial target
Pt
sentence, and use mean pooling over context states c?t = 1t l=1 cl as a summarization of the context
in source language. We concatenate r?t and c?t , and use a feed-forward network ?SM = fSM ([?
rt , c?t ])
to evaluate semantic information between the source sentence and the target sentence.
Context-Coverage (CC) Module: It is often observed that the more context covered in the attention
model, the better translation we will generate [14, 15]. Thus we build a context-coverage module to
measure the coverage of information used in the encoder-decoder framework. We argue that using
mean pooling over the context layer and the encoding states should give some effective knowledge.
? = 1 PTx hl , we use another feed-forward network ?CC = fCC ([?
? to
Similarly, denote h
ct , h])
l=1
Tx
process such information.
In the end, we concatenate both ?SM and ?CC and then use another fully connected layer with
sigmoid activation function to output a scalar as the value prediction. The whole architecture is shown
in Figure 1.
4.2
Training Data Generation
Based on the designed value network structure, we aim at finding a model that can correctly predict the
performance after the decoding ends. Popular value function learning algorithms include Monte-Carlo
methods and Temple-Difference methods, and both of them have been adopted in many challenging
tasks [13, 7, 11]. In this paper, we adopt the Monte-Carlo method to learn the value function. Given
a well-learnt NMT model ?? , the training of the value network for ?? is shown in Algorithm 1. For
randomly picked source sentence x in the training corpus, we generate a partial target sentence yp
using ?? with random early stop, i.e., we randomly terminate the decoding process before its end.
Then for the pair (x, yp ), we use ?? to finish the translation starting from yp and obtain a set S(yp )
of K complete target sentences, e.g., using beam search. In the end, we compute the BLEU score of
4
each complete target sentence and calculate the averaged BLEU score of (x, yp ):
avg_bleu(x, yp ) =
1
K
X
BLEU(y ? (x), y).
(6)
y?S(yp )
avg_bleu(x, yp ) can be considered as an estimation of the long-term reward of state (x, yp ) used in
value network training.
4.3
Learning
Algorithm 1 Value network training
1: Input: Bilingual corpus, a trained neural machine translation model ?? , hyperparameter K.
2: repeat
3:
t = t + 1.
4:
Randomly pick a source sentence x from the training dataset.
5:
Generate two partial translations yp,1 , yp,2 for x using ?? with random early stop.
6:
Generate K complete translations for each partial translation using ?? and beam search.
Denote this set of complete target sentences as S(yp,1 ) and S(yp,2 ).
7:
Compute the BLEU score for each sentence in S(yp,1 ) and S(yp,2 ).
8:
Calculate the average BLEU score for each partial translation according to Eqn.(6)
9:
Gradient Decent on stochastic loss defined in Eqn.(7).
10: until converge
11: Output: Value network with parameter ?
In conventional Monte-Carlo method for value function estimation, people usually use a regression
model to approximate the value function, i.e., learn a mapping from (x, yp ) ? avg_bleu(x, yp ) by
minimizing the mean square error (MSE). In this paper, we take an alternative objective function
which is shown to be more effective in experiments. We hope the value network we learn is accurate
and useful in differentiating good and bad examples. Thus we use pairwise ranking loss instead of
MSE loss.
To be concrete, we sample two partial sentences yp,1 and yp,2 for each x. We hope the predicted
score of (x, yp,1 ) can be larger than that of (x, yp,2 ) by certain margin if avg_bleu(x, yp,1 ) >
avg_bleu(x, yp,2 ). Denote ? as the parameter of the value function described in Section 4.1. We
design the loss function as follows:
X
L(?) =
ev? (x,yp,2 )?v? (x,yp,1 ) ,
(7)
(x,yp,1 ,yp,2 )
where avg_bleu(x, yp,1 ) > avg_bleu(x, yp,2 ).
Algorithm 2 Beam search with value network in NMT
1: Input: Testing example x, neural machine translation model P (y|x) with target vocabulary V ,
value network model v(x, y), beam search size K, maximum search depth L, weight ?.
2: Set S = ?, U = ? as candidate sets.
3: repeat
4:
t = t + 1.
5:
Uexpand ? {yi + {w}|yi ? U, w ? V }.
6:
U ? {top (K ? |S|) candidates that maximize
? ? 1t log P (y|x) + (1 ? ?) ? log v(x, y)|y ? Uexpand }
7:
Ucomplete ? {y|y ? U, yt = EOS}
8:
U ? U \ Ucomplete
9:
S ? S ? Ucomplete
10: until |S| = K or t = L
1
11: Output: y = argmaxy?S?U ? ? |y|
log P (y|x) + (1 ? ?) ? log v(x, y)
5
4.4
Inference
Since the value network estimates the long-term reward of a state, it will be helpful to enhance the
decoding process of NMT. For example, in a certain decoding step, the NMT model prefers word
w1 over w2 according to the conditional probability, but it does not know that picking w2 will be a
better choice for future decoding. As the value network provides sufficient information on the future
reward, if the value network outputs show that picking w2 is better than picking w1 , we can take both
NMT probability and future reward into consideration to choose a better action.
In this paper, we simply linearly combine the outputs of the NMT model and the value network,
which is motivated by the success of AlphaGo [11]. We first compute the normalized log probability
of each candidate, and then linearly combine it with the logarithmic value of the reward. In detail,
given a translation model P (y|x), a value network v(x, y) and a hyperparameter ? ? (0, 1), the score
of partial sequence y for x is computed by
??
1
log P (y|x) + (1 ? ?) ? log v(x, y),
|y|
(8)
where |y| is the length of y. The details of the decoding process are presented in Algorithm 2, and we
call our neural network-based decoding algorithm NMT-VNN for short.
5
Experiments
5.1
Settings
We compare our proposed NMT-VNN with two baselines. The first one is classic NMT with beam
search [2] (NMT-BS). The second one [16] trains a predictor that can evaluate the quality of any
partial sequence, e.g., partial BLEU score 1 , and then it uses the predictor to select words instead of
the probability. The main difference between [16] and ours is that they predict the local improvement
of BLEU for any single word, while ours aims at predicting the final BLEU score and use the predicted
score to select words. We refer their work as beam search optimization (we call it NMT-BSO). For
NMT-BS, we directly used the open source code [2]. NMT-BSO was implemented by ourselves
based on the open source code [2].
We tested our proposed algorithms and the baselines on three pairs of languages: English?French
(En?Fr), English?German (En?De), and Chinese?English (Zh?En). In detail, we used the same
bilingual corpora from WMT? 14 as used in [2] , which contains 12M, 4.5M and 10M training data
for each task. Following common practices, for En?Fr and En?De, we concatenated newstest2012
and newstest2013 as the validation set, and used newstest2014 as the testing set. For Zh?En, we
used NIST 2006 and NIST 2008 datasets for testing, and use NIST 2004 dataset for validation. For
all datasets in Chinese, we used a public tool for word segmentation. In all experiments, validation
sets were only used for early-stopping and hyperparameter tuning.
For NMT-VNN and NMT-BS, we need to train an NMT model first. We followed [2] to set
experimental parameters to train the NMT model. For each language, we constructed the vocabulary
with the most common 30K words in the parallel corpora, and out-of-vocabulary words were replaced
with a special token ?UNK". Each word was embedded into a vector space of 620 dimensions, and
the dimension of the recurrent unit was 1000. We removed sentences with more than 50 words from
the training set. Batch size was set as 80 with 20 batches pre-fetched and sorted by sentence lengths.
The NMT model was trained with asynchronized SGD on four K40m GPUs for about seven days.
For NMT-BSO, we implemented the algorithm and the model was trained in the same environment.
For the value network used in NMT-VNN, we set the same parameters for the encoder-decoder
layers as the NMT model. Additionally, in the SM module and CC module, we set function ?SM
and ?CC as single-layer feed forward networks with 1000 output nodes. In Algorithm 1, we set
the hyperparameter K = 20 to estimate the value of any partial sequence. We adapted mini-batch
training with batch size to be 80, and the value network model was trained with AdaDelta [21] on
one K40m GPU for about three days.
If the ground truth is y ? , the partial bleu on the partial sequence y<t at step t is defined as the BLEU score
?
on y<t and y<t
.
1
6
During testing, the hyperparameter ? for NMT-VNN was set by cross validation. For En?Fr, En?De
and Zh?En tasks, we found setting ? to be 0.85, 0.9 and 0.8 respectively are the best choices. We
used the BLEU score [8] as the evaluation metric, which is computed by the multi-bleu.perl script2 .
We set the beam search size to be 12 for all the algorithms following the common practice [12].
(a) En?Fr
(b) En?De
(c) Zh?En NIST 2006
(d) Zh?En NIST 2008
Figure 2: Translation results on the test sets of three tasks
5.2
Overall Results
We report the experimental results in this subsection. From Table 1 we can see that our NMT-VNN
algorithm outperforms the baseline algorithms on all tasks.
For English?French task and English?German task, NMT-VNN outperforms the baseline NMT-BS
by about 1.03/1.3 points. As the only difference between the two algorithms is that our NMT-VNN
additionally uses the outputs of value network to enhance decoding, we can conclude that such
additional knowledge provides useful information to help the NMT model. Our method outperforms
NMT-BSO by about 0.31/0.33 points. Since NMT-BSO only uses a local BLEU predictor to estimate
the partial BLEU score while ours predicts the future performance, our proposed value network which
considers long-term benefit is more powerful.
The performance of our NMT-VNN is much better than NMT-BS and NMT-BSO for
Chinese?English tasks. NMT-VNN outperforms the baseline NMT-BS by about 1.4/1.82 points on
NIST 2006 and NIST 2008, and outperforms NMT-BSO by about 1.01/0.72 points. We plot BLEU
scores with respect to the length of source sentences in Figure 2 for all the tasks. From the figures, we
can see that our NMT-VNN algorithm outperforms the baseline algorithms in almost all the ranges of
length.
Furthermore, we also test our value network on a deep NMT model in which the encoder and decoder
are both stacked 4-layer LSTMs. The result also shows that we can get 0.33 points improvement on
English?French task. These results demonstrate the effectiveness and robustness of our NMT-VNN
algorithm.
2
https://github.com/moses-smt/mosesdecoder/blob/master/scripts/generic/multi-bleu.perl. For final evaluation we use corpus-level BLEU, while for the value network training we use sentence-level BLEU as in
[1].
7
Table 1: Overall Performance
En?Fr En?De Zh?En NIST06 Zh?En NIST08 En?Fr Deep
NMT-BS 30.51
NMT-BSO 31.23
NMT-VNN 31.54
5.3
15.67
16.64
16.97
36.2
36.59
37.6
29.4
30.5
31.22
37.86
?
38.19
Analysis on Value Network
We further look into the learnt value network and conduct some analysis to better understand it.
First, as we use an additional component during decoding, it will affect the efficiency of the translation process. As the designed value network architecture is similar to the basic NMT model, the
computational complexity is similar to the NMT model and the two processes can be run in parallel.
Second, it has been observed that the accuracy of NMT is sometimes very sensitive to the size of
beam search on certain tasks. As the beam size grows, the accuracy will drop drastically. [14] argues
this is because the training of NMT favors short but inadequate translation candidates. We also
observe this phenomenon on English?German translation. However, we show that by using value
network, such shortage can be largely avoided. We tested the accuracy of our algorithm with different
beam sizes, as shown in Figure 3.(a). It can be seen that NMT-VNN is much more stable than the
original NMT without value network: its accuracy only differs a little for different beam sizes while
NMT-BS drops more than 0.5 point when the beam size is large.
(a)
(b)
Figure 3: (a). BLEU scores of En?De task w.r.t different beam size. (b). BLEU scores of En?De
task w.r.t different hyperparameter ?.
Third, we tested the performances of NMT-VNN using different hyperparameter ? during decoding
for English?German task. As can be seen from the figure, the performance is stable for the ? ranging
from 0.7 to 0.95, and slightly drops for a smaller ?. This shows that our proposed algorithm is robust
to the hyperparameter.
6
Conclusions and Future Work
In this work we developed a new decoding scheme that incorporates value networks for neural
machine translation. By introducing the value network, the new decoding scheme considers not
only the local conditional probability of a candidate word, but also its long-term reward for future
decoding. Experiments on three translation tasks verify the effectiveness of the new scheme. We plan
to explore the following directions in the future. First, it is interesting to investigate how to design
better structures for the value network. Second, the idea of using value networks is quite general, and
we will extend it to other sequence-to-sequence learning tasks, such as image captioning and dialog
systems.
8
Acknowledgments
This work was partially supported by National Basic Research Program of China (973 Program)
(grant no. 2015CB352502), NSFC (61573026). We would like to thank the anonymous reviewers for
their valuable comments on our paper.
References
[1] D. Bahdanau, P. Brakel, K. Xu, A. Goyal, R. Lowe, J. Pineau, A. Courville, and Y. Bengio. An
actor-critic algorithm for sequence prediction. ICLR, 2017.
[2] D. Bahdanau, K. Cho, and Y. Bengio. Neural machine translation by jointly learning to align
and translate. ICLR, 2015.
[3] S. Bengio, O. Vinyals, N. Jaitly, and N. Shazeer. Scheduled sampling for sequence prediction
with recurrent neural networks. In Advances in Neural Information Processing Systems, pages
1171?1179, 2015.
[4] K. Cho, B. van Merrienboer, C. Gulcehre, D. Bahdanau, F. Bougares, H. Schwenk, and
Y. Bengio. Learning phrase representations using rnn encoder?decoder for statistical machine
translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1724?1734, Doha, Qatar, October 2014. Association for
Computational Linguistics.
[5] D. He, Y. Xia, T. Qin, L. Wang, N. Yu, T. Liu, and W.-Y. Ma. Dual learning for machine
translation. In Advances in Neural Information Processing Systems, pages 820?828, 2016.
[6] S. Jean, K. Cho, R. Memisevic, and Y. Bengio. On using very large target vocabulary for
neural machine translation. In Proceedings of the 53rd Annual Meeting of the Association for
Computational Linguistics and the 7th International Joint Conference on Natural Language
Processing (Volume 1: Long Papers), pages 1?10, Beijing, China, July 2015. Association for
Computational Linguistics.
[7] V. Mnih, K. Kavukcuoglu, D. Silver, A. A. Rusu, J. Veness, M. G. Bellemare, A. Graves,
M. Riedmiller, A. K. Fidjeland, G. Ostrovski, et al. Human-level control through deep reinforcement learning. Nature, 518(7540):529?533, 2015.
[8] K. Papineni, S. Roukos, T. Ward, and W.-J. Zhu. Bleu: a method for automatic evaluation of
machine translation. In Proceedings of the 40th annual meeting on association for computational
linguistics, pages 311?318. Association for Computational Linguistics, 2002.
[9] M. Ranzato, S. Chopra, M. Auli, and W. Zaremba. Sequence level training with recurrent neural
networks. ICLR, 2016.
[10] S. Shen, Y. Cheng, Z. He, W. He, H. Wu, M. Sun, and Y. Liu. Minimum risk training for neural
machine translation. In ACL, 2016.
[11] D. Silver, A. Huang, C. J. Maddison, A. Guez, L. Sifre, G. Van Den Driessche, J. Schrittwieser,
I. Antonoglou, V. Panneershelvam, M. Lanctot, et al. Mastering the game of go with deep neural
networks and tree search. Nature, 529(7587):484?489, 2016.
[12] I. Sutskever, O. Vinyals, and Q. V. Le. Sequence to sequence learning with neural networks. In
Advances in neural information processing systems, pages 3104?3112, 2014.
[13] G. Tesauro. Td-gammon, a self-teaching backgammon program, achieves master-level play.
Neural computation, 6(2):215?219, 1994.
[14] Z. Tu, Y. Liu, L. Shang, X. Liu, and H. Li. Neural machine translation with reconstruction. In
AAAI, pages 3097?3103, 2017.
[15] Z. Tu, Z. Lu, Y. Liu, X. Liu, and H. Li. Coverage-based neural machine translation. CoRR,
abs/1601.04811, 2016.
9
[16] S. Wiseman and A. M. Rush. Sequence-to-sequence learning as beam-search optimization. In
EMNLP, 2016.
[17] L. Wu, Y. Xia, L. Zhao, F. Tian, T. Qin, J. Lai, and T.-Y. Liu. Adversarial neural machine
translation. arXiv preprint arXiv:1704.06933, 2017.
[18] Y. Wu, M. Schuster, Z. Chen, Q. V. Le, M. Norouzi, W. Macherey, M. Krikun, Y. Cao, Q. Gao,
K. Macherey, et al. Google?s neural machine translation system: Bridging the gap between
human and machine translation. arXiv preprint arXiv:1609.08144, 2016.
[19] Y. Xia, T. Qin, W. Chen, J. Bian, N. Yu, and T.-Y. Liu. Dual supervised learning. In ICML,
2017.
[20] Y. Xia, F. Tian, L. Wu, J. Lin, T. Qin, N. Yu, and T. Liu. Deliberation networks: Sequence
generation beyond one-pass decoding. In 31st Annual Conference on Neural Information
Processing Systems (NIPS), 2017.
[21] M. D. Zeiler. Adadelta: an adaptive learning rate method. arXiv preprint arXiv:1212.5701,
2012.
[22] J. Zhou, Y. Cao, X. Wang, P. Li, and W. Xu. Deep recurrent models with fast-forward
connections for neural machine translation. arXiv preprint arXiv:1606.04199, 2016.
10
| 6622 |@word middle:1 briefly:1 open:2 pick:1 sgd:1 liu:10 contains:2 score:21 qatar:1 ours:3 past:1 outperforms:6 current:1 com:3 contextual:1 activation:1 guez:1 attracted:1 gpu:1 concatenate:2 designed:2 plot:1 drop:3 krikun:1 short:6 provides:3 contribute:2 node:1 c2:3 constructed:1 become:1 htx:3 incorrect:1 combine:2 introduce:1 pairwise:1 expected:3 dialog:1 multi:2 inspired:1 td:1 little:1 considering:1 provided:1 estimating:1 maximizes:2 what:1 bso:8 developed:1 finding:2 ended:1 y3:1 every:1 expands:1 golden:1 tie:2 exactly:2 fcc:1 zaremba:1 facto:1 control:1 unit:4 grant:1 before:1 local:4 tends:1 encoding:1 nsfc:1 meet:1 might:1 acl:1 china:3 studied:1 challenging:1 ease:1 range:1 tian:2 averaged:1 acknowledgment:1 testing:8 practice:3 goyal:1 differs:1 x3:1 riedmiller:1 rnn:5 yan:2 empirical:1 significantly:1 liwei:1 outputted:3 matching:6 word:27 xtx:2 refers:1 pre:1 gammon:1 get:1 cannot:4 context:12 cb352502:1 risk:1 bellemare:1 conventional:2 reviewer:1 center:1 yt:8 exposure:1 attention:5 starting:1 go:1 shen:1 classic:1 handle:1 target:23 pt:1 play:1 decode:1 exact:1 us:4 jaitly:1 adadelta:2 predicts:3 observed:3 module:19 preprint:4 wang:2 calculate:2 connected:2 sun:1 ranzato:1 removed:1 valuable:1 mentioned:1 environment:1 complexity:2 reward:14 ideally:1 trained:4 weakly:1 efficiency:1 completely:1 joint:1 schwenk:1 tx:2 lu2:1 train:4 stacked:1 fast:1 describe:1 effective:2 monte:3 choosing:2 eos:2 quite:1 heuristic:1 widely:1 larger:1 jean:1 encoder:13 favor:1 ward:1 jointly:1 final:3 deliberation:1 sequence:21 blob:1 propose:3 reconstruction:1 coming:1 fr:6 qin:4 tu:2 cao:2 translate:2 achieve:1 papineni:1 sutskever:1 optimum:1 r1:2 captioning:1 generating:1 perfect:1 silver:2 help:3 recurrent:8 develop:4 h3:1 school:1 coverage:9 predicted:6 implemented:3 direction:2 shrunk:1 stochastic:1 human:2 public:1 alphago:2 anonymous:1 merrienboer:1 hold:1 considered:1 ground:3 exp:1 mapping:1 predict:4 achieves:1 adopt:1 early:3 estimation:3 label:1 currently:1 sensitive:1 tool:1 weighted:1 hope:3 aim:5 newstest2013:1 avoid:1 zhou:1 rusu:1 encode:1 focus:1 improvement:2 backgammon:1 likelihood:2 adversarial:1 baseline:7 summarizing:1 wang1:1 helpful:1 inference:1 stopping:1 entire:1 typically:1 hidden:3 tao:1 issue:4 wanglw:1 unk:1 overall:2 dual:2 plan:1 special:1 beach:1 sampling:2 veness:1 look:2 yu:3 icml:1 future:11 np:1 report:1 develops:1 randomly:3 national:1 replaced:1 argmax:2 ourselves:1 microsoft:3 ab:1 ostrovski:1 investigate:1 mnih:1 evaluation:4 analyzed:1 argmaxy:1 myopic:7 accurate:1 fsm:1 partial:21 conduct:2 tree:1 pku:2 rush:1 industry:1 temple:1 wiseman:1 phrase:1 introducing:1 predictor:4 successful:1 inadequate:1 eec:1 learnt:2 cho:3 st:2 lstm:2 international:1 memisevic:1 decoding:28 pool:1 enhance:2 picking:3 concrete:1 w1:2 aaai:1 choose:1 huang:1 emnlp:2 zhao:1 yp:29 li:3 de:8 summarized:1 ranking:2 script:1 try:1 lowe:1 h1:3 picked:1 characterizes:1 maintains:1 parallel:2 contribution:1 square:1 accuracy:5 largely:2 norouzi:1 kavukcuoglu:1 lu:1 carlo:3 cc:7 taoqin:1 ty:2 obvious:1 associated:1 di:1 stop:2 proved:1 dataset:3 popular:3 knowledge:3 subsection:1 organized:1 segmentation:1 feed:3 day:2 follow:1 methodology:1 bian:1 supervised:1 done:1 furthermore:1 until:3 eqn:2 lstms:1 google:1 french:3 pineau:1 quality:2 scheduled:2 newstest2014:1 believe:1 grows:1 usa:1 concept:2 y2:2 normalized:1 verify:1 read:1 laboratory:1 semantic:6 during:9 game:1 self:1 complete:6 demonstrate:2 vnn:15 argues:1 ranging:1 image:1 consideration:1 novel:1 sigmoid:1 common:3 rl:2 exponentially:2 volume:1 discussed:1 extend:1 association:5 he:3 bougares:1 mellon:1 refer:1 tuning:1 rd:1 doha:1 automatic:1 similarly:2 teaching:1 language:7 wmt:1 stable:2 actor:1 similarity:1 align:1 own:1 recent:1 tesauro:1 certain:5 success:2 meeting:2 yi:2 scoring:1 seen:2 minimum:1 additional:2 employed:1 converge:1 maximize:1 july:1 afterwards:1 full:1 offer:1 long:15 cross:1 lin:1 lai:1 paired:1 peking:2 impact:1 prediction:9 regression:1 basic:2 cmu:1 metric:1 arxiv:8 iteration:1 sometimes:2 represent:1 beam:26 receive:1 c1:3 source:20 w2:3 nmt:69 pooling:7 smt:1 comment:1 bahdanau:3 incorporates:1 effectiveness:3 call:4 chopra:1 leverage:1 bengio:5 decent:1 affect:1 finish:2 architecture:4 reduce:1 idea:1 cn:3 motivated:1 bridging:1 action:5 prefers:1 deep:6 ignored:1 useful:2 covered:1 shortage:1 locally:1 reduced:1 generate:6 http:1 exist:1 moses:1 estimated:1 correctly:2 carnegie:1 hyperparameter:8 key:1 four:1 kept:1 ptx:1 year:2 beijing:2 run:1 powerful:1 master:2 almost:1 reasonable:2 wu:4 fetched:1 lanctot:1 layer:8 hi:2 ct:5 followed:1 courville:1 cheng:1 annual:3 adapted:1 x2:2 generates:1 aspect:1 expanded:1 gpus:1 according:2 terminates:1 describes:1 slightly:1 y0:2 smaller:1 mastering:1 b:8 hl:1 den:1 gradually:1 discus:1 r3:2 mechanism:3 german:4 know:1 antonoglou:1 end:6 adopted:2 available:1 gulcehre:1 panneershelvam:1 observe:2 generic:1 appending:1 alternative:1 robustness:2 batch:4 original:1 top:5 remaining:1 denotes:1 include:1 completed:1 linguistics:5 zeiler:1 exploit:1 sighted:2 concatenated:1 build:3 chinese:3 seeking:1 objective:3 parametric:1 yty:1 rt:6 gradient:1 iclr:3 thank:1 fidjeland:1 decoder:15 w0:3 maddison:1 seven:1 mail:1 argue:1 considers:4 extent:1 collected:1 bleu:25 length:5 code:2 mini:1 minimizing:1 newstest2012:1 schrittwieser:1 unfortunately:2 october:1 negative:1 append:2 design:5 motivates:1 policy:3 summarization:2 gated:1 datasets:2 sm:7 nist:7 looking:1 y1:4 shazeer:1 auli:1 community:1 yingce:1 pair:3 moe:1 gru:2 c3:3 sentence:48 connection:1 learned:1 nip:2 address:2 beyond:1 usually:3 perception:1 mismatch:1 ev:1 program:3 perl:2 built:1 including:1 memory:1 endto:1 ranked:1 rely:1 natural:2 predicting:1 zhu:1 scheme:5 improve:2 github:1 technology:2 review:1 literature:2 zh:7 graf:1 embedded:1 loss:5 fully:4 asynchronized:1 macherey:2 generation:2 interesting:1 he1:1 validation:4 h2:2 sufficient:1 critic:1 roukos:1 translation:55 token:1 repeat:2 last:1 supported:1 english:9 drastically:1 bias:9 understand:1 institute:1 differentiating:1 benefit:1 van:2 xia:4 depth:2 vocabulary:4 dimension:2 k40m:2 computes:1 forward:6 commonly:1 reinforcement:5 adaptive:1 avoided:1 sifre:1 brakel:1 approximate:1 keep:2 corpus:5 conclude:1 xi:2 search:28 mosesdecoder:1 table:2 additionally:2 learn:4 reasonably:1 terminate:1 ca:1 robust:1 nature:2 mse:2 cl:1 main:2 linearly:2 big:1 motivation:1 bilingual:3 scored:1 whole:1 x1:2 xu:2 en:20 fashion:1 candidate:15 third:1 learns:1 bad:3 symbol:1 r2:2 cumulated:1 corr:1 ci:1 conditioned:3 margin:1 chen:2 gap:1 logarithmic:1 simply:1 likely:1 explore:1 gao:1 vinyals:2 partially:1 scalar:1 driessche:1 truth:4 relies:1 ma:1 conditional:10 sorted:1 replace:1 hard:1 specifically:1 shang:1 called:1 total:1 pas:1 experimental:3 select:3 people:1 ustc:1 phenomenon:1 evaluate:2 tested:3 schuster:1 |
6,216 | 6,623 | Parametric Simplex Method for Sparse Learning
?
Haotian Pang? Robert Vanderbei? Han Liu??
Tuo Zhao?
?
?
Princeton University Tencent AI Lab Northwestern University ? Georgia Tech?
Abstract
High dimensional sparse learning has imposed a great computational challenge to
large scale data analysis. In this paper, we are interested in a broad class of sparse
learning approaches formulated as linear programs parametrized by a regularization
factor, and solve them by the parametric simplex method (PSM). Our parametric
simplex method offers significant advantages over other competing methods: (1)
PSM naturally obtains the complete solution path for all values of the regularization
parameter; (2) PSM provides a high precision dual certificate stopping criterion;
(3) PSM yields sparse solutions through very few iterations, and the solution
sparsity significantly reduces the computational cost per iteration. Particularly,
we demonstrate the superiority of PSM over various sparse learning approaches,
including Dantzig selector for sparse linear regression, LAD-Lasso for sparse robust
linear regression, CLIME for sparse precision matrix estimation, sparse differential
network estimation, and sparse Linear Programming Discriminant (LPD) analysis.
We then provide sufficient conditions under which PSM always outputs sparse
solutions such that its computational performance can be significantly boosted.
Thorough numerical experiments are provided to demonstrate the outstanding
performance of the PSM method.
1
Introduction
A broad class of sparse learning approaches can be formulated as high dimensional optimization
problems. A well known example is Dantzig Selector, which minimizes a sparsity-inducing `1 norm
with an `1 norm constraint. Specifically, let X 2 Rn?d be a design matrix, y 2 Rn be a response
vector, and ? 2 Rd be the model parameter. Dantzig Selector can be formulated as the solution to the
following convex program,
?b = argmin k?k1 s.t. kX > (y X?)k1 ? .
(1.1)
?
where k ? k1 and k ? k1 denote the `1 and `1 norms respectively, and > 0 is a regularization factor.
Candes and Tao (2007) suggest to rewrite (1.1) as a linear program solved by linear program solvers.
Dantzig Selector motivates many other sparse learning approaches, which also apply a regularization
factor to tune the desired solution. Many of them can be written as a linear program in the following
generic form with either equality constraints:
max(c + c?)> x s.t. Ax = b + ?b, x 0,
(1.2)
x
or inequality constraints:
max(c + c?)> x
x
s.t. Ax ? b + ?b, x
0.
(1.3)
Existing literature usually suggests the popular interior point method (IPM) to solve (1.2) and (1.3).
The interior point method is famous for solving linear programs in polynomial time. Specifically,
the interior point method uses the log barrier to handle the constraints, and rewrites (1.2) or (1.3)
?
Correspondence to Tuo Zhao: [email protected].
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
as a unconstrained program, which is further solved by the Newton?s method. Since the log barrier
requires the Newton?s method to only iterate within the interior of the feasible region, IPM cannot
yield exact sparse iterates, and cannot take advantage of sparsity to boost the computation.
An alternative approach is the simplex method. From a geometric perspective, the classical simplex
method iterates over the vertices of a polytope. Algebraically, the algorithm involves moving from
one partition of the basic and nonbasic variables to another. Each partition deviates from the previous
in that one basic variable gets swapped with one nonbasic variable in a process called pivoting.
Different variants of the simplex method are defined by different rules of pivoting. The simplex
method has been shown to work well in practice, even though its worst-case iteration complexity has
been shown to scale exponentially with the problem scale in existing literature.
More recently, some researchers propose to use alternating direction methods of multipliers (ADMM)
to directly solve (1.1) without reparametrization as a linear program. Although these methods enjoy
O(1/T ) convergence rates based on variational inequality criteria, where T is the number of iterations.
ADMM can be viewed as an exterior point method, and always gives infeasible solutions within
finite number of iterations. We often observe that after ADMM takes a large number of iterations,
the solutions still suffer from significant feasibility violation. These methods work well only for
moderate scale problems (e.g., d < 1000). For larger d, ADMM becomes less competitive.
These methods, though popular, are usually designed for solving (1.2) and (1.3) for one single
regularization factor. This is not satisfactory, since an appropriate choice of is usually unknown.
Thus, one usually expects an algorithm to obtain multiple solutions tuned over a reasonable range of
values for . For each value of , we need to solve a linear program from scratch, and it is therefore
often very inefficient for high dimensional problems.
To overcome the above drawbacks, we propose to solve both (1.2) and (1.3) by a variant of the
parametric simplex method (PSM) in a principled manner (Murty, 1983; Vanderbei, 1995). Specifically, the parametric simplex method parametrizes (1.2) and (1.3) using the unknown regularization
factor as a ?parameter?. This eventually yields a piecewise linear solution path for a sequence of
regularization factors. Such a varying parameter scheme is also called homotopy optimization in
existing literature. PSM relies some special rules to iteratively choose the pair of variables to swap,
which algebraically calculates the solution path during each pivoting. PSM terminates at a value of
parameter, where we have successfully solved the full solution path to the original problem. Although
in the worst-case scenario, PSM can take an exponential number of pivots to find an optimal solution
path. Our empirical results suggest that the number of iterations is roughly linear in the number of
nonzero variables for large regularization factors with sparse optima. This means that the desired
sparse solutions can often be found using very few pivots.
Several optimization methods for solving (1.1) are closely related to PSM. However, there is a lack
of generic design in these methods. Their methods, for example, the simplex method proposed in
Yao and Lee (2014) can be viewed as a special example of our proposed PSM, where the perturbation
is only considered on the right-hand-side of the inequalities in the constraints. DASSO algorithm
computes the entire coefficient path of Dantzig selector by a simplex-like algorithm. Zhu et al. (2004)
propose a similar algorithm which takes advantage of the piece-wise linearity of the problem and
computes the whole solution path on `1 -SVM. These methods can be considered as similar algorithms
derived from PSM but only applied to special cases, where the entire solution path is computed but
an accurate dual certificate stopping criterion is not provided.
Notations: We denote all zero and all one vectors by 1 and 0 respectively.
P Given a vector a =
(a1 , ..., ad )> 2 Rd , we define the number of nonzero entries kak0 =
j 1(aj 6= 0), kak1 =
P
P 2
2
j |aj |, kak2 =
j aj , and kak1 = maxj |aj |. When comparing vectors, ? ? and ??? mean
component-wise comparison. Given a matrix A 2 Rd?d with entries aP
jk , we use |||A||| to denote
entry-wise norms and kAk to denote matrix norms. Accordingly |||A|||0 = j,k 1(ajk 6= 0), |||A|||1 =
P
P
P
|, kAk1 = maxk j |ajk |, kAk1 = maxj k |ajk |, kAk2 =
j,k |ajk |, |||A|||1 = maxj,k |ajkP
maxkak2 ?1 kAak2 , and kAk2F = j,k a2jk . We denote A\i\j as the submatrix of A with i-th row
and j-th column removed. We denote Ai\j as the i-th row of A with its j-th entry removed and
A\ij as the j-th column of A with its i-th entry removed. For any subset G of {1, 2, . . . , d}, we let
AG denote the submatrix of A 2 Rp?d consisting of the corresponding columns of A. The notation
A 0 means all of A?s entries are nonnegative. Similarly, for a vector a 2 Rd , we let aG denote the
subvector of a associated with the indices in G. Finally, Id denotes the d-dimensional identity matrix
2
and ei denotes vector that has a one in its i-th entry and zero elsewhere. In a large matrix, we leave a
submatrix blank when all of its entries are zeros.
2
Background
Many sparse learning approaches are formulated as convex programs in a generic form:
min L(?) + k?k1 ,
?
(2.1)
where L(?) is a convex loss function, and > 0 is a regularization factor controlling bias and
variance. Moreover, if L(?) is smooth, we can also consider an alternative formulation:
min k?k1 s.t. krL(?)k1 ? ,
(2.2)
?
where rL(?) is the gradient of L(?), and > 0 is a regularization factor. As will be shown later,
both (2.2) and (2.1) are naturally suited for our algorithm, when L(?) is piecewise linear or quadratic
respectively. Our algorithm yields a piecewise-linear solution path as a function of by varying
from large to small.
Before we proceed with our proposed algorithm, we first introduce the sparse learning problems of
our interests, including sparse linear regression, sparse linear classification, and undirected graph
estimation. Due to space limit, we only present three examples, and defer the others to the appendix.
Dantzig Selector: The first problem is sparse linear regression. Let y 2 Rn be a response vector
and X 2 Rn?d be the design matrix. We consider a linear model y = X?? + ?, where ?? 2 Rd is the
unknown regression coefficient vector, and ? is the observational noise vector. Here we are interested
in a high dimensional regime: d is much larger than n, i.e., d
n, and many entries in ?? are zero,
?
?
0
i.e., k? k0 = s ? n. To get a sparse estimator of ? , machine learning researchers and statisticians
have proposed numerous approaches including Lasso (Tibshirani, 1996), Dantzig Selector (Candes
and Tao, 2007) and LAD-Lasso (Wang et al., 2007).
The Dantzig selector is formulated as the solution to the following convex program:
min k?k1 subject to kX > (y X?)k1 ? .
?
By setting ? = ?+
linear program:
? with
min
1> (?+ + ? )
+
? ,?
s.t.
?
?j+
= ?j ? 1(?j > 0) and
X >X
X >X
X >X
X >X
??
?+
?
?j+
?
(2.3)
= ?j ? 1(?j < 0), we rewrite (2.3) as a
?
?
?
1 + X >y
,
1 X >y
?+ , ?
0. (2.4)
By complementary slackness, we can guarantee that the optimal ?j+ ?s and ?j ?s are nonnegative and
complementary to each other. Note that (2.4) fits into our parametric linear program as (1.3) with
? >
?
? > ?
? +?
X X
X >X
X y
?b = 1, c? = 0, x = ?
,
c
=
1,
.
A=
,
b
=
?
X >X X >X
X >y
Sparse Support Vector Machine: The second problem is Sparse SVM (Support Vector Machine),
which is a model-free discriminative modeling approach (Zhu et al., 2004). Given n independent
and identically distributed samples (x1 , y1 ), ..., (xn , yn ), where xi 2 Rd is the feature vector and
yi 2 {1, 1} is the binary label. Similar to sparse linear regression, we are interested in the high
dimensional regime. To obtain a sparse SVM classifier, we solve the following convex program:
n
X
min
[1 yi (?0 + ?> xi )]+ s.t. k?k1 ? ,
(2.5)
?0 ,?
i=1
where ?0 2 R and ? 2 Rd . Given a new sample z 2 Rd , Sparse SVM classifier predicts its label
by sign(?0 + ?> z). Let ti = 1 yi (?0 + ?> xi ) for i = 1, ..., n. Then ti can be expressed as
ti = t+
ti . Notice [1 yi (?0 + ?> xi )]+ can be represented by t+
i
i . We split ? and ?0 into
positive and negative parts as well: ? = ?+ ? and ?0 = ?0+ + ?0 and add slack variable w
to the constraint so that the constraint becomes equality: ?+ + ? + w = 1, w
0. Now we
cast the problem into the equality parametric simplex form (1.2). We identify each component
>
of (1.2) as the following: x = t+ t
2 R(n+1)?(2n+3d+2) , x
?+ ?
?0+ ?0 w
>
2 R2n+3d+2
, c? = 0 2 R2n+3d+2 , b? =
?
I
In Z
Z y
y
>
>
2
1> 0 2 Rn+1 , ?b = 0> 1 2 Rn+1 , and A = n
1> 1>
1>
R(n+1)?(2n+3d+2) , where Z = (y1 x1 , . . . , yn xn )> 2 Rn?d .
0, c =
1>
0>
0>
0>
0
0
0>
3
Differential Graph Estimation: The third problem is the differential graph estimation, which
aims to identify the difference between two undirected graphs (Zhao et al., 2013; Danaher et al.,
2013). The related applications in biological and medical research can be found in existing literature
(Hudson et al., 2009; Bandyopadhyaya et al., 2010; Ideker and Krogan, 2012). Specifically, given n1
i.i.d. samples x1 , ..., xn from Nd (?0X , ?0X ) and n2 i.i.d. samples y1 , ..., yn from Nd (?0Y , ?0Y ) We
are interested in estimating the difference of the precision matrices of two distributions:
0
= (?0X ) 1 (?0Y ) 1 .
P n1
We define the empirical covariance matrices as: SX = n11 j=1
(x
x
?)(xj x
?)> andSY =
P
P
Pn j
n2
n
1
1
1
>
y?)(yj x
?) , where x
? = n1 j=1 xj and y? = n2 j=1 yj . Then Zhao et al. (2013)
j=1 (yj
n2
propose to estimate 0 by solving the following problem:
min ||| |||1 s.t. |||SX SY SX + SY |||1 ? ,
(2.6)
where SX and SY are the empirical covariance matrices. As can be seen, (2.6) is essentially a special
example of a more general parametric linear program as follows,
min |||D|||1 s.t. |||XDZ Y |||1 ? ,
(2.7)
D
where D 2 R
d1 ?d2
,X 2R
m1 ?d1
, Z 2 Rd2 ?m2 and Y 2 Rm1 ?m2 are given data matrices.
Instead of directly solving (2.7), we consider a reparametrization by introducing an axillary variable
C = XD. Similar to CLIME, we decompose D = D+ + D , and eventually rewrite (2.7) as
min 1> (D+ + D )1 s.t. |||CZ
D + ,D
Y |||1 ? ,
X(D+
D ) = C,
D+ , D
0, (2.8)
Let vec(D+ ), vec(D ), vec(C) and vec(Y ) be the vectors obtained by stacking the columns of
matrices D+ , D C and Y , respectively. We write (2.8) as a parametric linear program,
0 0
1
X
X0
Im1 d2
A
Z0
I m1 m2
A=@
Z0
I m1 m2
0
1
0
1
X
z11 Im1
???
zd 2 1 I m 1
B
C
B
C
..
..
..
..
with X 0 = @
A 2 Rm1 d2 ?d1 d2 , Z 0 = @
A 2
.
.
.
.
X
z1m2 Im1 ? ? ?
zd 2 m 2 I m 1
Rm1 m2 ?m1 d2 , where zij denotes the (i, j) entry of matrix Z;
?
?>
x = vec(D+ ) vec(D ) vec(C) w 2 R2d1 d2 +m1 d2 +2m1 m2 ,
where w 2 R2m1 m2 is nonnegative slack variable vector used to make the inequality become
?
?>
an equality. Moreover, we also have b = 0> vec(Y )
2 Rm1 d2 +2m1 m2 , where
vec(Y )
the first m1 d2 components of vector b are 0 and the rest components are from matrix Y ; ?b =
>
0> 1> 1> 2 Rm1 d2 +2m1 m2 , where the first m1 d2 components of vector ?b are 0 and the rest
>
2m1 m2 components are 1; c =
1>
1> 0> 0> 2 R2d1 d2 +m1 d2 +2m1 m2 , where the first
2d1 d2 components of vector c are 1 and the rest m1 d2 + 2m1 m2 components are 0.
3
Homotopy Parametric Simplex Method
We first briefly review the primal simplex method for linear programming, and then derive the
proposed algorithm.
Preliminaries: We consider a standard linear program as follows,
max c> x
x
s.t. Ax = b,
x
0 x 2 Rn ,
(3.1)
where A 2 Rm?n , b 2 Rm and c 2 Rn are given. Without loss of generality, we assume that m ? n
and matrix A has full row rank m. Throughout our analysis, we assume that an optimal solution
exists (it needs not be unique). The primal simplex method starts from a basic feasible solution (to
be defined shortly?but geometrically can be thought of as any vertex of the feasible polytope) and
proceeds step-by-step (vertex-by-vertex) to the optimal solution. Various techniques exist to find the
first feasible solution, which is often referred to the Phase I method. See Vanderbei (1995); Murty
(1983); Dantzig (1951).
4
Algebraically, a basic solution corresponds to a partition of the indices {1, . . . , n} into m basic
indices denoted B and n m non-basic indices denoted N . Note that not all partitions are allowed?
the submatrix of A consisting of the columns of A associated with the basic indices, denoted AB ,
must be invertible. The submatrix of A corresponding to the nonbasic indices is denoted AN .
Suppressing the fact that the columns have been rearranged, we can write A = [AN , AB ]. If
we rearrange the rows
? of x and?c in the same way, we can introduce a corresponding partition of
x
c
these vectors: x = N , c = N . From the commutative property of addition, we rewrite the
xB
cB
constraint as AN xN + AB xB = b. Since the matrix AB is assumed to be invertible, we can express
xB in terms of xN as follows:
(3.2)
xB = x?B AB 1 AN xN ,
where we have written x?B as an abbreviation for AB 1 b. This rearrangement of the equality constraints
is called a dictionary because the basic variables are defined as functions of the nonbasic variables.
Denoting the objective c> x by ?, then we also can write:
>
?
? >
? = c> x = c>
(zN
) xN ,
B xB + c N xN = ?
1
1
1
?
>
?
?
>
where ? = cB AB b, xB = AB b and zN = (AB AN ) cB cN .
(3.3)
We call equations (3.2) and (3.3) the primal dictionary associated with the current basis B. Corresponding to each dictionary, there is a basic solution (also called a dictionary solution) obtained by setting the nonbasic variables to zero and reading off values of the basic variables:
xN = 0,
xB = x?B . This particular ?solution? satisfies the equality constraints of the problem by
construction. To be a feasible solution, one only needs to check that the values of the basic variables
are nonnegative. Therefore, we say that a basic solution is a basic feasible solution if x?B 0.
The dual of (3.1) is given by
max b> y
y
s.t. A> y
z = c,
z
0 z 2 Rn , y 2 Rm .
In this case, we separate variable z into basic and nonbasic parts as before: [z] =
corresponding dual dictionary is given by:
?
z N = zN
+ (AB 1 AN )> zB ,
?=
??
(x?B )> zB ,
(3.4)
?
zN
. The
zB
(3.5)
1
1
?
?
where ? denotes the objective function in the (3.4), ? ? = c>
B AB b, xB = AB b and zN =
1
>
(AB AN ) cB cN .
For each dictionary, we set xN and zB to 0 (complementarity) and read off the solutions to xB and
zN according to (3.2) and (3.5). Next, we remove one basic index and replacing it with a nonbasic
index, and then get an updated dictionary. The simplex method produces a sequence of steps to
adjacent bases such that the value of the objective function is always increasing at each step. Primal
feasibility requires that xB 0, so while we update the dictionary, primal feasibility must always be
satisfied. This process will stop when zN 0 (dual feasibility), since it satisfies primal feasibility,
dual feasibility and complementarity (i.e., the optimality condition).
Parametric Simplex Method: We derive the parametric simplex method used to find the full
solution path while solving the parametric linear programming problem only once. A few variants of
the simplex method are proposed with different rules for choosing the pair of variables to swap at each
iteration. Here we describe the rule used by the parametric simplex method: we add some positive
perturbations (?b and c?) times a positive parameter to both objective function and the right hand side
of the primal problem. The purpose of doing this is to guarantee the primal and dual feasibility when
is large. Since the problem is already primal feasible and dual feasible, there is no phase I stage
required for the parametric simplex method. Furthermore, if the i-th entry of b or the j-th entry of c
has already satisfied the feasibility condition (bi 0 or cj ? 0), then the corresponding perturbation
?bi or c?j to that entry is allowed to be 0. With these perturbations, (3.1) becomes:
max(c + c?)> x s.t. Ax = b + ?b, x 0 x 2 Rn .
(3.6)
x
We separate the perturbation vectors into basic and nonbasic parts as well and write down the the
dictionary with perturbations corresponding to (3.2),(3.3), and (3.5) as:
?
xB = (x?B + x
?B ) AB 1 AN xN , ? = ? ? (zN
+ z?N )> xN ,
(3.7)
5
where
x?B
?
zN = (zN
+ z?N ) + (AB 1 AN )> zB ,
1
= AB b,
?
zN
1
>
= (AB AN ) cB
? = ? ? (x?B + x
? B ) > zB ,
cN , x
?B = AB 1?b and z?N = (AB 1 AN )> c?B
(3.8)
c?N .
?
When is large, the dictionary will be both primal and dual feasible (x?B + x
?B 0 and zN
+ z?N
0). The corresponding primal solution is simple: xB = x?B + x
?B and xN = 0. This solution is
valid until hits a lower bound which breaks the feasibility. The smallest value of without break
any feasibility is given by
?
?
= min{ : zN
+ z?N
0 and x?B + x
?B
(3.9)
0}.
In other words, the dictionary and its corresponding solution xB =
+ x
?B and xN = 0 is optimal
for the value of 2 [ ? , max ], where
?
?
?
zN
x?
?
= max
maxj2N , z?Nj >0 z?Nj , maxi2B,?xBi >0 x?BBi ,
(3.10)
x?B
j
max
= min
?
minj2N , z?Nj <0
?
zN
i
j
z?Nj
,
mini2B,?xBi <0
x?
Bi
x
? Bi
?
.
(3.11)
Note that although initially the perturbations are nonnegative, as the dictionary gets updated, the
perturbation does not necessarily maintain nonnegativity. For each dictionary, there is a corresponding
interval of given by (3.10) and (3.11). We have characterized the optimal solution for this interval,
and these together give us the solution path of the original parametric linear programming problem.
Next, we show how the dictionary gets updated as the leaving variable and entering variable swap.
We expect that after swapping the entering variable j and leaving variable i, the new solution in the
dictionary (3.7) and (3.8) would slightly change to:
x?j = t, x
??j = t?, zi? = s, z?i? = s?, x?B
x?B t xB , x
?B
x
?B t? xB ,
?
zN
?
zN
s zN , z?N
z?N s? zN ,
?
where t and t are the primal step length for the primal basic variables and perturbations, s and s? are
the dual step length for the dual nonbasic variables and perturbations, xB and zN are the primal
and dual step directions, respectively. We explain how to find these values in details now.
?
There is either a j 2 N for which zN
+ z?N = 0 or an i 2 B for which x?B + x
?B = 0 in (3.9). If
it corresponds to a nonbasic index j, then we do one step of the primal simplex. In this case, we
declare j as the entering variable, then we need to find the primal step direction xB . After the
entering variable j has been selected, xN changes from 0 to tej , where t is the primal step length.
Then according to (3.7), we have that xB = (x?B + x
?B ) AB 1 AN tej . The step direction xB
is given by xB = AB 1 AN ej . We next select the leaving variable. In order to maintain primal
feasibility, we need to keep xB
0, therefore, the leaving variable i is selected such that i 2 B
achieves the maximal value of x? + x?i x?i . It only remains to show how zN changes. Since i is the
i
leaving variable, according to (3.8), we have zN = (AB 1 AN )> ei . After we know the entering
variables, the primal and dual step directions, the primal and dual step lengths can be found as
z?
x?
z?
t = xi i , t? = x?xi i , s = zj j , s? = zj j .
If, on the other hand, the constraint in (3.9) corresponds to a basic index i, we declare i as the leaving
variable, then similar calculation can be made based on the dual simplex method (apply the primal
simplex method to the dual problem). Since it is very similar to the primal simplex method, we omit
the detailed description.
The algorithm will terminate whenever ? ? 0. The corresponding solution is optimal since our
dictionary always satisfies primal feasibility, dual feasibility and complementary slackness condition.
The only concern during the entire process of the parametric simplex method is that does not equal
to zero, so as long as can be set to be zero, we have the optimal solution to the original problem.
We summarize the parametric simplex method in Algorithm 1:
The following theorem shows that the updated basic and nonbasic partition gives the optimal solution.
Theorem 3.1. For a given dictionary with parameter in the form of (3.7) and (3.8), let B be a basic
index set and N be an nonbasic index set. Assume this dictionary is optimal for 2 [ ? , max ],
where ? and max are given by (3.10) and (3.11), respectively. The updated dictionary with basic
index set B ? and nonbasic index set N ? is still optimal at = ? .
6
Write down the dictionary as in (3.7) and (3.8);
Find ? given by (3.10);
while ? > 0 do
if the constraint in (3.10) corresponds to an index j 2 N then
Declare xj as the entering variable;
Compute primal step direction. xB = AB 1 AN ej ;
Select leaving variable. Need to find i 2 B that achieves the maximal value of
1
x?
i+
Compute dual step direction. It is given by zN = (AB AN )> ei ;
else if the constraint in (3.10) corresponds to an index i 2 B then
Declare zi as the leaving variable;
Compute dual step direction. zN = (AB 1 AN )> ei ;
Select entering variable. Need to find j 2 N that achieves the maximal value of
1
xi
?x
?
zj? +
i
;
zj
?z
?j
;
Compute primal step direction. It is given by xB = AB AN ej ;
Compute the dual and primal step lengths for both variables and perturbations:
t=
x?i
,
xi
x
?i
t? =
,
xi
s=
zj?
,
zj
s? =
z?j
.
zj
Update the primal and dual solutions:
x?j = t,
x?B
x
?j = t?,
x?B
zi? = s,
?
zN
z?i = s?,
?
zN
t xB , x
?B
x
?B t? xB
s zN , z?N
z?N s? zN .
Update the basic and nonbasic index sets B := B \ {i} \ {j} and N := N \ {j} \ {i}. Write
down the new dictionary and compute ? given by (3.10);
end
Set the nonbasic variables as 0s and read the values of the basic variables.
Algorithm 1: The parametric simplex method
During each iteration, there is an optimal solution corresponding to 2 [ ? , max ]. Notice each of
these ?s range is determined by a partition between basic and nonbasic variables, and the number
of the partition into basic and nonbasic variables is finite. Thus after finite steps, we must find the
optimal solution corresponding to all values.
Theory: We present our theoretical analysis on solving Dantzig selector using PSM. Specifically,
given X 2 Rn?d , y 2 Rn , we consider a linear model y = X?? + ?, where ?? is the unknown sparse
regression coefficient vector with k?? k0 = s? , and ? ? N (0, 2 In ). We show that PSM always
maintains a pair of sparse primal and dual solutions. Therefore, the computation cost within each
iteration of PSM can be significantly reduced. Before we proceed with our main result, we introduce
two assumptions. The first assumption requires the regularization factor to be sufficiently large.
Assumption 3.2. Suppose that PSM solves (2.3) for a regularization sequence { K }N
K=0 . The
smallest regularization factor N satisfies
r
log d
4kX > ?k1 for some generic constant C.
N =C
n
Existing literature has extensively studied Assumption 3.2 for high dimensional statistical theories.
Such an assumption enforces all regularization parameters to be sufficiently large in order to eliminate
irrelevant coordinates along the regularization path. Note that Assumption 3.2 is deterministic
for any given N . Existing literature has verified that for sparse linear regression models, given
? ? N (0, 2 In ), Assumption 3.2 holds with overwhelming probability.
Before we present the second assumption, we define the largest and smallest s-sparse eigenvalues of
n 1 X > X respectively as follows.
Definition 3.3. Given an integer s 1, we define
X >X
nk k22
T
?+ (s) = sup
k k0 ?s
and
7
X >X
nk k22
T
? (s) =
inf
k k0 ?s
.
Assumption 3.4. Given k?? k0 ? s? , there exists an integer se such that
se 100?s? , ?+ (s? + se) < +1, and ?e (s? + se) > 0,
where ? is defined as ? = ?+ (s? + se)/e
? (s? + se).
Assumption 3.4 guarantees that n 1 X > X satisfies the sparse eigenvalue conditions as long as the
e along the solution path. That is closely related to
number of active irrelevant blocks never exceeds 2s
the restricted isometry property (RIP) and restricted eigenvalue (RE) conditions, which have been
extensively studied in existing literature.
We then characterize the sparsity of the primal and dual solutions within each iteration.
Theorem 3.5 (Primal and Dual Sparsity). Suppose that Assumptions 3.2 and 3.4 hold. We consider
an alterantive formulation to the Dantzig selector,
0
?b = argmin k?k1 subject to rj L(?) ? 0 , rj L(?) ? 0 .
(3.12)
?
0
0
0
0
0
Let ?
b = [b
?1 , ..., ?
bd , bd+1 , ..., b2d ]> denote the optimal dual variables to (3.12). For any
0
0
we have kb
? k0 + kb k0 ? s? + se. Moreover, given design matrix satisfying
kXS> XS (XS> XS )
where ? > 0 is a generic constant, S = {j |
?j?
1
k1 ? 1
0
,
?,
0
6= 0} and S = {j | ?j? = 0}, we have k?b k0 ? s? .
The proof of Theorem 3.5 is provided in Appendix B. Theorem 3.5 shows that within each iteration,
both primal and dual variables are sparse, i.e., the number of nonzero entries are far smaller than d.
Therefore, the computation cost within each iteration of PSM can be significantly reduced by a factor
of O(d/s? ). This partially justifies the superior performance of PSM in sparse learning.
4
Numerical Experiments
1
Flare
PSM
?
?
Infeasibility
200
300
400
300
100
100
?1
0
?
200
Values of Lambda along the Path
?
?2
Nonzero Entries of the Response Vector
True Value
Estimated Path
400
?
?
?
?
500
2
In this section, we present some numerical experiments and give some insights about how the
parametric simplex method solves different linear programming problems. We verify the following
assertions: (1) The parametric simplex method requires very few iterations to identify the nonzero
component if the original problem is sparse. (2) The parametric simplex method is able to find the
full solution path with high precision by solving the problem only once in an efficient and scalable
manner. (3) The parametric simplex method maintains the feasibility of the problem up to machine
precision along the solution path.
0
0
?3
?
0
5
10
Iteration
15
0
5
10
Iteration
15
0
100
200
300
400
500
Iteration
(a) Solution Path
(b) Parameter Path (Rescaled by n)
(c) Feasibility Violation
Figure 1: Dantzig selector method: (a) The solution path of the parametric simplex method; (b) The parameter
path of the parametric simplex method; (c) Feasibility violation along the solution path.
Solution path of Dantzig selector: We start with a simple example that illustrates how the recovered solution path of the Dantzig selector model changes as the parametric simplex method iterates.
We adopt the example used in Candes and Tao (2007). The design matrix X has n = 100 rows and
d = 250 columns. The entries of X are generated from an array of independent Gaussian random
variables that are then Gaussianized so that each column has a given norm. We randomly select s = 8
entries from the response vector ?0 , and set them as ?i0 = si (1 + ai ), where si = 1 or 1, with
probability 1/2 and ai ? N (0, 1). The other entries of ?0 are set to zero. We form y =pX?0 + ?,
where ?i ? N (0, ), with = 1. We stop the parametric simplex method when ? n log d/n.
The solution path of the result is shown in Figure 1(a). We see that our method correctly identifies all
nonzero entries of ? in less than 10 iterations. Some small overestimations occur in a few iterations
after all nonzero entries have been identified. We also show how the parameter evolves as the
parametric simplex method iterates in Figure 1(b). As we see, decreases sharply to less than 5
8
after all nonzero components have been identified. This reconciles with the theorem we developed.
The algorithm itself only requires a very small number of iterations to correctly identify the nonzero
entries of ?. In our example, each iteration in the parametric simplex method identifies one or two
non-sparse entries in ?.
Feasibility of Dantzig Selector: Another advantage of the parametric simplex method is that the
solution is always feasible along the path while other estimating methods usually generate infeasible
solutions along the path. We compare our algorithm with ?flare? (Li et al., 2015) which uses the
Alternating Direction Method of Multipliers (ADMM) using the same example described above. We
i
compute the values of kX > X?i X > yk1
i along the solution path, where ? is the i-th basic
solution (with corresponding i ) obtained while the parametric simplex method is iterating. Without
any doubts, we always obtain 0s during each iteration. We plug the same list of i into ?flare? and
compute the solution path for this list as well. As shown in Table 1, the parametric simplex method is
always feasible along the path since it is solving each iteration up to machine precision; while the
solution path of the ADMM is almost always breaking the feasibility by a large amount, especially in
the first few iterations which correspond to large values. Each experiment is repeated for 100 times.
Table 1: Average feasibility violation with standard errors along the solution path
ADMM
PSM
Maximum violation
498(122)
0(0)
Minimum Violation
143(73.2)
0(0)
Performance Benchmark of Dantzig Selector: In this part, we compare the timing performance
of our algorithm with R package ?flare?. We fix the sample size n to be 200 and vary the data
dimension d from 100 to 5000. Again, each entries of X is independent Gaussian and Gaussianized
such that the column has uniform norm. We randomly select 2% entries from vector ? to be nonzero
and each entry is chosen as ? N (0, 1). We compute y = X? + ?, with ?i ?
pN (0, 1) and try to
recover vector ?, given X and y. Our method stops when is less than 2
log d/n, such that
the full solution path for all the values of up to this value is computed by the parametric simplex
method. In ?flare?, we estimate ? when is equal to the value in the Dantzig selector model. This
means ?flare? has much less computation task than the parametric simplex method. As we can see in
Table 2, our method has a much better performance than ?flare? in terms of speed. We compare and
present the timing performance of the two algorithms in seconds and each experiment is repeated for
100 times. In practice, only very few iterations is required when the response vector ? is sparse.
Table 2: Average timing performance (in seconds) with standard errors in the parentheses on Dantzig selector
Flare
PSM
500
19.5(2.72)
2.40(0.220)
1000
44.4(2.54)
29.7(1.39)
2000
142(11.5)
47.5(2.27)
5000
1500(231)
649(89.8)
Performance Benchmark of Differential Network: We now apply this optimization method to
the Differential Network model. We need the difference between two inverse covariance matrices
to be sparse. We generate ?0x = U > ?U , where ? 2 Rd?d is a diagonal matrix and its entries are
i.i.d. and uniform on [1, 2], and U 2 Rd?d is a random matrix with i.i.d. entries from N (0, 1). Let
D1 2 Rd?d be a random sparse symmetric matrix with a certain sparsity level. Each entry of D1
is i.i.d. and from N (0, 1). We set D = D1 + 2| min (D1 )|Id in order to guarantee the positive
definiteness of D, where min (D1 ) is the smallest eigenvalue of D1 . Finally, we let ?0x = (?0x ) 1
and ?0y = ?0x + D.
We then generate data of sample size n = 100. The corresponding sample covariance matrices SX
and SY are also computed based on the data. We are not able to find other software which can
efficiently solve this problem, so we only list the timing performance of our algorithm as dimension
d varies from 25 to 200 in Table 3. We stop our algorithm whenever the solution achieved the desired
sparsity level. When d = 25, 50 and 100, the sparsity level of D1 is set to be 0.02 and when d = 150
and 200, the sparsity level of D1 is set to be 0.002. Each experiment is repeated for 100 times.
Table 3: Average timing performance (in seconds) and iteration numbers with standard errors in the parentheses
on differential network
Timing
Iteration Number
25
0.0185(0.00689)
15.5(7.00)
50
0.376(0.124)
55.3(18.8)
9
100
6.81(2.38)
164(58.2)
150
13.41(1.26)
85.8(16.7)
200
46.88(7.24)
140(26.2)
References
BANDYOPADHYAYA , S., M EHTA , M., K UO , D., S UNG , M.-K., C HUANG , R., JAEHNIG , E. J.,
B ODENMILLER , B., L ICON , K., C OPELAND , W., S HALES , M., F IEDLER , D., D UTKOWSKI ,
J., G U?NOL? , A., ATTIKUM , H. V., S HOKAT, K. M., KOLODNER , R. D., H UH , W.-K.,
A EBERSOLD , R., K EOGH , M.-C. and K ROGAN , N. J. (2010). Rewiring of genetic networks in
response to dna damage. Science Signaling 330 1385?1389.
B ?HLMANN , P. and VAN D E G EER , S. (2011). Statistics for high-dimensional data: methods,
theory and applications. Springer Science & Business Media.
C AI , T. and L IU , W. (2011). A direct estimation approach to sparse linear discriminant analysis.
Journal of the American Statistical Association 106 1566?1578.
C AI , T., L IU , W. and L UO , X. (2011). A constrained l1 minimization approach to sparse precision
matrix estimation. Journal of the American Statistical Association 106 594?607.
C ANDES , E. and TAO , T. (2007). The dantzig selector: Statistical estimation when p is much larger
than n. The Annals of Statistics 35 2313?2351.
DANAHER , P., WANG , P. and W ITTEN , D. M. (2013). The joint graphical lasso for inverse
covariance estimation across multiple classes. Journal of the Royal Statistical Society Series B 7
373?397.
DANTZIG , G. (1951). Linear Programming and Extensions. Princeton University Press.
D EMPSTER , A. (1972). Covariance selection. Biometrics 28 157?175.
G AI , Y., Z HU , L. and L IN , L. (2013). Model selection consistency of dantzig selector. Statistica
Sinica 615?634.
H UDSON , N. J., R EVERTER , A. and DALRYMPLE , B. P. (2009). A differential wiring analysis of
expression data correctly identifies the gene containing the causal mutation. PLoS Computational
Biology. 5.
I DEKER , T. and K ROGAN , N. (2012). Differential network biology. Molecular Systems Biology 5
565.
L I , X., Z HAO , T., Y UAN , X. and L IU , H. (2015). The flare package for hign dimensional linear
regression and precision matrix estimation in r. Journal of Machine Learning Research 16 553?557.
M URTY, K. (1983). Linear Programming. Wiley, New York, NY.
T IBSHIRANI , R. (1996). Regression shrinkage and selection via the lasso. Journal of the Royal
Statistical Society 101 267?288.
VANDERBEI , R. (1995). Linear Programming, Foundations and Extensions. Kluwer.
WANG , H., L I , G. and J IANG , G. (2007). Robust regression shrinkage and consistent variable
selection through the lad-lasso. Journal of Business & Economic Statistics 25 347?355.
YAO , Y. and L EE , Y. (2014). Another look at linear programming for feature selection via methods
of regularization. Statistics and Computing 24 885?905.
Z HAO , S. D., C AI , T. and L I , H. (2013). Direct estimation of differential networks. Biometrika 58
253?268.
Z HOU , H. and H ASTIE , T. (2005). Regularization and variable selection via the elastic net. Journal
of the Royal Statistical Society 67 301?320.
Z HU , J., ROSSET, S., H ASTIE , T. and T IBSHIRANI , R. (2004). 1-norm support vector machines.
Advances in Neural Information Processing Systems 16.
10
| 6623 |@word briefly:1 polynomial:1 norm:8 nd:2 d2:15 hu:2 covariance:6 ipm:2 liu:1 series:1 zij:1 tuned:1 denoting:1 suppressing:1 genetic:1 amp:1 existing:7 blank:1 comparing:1 current:1 recovered:1 si:2 written:2 must:3 bd:2 hou:1 numerical:3 partition:8 remove:1 designed:1 update:3 rd2:1 selected:2 flare:9 accordingly:1 provides:1 certificate:2 iterates:4 along:10 direct:2 differential:9 become:1 introduce:3 manner:2 x0:1 roughly:1 overwhelming:1 solver:1 increasing:1 becomes:3 provided:3 estimating:2 linearity:1 notation:2 moreover:3 medium:1 argmin:2 minimizes:1 developed:1 ag:2 nj:4 guarantee:4 thorough:1 ti:4 xd:1 biometrika:1 classifier:2 rm:3 hit:1 uo:2 medical:1 superiority:1 omit:1 enjoy:1 yn:3 before:4 positive:4 hudson:1 declare:4 timing:6 limit:1 id:2 path:33 ap:1 dantzig:21 studied:2 xbi:2 suggests:1 nol:1 range:2 bi:4 unique:1 enforces:1 yj:3 practice:2 block:1 signaling:1 empirical:3 significantly:4 murty:2 thought:1 word:1 eer:1 suggest:2 get:5 cannot:2 interior:4 selection:6 imposed:1 deterministic:1 convex:5 m2:12 rule:4 estimator:1 insight:1 array:1 handle:1 coordinate:1 updated:5 annals:1 controlling:1 construction:1 suppose:2 rip:1 exact:1 programming:9 us:2 complementarity:2 satisfying:1 particularly:1 jk:1 predicts:1 yk1:1 solved:3 wang:3 worst:2 region:1 andes:1 decrease:1 removed:3 rescaled:1 plo:1 principled:1 complexity:1 overestimation:1 rewrite:5 solving:9 swap:3 basis:1 uh:1 joint:1 k0:8 various:2 represented:1 tej:2 describe:1 choosing:1 larger:3 solve:7 say:1 statistic:4 itself:1 advantage:4 sequence:3 eigenvalue:4 net:1 propose:4 rewiring:1 maximal:3 kak1:4 description:1 inducing:1 convergence:1 optimum:1 produce:1 leave:1 derive:2 ij:1 solves:2 involves:1 direction:10 drawback:1 closely:2 kb:2 observational:1 fix:1 homotopy:2 decompose:1 preliminary:1 biological:1 extension:2 hold:2 sufficiently:2 considered:2 great:1 cb:5 dictionary:21 achieves:3 smallest:4 adopt:1 vary:1 purpose:1 estimation:11 label:2 largest:1 successfully:1 minimization:1 always:10 gaussian:2 aim:1 pn:2 ej:3 shrinkage:2 boosted:1 varying:2 gatech:1 ax:4 derived:1 rank:1 check:1 tech:1 stopping:2 i0:1 entire:3 eliminate:1 initially:1 interested:4 tao:4 iu:3 dual:26 classification:1 denoted:4 constrained:1 special:4 equal:2 once:2 never:1 ung:1 beach:1 biology:3 broad:2 look:1 simplex:43 parametrizes:1 others:1 piecewise:3 few:7 randomly:2 maxj:3 phase:2 consisting:2 statistician:1 n1:3 maintain:2 ab:26 rearrangement:1 interest:1 violation:6 primal:31 rearrange:1 swapping:1 xb:25 accurate:1 biometrics:1 desired:3 re:1 causal:1 theoretical:1 column:9 modeling:1 assertion:1 zn:29 hlmann:1 cost:3 introducing:1 vertex:4 entry:28 expects:1 subset:1 stacking:1 uniform:2 characterize:1 varies:1 rosset:1 st:1 lee:1 off:2 invertible:2 together:1 yao:2 again:1 satisfied:2 containing:1 choose:1 huang:1 lambda:1 american:2 zhao:5 inefficient:1 li:1 doubt:1 coefficient:3 ad:1 piece:1 later:1 break:2 try:1 lab:1 doing:1 sup:1 competitive:1 start:2 maintains:2 reparametrization:2 recover:1 candes:3 defer:1 clime:2 mutation:1 pang:1 variance:1 efficiently:1 sy:4 yield:4 identify:4 correspond:1 famous:1 researcher:2 icon:1 explain:1 whenever:2 definition:1 iang:1 naturally:2 associated:3 vanderbei:4 proof:1 stop:4 popular:2 cj:1 response:6 formulation:2 though:2 generality:1 furthermore:1 stage:1 until:1 hand:3 ei:4 replacing:1 lack:1 slackness:2 aj:4 usa:1 k22:2 verify:1 multiplier:2 true:1 regularization:17 equality:6 alternating:2 read:2 iteratively:1 satisfactory:1 nonzero:10 entering:7 symmetric:1 adjacent:1 wiring:1 during:4 a2jk:1 kak:1 criterion:3 complete:1 demonstrate:2 l1:1 variational:1 wise:3 krl:1 recently:1 pivoting:3 superior:1 rl:1 exponentially:1 association:2 m1:15 kluwer:1 lad:3 significant:2 vec:9 ai:8 rd:11 unconstrained:1 b2d:1 consistency:1 similarly:1 moving:1 han:1 add:2 base:1 isometry:1 perspective:1 moderate:1 irrelevant:2 inf:1 scenario:1 certain:1 inequality:4 binary:1 yi:4 seen:1 minimum:1 algebraically:3 ibshirani:2 multiple:2 full:5 rj:2 reduces:1 danaher:2 z11:1 smooth:1 exceeds:1 characterized:1 calculation:1 offer:1 long:3 plug:1 molecular:1 a1:1 feasibility:19 calculates:1 n11:1 variant:3 regression:11 basic:26 scalable:1 essentially:1 parenthesis:2 iteration:27 cz:1 achieved:1 background:1 addition:1 interval:2 else:1 leaving:8 swapped:1 rest:3 subject:2 undirected:2 call:1 integer:2 ee:1 split:1 identically:1 iterate:1 xj:3 fit:1 zi:3 competing:1 lasso:6 identified:2 economic:1 cn:3 pivot:2 expression:1 lpd:1 suffer:1 proceed:2 york:1 iterating:1 detailed:1 se:7 tune:1 amount:1 extensively:2 rearranged:1 reduced:2 generate:3 dna:1 exist:1 zj:7 notice:2 sign:1 estimated:1 per:1 tibshirani:1 correctly:3 zd:2 write:6 express:1 im1:3 verified:1 graph:4 geometrically:1 package:2 psm:23 inverse:2 throughout:1 reasonable:1 almost:1 appendix:2 submatrix:5 bound:1 correspondence:1 quadratic:1 nonnegative:5 uan:1 occur:1 constraint:13 sharply:1 software:1 speed:1 min:12 optimality:1 px:1 according:3 terminates:1 slightly:1 smaller:1 across:1 evolves:1 restricted:2 equation:1 remains:1 slack:2 eventually:2 know:1 end:1 apply:3 observe:1 generic:5 appropriate:1 alternative:2 shortly:1 rp:1 original:4 denotes:4 graphical:1 newton:2 k1:13 especially:1 classical:1 society:3 objective:4 already:2 parametric:34 damage:1 kak2:2 diagonal:1 gradient:1 separate:2 parametrized:1 kak2f:1 polytope:2 discriminant:2 length:5 index:17 sinica:1 robert:1 hao:2 negative:1 design:5 motivates:1 nonbasic:17 unknown:4 ideker:1 benchmark:2 finite:3 maxk:1 y1:3 rn:13 perturbation:11 tuo:3 pair:3 subvector:1 cast:1 required:2 boost:1 nip:1 able:2 proceeds:1 usually:5 regime:2 sparsity:9 challenge:1 reading:1 bbi:1 program:17 summarize:1 including:3 max:11 royal:3 dasso:1 business:2 zhu:2 scheme:1 numerous:1 identifies:3 deviate:1 review:1 literature:7 geometric:1 loss:2 expect:1 northwestern:1 foundation:1 sufficient:1 consistent:1 r2n:2 row:5 elsewhere:1 free:1 infeasible:2 side:2 bias:1 barrier:2 empster:1 sparse:41 distributed:1 van:1 overcome:1 dimension:2 xn:15 valid:1 computes:2 made:1 far:1 obtains:1 selector:19 keep:1 gene:1 active:1 astie:2 assumed:1 krogan:1 discriminative:1 xi:9 infeasibility:1 table:6 gaussianized:2 terminate:1 robust:2 ca:1 exterior:1 elastic:1 tencent:1 necessarily:1 reconciles:1 main:1 ehta:1 statistica:1 whole:1 noise:1 n2:4 allowed:2 complementary:3 repeated:3 x1:3 referred:1 georgia:1 definiteness:1 ny:1 wiley:1 precision:8 nonnegativity:1 exponential:1 isye:1 breaking:1 third:1 z0:2 down:3 theorem:6 hale:1 list:3 x:3 svm:4 concern:1 exists:2 justifies:1 commutative:1 illustrates:1 kx:4 sx:5 nk:2 suited:1 expressed:1 partially:1 springer:1 corresponds:5 satisfies:5 relies:1 abbreviation:1 kak0:1 viewed:2 formulated:5 identity:1 rm1:5 admm:7 feasible:11 ajk:4 change:4 specifically:5 determined:1 zb:6 called:4 kxs:1 select:5 support:3 outstanding:1 princeton:2 d1:12 scratch:1 |
6,217 | 6,624 | Group Sparse Additive Machine
1
Hong Chen1 , Xiaoqian Wang1 , Cheng Deng2 , Heng Huang1?
Department of Electrical and Computer Engineering, University of Pittsburgh, USA
2
School of Electronic Engineering, Xidian University, China
[email protected],[email protected]
[email protected],[email protected]
Abstract
A family of learning algorithms generated from additive models have attracted
much attention recently for their flexibility and interpretability in high dimensional
data analysis. Among them, learning models with grouped variables have shown
competitive performance for prediction and variable selection. However, the
previous works mainly focus on the least squares regression problem, not the
classification task. Thus, it is desired to design the new additive classification
model with variable selection capability for many real-world applications which
focus on high-dimensional data classification. To address this challenging problem,
in this paper, we investigate the classification with group sparse additive models
in reproducing kernel Hilbert spaces. A novel classification method, called as
group sparse additive machine (GroupSAM), is proposed to explore and utilize
the structure information among the input variables. Generalization error bound is
derived and proved by integrating the sample error analysis with empirical covering
numbers and the hypothesis error estimate with the stepping stone technique. Our
new bound shows that GroupSAM can achieve a satisfactory learning rate with
polynomial decay. Experimental results on synthetic data and seven benchmark
datasets consistently show the effectiveness of our new approach.
1
Introduction
The additive models based on statistical learning methods have been playing important roles for
the high-dimensional data analysis due to their well performance on prediction tasks and variable
selection (deep learning models often don?t work well when the number of training data is not
large). In essential, additive models inherit the representation flexibility of nonlinear models and
the interpretability of linear models. For a learning approach under additive models, there are two
key components: the hypothesis function space and the regularizer to address certain restrictions
on estimator. Different from traditional learning methods, the hypothesis space used in additive
models is relied on the decomposition of input vector. Usually, each input vector X ? Rp is divided
into p parts directly [17, 30, 6, 28] or some subgroups according to prior structural information
among input variables [27, 26]. The component function is defined on each decomposed input and
the hypothesis function is constructed by the sum of all component functions. Typical examples of
hypothesis space include the kernel-based function space [16, 6, 11] and the spline-based function
space [13, 15, 10, 30]. Moreover, the Tikhonov regularization scheme has been used extensively for
constructing the additive models, where the regularizer is employed to control the complexity of
hypothesis space. The examples of regularizer include the kernel-norm regularization associated with
the reproducing kernel Hilbert space (RKHS) [5, 6, 11] and various sparse regularization [17, 30, 26].
More recently several group sparse additive models have been proposed to tackle the high-dimensional
regression problem due to their nice theoretical properties and empirical effectiveness [15, 10,
?
Corresponding author
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
26]. However, most existing additive model based learning approaches are mainly limited to the
least squares regression problem and spline-based hypothesis spaces. Surprisingly, there is no any
algorithmic design and theoretical analysis for classification problem with group sparse additive
models in RKHS. This paper focuses on filling in this gap on algorithmic design and learning theory
for additive models. A novel sparse classification algorithm, called as group sparse additive machine
(GroupSAM), is proposed under a coefficient-based regularized framework, which is connected to
the linear programming support vector machine (LPSVM) [22, 24]. By incorporating the grouped
variables with prior structural information and the `2,1 -norm based structured sparse regularizer, the
new GroupSAM model can conduct the nonlinear classification and variable selection simultaneously.
Similar to the sparse additive machine (SAM) in [30], our GroupSAM model can be efficiently solved
via proximal gradient descent algorithm. The main contributions of this paper can summarized in
two-fold:
? A new group sparse nonlinear classification algorithm (GroupSAM) is proposed by extending
the previous additive regression models to the classification setting, which contains the
LPSVM with additive kernel as its special setting. To the best of our knowledge, this is the
first algorithmic exploration of additive classification models with group sparsity.
? Theoretical analysis and empirical evaluations on generalization ability are presented to support the effectiveness of GroupSAM. Based on constructive analysis on the hypothesis error,
we get the estimate on the excess generalization error, which shows that our GroupSAM
model can achieve the fast convergence rate O(n?1 ) under mild conditions. Experimental
results demonstrate the competitive performance of GroupSAM over the related methods on
both simulated and real data.
Before ending this section, we discuss related works. In [5], support vector machine (SVM) with
additive kernels was proposed and its classification consistency was established. Although this
method can also be used for grouped variables, it only focuses on the kernel-norm regularizer without
addressing the sparseness for variable selection. In [30], the SAM was proposed to deal with the
sparse representation on the orthogonal basis of hypothesis space. Despite good computation and
generalization performance, SAM does not explore the structure information of input variables and
ignores the interactions among variables. More important, different from finite splines approximation
in [30], our approach enables us to estimate each component function directly in RKHS. As illustrated
in [20, 14], the RKHS-based method is flexible and only depends on few tuning parameters, but the
commonly used spline methods need specify the number of basis functions and the sequence of knots.
It should be noticed that the group sparse additive models (GroupSpAM in [26]) also address the
sparsity on the grouped variables. However, there are key differences between GroupSAM and
GroupSpAM: 1) Hypothesis space. The component functions in our model are obtained by searching
in kernel-based data dependent hypothesis spaces, but the method in [26] uses data independent
hypothesis space (not associated with kernel). As shown in [19, 18, 4, 25], the data dependent
hypothesis space can provide much more adaptivity and flexibility for nonlinear prediction. The
advantage of kernel-based hypothesis space for additive models is also discussed in [14]. 2) Loss
function. The hinge loss used in our classification model is different from the least-squares loss in
[26]. 3) Optimization. Our GroupSAM only needs to construct one component function for each
variable group, but the model in [26] needs to find the component functions for each variable in a
group. Thus, our method is usually more efficient. Due to the kernel-based component function and
non-smooth hinge loss, the optimization of GroupSpAM can not be extended to our model directly. 4)
Learning theory. We establish the generalization bound of GroupSAM by the error estimate technique
with data dependent hypothesis spaces, while the error bound is not covered in [26].
Now, we present a brief summary in Table 1 to better illustrate the differences of our GroupSAM
with other methods.
The rest of this paper is organized as follows. In next section, we revisit the related classification
formulations and propose the new GroupSAM model. Theoretical analysis on generalization error
bound is established in Section 3. In Section 4, experimental results on both simulated examples and
real data are presented and discussed. Finally, Section 5 concludes this paper.
2
Table 1: Properties of different additive models.
SAM [30]
Group Lasso[27] GroupSpAM [26] GroupSAM
Hypothesis space
data-independent data-independent data-independent data-dependent
Loss function
hinge loss
least-square
least-square
hinge loss
No
Yes
Yes
Yes
Group sparsity
Generalization bound
Yes
No
No
Yes
2
Group sparse additive machine
In this section, we first revisit the basic background of binary classification and additive models, and
then introduce our new GroupSAM model.
Let Z := (X , Y) ? Rp+1 , where X ? Rp is a compact input space and Y = {?1, 1} is the set
of labels. We assume that the training samples z := {zi }ni=1 = {(xi , yi )}ni=1 are independently
drawn from an unknown distribution ? on Z, where each xi ? X and yi ? {?1, 1}. Let?s denote
the marginal distribution of ? on X as ?X and denote its conditional distribution for given x ? X as
?(?|x).
For a real-valued function f : X ? R, we define its induced classifier as sgn(f ), where sgn(f )(x) =
1 if f (x) ? 0 and sgn(f )(x) = ?1 if f (x) < 0. The prediction performance of f is measured by
the misclassification error:
Z
R(f ) = Prob{Y f (X) ? 0} =
Prob(Y 6= sgn(f )(x)|x)d?X .
(1)
X
It is well known that the minimizer of R(f ) is the Bayes rule:
Z
fc (x) = sgn
yd?(y|x) = sgn Prob(y = 1|x) ? Prob(y = ?1|x) .
Y
Since the Bayes rule involves the unknown distribution ?, it can not be computed directly. In machine
learning literature, the classification algorithm usually aims to find a good approximation of fc by
minimizing the empirical misclassification risk:
n
1X
I(yi f (xi ) ? 0) ,
n i=1
Rz (f ) =
(2)
where I(A) = 1 if A is true and 0 otherwise. However, the minimization problem associated with
Rz (f ) is NP-hard due to the 0 ? 1 loss I. To alleviate the computational difficulty, various convex
losses have been introduced to replace the 0 ? 1 loss, e.g., the hinge loss, the least square loss, and
the exponential loss [29, 1, 7]. Among them, the hinge loss is the most popular error metric for
classification problem due to its nice theoretical properties. In this paper, following [5, 30], we use
the hinge loss:
`(y, f (x)) = (1 ? yf (x))+ = max{1 ? yf (x), 0}
to measure the misclassification cost.The expected and empirical risks associated with the hinge loss
are defined respectively as:
Z
E(f ) =
(1 ? yf (x))+ d?(x, y) ,
Z
and
n
Ez (f ) =
1X
(1 ? yi f (xi ))+ .
n i=1
In theory, the excess misclassification error R(sgn(f )) ? R(fc ) can be bounded by the excess convex
risk E(f ) ? E(fc ) [29, 1, 7]. Therefore, the classification algorithm usually is constructed under
structural risk minimization [22] associated with Ez (f ).
3
In this paper, we propose a novel group sparse additive machine (GroupSAM) for nonlinear classification. Let {1, ? ? ? , p} be partitioned into d groups. For each j ? {1, ..., d}, we set X (j) as the
grouped input space and denote f (j) : X (j) ? R as the corresponding component function. Usually,
the groups can be obtained by prior knowledge [26] or be explored by considering the combinations
of input variables [11].
Let each K (j) : X (j) ? X (j) ? R be a Mercer kernel and let HK (j) be the corresponding RKHS
with norm k ? kK (j) . It has been proved in [5] that
H=
d
nX
o
f (j) : f (j) ? HK (j) , 1 ? j ? d
j=1
with norm
kf k2K = inf
d
nX
kf (j) k2K (j) : f =
j=1
d
X
f (j)
o
j=1
is an RKHS associated with the additive kernel K =
Pd
j=1
K (j) .
For any given training set z = {(xi , yi )}ni=1 , the additive model in H can be formulated as:
f?z =
arg min
P
f=
d
j=1
f (j) ?H
d
o
n
X
Ez (f ) + ?
?j kf (j) k2K (j) ,
(3)
j=1
where ? = ?(n) is a positive regularization parameter and {?j } are positive bounded weights for
different variable groups.
The solution f?z in (3) has the following representation:
f?z (x) =
d
X
(j)
f?z (x(j) ) =
j=1
d X
n
X
(j)
(j)
(j)
?
? z,i yi K (j) (xi , x(j) ), ?
? z,i ? R, 1 ? i ? n, 1 ? j ? d .
j=1 i=1
(j)
(j)
(j)
Observe that f?z (x) ? 0 is equivalent to ?
? z,i = 0 for all i. Hence, we expect k?
?z k2 = 0 for
(j)
(j)
??z (j) = (?
?z,1 , ? ? ? , ?
? z,n )T ? Rn if the j-th variable group is not truly informative. This motivation
pushes us to consider the sparsity-induced penalty:
?(f ) = inf
d
nX
?j k?(j) k2 : f =
j=1
d X
n
X
o
(j)
(j)
?i yi K (j) (xi , ?) .
j=1 i=1
This group sparse penalty aims at the variable selection [27] and was introduced into the additive
regression model [26].
Inspired by learning with data dependent hypothesis spaces [19], we introduce the following hypothesis spaces associated with training samples z:
d
n
o
X
Hz = f =
f (j) : f (j) ? Hz(j) ,
(4)
j=1
where
n
n
o
X
(j)
(j)
(j)
Hz(j) = f (j) =
?i K (j) (xi , ?) : ?i ? R .
i=1
Under the group sparse penalty and data dependent hypothesis space, the group sparse additive
machine (GroupSAM) can be written as:
fz = arg min
f ?Hz
n
n1 X
n
o
(1 ? yi f (xi ))+ + ??(f ) ,
i=1
4
(5)
where ? > 0 is a regularization parameter.
(j)
(j)
(j)
Let?s denote ?(j) = (?1 , ? ? ? , ?n )T and Ki
GroupSAM in (5) can be rewritten as:
fz =
d
X
fz(j) =
(j)
(j)
(j)
(j)
= (K (j) (x1 , xi ), ? ? ? , K (j) (xn , xi ))T . The
d X
n
X
(j)
(j)
?z,t K (j) (xt , ?) ,
j=1 t=1
j=1
with
{?z(j) } =
n
n1 X
arg min
?(j) ?Rn ,1?j?d
n
1 ? yi
i=1
d
d
o
X
X
(j)
(Ki )T ?(j) + + ?
?j k?(j) k2 .
j=1
(6)
j=1
The formulation (6) transforms the function-based learning problem (5) into a coefficient-based
learning problem in a finite dimensional vector space. The solution of (5) is spanned naturally by
(j)
the kernelized functions {K (j) (?, xi ))}, rather than B-Spline basis functions [30]. When d = 1,
our GroupSAM model degenerates to the special case which includes the LPSVM loss and the
sparsity regularization term. Compared with LPSVM [22, 24] and SVM with additive kernels [5], our
GroupSAM model imposes the sparsity on variable groups to improve the prediction interpretation of
additive classification model.
For given {?j }, the optimization problem of GroupSAM can be computed efficiently via an accelerated proximal gradient descent algorithm developed in [30]. Due to space limitation, we don?t recall
the optimization algorithm here again.
3
Generalization error bound
In this section, we will derive the estimate on the excess misclassification error R(sgn(fz )) ? R(fc ).
Before providing the main theoretical result, we introduce some necessary assumptions for learning
theory analysis.
Assumption A. The intrinsic distribution ? on Z := X ? Y satisfies the Tsybakov noise condition
with exponent 0 ? q ? ?. That is to say, for some q ? [0, ?) and ? > 0,
?X {x ? X : |Prob(y = 1|x) ? Prob(y = ?1|x)| ? ?t} ? tq , ?t > 0.
(7)
The Tsybakov noise condition was proposed in [21] and has been used extensively for theoretical
analysis of classification algorithms [24, 7, 23, 20]. Indeed, (7) holds with exponent q = 0 for any
distribution and with q = ? for well separated classes.
Now we introduce the empirical covering numbers [8] to measure the capacity of hypothesis space.
Definition 1 Let F be a set of functions on Z with u = {ui }ki=1 ? Z. Define the `2 -empirical
1
Pk
metric as `2,u (f, g) = n1 t=1 (f (ut ) ? g(ut ))2 2 . The covering number of F with `2 -empirical
metric is defined as N2 (F, ?) = supn?N supu?X n N2,u (F, ?), where
l
n
o
[
N2,u (F, ?) = inf l ? N : ?{fi }li=1 ? F s. t. F =
{f ? F : `2,u (f, fi ) ? ?} .
i=1
(j)
Let Br = {f ? HK : kf kK ? r} and Br = {f (j) ? HK (j) : kf (j) kK (j) ? r}.
p
Pd
K (j) (x(j) , x(j) ) < ? and for some s ?
Assumption B. Assume that ? =
j=1 supx(j)
(0, 2), cs > 0,
(j)
log N2 (B1 , ?) ? cs ??s , ?? > 0, j ? {1, ..., d}.
It has been asserted in [6] that under Assumption B the following holds:
log N2 (B1 , ?) ? cs d1+s ??s , ?? > 0.
5
It is worthy noticing that the empirical covering number has been studied extensively in learning
theory literatures [8, 20]. Detailed examples have been provided in Theorem 2 of [19], Lemma 3 of
[18], and Examples 1, 2 of [9]. The capacity condition of additive assumption space just depends
on the dimension of subspace X (j) . When K (j) ? C ? (X (j) ? X (j) ) for every j ? {1, ? ? ? , d}, the
theoretical analysis in [19] assures that Assumption B holds true for:
? 2d0
? d0 +2? , ? ? (0, 1];
2d0
s=
, ? ? [1, 1 + d0 /2];
? dd00 +?
? ? (1 + d0 /2, ?).
? ,
Here d0 denotes the maximum dimension among {X (j) }.
With respect to (3), we introduce the data-free regularized function f? defined by:
f? =
arg min
P
(j) ?H
f= d
j=1 f
d
n
o
X
E(f ) + ?
?j kf (j) k2K (j) .
(8)
j=1
Inspired by the analysis in [6], we define:
D(?) = E(f? ) ? E(fc ) + ?
d
X
?j kf?(j) k2K (j)
(9)
j=1
as the approximation error, which reflects the learning ability of hypothesis space H under Tikhonov
regularization scheme.
The following approximation condition has been studied and used extensively for classification
problems, such as [3, 7, 24, 23]. Please see Examples 3 and 4 in [3] for the explicit version for Soblov
kernel and Gaussian kernel induced reproducing kernel Hilbert space.
Assumption C. There exists an exponent ? ? (0, 1) and a positive constant c? such that:
D(?) ? c? ? ? , ?? > 0.
Now we introduce our main theoretical result on the generalization bound as follows.
Theorem 1 Let 0 < min ?j ? max ?j ? c0 < ? and Assumptions A-C hold true. Take ? = n?? in
j
j
3+5?
(5) for 0 < ? ? min{ 2?s
2s , 2?2? }. For any ? ? (0, 1), there exists a constant C independent of n, ?
such that
R(sgn(fz )) ? R(fc ) ? C log(3/?)n??
with confidence 1 ? ?, where
n q + 1 ?(2? + 1) (q + 1)(2 ? s ? 2s?) 3 + 5? + 2?? ? 2? o
? = min
,
,
,
.
q + 2 2? + 2
4 + 2q + sq
4 + 4?
Theorem 1 demonstrates that GroupSAM in (5) can achieve the convergence rate with polynomial
decay under mild conditions in hypothesis function space. When q ? ?, ? ? 1, and each
1+2?
K (j) ? C ? , the error decay rate of GroupSAM can arbitrarily close to O(n? min{1, 4 } ). Hence,
the fast convergence rate O(n?1 ) can be obtained under proper selections on parameters. To verify
the optimal bound, we need provide the lower bound for the excess misclassification error. This is
beyond the main focus of this paper and we leave it for future study.
Additionally, the consistency of GroupSAM can be guaranteed with the increasing number of training
samples.
Corollary 1 Under conditions in Theorem 1, there holds R(sgn(fz )) ? R(fc ) ? 0 as n ? ?.
To better understand our theoretical result, we compare it with the related works as below:
6
1) Compared with group sparse additive models. Although the asymptotic theory of group sparse
additive models has been well studied in [15, 10, 26], all of them only consider the regression task under the mean square error criterion and basis function expansion. Due to the kernel-based component
function and non-smooth hinge loss, the previous analysis cannot be extended to GroupSAM directly.
2) Compared with classification with additive models. In [30], the convergence rate is presented for
sparse additive machine (SAM), where the input space X is divided into p subspaces directly without
considering the interactions among variables. Different to the sparsity on variable groups in this
paper, SAM is based on the sparse representation of orthonormal basis similar with [15]. In [5], the
consistency of SVM with additive kernel is established, where the kernel-norm regularizer is used.
However, the sparsity on variables and the learning rate are not investigated in previous articles.
3) Compared with the related analysis techniques. While the analysis technique used here is inspired
from [24, 23], it is the first exploration for additive classification model with group sparsity. In
particular, the hypothesis error analysis develops the stepping stone technique from the `1 -norm
regularizer to the group sparse `2,1 -norm regularizer. Our analysis technique also can be applied to
other additive models. For example, we can extend the shrunk additive regression model in [11] to
the sparse classification setting and investigate its generalization bound by the current technique.
Proof sketches of Theorem 1
To get tight error estimate, we introduce the clipping operator ?(f )(x) = max{?1, min{f (x), 1}},
which has been widely used in learning theory literatures, such as [7, 20, 24, 23]. Since R(sgn(fz ))?
R(fc ) can be bounded by E(?(fz )) ? E(fc ), we focus on bounding the excess convex risk.
Using f? as the intermediate function, we can obtain the following error decomposition.
Proposition 1 For fz defined in (5), there holds
R(sgn(fz )) ? R(fc ) ? E(?(fz )) ? E(fc ) ? E1 + E2 + E3 + D(?),
where D(?) is defined in (9),
E1
=
E2
=
E(?(fz )) ? E(fc ) ? Ez (?(fz )) ? Ez (fc ) ,
Ez (f? ) ? Ez (fc ) ? Ez (f? ) ? E(fc ) ,
and
E3 = Ez (?(fz )) + ??(fz ) ? Ez (f? ) + ?
d
X
?j kf?(j) k2K (j) .
j=1
In learning theory literature, E1 + E2 is called as the sample error and E3 is named as the hypothesis
error. Detailed proofs for these error terms are provided in the supplementary materials.
The upper bound of hypothesis error demonstrates that the divergence induced from regularization
and hypothesis space tends to zero as n ? ? under proper selected parameters. To estimate the
hypothesis error E3 , we choose f?z as the stepping stone function to bridge Ez (?(fz )) + ??(fz )
Pd
(j)
and Ez (f? ) + ? j=1 ?j kf? k2K (j) . The proof is inspired from the stepping stone technique for
support vector machine classification [24]. Notice that our analysis is associated with the `2,1 -norm
regularizer while the previous analysis just focuses on the `1 -norm regularization.
The error term E1 reflects the divergence between the expected excess risk E(?(fz )) ? E(fc ) and
the empirical excess risk Ez (?(fz )) ? Ez (fc ). Since fz involves any given z = {(xi , yi )}ni=1 , we
introduce the concentration inequality in [23] to bound E1 . We also bound the error term E2 in terms
of the one-side Bernstein inequality [7].
4
Experiments
To evaluate the performance of our proposed GroupSAM model, we compare our model with the
following methods: SVM (linear SVM with `2 -norm regularization), L1SVM (linear SVM with `1 norm regularization), GaussianSVM (nonlinear SVM using Gaussian kernel), SAM (Sparse Additive
Machine) [30], and GroupSpAM (Group Sparse Additive Models) [26] which is adapted to the
classification setting.
7
Table 2: Classification accuracy comparison on the synthetic data. The upper half shows the results
with 24 features groups, while the lower half corresponds to the results with 300 feature groups. The
table shows the average classification accuracy and the standard deviation in 2-fold cross validation.
SVM
GaussianSVM L1SVM
SAM
GroupSpAM GroupSAM
? = 0.8 0.943?0.011 0.935?0.028 0.925?0.035 0.895?0.021 0.880?0.021 0.953?0.018
? = 0.85 0.943?0.004 0.938?0.011 0.938?0.004 0.783?0.088 0.868?0.178 0.945?0.000
? = 0.9 0.935?0.014 0.925? 0.007 0.938?0.011 0.853? 0.117 0.883?0.011 0.945?0.007
? = 0.8 0.975?0.035 0.975?0.035 0.975?0.035 0.700?0.071 0.275?0.106 1.000?0.000
? = 0.85 0.975?0.035 0.975?0.035 0.975?0.035 0.600?0.141 0.953?0.004 1.000?0.000
? = 0.9 0.975?0.035 0.975?0.035 0.975?0.035 0.525?0.035 0.983?0.004 1.000?0.000
As for evaluation metric, we calculate the classification accuracy, i.e., percentage of correctly labeled
samples in the prediction. In comparison, we adopt 2-fold cross validation and report the average
performance of each method.
We implement SVM, L1SVM and GaussianSVM using the LIBSVM toolbox [2]. We determine the
hyper-parameter of all models, i.e., parameter C of SVM, L1SVM and GaussianSVM, parameter
? of SAM, parameter ? of GroupSpAM, parameter ? in Eq. (6) of GroupSAM, in the range of
{10?3 , 10?2 , . . . , 103 }. We tune the hyper-parameters via 2-fold cross validation on the training
data and report the best parameter w.r.t. classification accuracy of each method. In the accelerated
proximal gradient descent algorithm for both SAM and GroupSAM, we set ? = 0.5, and the number
of maximum iterations as 2000.
4.1
Performance comparison on synthetic data
We first examine the classification performance on the synthetic data as a sanity check. Our synthetic
data is randomly generated as a mixture of Gaussian distributions. In each class, data points are
sampled i.i.d. from a multivariate Gaussian distribution with the covariance being ?I, with I as
the identity matrix. This setting indicates independent covariates of the data. We set the number of
classes to be 4, the number of samples to be 400, and the number of dimensions to be 24. We set
the value of ? in the range of {0.8, 0.85, 0.9} respectively. Following the experimental setup in
[31], we make three replicates for each feature in the data to form 24 feature groups (each group
has three replicated features). We randomly pick 6 feature groups to generate the data such that we
can evaluate the capability of GroupSAM in identifying truly useful feature groups. To make the
classification task more challenging, we add random noise drawn from uniform distribution U(0, ?)
where ? is 0.8 times the maximum value in the data. In addition, we test on a high-dimensional case
by generating 300 feature groups (e.g., a total of 900 features) with 40 samples in a similar approach.
We summarize the classification performance comparison on the synthetic data in Table 2. From
the experimental results we notice that GroupSAM outperforms other approaches under all settings.
This comparison verifies the validity of our method. We can see that GroupSAM significantly
improves the performance of SAM, which shows that the incorporation of group information is
indeed beneficial for classification. Moreover, we can notices the superiority of GroupSAM over
GroupSpAM, which illustrates that our GroupSAM model is more suitable for classiciation. We also
present the comparison of feature groups in Table 3. For illustration purpose, we use the case with 24
feature groups as an example. Table 3 shows that the feature groups identified by GroupSAM are
exactly the same as the ground truth feature groups used for synthetic data generation. Such results
further demonstrate the effectiveness of GroupSAM method, from which we know GroupSAM is
able to select the truly informative feature groups thus improve the classification performance.
4.2
Performance comparison on benchmark data
In this subsection, we use 7 benchmark data from UCI repository [12] to compare the classification
performance of different methods. The 7 benchmark data includes: Ecoli, Indians Diabetes, Breast
Cancer, Stock, Balance Scale, Contraceptive Method Choice (CMC) and Fertility. Similar to the
settings in synthetic data, we construct feature groups by replicating each feature for 3 times. In each
8
Table 3: Comparison between the true feature group ID (used for data generation) and the selected
feature group ID by our GroupSAM method on the synthetic data. Order of the true feature group ID
does not represent the order of importance.
? = 0.8
? = 0.85
? = 0.9
True Feature Group IDs
2,3,4,8,10,17
1,5,10,12,17,21
2,6,7,9,12,22
Selected Feature Group IDs via GroupSAM
3,10,17,8,2,4
5,12,17,21,1,10
6,22,7,9,2,12
Table 4: Classification accuracy comparison on the benchmark data. The table shows the average
classification accuracy and the standard deviation in 2-fold cross validation.
Ecoli
Indians
Diabetes
Breast
Cancer
Stock
Balance
Scale
CMC
Fertility
SVM
GaussianSVM
L1SVM
SAM
GroupSpAM GroupSAM
0.815?0.054 0.818?0.049 0.711?0.051 0.816?0.039 0.771?0.009 0.839?0.028
0.651?0.000
0.652?0.002 0.638?0.018 0.652?0.000 0.643?0.004 0.660?0.013
0.968?0.017
0.965?0.017 0.833?0.008 0.833?0.224 0.958?0.027 0.966?0.014
0.913?0.001
0.911?0.002 0.873?0.001 0.617?0.005 0.875?0.005 0.917?0.005
0.864? 0.003 0.869?0.004 0.870?0.003 0.763?0.194 0.848?0.003 0.893?0.003
0.420? 0.011 0.445?0.015 0.437?0.014 0.427?0.000 0.433?0.003 0.456?0.003
0.880? 0.000 0.880?0.000 0.750?0.184 0.860?0.028 0.780?0.000 0.880?0.000
feature group, we add random noise drawn from uniform distribution U(0, ?) where ? is 0.3 times the
maximum value in each data.
We display the comparison results in Table 4. We find that GroupSAM performs equal or better than
the compared methods in all benchmark datasets. Compared with SVM and L1SVM, our method
uses additive model to incorporate nonlinearity thus is more appropriate to find the complex decision
boundary. Moreover, the comparison with Gaussian SVM and SAM illustrates that by involving
the group information in classification, GroupSAM makes better use of the structure information
among features such that the classification ability can be enhanced. Compared with GroupSpAM, our
GroupSAM model is proposed in data dependent hypothesis spaces and employs hinge loss in the
objective, thus is more suitable for classification.
5
Conclusion
In this paper, we proposed a novel group sparse additive machine (GroupSAM) by incorporating the
group sparsity into the additive classification model in reproducing kernel Hilbert space. By developing the error analysis technique with data dependent hypothesis space, we obtain the generalization
error bound of the proposed GroupSAM, which demonstrates our model can achieve satisfactory
learning rate under mild conditions. Experimental results on both synthetic and real-world benchmark
datasets validate the algorithmic effectiveness and support our learning theory analysis. In the future,
it is interesting to investigate the learning performance of robust group sparse additive machines with
loss functions induced by quantile regression [6, 14].
Acknowledgments
This work was partially supported by U.S. NSF-IIS 1302675, NSF-IIS 1344152, NSF-DBI 1356628,
NSF-IIS 1619308, NSF-IIS 1633753, NIH AG049371. Hong Chen was partially supported by
National Natural Science Foundation of China (NSFC) 11671161. We are grateful to the anonymous
NIPS reviewers for the insightful comments.
9
References
[1] P. L. Bartlett, M. I. Jordan, and J. D. Mcauliffe. Convexity, classification and risk bounds. J.
Amer. Statist. Assoc., 101(473):138?156, 2006.
[2] C.-C. Chang and C.-J. Lin. LIBSVM: A library for support vector machines. ACM Transactions
on Intelligent Systems and Technology, 2(27):1?27, 2011.
[3] D. R. Chen, Q. Wu, Y. Ying, and D. X. Zhou. Support vector machine soft margin classifiers:
error analysis. J. Mach. Learn. Res., 5:1143?1175, 2004.
[4] H. Chen, Z. Pan, L. Li, and Y. Tang. Learning rates of coefficient-based regularized classifier
for density level detection. Neural Comput., 25(4):1107?1121, 2013.
[5] A. Christmann and R. Hable. Consistency of support vector machines using additive kernels for
additive models. Comput. Stat. Data Anal., 56:854?873, 2012.
[6] A. Christmann and D. X. Zhou. Learning rates for the risk of kernel based quantile regression
estimators in additive models. Anal. Appl., 14(3):449?477, 2016.
[7] F. Cucker and D. X. Zhou. Learning Theory: An Approximation Theory Viewpoint. Cambridge
Univ. Press, Cambridge, U.K., 2007.
[8] D. Edmunds and H. Triebel. Function Spaces, Entropy Numbers, Differential Operators.
Cambridge Univ. Press, Cambridge, U.K., 1996.
[9] Z. Guo and D. X. Zhou. Concentration estimates for learning with unbounded sampling. Adv.
Comput. Math., 38(1):207?223, 2013.
[10] J. Huang, J. Horowitz, and F. Wei. Variable selection in nonparametric additive models. Ann.
Statist., 38(4):2282?2313, 2010.
[11] K. Kandasamy and Y. Yu. Additive approximation in high dimensional nonparametric regression
via the salsa. In ICML, 2016.
[12] M. Lichman. UCI machine learning repository, 2013.
[13] Y. Lin and H. H. Zhang. Component selection and smoothing in smoothing spline analysis of
variance models. Ann. Statist., 34(5):2272?2297, 2006.
[14] S. Lv, H. Lin, H. Lian, and J. Huang. Oracle inequalities for sparse additive quantile regression
in reproducing kernel hilbert space. Ann. Statist., preprint, 2017.
[15] L. Meier, S. van de Geer, and P. Buehlmann. High-dimensional additive modeling. Ann. Statist.,
37(6B):3779?3821, 2009.
[16] G. Raskutti, M. Wainwright, and B. Yu. Minimax-optimal rates for sparse additive models over
kernel classes via convex programming. J. Mach. Learn. Res., 13:389?427, 2012.
[17] P. Ravikumar, J. Lafferty, H. Liu, and L. Wasserman. Sparse additive models. J. Royal. Statist.
Soc B., 71:1009?1030, 2009.
[18] L. Shi. Learning theory estimates for coefficient-based regularized regression. Appl. Comput.
Harmon. Anal., 34(2):252?265, 2013.
[19] L. Shi, Y. Feng, and D. X. Zhou. Concentration estimates for learning with `1 -regularizer and
data dependent hypothesis spaces. Appl. Comput. Harmon. Anal., 31(2):286?302, 2011.
[20] I. Steinwart and A. Christmann. Support Vector Machines. Springer, 2008.
[21] A. B. Tsybakov. Optimal aggregation of classifiers in statistical learning. Ann. Statis., 32:135?
166, 2004.
[22] V. Vapnik. Statistical Learning Theory. John Wiley and Sons, 1998.
10
[23] Q. Wu, Y. Ying, and D. X. Zhou. Multi-kernel regularized classfiers. J. Complexity, 23:108?134,
2007.
[24] Q. Wu and D. X. Zhou. Svm soft margin classifiers: linear programming versus quadratic
programming. Neural Comput., 17:1160?1187, 2005.
[25] L. Yang, S. Lv, and J. Wang. Model-free variable selection in reproducing kernel hilbert space.
J. Mach. Learn. Res., 17:1?24, 2016.
[26] J. Yin, X. Chen, and E. Xing. Group sparse additive models. In ICML, 2012.
[27] M. Yuan and Y. Lin. Model selection and estimation in regression with grouped variabels. J.
Royal. Statist. Soc B., 68(1):49?67, 2006.
[28] M. Yuan and D. X. Zhou. Minimax optimal rates of estimation in high dimensional additive
models. Ann. Statist., 44(6):2564?2593, 2016.
[29] T. Zhang. Statistical behavior and consistency of classification methods based on convex risk
minimization. Ann. Statist., 32:56?85, 2004.
[30] T. Zhao and H. Liu. Sparse additive machine. In AISTATS, 2012.
[31] L. W. Zhong and J. T. Kwok. Efficient sparse modeling with automatic feature grouping. In
ICML, 2011.
11
| 6624 |@word mild:3 repository:2 version:1 polynomial:2 norm:12 c0:1 decomposition:2 covariance:1 pick:1 liu:2 contains:1 lichman:1 rkhs:6 outperforms:1 existing:1 current:1 com:1 gmail:1 attracted:1 written:1 john:1 additive:59 informative:2 enables:1 statis:1 half:2 selected:3 kandasamy:1 math:1 zhang:2 unbounded:1 constructed:2 differential:1 yuan:2 introduce:8 indeed:2 expected:2 behavior:1 examine:1 multi:1 inspired:4 decomposed:1 considering:2 increasing:1 provided:2 moreover:3 bounded:3 developed:1 every:1 buehlmann:1 tackle:1 exactly:1 classifier:5 k2:3 demonstrates:3 control:1 assoc:1 superiority:1 mcauliffe:1 before:2 positive:3 engineering:2 tends:1 despite:1 mach:3 id:5 nsfc:1 xidian:2 yd:1 china:2 studied:3 challenging:2 appl:3 limited:1 range:2 acknowledgment:1 implement:1 supu:1 sq:1 empirical:10 significantly:1 confidence:1 integrating:1 get:2 cannot:1 close:1 selection:11 operator:2 risk:10 restriction:1 equivalent:1 reviewer:1 shi:2 attention:1 independently:1 convex:5 identifying:1 wasserman:1 estimator:2 rule:2 dbi:1 spanned:1 orthonormal:1 searching:1 enhanced:1 programming:4 us:2 hypothesis:30 diabetes:2 labeled:1 role:1 preprint:1 electrical:1 solved:1 wang:1 calculate:1 connected:1 adv:1 pd:3 convexity:1 complexity:2 ui:1 covariates:1 grateful:1 tight:1 basis:5 stock:2 various:2 regularizer:10 separated:1 univ:2 fast:2 hyper:2 sanity:1 huang1:1 valued:1 widely:1 say:1 supplementary:1 otherwise:1 ability:3 sequence:1 advantage:1 propose:2 interaction:2 uci:2 flexibility:3 achieve:4 degenerate:1 validate:1 convergence:4 extending:1 generating:1 leave:1 illustrate:1 derive:1 stat:1 measured:1 school:1 eq:1 soc:2 c:3 involves:2 cmc:2 christmann:3 shrunk:1 exploration:2 sgn:12 material:1 generalization:11 alleviate:1 anonymous:1 proposition:1 hold:6 ground:1 algorithmic:4 pitt:1 adopt:1 purpose:1 estimation:2 label:1 bridge:1 grouped:6 reflects:2 minimization:3 gaussian:5 aim:2 rather:1 zhou:8 zhong:1 edmunds:1 corollary:1 derived:1 focus:7 consistently:1 check:1 mainly:2 indicates:1 hk:4 l1svm:6 wang1:1 dependent:9 kernelized:1 arg:4 among:8 classification:45 flexible:1 exponent:3 smoothing:2 special:2 marginal:1 equal:1 construct:2 beach:1 sampling:1 yu:2 icml:3 filling:1 future:2 report:2 np:1 spline:6 develops:1 employ:1 few:1 intelligent:1 randomly:2 simultaneously:1 divergence:2 national:1 ag049371:1 n1:3 tq:1 detection:1 investigate:3 evaluation:2 replicates:1 truly:3 mixture:1 asserted:1 necessary:1 orthogonal:1 harmon:2 conduct:1 desired:1 re:3 theoretical:10 soft:2 modeling:2 clipping:1 cost:1 addressing:1 deviation:2 uniform:2 supx:1 proximal:3 synthetic:10 st:1 density:1 cucker:1 again:1 huang:3 choose:1 horowitz:1 zhao:1 li:2 de:1 summarized:1 includes:2 coefficient:4 depends:2 competitive:2 relied:1 bayes:2 capability:2 aggregation:1 xing:1 contribution:1 square:7 ni:4 accuracy:6 variance:1 efficiently:2 yes:5 knot:1 ecoli:2 definition:1 e2:4 naturally:1 associated:8 proof:3 sampled:1 proved:2 popular:1 recall:1 knowledge:2 ut:2 improves:1 subsection:1 hilbert:6 organized:1 specify:1 wei:1 formulation:2 amer:1 just:2 sketch:1 steinwart:1 nonlinear:6 yf:3 usa:2 validity:1 verify:1 true:6 regularization:11 hence:2 satisfactory:2 illustrated:1 deal:1 please:1 covering:4 hong:2 criterion:1 stone:4 demonstrate:2 performs:1 novel:4 recently:2 fi:2 nih:1 raskutti:1 stepping:4 discussed:2 interpretation:1 extend:1 cambridge:4 tuning:1 automatic:1 consistency:5 nonlinearity:1 replicating:1 add:2 multivariate:1 inf:3 salsa:1 tikhonov:2 certain:1 inequality:3 binary:1 arbitrarily:1 yi:10 employed:1 determine:1 ii:4 hzau:1 d0:6 smooth:2 constructive:1 cross:4 long:1 lin:4 divided:2 e1:5 ravikumar:1 prediction:6 involving:1 regression:13 basic:1 breast:2 metric:4 iteration:1 kernel:28 represent:1 background:1 addition:1 rest:1 comment:1 induced:5 hz:4 lafferty:1 effectiveness:5 jordan:1 structural:3 yang:1 intermediate:1 bernstein:1 zi:1 lasso:1 identified:1 cn:2 br:2 triebel:1 bartlett:1 penalty:3 e3:4 deep:1 useful:1 covered:1 detailed:2 tune:1 transforms:1 nonparametric:2 tsybakov:3 extensively:4 statist:9 generate:1 fz:20 percentage:1 nsf:5 revisit:2 notice:3 correctly:1 group:54 key:2 drawn:3 libsvm:2 utilize:1 sum:1 prob:6 noticing:1 named:1 family:1 electronic:1 wu:3 decision:1 bound:16 ki:3 guaranteed:1 display:1 cheng:1 fold:5 quadratic:1 oracle:1 adapted:1 incorporation:1 min:9 department:1 structured:1 according:1 developing:1 combination:1 beneficial:1 pan:1 sam:13 son:1 partitioned:1 assures:1 discus:1 know:1 xiaoqian:1 rewritten:1 observe:1 kwok:1 appropriate:1 rp:3 rz:2 denotes:1 include:2 hinge:10 quantile:3 establish:1 feng:1 objective:1 noticed:1 concentration:3 traditional:1 gradient:3 supn:1 subspace:2 simulated:2 capacity:2 nx:3 seven:1 mail:2 kk:3 providing:1 minimizing:1 illustration:1 balance:2 ying:2 setup:1 design:3 anal:4 proper:2 unknown:2 upper:2 datasets:3 benchmark:7 finite:2 descent:3 extended:2 rn:2 worthy:1 reproducing:6 introduced:2 meier:1 toolbox:1 chen1:1 established:3 subgroup:1 nip:2 address:3 beyond:1 able:1 usually:5 below:1 sparsity:10 summarize:1 interpretability:2 max:3 royal:2 wainwright:1 misclassification:6 suitable:2 difficulty:1 natural:1 regularized:5 xqwang1991:1 minimax:2 scheme:2 improve:2 technology:1 brief:1 library:1 concludes:1 fertility:2 prior:3 nice:2 literature:4 kf:9 asymptotic:1 loss:20 expect:1 adaptivity:1 generation:2 limitation:1 interesting:1 versus:1 lv:2 validation:4 foundation:1 imposes:1 mercer:1 article:1 viewpoint:1 heng:2 playing:1 cancer:2 summary:1 surprisingly:1 supported:2 free:2 side:1 understand:1 sparse:34 van:1 boundary:1 dimension:3 xn:1 world:2 ending:1 ignores:1 author:1 commonly:1 replicated:1 transaction:1 excess:8 compact:1 b1:2 pittsburgh:1 xi:13 don:2 table:11 additionally:1 learn:3 robust:1 ca:1 expansion:1 investigated:1 complex:1 constructing:1 inherit:1 pk:1 main:4 aistats:1 k2k:7 motivation:1 noise:4 bounding:1 n2:5 verifies:1 x1:1 wiley:1 explicit:1 exponential:1 comput:6 tang:1 theorem:5 xt:1 insightful:1 explored:1 decay:3 chenh:1 svm:14 grouping:1 essential:1 incorporating:2 intrinsic:1 exists:2 vapnik:1 importance:1 illustrates:2 push:1 sparseness:1 margin:2 gap:1 chen:4 entropy:1 yin:1 fc:18 explore:2 ez:14 partially:2 chang:1 springer:1 corresponds:1 minimizer:1 satisfies:1 truth:1 acm:1 conditional:1 identity:1 formulated:1 ann:7 replace:1 hard:1 typical:1 lemma:1 called:3 total:1 geer:1 experimental:6 select:1 support:9 guo:1 accelerated:2 indian:2 incorporate:1 evaluate:2 lian:1 d1:1 |
6,218 | 6,625 | Uprooting and Rerooting Higher-Order Graphical
Models
Mark Rowland?
University of Cambridge
[email protected]
Adrian Weller?
University of Cambridge and Alan Turing Institute
[email protected]
Abstract
The idea of uprooting and rerooting graphical models was introduced specifically
for binary pairwise models by Weller [19] as a way to transform a model to any
of a whole equivalence class of related models, such that inference on any one
model yields inference results for all others. This is very helpful since inference, or
relevant bounds, may be much easier to obtain or more accurate for some model
in the class. Here we introduce methods to extend the approach to models with
higher-order potentials and develop theoretical insights. In particular, we show
that the triplet-consistent polytope TRI is unique in being ?universally rooted?.
We demonstrate empirically that rerooting can significantly improve accuracy of
methods of inference for higher-order models at negligible computational cost.
1
Introduction
Undirected graphical models with discrete variables are a central tool in machine learning. In this
paper, we focus on three canonical tasks of inference: identifying a configuration with highest
probability (termed maximum a posteriori or MAP inference), computing marginal probabilities of
subsets of variables (marginal inference) and calculating the normalizing constant (partition function).
All three tasks are typically computationally intractable, leading to much work to identify settings
where exact polynomial-time methods apply, or to develop approximate algorithms that perform well.
Weller [19] introduced an elegant method which first uproots and then reroots a given model M to any
of a whole class of rerooted models {Mi }. The method relies on specific properties of binary pairwise
models and makes use of an earlier construction which reduced MAP inference to the MAXCUT
problem on the suspension graph ?G (1; 2; 12; 19, see ?3 for details). For many important inference
tasks, the rerooted models are equivalent in the sense that results for any one model yield results for
all others with negligible computational cost. This can be very helpful since various models in the
class may present very different computational difficulties for inference.
Here we show how the idea may be generalized to apply to models with higher-order potentials over
any number of variables. Such models have many important applications, for example in computer
vision [6] or modeling protein interactions [5]. As for pairwise models, we again obtain significant
benefits for inference. We also develop a deeper theoretical understanding and derive important new
results. We highlight the following contributions:
? In ?3-?4, we show how to achieve efficient uprooting and rerooting of binary graphical models
with potentials of any order, while still allowing easy recovery of inference results.
? In ?5, to simplify the subsequent analysis, we introduce pure k-potentials for any order k, which
may be of independent interest. We show that there is essentially only one pure k-potential which
we call the even k-potential, and that even k-potentials form a basis for all model potentials.
? In ?6, we carefully analyze the effect of uprooting and rerooting on Sherali-Adams [11] relaxations
Lr of the marginal polytope, for any order r. One surprising observation in ?6.2 is that L3 (the
?
Authors contributed equally.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
triplet-consistent polytope or TRI) is unique in being universally rooted, in the sense that there is
an affine score-preserving bijection between L3 for a model and L3 for each of its rerootings.
? In ?7, our empirical results demonstrate that rerooting can significantly improve accuracy of
inference in higher-order models. We introduce effective heuristics to choose a helpful rerooting.
Our observations have further implications for the many variational methods of marginal inference
which optimize the sum of score and an entropy approximation over a Sherali-Adams polytope
relaxation. These include the Bethe approximation (intimately related to belief propagation) and
cluster extensions, tree-reweighted (TRW) approaches and logdet methods [12; 14; 16; 22; 24].
1.1
Background and discussion of theoretical contributions
Based on earlier connections in [2], [19] showed the remarkable result for pairwise models that the
triplet-consistent polytope (L3 or TRI) is universally rooted (in the restricted sense defined in [19,
Theorem 3]). This observation allowed straightforward strengthening of previously known results,
for example: it was previously shown [23] that the LP relaxation on TRI (LP+TRI) is always tight
for an ?almost-balanced? binary pairwise model, that is a model which can be rendered balanced
by removing one variable [17]. Given [19, Theorem 3], this earlier result could immediately be
significantly strengthened to [19, Theorem 4], which showed that LP+TRI is tight for a binary
pairwise model provided only that some rerooting exists such that the rerooted model is almost
balanced.
Following [19], it was natural to suspect that the universal rootedness property might hold for all
(or at least some) Lr , r ? 3. This would have impact on work such as [10] which examines which
signed minors must be forbidden to guarantee tightness of LP+L4 . If L4 were universally rooted,
then it would be possible to simplify significantly the analysis in [10].
Considering this issue led to our analysis of the mappings to symmetrized uprooted polytopes given
in our Theorem 17. We believe this is the natural generalization of the lower order relationships of
L2 and L3 to RMET and MET described in [2], though this direction was not clear initially.
With this formalism, together with the use of even potentials, we demonstrate our Theorems 20 and
21, showing that in fact TRI is unique in being universally rooted (and indeed in a stronger sense
than given in [19]). We suggest that this result is surprising and may have further implications.
As a consequence, it is not possible to generate some quick theoretical wins by generalizing previous
results as [19] did to derive their Theorem 4, but on the other hand we observe that rerooting may be
helpful in practice for any approach using a Sherali-Adams relaxation other than L3 . We verify the
potential for significant benefits experimentally in ?7.
2
Graphical models
A discrete graphical model M [G(V, E), (?E )E?E ] consists of: a hypergraph G = (V, E), which has
n vertices V = {1, . . . , n} corresponding to the variables of the model, and hyperedges E ? P(V ),
where P(V ) is the powerset of V ; together with potential functions (?E )E?E over the hyperedges
E ? E. We consider binary random variables (Xv )v?V with each Xv ? Xv = {0, 1}. For a subset
U ? V , xU ? {0, 1}U is a configuration of those variables (Xv )v?U . We write xU for the flipping
of xU , defined by xi = 1 ? xi P
?i ? U . The joint probability mass function factors as follows, where
the normalizing constant Z = xV ?{0,1}V exp(score(xV )) is the partition function:
X
1
p(xV ) = exp (score(xV )) ,
score(xV ) =
?E (xE ).
(1)
Z
E?E
3
Uprooting and rerooting
Our goal is to map a model M to any of a whole family of models {Mi } in such a way that inference
on any Mi will allow us easily to recover inference results on the original model M . In this section
we provide our mapping, then in ?4 we explain how to recover inference results for M .
The uprooting mechanism used by Weller [19] first reparametrizes edge potentials to the form
?ij (xi , xj ) = ? 21 Wij 1[xi 6= xj ], where 1[?] is the indicator function (a reparameterization modifies
2
1
2
1
2
3
4
3
4
M = M0
1
0
3
M+
2
1
0
M4 = M + |X4 =0
3
4
0
M2 = M + |X2 =0
Figure 1: Left: The hypergraph G of a graphical model M over 4 variables, with potentials on the hyperedges
{1, 2}, {1, 3, 4}, and {2, 4}. Center-left: The suspension hypergraph ?G of the uprooted model M + . Centerright: The hypergraph ?G \ {4} of the rerooted model M4 = M + |X4 =0 , i.e. M + with X4 clamped to 0. Right:
The hypergraph ?G \ {2} of the rerooted model M2 = M + |X2 =0 , i.e. M + with X2 clamped to 0.
potential functions such that the complete score of each configuration is unchanged, see 15 for details).
Next, singleton potentials are converted to edge potentials with this same form by connecting to an
added variable X0 . This mechanism had been used previously to reduce MAP inference on M to
MAXCUT on the converted model [1; 12], and applies specifically only to binary pairwise models.
We introduce a generalized construction which applies to models with potentials of any order. We first
uproot a model M to a highly symmetric uprooted model M + where an extra variable X0 is added,
in such a way that the original model M is exactly M + with X0 clamped to the value 0. Since X0 is
clamped to retrieve M , we may write M = M0 := M + |X0 =0 . Alternatively, we can choose instead
to clamp a different variable Xi in M + which will lead to the rerooted model Mi := M + |Xi =0 .
Definition 1 (Clamping). For a graphical model M [G = (V, E), (?E )E?E ], and i ? V , the model
M |Xi =a obtained by clamping the variable Xi to the value a ? Xi is given by: the hypergraph
(V \ {i}, Ei ), where Ei = {E \ {i}|E ? E}; and potentials which are unchanged for hyperedges
which do not contain i, while if i ? E then ?E\{i} (xE\{i} ) = ?E (xE\{i} , xi = a).
Definition 2 (Uprooting, suspension hypergraph). Given a model M [G(V, E), (?E )E?E ], the
uprooted model M + adds a variable X0 , which is added to every hyperedge of the original model. M +
has hypergraph ?G, with vertex set V + = V ?{0} and hyperedge set E + = {E + = E ?{0}|E ? E}.
+
?G is the suspension hypergraph of G. M + has potential functions (?E?{0}
)E?E given by
? (x ) if x0 = 0
+
?E?{0}
(xE?{0} ) = E E
?E (xE ) if x0 = 1.
With this definition, all uprooted potentials are symmetric in that ?E++ (xE + ) = ?E++ (xE + ) ?E + ? E + .
Definition 3 (Rerooting). From Definition 2, we see that given a model M , if we uproot to M +
then clamp X0 = 0, we recover the original model M . If instead in M + we clamp Xi = 0 for any
i = 1, . . . , n, then we obtain the rerooted model Mi := M + |Xi =0 .
See Figure 1 and Table 1 for examples of uprooting and rerooting. We explore the question of how to
choose a good variable for rerooting (i.e. how to choose a good variable to clamp in M + ) in ?7.
4
Recovery of inference tasks
Here we demonstrate that the partition function, MAP score and configuration, and marginal distributions for a model M , can all be recovered from its uprooted model M + or any rerooted model
Mi i ? V , with negligible computational cost. We write Vi = {0, 1, . . . , n} \ {i} for the variable set
of rerooted model Mi ; scorei (xVi ) for the score of xVi in Mi ; and pi for the probability distribution
for Mi . We use superscript + to indicate the uprooted
P model. For example, the probability distribution for M + is given by p+ (xV + ) = Z1+ exp
E?E + ?E (xE ) . From the definitions of ?3, we
obtain the following key lemma, which is critical to enable recovery of inference results.
Lemma 4 (Score-preserving map). Each configuration xV of M maps to 2 configurations of
the uprooted M + with the same score, i.e. from M, xV ? in M + , both of (x0 = 0, xV ) and
(x0 = 1, xV ) with score(xV ) = score+ (x0 = 0, xV ) = score+ (x0 = 1, xV ). For any i ? V + ,
exactly one of the two uprooted configurations has xi = 0, and just this one will be selected in Mi .
Hence, there is a score-preserving bijection between configurations of M and those of Mi :
(x0 = 0, xV \{i} ) if xi = 0
For any i ? V + :
in M, xV ? in Mi ,
(2)
(x0 = 1, xV \{i} ) if xi = 1.
3
M
x1
0
0
0
0
1
1
1
1
config
x3 x4
0
0
0
1
1
0
1
1
0
0
0
1
1
0
1
1
M + configuration
x0 x1 x3 x4
0
0
0
0
0
0
0
1
0
0
1
0
0
0
1
1
0
1
0
0
0
1
0
1
0
1
1
0
0
1
1
1
1
0
0
0
1
0
0
1
1
0
1
0
1
0
1
1
1
1
0
0
1
1
0
1
1
1
1
0
1
1
1
1
M4 config
x0 x1 x3
0
0
0
0
0
1
0
1
0
0
1
1
1
0
0
1
0
1
1
1
0
1
1
1
Table 1: An illustration of how scores of potential ?134 on hyperedge {1, 3, 4} in an original model M map to
potential ?0134 in M + and then to ?013 in M4 . See Figure 1 for the hypergraphs. Each color indicates a value
of ?134 (x1 , x3 , x4 ) for a different configuration (x1 , x3 , x4 ). Note that M + has 2 rows of each color, while
after rerooting to M4 , we again have exactly one row of each color. The 1-1 score preserving map between
configurations of M and any Mi is critical to enable recovery of inference results; see Lemma 4.
Table 1 illustrates this perhaps surprising result, from which the next two propositions follow.
Proposition 5 (Recovering the partition function). Given a model M [G(V, E), (?E )E?E ] with
partition function Z as in (1), the partition function Z + of the uprooted model M + is twice Z, and
the partition function of each rerooted model Mi is exactly Z, for any i ? V .
Proposition 6 (Recovering a MAP configuration). From M + : xV is an arg max for p iff (x0 =
0, xV ) is an arg max for p+ iff (x0 = 1, xV ) is an arg max for p+ . From a rerooted model Mi :
(xV \{i} , xi = 0) is an arg max for p iff (x0 = 0, xV \{i} ) is an arg max for pi ; (xV \{i} , xi = 1) is an
arg max for p iff (x0 = 1, xV \{i} ) is an arg max for pi .
We can recover marginals as shown in the following proposition, proof in the Appendix ?9.1.
Proposition 7 (Recovering marginals). For a subset ? 6= U ? V , we can recover from M + :
p(xU ) = p+ (x0 = 0, xU ) + p+ (x0 = 1, xU )
= 2p+ (x0 = 0, xU ) = 2p+ (x0 = 1, xU ).
To recover from a rerooted Mi : (i) For any i ? V \ U , p(xU ) = pi (x0 = 0, xU ) + pi (x0 = 1, xU ).
pi (x0 = 0, xU \{i} ) xi = 0
(ii) For any i ? U , p(xU ) =
pi (x0 = 1, xU \{i} ) xi = 1.
In ?6, we provide a careful analysis of the impact of uprooting and rerooting on the Sherali-Adams
hierarchy of relaxations of the marginal polytope [11]. We first introduce a way to parametrize
potentials which will be particularly useful, and which may be of independent interest.
5
Pure k-potentials
We introduce the notion of pure k-potentials. These allow the specification of interactions which act
?purely? over a set of variables of a given size k, without influencing the distribution of any subsets.
We show that in fact, there is essentially only one pure k-potential. Further, we show that one can
express any ?E potential in terms of pure potentials over E and subsets of E, and that pure potentials
have appealing properties when uprooted and rerooted which help our subsequent analysis.
We say that a potential is a k-potential if k is the smallest number such that the score of the potential
may be determined by considering the configuration of k variables. Usually a potential ?E is a
k-potential with k = |E|. For example, typically a singleton potential is a 1-potential, and an edge
potential is a 2-potential. However, note that k < |E| is possible if one or more variables in E are not
needed to establish the score (a simple example is ?12 (x1 , x2 ) = x1 , which clearly is a 1-potential).
4
In general, a k-potential will affect the marginal distributions of all subsets of the k variables. For
example, one popular form of 2-potential is ?ij (xi , xj ) = Wij xi xj , which tends to pull Xi and Xj
toward the same value, but also tends to increase each of p(Xi = 1) and p(Xj = 1). For pairwise
models, a different reparameterization of potentials instead writes the score as
X
1 X
score(xV ) =
? i xi +
Wij 1[xi = xj ].
(3)
2
i?V
(i,j)?E
Expression (3) has the desirable feature that the ?ij (xi , xj ) = 12 Wij 1[xi = xj ] edge potentials affect
only the pairwise marginals, without disturbing singleton marginals. This motivates the following
definition.
Definition 8. Let k ? 2, and let U be a set of size k. We say that a k-potential ?U : {0, 1}U ? R
is a pure k-potential if the distribution induced by the potential, p(xU ) ? exp(?U (xU )), has the
property that for any ? =
6 W ( U , the marginal distribution p(xW ) is uniform.
We shall see in Proposition 10 that a pure k-potential must essentially be an even k-potential.
Definition 9. Let k ? N, and |U | = k. An even k-potential is a k-potential ?U : {0, 1}U ? R of the
form ?U (xU ) = a1[ |{i ? U |xi = 1}| is even], for some a ? R which is its coefficient. In words,
?U (xU ) takes value a if xU has an even number of 1s, else it takes value 0.
As an example, the 2-potential ?ij (xi , xj ) = 12 Wij 1[xi = xj ] in (3) is an even 2-potential with
U = {i, j} and coefficient Wij /2. The next two propositions are proved in the Appendix ?9.2.
Proposition 10 (All pure potentials are essentially even potentials). Let k ? 2, and |U | = k. If
?U : {0, 1}U ? R is a pure k-potential then ?U must be an affine function of the even k-potential, i.e.
? a, b ? R s.t. ?U (xU ) = a1[ |{i ? U |xi = 1}| is even] + b.
Proposition 11 (Even k-potentials
form a basis). For a finite set U , the set of even k-potentials
1[ |{i ? W |Xi = 1}| is even] W ?U , indexed by subsets W ? U , forms a basis for the vector
space of all potential functions ? : {0, 1}U ? R.
Any constant in a potential will be absorbed into the partition function Z and does not affect the
probability distribution, see (1). An even 2-potential with positive coefficient, e.g. as in (3) if
Wij > 0, is supermodular. Models with only supermodular potentials (equivalently, submodular cost
functions) typically admit easier inference [3; 7]; if such a model is binary pairwise then it is called
attractive. However, for k > 2, even k-potentials ?E are neither supermodular nor submodular. Yet if
k is an even number, observe that ?E (xE ) = ?E (xE ). We discuss this further in Appendix ?10.4.
When a k-potential is uprooted, in general it may become a (k + 1)-potential (recall Definition 2).
The following property of even k-potentials is helpful for our analysis in ?6, and is easily checked.
Lemma 12 (Uprooting an even k-potential). When an even k-potential ?E with |E| = k is uprooted:
if k is an even number, then the uprooted potential is exactly the same even k-potential; if k is odd,
then we obtain the even (k + 1)-potential over E ? {0} with the same coefficient as the original ?E .
6
Marginal polytope and Sherali-Adams relaxations
We saw in Lemma 4 that there is a score-preserving 1-2 mapping from configurations of M to those
of M + , and a bijection between configurations of M and any Mi . Here we examine the extent to
which these score-preserving mappings extend to (pseudo-)marginal probability distributions over
variables by considering the Sherali-Adams relaxations [11] of the respective marginal polytopes.
These relaxations feature prominently in many approaches for MAP and marginal inference.
For U ? V , we write ?U for a probability distribution in P({0, 1}U ), the set of all probability
distributions on {0, 1}U . Bold ? will represent a collection of measures over various subsets of
variables. Given (1), to compute an expected score, we need (?E )E?E . This motivates the following.
Definition 13. The marginal polytope M(G(V, E)) = {(?E )E?E ??V s.t. ?V E = ?E ?E ? E},
where for U1 ? U2 ? V , ?U2 ?U1 denotes the marginalization of ?U2 ? P({0, 1}U2 ) onto {0, 1}U1 .
M(G) consists of marginal distributions for every hyperedge E ? E such that all the marginals are
consistent with a global distribution over all variables V . Methods of variational inference typically
5
optimize either the score (for MAP inference) or the score plus an entropy term (for marginal
inference) over a relaxation of the marginal polytope [15]. This is because M(G) is computationally
intractable, with an exponential number of facets [2]. Relaxations from the Sherali-Adams hierarchy
[11] are often used, requiring consistency only over smaller clusters of variables.
Definition 14. Given an integer r ? 2, if a hypergraph G(V, E) satisfies maxE?E |E| ? r ? |V |,
then we say that G is r-admissible, and define the Sherali-Adams polytope of order r on G by
Lr (G) = (?E )E?E ?(?U ) U ?V locally consistent, s.t. ?U ?E = ?E ? E ? U ? V, |U | = r ,
|U |=r
where a collection of measures (?A )A?I (for some set I of subsets of V ) is locally consistent, or l.c.,
if for any A1 , A2 ? I, we have ?A1 ?A1 ?A2 = ?A2 ?A1 ?A2 . Each element of Lr (G) is a set of locally
consistent probability measures over the hyperedges. Note that M(G) ? Lr (G) ? Lr?1 (G). The
pairwise relaxation L2 (G) is commonly used but higher-order relaxations achieve greater accuracy,
have received significant attention [10; 13; 18; 22; 23], and are required for higher-order potentials.
6.1
The impact of uprooting and rerooting on Sherali-Adams polytopes
We introduce two variants of the Sherali-Adams polytopes which will be helpful in analyzing
uprooted models. For a measure ?U ? P({0, 1}U ), we define the flipped measure ?U as ?U (xU ) =
?U (xU ) ?xU ? {0, 1}U . A measure ?U is flipping-invariant if ?U = ?U .
Definition 15. The symmetrized Sherali-Adams polytopes for an uprooted hypergraph ?G(V + , E + )
(as given in Definition 2), is:
e r (?G) = (?E )
?E = ?E ?E ? E + .
L
?
L
(?G)
+
r
E?E
Definition 16. For any i ? V + , and any integer r ? 2 such that maxE?E + |E| ? r ? |V + |, we
define the symmetrized Sherali-Adams polytope of order r uprooted at i to be
?U ?E = ?E ? E ? U ? V, |U | = r, i ? U
e i (?G) = (?E )
+
L
?(?
)
l.c.,
s.t.
.
+
U i?U ?V
r
E?E
?U = ?U ?U ? V, |U | = r, i ? U
|U |=r
e i (?G), there exist corresponding flippingThus, for each collection of measures over hyperedges in L
r
invariant, locally consistent measures on sets of size r which contain i (and their subsets). Note
e r+1 (?G) ? L
e i (?G) ? L
e r (?G).
that for any hypergraph G(V, E) and any i ? V + , we have L
r+1
We next extend the correspondence of Lemma 4 to collections of locally-consistent probability
distributions on the hyperedges of G, see the Appendix ?9.3 for proof.
Theorem 17. For a hypergraph G(V, E), and integer r such that maxE?E |E| ? r ? |V |, there is
an affine score-preserving bijection
Uproot
Lr (G)
RootAt0
e 0 (?G) .
L
r+1
Theorem 17 establishes the following diagram of polytope inclusions and affine bijections:
For M = M0 :
For M
+
:
Lr+1
?x(G)
??
Uprooty?RootAt0
e 0 (?G)
L
r+2
?
Unnamed
?x
??
Uprooty?RootAt0
e r+1 (?G)
? L
?
L?r (G)
x
??
Uprooty?RootAt0
e 0 (?G) .
? L
r+1
(4)
A question of theoretical interest and practical importance is which of the inclusions in (4) are
strict. Our perspective here generalizes earlier work. Using different language, Deza and Laurent
e 0 (?G), which was termed RMET, the rooted semimetric polytope; and
[2] identified L2 (G) with L
3
e 3 (?G) with MET, the semimetric polytope. Building on this, Weller [19] considered L3 (G), the
L
triplet-consistent polytope or TRI, though only in the context of pairwise potentials, and showed that
L3 (G) has the remarkable property that if it is used to optimize an LP for a model M on G, the exact
same optimum is achieved for L3 (Gi ) for any rerooting Mi . It was natural to conjecture that Lr (G)
might have this same property for all r > 3, yet this was left as an open question.
6
6.2
L3 is unique in being universally rooted
We shall first strengthen [19] to show that L3 is universally rooted in the following stronger sense.
Definition 18. We say that the rth -order Sherali-Adams relaxation is universally rooted (and write
?Lr is universally rooted? for short) if for all admissible hypergraphs G, there is an affine scorepreserving bijection between Lr (G) and Lr (Gi ), for each rerooted hypergraph (Gi )i?V .
If Lr is universally rooted, this applies for potentials over up to r variables (the maximum which
makes sense in this context), and clearly it implies that optimizing score over any rerooting (as in
MAP inference) will attain the same objective. The following result is proved in the Appendix ?9.3.
Lemma 19. If Lr is universally rooted for hypergraphs of maximum hyperedge degree p < r with p
even, then Lr is also universally rooted for r-admissible hypergraphs with maximum degree p + 1.
e 0 (?G). Then by considering
The proof relies on mapping to the symmetrized uprooted polytope L
r+1
marginals using a basis equivalent to that described in Proposition 11 for even k-potentials, we
observe that the symmetry of the polytope enforces only one possible marginal for (p + 1)-clusters.
Combining Lemma 19 with arguments which extend those used by [19] demonstrates the following
result, proved in the Appendix.
Theorem 20. L3 is universally rooted.
We next provide a striking and rather surprising result, see the Appendix for proof and details.
Theorem 21. L3 is unique in being universally rooted. Specifically, for any integer r > 1 other than
r = 3, we constructively demonstrate a hypergraph G(V, E) with |V | = r + 1 variables for which
e 0 (?G) 6= L
e i (?G) for any i ? V .
L
r+1
r+1
e 0 (?G) and L
e i (?G), which by Theorem 17 are the uprooted equivalents
Theorem 21 examines L
r+1
r+1
of Lr (G) and Lr (Gi ). It might appear more satisfying to try to demonstrate the result directly for
the rooted polytopes, i.e. to show Lr (G) 6= Lr (Gi ). However, in general the rooted polytopes
are not comparable: an r-potential in M can map to an (r + 1)-potential in M + and then to an
(r + 1)-potential in Mi which cannot be evaluated for an Lr polytope.
Theorem 21 shows that we may hope for benefits from rerooting for any inference method based on a
Sherali-Adams relaxed polytope Lr , unless r = 3.
7
Experiments
Here we show empirically the benefits of uprooting and rerooting for approximate inference methods
in models with higher-order potentials. We introduce an efficient heuristic which can be used in
practice to select a variable for rerooting, and demonstrate its effectiveness.
We compared performance after different rerootings of marginal inference (to guarantee convergence
we used the double loop method of Heskes et al. [4], which relates to generalized belief propagation,
24) and MAP inference (using loopy belief propagation, LBP [9]). For true values, we used the
junction tree algorithm. All methods were implemented using libDAI [8]. We ran experiments
on complete hypergraphs (with 8 variables) and toroidal grid models (5 ? 5 variables). Potentials
up to order 4 were selected randomly, by drawing even k-potentials from Unif([?Wmax , Wmax ])
distributions for a variety of Wmax parameters, as shown in Figure 2, which highlights results for
estimating log Z. For each regime of maximum potential values, we plot results averaged over 20
runs. For additional details and results, including marginals, other potential choices and larger models,
see Appendix ?10.
We display average error of the inference method applied to: the original model M ; the uprooted
model M + ; then rerootings at: the worst variable, the best variable, the K heuristic variable, and
the G heuristic variable. Best and worst always refer to the variable at which rerooting gave with
hindsight the best and worst error for the partition function (even in plots for other measures).
7
7.1
Heuristics to pick a good variable for rerooting
From our Definition 3, a rerooted model Mi is obtained by clamping the uprooted model M + at
variable Xi . Hence, selecting a good variable for rerooting is exactly the choice of a good variable
to clamp in M + . Considering pairwise models, Weller [19] refined the maxW method [20; 21] to
introduce the maxtW heuristic, and showed that it was very effective empirically. maxtW selects
P
W
the variable Xi with max j?N (i) tanh | 4ij |, where N (i) is the set of neighbors of i in the model
graph, and Wij is the strength of the pairwise interaction.
The intuition for maxtW is as follows. Pairwise methods of approximate inference such as Bethe
are exact for models with no cycles. If we could, we would like to ?break? tight cycles with strong
edge weights, since these lead to error. When a variable is clamped, it is effectively removed from
the model. Hence, we would like to reroot at a variable that sits on many cycles with strong edge
weights. Identifying such cycles is NP-hard, but the maxtW heuristic attempts to do this by looking
only locally around each variable. Further, the effect of a strong edge weight saturates [21]: a very
strong edge weight Wij effectively ?locks? its end variables (either together or opposite depending on
the sign of Wij ), and this effect cannot be significantly increased even by an extremely strong edge.
Hence the tanh function was introduced to the earlier maxW method, leading to the maxtW heuristic.
As observed in ?5, if we express our model potentials in terms of pure k-potentials, then the uprooted
model will only have pure k-potentials for various values of k which are even numbers. Intuitively,
the higher the coefficients on these potentials, the more tightly connected is the model leading to more
challenging inference. Hence, a natural way to generalize the maxtW approach to handle higher-order
potentials is to pick a variable Xi in M + which maximizes the following measure:
X
X
clamp-heuristic-measure(i) =
c2 tanh |t2 aE | +
c4 tanh |t4 aE |,
(5)
i?E:|E|=2
i?E:|E|=4
where aE is the coefficient (weight) of the relevant pure k-potential, see Definition 9, and the
{c2 , t2 }, {c4 , t4 } terms are constants for pure 2-potentials and for pure 4-potentials respectively. This
approach extends to potentials of higher orders by adding similar further terms. Since our goal is to
rank the measures for each i ? V + , without loss of generality we take c2 = 1. We fit the t2 , c4 and t4
constants to the data from our experimental runs, see the Appendix for details. Our K heuristic was fit
only to runs for complete hypergraphs while the G heuristic was fit only to runs for models on grids.
7.2
Observations on results
Considering all results across models and approximate methods for estimating log Z, marginals and
MAP inference (see Figure 2 and Appendix ?10.3), we make the following observations. Both K and
G heuristics perform well (in and out of sample): they never hurt materially and often significantly
improve accuracy, attaining results close to the best possible rerooting. Since our two heuristics
achieve similar performance, sensitivity to the exact constants in (5) appears low. We verified this by
comparing to maxtW for pairwise models as in [19]: both K and G heuristics performed just slightly
better than maxtW. For all our runs, inference on rerooted models took similar time as on the original
model (time required to reroot and later to map back inference results is negligible), see ?10.3.1.
Observe that stronger 1-potentials tend to make inference easier, pulling each variable toward a
specific setting, and reducing the benefits from rerooting (left column of Figure 2). Stronger pure
k-potentials for k > 1 intertwine variables more tightly: this typically makes inference harder and
increases the gains in accuracy from rerooting. The pure k-potential perspective facilitates this
analysis.
When we examine larger models, or models with still higher order potentials, we observe qualitatively
similar results, see Appendix ?10.3.4 and 10.3.6.
8
Conclusion
We introduced methods which broaden the application of the uprooting and rerooting approach
to binary models with higher-order potentials of any order. We demonstrated several important
theoretical insights, including Theorems 20 and 21 which show that L3 is unique in being universally
rooted. We developed the helpful tool of even k-potentials in ?5, which may be of independent
8
Average abs(error) in log Z for K8 complete hypergraphs (fully connected) on 8 variables.
Average abs(error) in log Z for Grids on 5 ? 5 variables (toroidal). Legends are consistent across all plots.
vary Wmax for 1-pots
vary Wmax for 2-pots
vary Wmax for 3-pots
vary Wmax for 4-pots
Figure 2: Error in estimating log Z for random models with various pure k-potentials over 20 runs. If not
shown, Wmax max coefficients for pure k-potentials are 0 for k = 1, 8 for k = 2, 0 for k = 3, 8 for k = 4.
Where the red K heuristic curve is not visible, it coincides with the green G heuristic. Both K and G heuristics
for selecting a rerooting work well: they never hurt and often yield large benefits. See ?7 for details.
interest. We empirically demonstrated significant benefits for rerooting in higher-order models ?
particularly for the hard case of strong cluster potentials and weak 1-potentials ? and provided an
efficient heuristic to select a variable for rerooting. This heuristic is also useful to indicate when
rerooting is unlikely to be helpful for a given model (if (5) is maximized by taking i = 0).
It is natural to compare the effect of rerooting M to Mi , against simply clamping Xi in the original
model M . A key difference is that rerooting achieves the clamping at Xi for negligible computational
cost. In contrast, if Xi is clamped in the original model then the inference method will have to
be run twice: once clamping Xi = 0, and once clamping Xi = 1, then results must be combined.
This is avoided with rerooting given the symmetry of M + . Rerooting effectively replaces what may
be a poor initial implicit choice of clamping at X0 with a carefully selected choice of clamping
variable almost for free. This is true even for large models where it may be advantageous to clamp a
series of variables: by rerooting, one of the series is obtained for free, potentially gaining significant
benefit with little work required. Note that each separate connected component may be handled
independently, with its own added variable. This could be useful for (repeatedly) composing clamping
and then rerooting each separated component to obtain an almost free clamping in each.
Acknowledgements
We thank the anonymous reviewers for helpful comments. MR acknowledges support by the UK
Engineering and Physical Sciences Research Council (EPSRC) grant EP/L016516/1 for the University
of Cambridge Centre for Doctoral Training, the Cambridge Centre for Analysis. AW acknowledges
support by the Alan Turing Institute under the EPSRC grant EP/N510129/1, and by the Leverhulme
Trust via the CFI.
References
[1] F. Barahona, M. Gr?tschel, M. J?nger, and G. Reinelt. An application of combinatorial optimization to
statistical physics and circuit layout design. Operations Research, 36(3):493?513, 1988.
[2] M. Deza and M. Laurent. Geometry of Cuts and Metrics. Springer Publishing Company, Incorporated, 1st
edition, 1997. ISBN 978-3-642-04294-2.
[3] J. Djolonga and A. Krause. Scalable variational inference in log-supermodular models. In ICML, pages
1804?1813, 2015.
[4] T. Heskes, K. Albers, and B. Kappen. Approximate inference and constrained optimization. In UAI, pages
313?320, 2003.
9
[5] A. Jaimovich, G. Elidan, H. Margalit, and N. Friedman. Towards an integrated protein?protein interaction
network: A relational Markov network approach. Journal of Computational Biology, 13(2):145?164, 2006.
[6] P. Kohli, L. Ladicky, and P. Torr. Robust higher order potentials for enforcing label consistency. International Journal of Computer Vision, 82(3):302?324, 2009.
[7] V. Kolmogorov, J. Thapper, and S. ?ivn?. The power of linear programming for general-valued CSPs.
SIAM Journal on Computing, 44(1):1?36, 2015.
[8] J. Mooij. libDAI: A free and open source C++ library for discrete approximate inference in graphical
models. Journal of Machine Learning Research, 11:2169?2173, August 2010. URL http://www.jmlr.
org/papers/volume11/mooij10a/mooij10a.pdf.
[9] J. Pearl. Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference. Morgan
Kaufmann, 1988.
[10] M. Rowland, A. Pacchiano, and A. Weller. Conditions beyond treewidth for tightness of higher-order LP
relaxations. In Artifical Intelligence and Statistics (AISTATS), 2017.
[11] H. Sherali and W. Adams. A hierarchy of relaxations between the continuous and convex hull representations for zero-one programming problems. SIAM Journal on Discrete Mathematics, 3(3):411?430,
1990.
[12] D. Sontag. Cutting plane algorithms for variational inference in graphical models. Master?s thesis, MIT,
EECS, 2007.
[13] D. Sontag and T. Jaakkola. New outer bounds on the marginal polytope. In NIPS, 2007.
[14] M. Wainwright and M. Jordan. Log-determinant relaxation for approximate inference in discrete Markov
random fields. IEEE Transactions on Signal Processing, 2006.
[15] M. Wainwright and M. Jordan. Graphical models, exponential families and variational inference. Foundations and Trends in Machine Learning, 1(1-2):1?305, 2008.
[16] M. Wainwright, T. Jaakkola, and A. Willsky. A new class of upper bounds on the log partition function.
IEEE Transactions on Information Theory, 51(7):2313?2335, 2005.
[17] A. Weller. Revisiting the limits of MAP inference by MWSS on perfect graphs. In AISTATS, 2015.
[18] A. Weller. Characterizing tightness of LP relaxations by forbidding signed minors. In UAI, 2016.
[19] A. Weller. Uprooting and rerooting graphical models. In International Conference on Machine Learning
(ICML), 2016.
[20] A. Weller and J. Domke. Clamping improves TRW and mean field approximations. In Artificial Intelligence
and Statistics (AISTATS), 2016.
[21] A. Weller and T. Jebara. Clamping variables and approximate inference. In Neural Information Processing
Systems (NIPS), 2014.
[22] A. Weller, K. Tang, D. Sontag, and T. Jebara. Understanding the Bethe approximation: When and how can
it go wrong? In Uncertainty in Artificial Intelligence (UAI), 2014.
[23] A. Weller, M. Rowland, and D. Sontag. Tightness of LP relaxations for almost balanced models. In
Artificial Intelligence and Statistics (AISTATS), 2016.
[24] J. Yedidia, W. Freeman, and Y. Weiss. Constructing free-energy approximations and generalized belief
propagation algorithms. IEEE Trans. Information Theory, pages 2282?2312, 2005.
10
| 6625 |@word kohli:1 determinant:1 polynomial:1 stronger:4 advantageous:1 adrian:1 open:2 unif:1 barahona:1 pick:2 harder:1 kappen:1 initial:1 configuration:15 series:2 score:27 selecting:2 sherali:15 recovered:1 comparing:1 surprising:4 yet:2 forbidding:1 must:4 subsequent:2 partition:10 visible:1 plot:3 intelligence:4 selected:3 plane:1 short:1 lr:21 bijection:5 sits:1 org:1 c2:3 become:1 consists:2 introduce:9 x0:30 pairwise:16 expected:1 indeed:1 nor:1 examine:2 freeman:1 company:1 little:1 considering:6 provided:2 estimating:3 maximizes:1 mass:1 circuit:1 what:1 developed:1 hindsight:1 guarantee:2 pseudo:1 every:2 act:1 exactly:6 demonstrates:1 toroidal:2 uk:3 wrong:1 grant:2 appear:1 positive:1 negligible:5 influencing:1 engineering:1 xv:28 tends:2 consequence:1 limit:1 analyzing:1 laurent:2 might:3 signed:2 twice:2 plus:1 doctoral:1 equivalence:1 challenging:1 averaged:1 unique:6 practical:1 enforces:1 practice:2 x3:5 writes:1 cfi:1 empirical:1 universal:1 significantly:6 attain:1 word:1 protein:3 suggest:1 onto:1 cannot:2 close:1 context:2 optimize:3 equivalent:3 map:18 quick:1 center:1 demonstrated:2 modifies:1 straightforward:1 attention:1 reviewer:1 independently:1 layout:1 convex:1 go:1 identifying:2 recovery:4 pure:20 immediately:1 m2:2 insight:2 examines:2 pull:1 retrieve:1 reparameterization:2 handle:1 notion:1 hurt:2 construction:2 hierarchy:3 strengthen:1 exact:4 programming:2 element:1 trend:1 satisfying:1 particularly:2 cut:1 observed:1 epsrc:2 ep:2 worst:3 revisiting:1 cycle:4 connected:3 highest:1 removed:1 ran:1 balanced:4 intuition:1 hypergraph:15 cam:2 tight:3 purely:1 basis:4 easily:2 joint:1 various:4 kolmogorov:1 separated:1 effective:2 artificial:3 refined:1 heuristic:19 larger:2 valued:1 plausible:1 say:4 tightness:4 drawing:1 statistic:3 gi:5 transform:1 superscript:1 isbn:1 took:1 clamp:7 interaction:4 strengthening:1 relevant:2 combining:1 loop:1 iff:4 achieve:3 convergence:1 cluster:4 optimum:1 double:1 adam:15 perfect:1 help:1 derive:2 develop:3 ac:2 depending:1 ij:5 odd:1 minor:2 received:1 albers:1 strong:6 recovering:3 implemented:1 pot:4 indicate:2 implies:1 met:2 treewidth:1 direction:1 hull:1 enable:2 unnamed:1 generalization:1 anonymous:1 proposition:10 extension:1 hold:1 around:1 considered:1 exp:4 mapping:5 m0:3 vary:4 achieves:1 smallest:1 a2:4 combinatorial:1 tanh:4 label:1 saw:1 council:1 establishes:1 tool:2 hope:1 mit:1 clearly:2 always:2 rather:1 jaakkola:2 focus:1 rank:1 indicates:1 contrast:1 sense:6 helpful:9 inference:49 posteriori:1 typically:5 unlikely:1 integrated:1 initially:1 margalit:1 wij:10 selects:1 issue:1 arg:7 constrained:1 marginal:19 field:2 libdai:2 never:2 once:2 beach:1 x4:7 flipped:1 biology:1 icml:2 djolonga:1 others:2 np:1 simplify:2 t2:3 intelligent:1 randomly:1 tightly:2 m4:5 powerset:1 geometry:1 attempt:1 ab:2 friedman:1 interest:4 highly:1 implication:2 accurate:1 edge:9 reroot:2 respective:1 unless:1 tree:2 indexed:1 theoretical:6 increased:1 formalism:1 earlier:5 modeling:1 facet:1 column:1 loopy:1 cost:5 vertex:2 subset:10 uniform:1 gr:1 weller:14 aw:1 eec:1 combined:1 st:2 international:2 sensitivity:1 siam:2 probabilistic:1 physic:1 together:3 connecting:1 again:2 central:1 suspension:4 thesis:1 choose:4 admit:1 leading:3 potential:103 converted:2 singleton:3 attaining:1 bold:1 coefficient:7 vi:1 performed:1 try:1 break:1 later:1 analyze:1 red:1 recover:6 contribution:2 accuracy:5 kaufmann:1 maximized:1 yield:3 identify:1 generalize:1 weak:1 explain:1 checked:1 definition:18 against:1 energy:1 semimetric:2 proof:4 mi:21 gain:1 proved:3 popular:1 recall:1 color:3 improves:1 carefully:2 back:1 trw:2 appears:1 higher:16 supermodular:4 follow:1 wei:1 evaluated:1 though:2 generality:1 just:2 implicit:1 hand:1 ei:2 wmax:8 trust:1 propagation:4 perhaps:1 pulling:1 believe:1 usa:1 effect:4 building:1 verify:1 contain:2 requiring:1 true:2 www:1 hence:5 symmetric:2 reweighted:1 attractive:1 rooted:18 scorei:1 coincides:1 generalized:4 pdf:1 complete:4 demonstrate:7 reasoning:1 variational:5 empirically:4 physical:1 extend:4 hypergraphs:7 rth:1 marginals:8 significant:5 refer:1 cambridge:4 grid:3 consistency:2 heskes:2 inclusion:2 maxcut:2 centre:2 mathematics:1 submodular:2 language:1 had:1 l3:14 specification:1 add:1 own:1 showed:4 forbidden:1 perspective:2 optimizing:1 csps:1 termed:2 binary:9 hyperedge:5 xe:10 preserving:7 morgan:1 greater:1 relaxed:1 additional:1 mr:1 elidan:1 signal:1 ii:1 relates:1 desirable:1 alan:2 long:1 equally:1 a1:6 impact:3 variant:1 scalable:1 ae:3 vision:2 essentially:4 metric:1 represent:1 achieved:1 background:1 lbp:1 krause:1 else:1 diagram:1 hyperedges:7 source:1 extra:1 tri:8 comment:1 strict:1 suspect:1 induced:1 elegant:1 undirected:1 tend:1 facilitates:1 legend:1 effectiveness:1 jordan:2 call:1 integer:4 config:2 easy:1 variety:1 xj:11 affect:3 marginalization:1 gave:1 fit:3 identified:1 opposite:1 reduce:1 idea:2 expression:1 handled:1 url:1 sontag:4 logdet:1 repeatedly:1 useful:3 n510129:1 clear:1 locally:6 bijections:1 reduced:1 generate:1 http:1 exist:1 canonical:1 sign:1 leverhulme:1 discrete:5 write:5 shall:2 express:2 key:2 neither:1 verified:1 graph:3 relaxation:18 sum:1 run:7 turing:2 master:1 uncertainty:1 striking:1 extends:1 almost:5 family:2 appendix:11 comparable:1 bound:3 display:1 correspondence:1 replaces:1 strength:1 k8:1 ladicky:1 x2:4 u1:3 argument:1 extremely:1 aw665:1 rendered:1 conjecture:1 poor:1 smaller:1 across:2 slightly:1 intimately:1 lp:8 appealing:1 intuitively:1 restricted:1 invariant:2 computationally:2 previously:3 discus:1 mechanism:2 needed:1 end:1 parametrize:1 generalizes:1 junction:1 operation:1 yedidia:1 apply:2 observe:5 symmetrized:4 original:10 broaden:1 denotes:1 include:1 publishing:1 graphical:12 lock:1 xw:1 calculating:1 establish:1 unchanged:2 objective:1 added:4 question:3 flipping:2 win:1 separate:1 thank:1 outer:1 polytope:20 extent:1 reinelt:1 toward:2 enforcing:1 willsky:1 relationship:1 illustration:1 equivalently:1 potentially:1 constructively:1 design:1 motivates:2 perform:2 allowing:1 contributed:1 upper:1 observation:5 markov:2 finite:1 saturates:1 looking:1 incorporated:1 relational:1 august:1 jebara:2 introduced:4 required:3 connection:1 z1:1 c4:3 xvi:2 polytopes:7 pearl:1 nip:3 trans:1 beyond:1 usually:1 regime:1 max:9 including:2 green:1 belief:4 gaining:1 power:1 critical:2 wainwright:3 difficulty:1 natural:5 indicator:1 improve:3 library:1 acknowledges:2 understanding:2 l2:3 acknowledgement:1 mooij:1 loss:1 fully:1 highlight:2 remarkable:2 foundation:1 ivn:1 degree:2 affine:5 consistent:11 pi:7 row:2 deza:2 free:5 allow:2 deeper:1 institute:2 neighbor:1 taking:1 characterizing:1 benefit:8 curve:1 author:1 collection:4 disturbing:1 universally:15 commonly:1 qualitatively:1 avoided:1 rowland:3 transaction:2 approximate:8 cutting:1 global:1 uai:3 xi:40 alternatively:1 continuous:1 triplet:4 table:3 bethe:3 robust:1 ca:1 composing:1 symmetry:2 tschel:1 constructing:1 jaimovich:1 did:1 aistats:4 whole:3 edition:1 allowed:1 xu:23 x1:7 strengthened:1 exponential:2 prominently:1 clamped:6 jmlr:1 admissible:3 tang:1 theorem:14 removing:1 specific:2 showing:1 normalizing:2 intractable:2 exists:1 adding:1 effectively:3 importance:1 illustrates:1 t4:3 clamping:13 easier:3 entropy:2 generalizing:1 led:1 simply:1 explore:1 absorbed:1 u2:4 maxw:2 applies:3 springer:1 satisfies:1 relies:2 goal:2 careful:1 towards:1 experimentally:1 hard:2 specifically:3 determined:1 reducing:1 torr:1 domke:1 lemma:8 called:1 experimental:1 l4:2 maxe:3 select:2 mark:1 support:2 artifical:1 |
6,219 | 6,626 | The Unreasonable Effectiveness of Structured
Random Orthogonal Embeddings
Krzysztof Choromanski ?
Google Brain Robotics
[email protected]
Mark Rowland ?
University of Cambridge
[email protected]
Adrian Weller
University of Cambridge and Alan Turing Institute
[email protected]
Abstract
We examine a class of embeddings based on structured random matrices with
orthogonal rows which can be applied in many machine learning applications
including dimensionality reduction and kernel approximation. For both the JohnsonLindenstrauss transform and the angular kernel, we show that we can select matrices
yielding guaranteed improved performance in accuracy and/or speed compared to
earlier methods. We introduce matrices with complex entries which give significant
further accuracy improvement. We provide geometric and Markov chain-based
perspectives to help understand the benefits, and empirical results which suggest
that the approach is helpful in a wider range of applications.
1
Introduction
Embedding methods play a central role in many machine learning applications by projecting feature
vectors into a new space (often nonlinearly), allowing the original task to be solved more efficiently.
The new space might have more or fewer dimensions depending on the goal. Applications include
the Johnson-Lindenstrauss Transform for dimensionality reduction (JLT, Johnson and Lindenstrauss,
1984) and kernel methods with random feature maps (Rahimi and Recht, 2007). The embedding can
be costly hence many fast methods have been developed, see ?1.1 for background and related work.
We present a general class of random embeddings based on particular structured random matrices
with orthogonal rows, which we call random ortho-matrices (ROMs); see ?2. We show that ROMs
may be used for the applications above, in each case demonstrating improvements over previous
methods in statistical accuracy (measured by mean squared error, MSE), in computational efficiency
(while providing similar accuracy), or both. We highlight the following contributions:
? In ?3: The Orthogonal Johnson-Lindenstrauss Transform (OJLT) for dimensionality reduction.
We prove this has strictly smaller MSE than the previous unstructured JLT mechanisms. Further,
OJLT is as fast as the fastest previous JLT variants (which are structured).
? In ?4: Estimators for the angular kernel (Sidorov et al., 2014) which guarantee better MSE. The
angular kernel is important for many applications, including natural language processing (Sidorov
et al., 2014), image analysis (J?gou et al., 2011), speaker representations (Schmidt et al., 2014)
and tf-idf data sets (Sundaram et al., 2013).
? In ?5: Two perspectives on the effectiveness of ROMs to help build intuitive understanding.
In ?6 we provide empirical results which support our analysis, and show that ROMs are effective for
a still broader set of applications. Full details and proofs of all results are in the Appendix.
?
equal contribution
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
1.1
Background and related work
Our ROMs can have two forms (see ?2 for details): (i) a Gort is a random Gaussian matrix conditioned on rows being orthogonal; or (ii) an SD-product matrix is formed by multiplying some
number k of SD blocks, each of which is highly structured, typically leading to fast computation
of products. Here S is a particular structured matrix, and D is a random diagonal matrix; see ?2
for full details. Our SD block generalizes an HD block, where H is a Hadamard matrix, which
received previous attention. Earlier approaches to embeddings have explored using various structured
matrices, including particular versions of one or other of our two forms, though in different contexts.
For dimensionality reduction, Ailon and Chazelle (2006) used a single HD block as a way to spread
out the mass of a vector over all dimensions before applying a sparse Gaussian matrix. Choromanski
and Sindhwani (2016) also used just one HD block as part of a larger structure. Bojarski et al. (2017)
discussed using k = 3 HD blocks for locality-sensitive hashing methods but gave no concrete results
for their application to dimensionality reduction or kernel approximation. All these works, and other
earlier approaches (Hinrichs and Vyb?ral, 2011; Vyb?ral, 2011; Zhang and Cheng, 2013; Le et al.,
2013; Choromanska et al., 2016), provided computational benefits by using structured matrices with
less randomness than unstructured iid Gaussian matrices, but none demonstrated accuracy gains.
Yu et al. (2016) were the first to show that Gort -type matrices can yield improved accuracy, but their
theoretical result applies only asymptotically for many dimensions, only for the Gaussian kernel and
for just one specific orthogonal transformation, which is one instance of the larger class we consider.
Their theoretical result does not yield computational benefits. Yu et al. (2016) did explore using a
number k of HD blocks empirically, observing good computational and statistical performance for
k = 3, but without any theoretical accuracy guarantees. It was left as an open question why matrices
formed by a small number of HD blocks can outperform non-discrete transforms.
In contrast, we are able to prove that ROMs yield improved MSE in several settings and for many of
them for any number of dimensions. In addition, SD-product matrices can deliver computational
speed benefits. We provide initial analysis to understand why k = 3 can outperform the state-ofthe-art, why odd k yields better results than even k, and why higher values of k deliver decreasing
additional benefits (see ?3 and ?5).
2
The family of Random Ortho-Matrices (ROMs)
Random ortho-matrices (ROMs) are taken from two main classes of distributions defined below that
require the rows of sampled matrices to be orthogonal. A central theme of the paper is that this
orthogonal structure can yield improved statistical performance. We shall use bold uppercase (e.g.
M) to denote matrices and bold lowercase (e.g. x) for vectors.
Gaussian orthogonal matrices. Let G be a random matrix taking values in Rm?n with iid N (0, 1)
elements, which we refer to as an unstructured Gaussian matrix. The first ROM distribution we
consider yields the random matrix Gort , which is defined as a random Rn?n matrix given by first
taking the rows of the matrix to be a uniformly random orthonormal basis, and then independently
scaling each row, so that the rows marginally have multivariate Gaussian N (0, I) distributions. The
random variable Gort can then be extended to non-square matrices by either stacking independent
copies of the Rn?n random matrices, and deleting superfluous rows if necessary. The orthogonality
of the rows of this matrix has been observed to yield improved statistical properties for randomized
algorithms built from the matrix in a variety of applications.
SD-product matrices. Our second class of distributions is motivated by the desire to obtain similar
statistical benefits of orthogonality to Gort , whilst gaining computational efficiency by employing
more structured matrices. We call this second class SD-product matrices. These take the more
Qk
structured form i=1 SDi , where S = {si,j } ? Rn?n has orthogonal rows, |si,j | = ?1n ?i, j ?
Qk
{1, . . . , n}; and the (Di )ki=1 are independent diagonal matrices described below. By i=1 SDi , we
mean the matrix product (SDk ) . . . (SD1 ). This class includes as particular cases several recently
introduced random matrices (e.g. Andoni et al., 2015; Yu et al., 2016), where good empirical
performance was observed. We go further to establish strong theoretical guarantees, see ?3 and ?4.
2
A prominent example of an S matrix is thenormalized Hadamard
matrix H, defined recursively by
H
H
i?1
i?1
1
H1 = (1), and then for i > 1, Hi = ?2
. Importantly, matrix-vector products
Hi?1 ?Hi?1
with H are computable in O(n log n) time via the fast Walsh-Hadamard transform, yielding large
computational savings. In addition, H matrices enable a significant space advantage: since the
fast Walsh-Hadamard transform can be computed without explicitly storing H, only O(n) space is
required to store the diagonal elements of (Di )ki=1 . Note that these Hn matrices are defined only for
n a power of 2, but if needed, one can always adjust data by padding with 0s to enable the use of ?the
next larger? H, doubling the number of dimensions in the worst case.
Matrices H are representatives of a much larger family in S which also attains computational savings.
These are L2 -normalized versions of Kronecker-product matrices of the form A1 ? ... ? Al ? Rn?n
for l ? N, where ? stands for a Kronecker product and blocks Ai ? Rd?d have entries of the
same magnitude and pairwise orthogonal rows each. For these matrices, matrix-vector products are
computable in O(n(2d ? 1) logd (n)) time (Zhang et al., 2015).
S includes also the Walsh matrices W = {wi,j } ? Rn?n , where wi,j = ?1n (?1)iN ?1 j0 +...+i0 jN ?1
and iN ?1 ...i0 , jN ?1 ...j0 are binary representations of i and j respectively.
For diagonal (Di )ki=1 , we mainly consider Rademacher entries leading to the following matrices.
(R) k
)i=1
Definition 2.1. The S-Rademacher random matrix with k ? N blocks is below, where (Di
are diagonal with iid Rademacher random variables [i.e. Unif({?1})] on the diagonals:
(k)
MSR
=
k
Y
(R)
SDi
.
(1)
i=1
Having established the two classes of ROMs, we next apply them to dimensionality reduction.
3
The Orthogonal Johnson-Lindenstrauss Transform (OJLT)
Let X ? Rn be a dataset of n-dimensional real vectors. The goal of dimensionality reduction via
F
random projections is to transform linearly each x ? X by a random mapping x 7? x0 , where:
n
m
0 > 0
F : R ? R for m < n, such that for any x, y ? X the following holds: (x ) y ? x> y. If
we furthermore have E[(x0 )> y0 ] = x> y then the dot-product estimator is unbiased. In particular,
this dimensionality reduction mechanism should in expectation preserve information about vectors?
norms, i.e. we should have: E[kx0 k22 ] = kxk22 for any x ? X .
The standard JLT mechanism uses the randomized linear map F = ?1m G, where G ? Rm?n is as
in ?2, requiring mn multiplications to evaluate. Several fast variants (FJLTs) have been proposed by
replacing G with random structured matrices, such as sparse or circulant Gaussian matrices (Ailon
and Chazelle, 2006; Hinrichs and Vyb?ral, 2011; Vyb?ral, 2011; Zhang and Cheng, 2013). The fastest
of these variants has O(n log n) time complexity, but at a cost of higher MSE for dot-products.
Our Orthogonal Johnson-Lindenstrauss Transform (OJLT) is obtained by replacing the unstructured
(k),sub
random matrix G with a sub-sampled ROM from ?2: either Gort , or a sub-sampled version MSR
of the S-Rademacher ROM, given by sub-sampling rows from the left-most S matrix in the product.
We sub-sample since m < n. We typically assume uniform sub-sampling without replacement. The
resulting dot-product estimators for vectors x, y ? X are given by:
b base (x, y) = 1 (Gx)> (Gy) [unstructured iid baseline, previous state-of-the-art accuracy],
K
m
m
>
(k),sub
b ort (x, y) = 1 (Gort x)> (Gort y),
b (k) (x, y) = 1 M(k),sub x
K
K
MSR y . (2)
m
m
SR
m
m
We contribute the following closed-form expressions, which exactly quantify the mean-squared error
b
(MSE) for these three estimators. Precisely, the MSE of an
h estimator K(x, y) iof the inner product
b
b
hx, yi for x, y ? X is defined to be MSE(K(x, y)) = E (K(x, y) ? hx, yi2 ) . See the Appendix
for detailed proofs of these results and all others in this paper.
3
b base of x, y ? Rn using mLemma 3.1. The MSE of the unstructured JLT dot-product estimator K
m
base
b m (x, y)) = 1 ((x> y)2 + kxk2 kyk2 ).
dimensional random feature maps is unbiased, with MSE(K
2
2
m
b ort is unbiased and satisfies, for n ? 4:
Theorem 3.2. The estimator K
m
b ort (x, y))
MSE(K
m
base
b
=MSE(Km
(x, y)) +
m
kxk22 kyk22 n2
1
1
1
2
?
(I(n ? 3) ? I(n ? 1))I(n ? 4) cos (?) +
+
m ? 1 4I(n ? 3)I(n ? 4)
n n+2
2
1
1
1
I(n ? 1) (I(n ? 4) ? I(n ? 2))
?
cos2 (?) ?
? hx, yi2 ,
n?2 n
2
(3)
?
R?
where I(n) = 0 sinn (x)dx = ??((n+1)/2)
?(n/2+1) .
(k)
bm
Theorem 3.3 (Key result). The OJLT estimator K
(x, y) with k blocks, using m-dimensional
random feature maps and uniform sub-sampling policy without replacement, is unbiased with
1 n?m
(k)
b
MSE(Km (x, y)) =
((x> y)2 + kxk2 kyk2 ) +
(4)
m n?1
k?1
n
X (?1)r 2r
(?1)k 2k X 2 2
>
2
2
2
(2(x y) + kxk kyk ) +
x y .
nr
nk?1 i=1 i i
r=1
Proof (Sketch). For k = 1, the random projection matrix is given by sub-sampling rows from SD1 ,
and the computation can be carried out directly. For k ? 1, the proof proceeds by induction.
The random projection matrix in the general case is given by sub-sampling rows of the matrix
SDk ? ? ? SD1 . By writing the MSE as an expectation and using the law of conditional expectations
conditioning on the value of the first k ? 1 random matrices Dk?1 , . . . , D1 , the statement of the
theorem for 1 SD block and for k ? 1 SD blocks can be neatly combined to yield the result.
To our knowledge, it has not previously been possible to provide theoretical guarantees that
SD-product matrices outperform iid matrices. Combining Lemma 3.1 with Theorem 3.3 yields the
following important result.
(k)
bm
(subsampling
Corollary 3.4 (Theoretical guarantee of improved performance). Estimators K
b base .
without replacement) yield guaranteed lower MSE than K
m
(k)
ort
bm
bm
It is not yet clear when K
is better or worse than K
; we explore this empirically in ?6.
Theorem 3.3 shows that there are diminishing MSE benefits to using a large number k of SD
(2k?1)
bm
blocks. Interestingly, odd k is better than even: it is easy to observe that MSE(K
(x, y)) <
(2k)
(2k+1)
b m (x, y)) > MSE(K
bm
MSE(K
(x, y)). These observations, and those in ?5, help to understand why empirically k = 3 was previously observed to work well (Yu et al., 2016).
If we take S to be a normalized Hadamard matrix H, then even though we are using sub-sampling,
and hence the full computational benefits of the Walsh-Hadamard transform are not available, still
(k)
bm
K
achieves improved MSE compared to the base method with less computational effort, as follows.
Lemma 3.5. There exists an algorithm (see Appendix for details) which computes an embedding for
(k)
bm
a given datapoint x using K
with S set to H and uniform sub-sampling policy in expected time
min{O((k ? 1)n log(n) + nm ? (m?1)m
, kn log(n)}.
2
Note that for m = ?(k log(n)) or if k = 1, the time complexity is smaller than the brute force
?(nm). The algorithm uses a simple observation that one can reuse calculations conducted for the
upper half of the Hadamard matrix while performing computations involving rows from its other half,
instead of running these calculations from scratch (details in the Appendix).
An alternative to sampling without replacement is deterministically to choose the first m rows. In our
experiments in ?6, these two approaches yield the same empirical performance, though we expect
4
that the deterministic method could perform poorly on adversarially chosen data. The first m rows
approach can be realized in time O(n log(m) + (k ? 1)n log(n)) per datapoint.
Theorem 3.3 is a key result in this paper, demonstrating that SD-product matrices yield both statistical
and computational improvements compared to the base iid procedure, which is widely used in practice.
We next show how to obtain further gains in accuracy.
3.1
Complex variants of the OJLT
We show that the MSE benefits of Theorem 3.3 may be markedly improved by using SD-product
(k)
matrices with complex entries MSH . Specifically, we consider the variant S-Hybrid random matrix
(U )
below, where Dk is a diagonal matrix with iid Unif(S 1 ) random variables on the diagonal, inde(R) k?1
pendent of (Di )i=1 , and S 1 is the unit circle of C. We use the real part of the Hermitian product
between projections as a dot-product estimator; recalling the definitions of ?2, we use:
(k)
MSH
=
(U )
SDk
k?1
Y
(R)
SDi
>
1
(k),sub
(k),sub
H,(k)
b
Km (x, y) =
Re MSH x
MSH y .
m
,
i=1
(5)
Remarkably, this complex variant yields exactly half the MSE of the OJLT estimator.
H,(k)
b m (x, y), applying uniform sub-sampling without replacement, is
Theorem 3.6. The estimator K
H,(k)
(k)
bm
bm
(x, y)).
unbiased and satisfies: MSE(K
(x, y)) = 12 MSE(K
(k)
b m . However,
This large factor of 2 improvement could instead be obtained by doubling m for K
this would require doubling the number of parameters for the transform, whereas the S-Hybrid
(U )
estimator requires additional storage only for the complex parameters in the matrix Dk . Strikingly,
it is straightforward to extend the proof of Theorem 3.6 (see Appendix) to show that rather than
(k),sub
taking the complex random variables in MSH
to be Unif(S 1 ), it is possible to take them to be
Unif({1, ?1, i, ?i}) and still obtain exactly the same benefit in MSE.
(U )
H,(k)
bm
Theorem 3.7. For the estimator K
defined in Equation (5): replacing the random matrix Dk
(which has iid Unif(S 1 ) elements on the diagonal) with instead a random diagonal matrix having iid
Unif({1, ?1, i, ?i}) elements on the diagonal, does not affect the MSE of the estimator.
It is natural to wonder if using an SD-product matrix with more complex random variables (for all
SD blocks) would improve performance still further. However, interestingly, this appears not to be
the case; details are provided in the Appendix ?8.7.
3.2
Sub-sampling with replacement
Our results above focus on SD-product matrices where rows have been sub-sampled without
replacement. Sometimes (e.g. for parallelization) it can be convenient instead to sub-sample with
replacement. As might be expected, this leads to worse MSE, which we can quantify precisely.
(k)
H,(k)
bm
bm
Theorem 3.8. For each of the estimators K
and K
, if uniform sub-sampling with (rather
n?1
than without) replacement is used then the MSE is worsened by a multiplicative constant of n?m
.
4
Kernel methods with ROMs
ROMs can also be used to construct high-quality random feature maps for non-linear kernel
approximation. We analyze here the angular kernel, an important example of a Pointwise Nonlinear
Gaussian kernel (PNG), discussed in more detail at the end of this section.
Definition 4.1. The angular kernel K ang is defined on Rn by K ang (x, y) = 1 ?
is the angle between x and y.
5
2?x,y
? ,
where ?x,y
To employ random feature style approximations to this kernel, we first observe it may be rewritten as
K ang (x, y) = E [sign(Gx)sign(Gy)] ,
where G ? R1?n is an unstructured isotropic Gaussian vector. This motivates approximations of the
form:
b ang m(x, y) = 1 sign(Mx)> sign(My),
K
(6)
m
where M ? Rm?n is a random matrix, and the sign function is applied coordinate-wise. Such
kernel estimation procedures are heavily used in practice (Rahimi and Recht, 2007), as they allow
fast approximate linear methods to be used (Joachims, 2006) for inference tasks. If M = G, the
unstructured Gaussian matrix, then we obtain the standard random feature estimator. We shall contrast
this approach against the use of matrices from the ROMs family.
When constructing random feature maps for kernels, very often m > n. In this case, our structured
mechanism can be applied by concatenating some number of independent structured blocks. Our
theoretical guarantees will be given just for one block, but can easily be extended to a larger number
of blocks since different blocks are independent.
ang,base
bm
The standard random feature approximation K
for approximating the angular kernel is
defined by taking M to be G, the unstructured Gaussian matrix, in Equation (6), and satisfies the
following.
ang,base
ang,base
bm
bm
Lemma 4.2. The estimator K
is unbiased and MSE(K
(x, y)) =
4?x,y (???x,y )
.
m? 2
b ang (x, y) of the true angular kernel K ang (x, y) is defined analogously
The MSE of an estimator K
to the MSE of an estimator of the dot product, given in ?3. Our main result regarding angular kernels
ang,ort
bm
states that if we instead take M = Gort in Equation (6), then we obtain an estimator K
with
strictly smaller MSE, as follows.
ang,ort
bm
Theorem 4.3. Estimator K
is unbiased and satisfies:
ang,ort
ang,base
bm
bm
MSE(K
(x, y)) < MSE(K
(x, y)).
ang,M
bm
We also derive a formula for the MSE of an estimator K
of the angular kernel which replaces G
with an arbitrary random matrix M and uses m random feature maps. The formula is helpful to see
how the quality of the estimator depends on the probabilities that the projections of the rows of M are
contained in some particular convex regions of the 2-dimensional space Lx,y spanned by datapoints
x and y. For an illustration of the geometric definitions introduced in this Section, see Figure 1. The
formula depends on probabilities involving events Ai = {sgn((ri )T x) 6= sgn((ri )T y)}, where
ri stands for the ith row of the structured matrix. Notice that Ai = {riproj ? Cx,y }, where riproj
stands for the projection of ri into Lx,y and Cx,y is the union of two cones in Lx,y , each of angle ?x,y .
ang,M
bm
Theorem 4.4. Estimator K
satisfies the following, where: ?i,j = P[Ai ? Aj ] ? P[Ai ]P[Aj ]:
?
?
"
#
m
m
X
X
X
1
4
?x,y 2
ang,M
bm
(x, y)) = 2 m ?
MSE(K
(1 ? 2P[Ai ])2 + 2 ? (P[Ai ] ?
) +
?i,j ? .
m
m
?
i=1
i=1
i6=j
Note that probabilities P[Ai ] and ?i,j depend on the choice of M. It is easy to prove that for
?
unstructured G and Gort we have: P[Ai ] = x,y
? . Further, from the independence of the rows of
G, ?i,j = 0 for i 6= j. For unstructured G we obtain Lemma 4.2. Interestingly, we see that to
prove Theorem 4.3, it suffices to show ?i,j < 0, which is the approach we take (see Appendix). If
(k)
?
we replace G with MSR , then the expression = P[Ai ] ? x,y
? does not depend on i. Hence, the
angular
kernel
estimator
based
on
Hadamard
matrices
gives
smaller
MSE estimator if and only if
P
2
i6=j ?i,j + m < 0. It is not yet clear if this holds in general.
As alluded to at the beginning of this section, the angular kernel may be viewed as a member of a wie
family of kernels known as Pointwise Nonlinear Gaussian kernels.
6
Figure 1: Left part: Left: g1 is orthogonal to Lx,y . Middle: g1 ? Lx,y . Right: g1 is close to orthogonal to
Lx,y . Right part: Visualization of the Cayley graph explored by the Hadamard-Rademacher process in two
dimensions. Nodes are colored red, yellow, light blue, dark blue, for Cayley distances of 0, 1, 2, 3 from the
identity matrix respectively. See text in ?5.
f
Definition 4.5. For a given
function f , the
Pointwise Nonlinear Gaussian kernel (PNG) K is
defined by K f (x, y) = E f (gT x)f (gT y) , where g is a Gaussian vector with i.i.d N (0, 1) entries.
Many prominent examples of kernels (Williams, 1998; Cho and Saul, 2009) are PNGs. Wiener?s
tauberian theorem shows that all stationary kernels may be approximated arbitrarily well by sums of
PNGs (Samo and Roberts, 2015). In future work we hope to explore whether ROMs can be used to
achieve statistical benefit in estimation tasks associated with a wider range of PNGs.
5
Understanding the effectiveness of orthogonality
Here we build intuitive understanding for the effectiveness of ROMs. We examine geometrically the
angular kernel (see ?4), then discuss a connection to random walks over orthogonal matrices.
Angular kernel. As noted above for the Gort -mechanism, smaller MSE than that for unstructured
G is implied by the inequality P[Ai ? Aj ] < P[Ai ]P[Aj ], which is equivalent to: P[Aj |Ai ] < P[Aj ].
Now it becomes clear why orthogonality is crucial. Without loss of generality take: i = 1, j = 2, and
let g1 and g2 be the first two rows of Gort .
Consider first the extreme case (middle of left part of Figure 1), where all vectors are 2-dimensional.
Recall definitions from just after Theorem 4.3. If g1 is in Cx,y then it is much less probable for
g2 also to belong to Cx,y . In particular, if ? < ?2 then the probability is zero. That implies the
inequality. On the other hand, if g1 is perpendicular to Lx,y then conditioning on Ai does not have
any effect on the probability that g2 belongs to Cx,y (left subfigure of Figure 1). In practice, with high
probability the angle ? between g1 and Lx,y is close to ?2 , but is not exactly ?2 . That again implies
that conditioned on the projection gp1 of g1 into Lx,y to be in Cx,y , the more probable directions of
gp2 are perpendicular to gp1 (see: ellipsoid-like shape in the right subfigure of Figure 1 which is the
projection of the sphere taken from the (n ? 1)-dimensional space orthogonal to g1 into Lx,y ). This
makes it less probable for gp2 to be also in Cx,y . The effect is subtle since ? ? ?2 , but this is what
provides superiority of the orthogonal transformations over state-of-the-art ones in the angular kernel
approximation setting.
Markov chain perspective. We focus on Hadamard-Rademacher random matrices HDk ...HD1 ,
a special case of the SD-product matrices described in Section 2. Our aim is to provide intuition
for how the choice of k affects the quality of the random matrix, following our earlier observations
just after Corollary 3.4, which indicated that for SD-product matrices, odd values of k yield greater
benefits than even values, and that there are diminishing benefits from higher values of k. We proceed
by casting the random matrices into the framework of Markov chains.
Definition 5.1. The Hadamard-Rademacher process in n dimensions is the Markov chain (Xk )?
k=0
taking values in the orthogonal group O(n), with X0 = I almost surely, and Xk = HDk Xk?1
almost surely, where H is the normalized Hadamard matrix in n dimensions, and (Dk )?
k=1 are iid
diagonal matrices with independent Rademacher random variables on their diagonals.
Constructing an estimator based on Hadamard-Rademacher matrices is equivalent to simulating
several time steps from the Hadamard-Rademacher process. The quality of estimators based on
Hadamard-Rademacher random matrices comes from a quick mixing property of the corresponding
7
(a) g50c - pointwise evalu- (b) random - angular kernel (c) random - angular kernel (d) g50c - inner product esation MSE for inner product
with true angle ?/4
timation MSE for variants of
estimation
3-block SD-product matrices.
(e) LETTER - dot-product
(f) USPS - dot-product
(g) LETTER - angular kernel (h) USPS - angular kernel
Figure 2: Top row: MSE curves for pointwise approximation of inner product and angular kernels on the
g50c dataset, and randomly chosen vectors. Bottom row: Gram matrix approximation error for a variety of
data sets, projection ranks, transforms, and kernels. Note that the error scaling is dependent on the application.
Markov chain. The following demonstrates attractive properties of the chain in low dimensions.
Proposition 5.2. The Hadamard-Rademacher process in two dimensions: explores a state-space of
16 orthogonal matrices, is ergodic with respect to the uniform distribution on this set, has period 2,
the diameter of the Cayley graph of its state space is 3, and the chain is fully mixed after 3 time steps.
This proposition, and the Cayley graph corresponding to the Markov chain?s state space (Figure 1
right), illustrate the fast mixing properties of the Hadamard-Rademacher process in low dimensions;
this agrees with the observations in ?3 that there are diminishing returns associated with using a large
number k of HD blocks in an estimator. The observation in Proposition 5.2 that the Markov chain
has period 2 indicates that we should expect different behavior for estimators based on odd and even
numbers of blocks of HD matrices, which is reflected in the analytic expressions for MSE derived in
Theorems 3.3 and 3.6 for the dimensionality reduction setup.
6
Experiments
We present comparisons of estimators introduced in ?3 and ?4, illustrating our theoretical results, and
further demonstrating the empirical success of ROM-based estimators at the level of Gram matrix
approximation. We compare estimators based on: unstructured Gaussian matrices G, matrices Gort ,
S-Rademacher and S-Hybrid matrices with k = 3 and different sub-sampling strategies. Results
for k > 3 do not show additional statistical gains empirically. Additional experimental results,
including a comparison of estimators using different numbers of SD blocks, are in the Appendix ?10.
Throughout, we use the normalized Hadamard matrix H for the structured matrix S.
6.1
Pointwise kernel approximation
Complementing the theoretical results of ?3 and ?4, we provide several salient comparisons of the
various methods introduced - see Figure 2 top. Plots presented here (and in the Appendix) compare
MSE for dot-product and angular and kernel. They show that estimators based on Gort , S-Hybrid
and S-Rademacher matrices without replacement, or using the first m rows, beat the state-of-the-art
unstructured G approach on accuracy for all our different datasets in the JLT setup. Interestingly, the
latter two approaches give also smaller MSE than Gort -estimators. For angular kernel estimation,
where sampling is not relevant, we see that Gort and S-Rademacher approaches again outperform
the ones based on matrices G.
8
6.2
Gram matrix approximation
Moving beyond the theoretical guarantees established in ?3 and ?4, we show empirically that the
superiority of estimators based on ROMs is maintained at the level of Gram matrix approximation.
We compute Gram matrix approximations (with respect to both standard dot-product, and angular
b 2 /kKk2
kernel) for a variety of datasets. We use the normalized Frobenius norm error kK ? Kk
as our metric (as used by Choromanski and Sindhwani, 2016), and plot the mean error based on
1,000 repetitions of each random transform - see Figure 2 bottom. The Gram matrices are computed
on a randomly selected subset of 550 data points from each dataset. As can be seen, the S-Hybrid
estimators using the ?no-replacement? or ?first m rows? sub-sampling strategies outperform even
the orthogonal Gaussian ones in the dot-product case. For the angular case, the Gort -approach and
S-Rademacher approach are practically indistinguishable.
7
Conclusion
We defined the family of random ortho-matrices (ROMs). This contains the SD-product matrices,
which include a number of recently proposed structured random matrices. We showed theoretically
and empirically that ROMs have strong statistical and computational properties (in several cases
outperforming previous state-of-the-art) for algorithms performing dimensionality reduction and
random feature approximations of kernels. We highlight Corollary 3.4, which provides a theoretical
guarantee that SD-product matrices yield better accuracy than iid matrices in an important dimensionality reduction application (we believe the first result of this kind). Intriguingly, for dimensionality
reduction, using just one complex structured matrix yields random features of much better quality.
We provided perspectives to help understand the benefits of ROMs, and to help explain the behavior
of SD-product matrices for various numbers of blocks. Our empirical findings suggest that our
theoretical results might be further strengthened, particularly in the kernel setting.
Acknowledgements
We thank Vikas Sindhwani at Google Brain Robotics and Tamas Sarlos at Google Research for
inspiring conversations that led to this work. We thank Matej Balog, Maria Lomeli, Jiri Hron and
Dave Janz for helpful comments. MR acknowledges support by the UK Engineering and Physical
Sciences Research Council (EPSRC) grant EP/L016516/1 for the University of Cambridge Centre
for Doctoral Training, the Cambridge Centre for Analysis. AW acknowledges support by the Alan
Turing Institute under the EPSRC grant EP/N510129/1, and by the Leverhulme Trust via the CFI.
9
References
N. Ailon and B. Chazelle. Approximate nearest neighbors and the fast Johnson-Lindenstrauss transform. In
STOC, 2006.
A. Andoni, P. Indyk, T. Laarhoven, I. Razenshteyn, and L. Schmidt. Practical and optimal LSH for angular
distance. In NIPS, 2015.
M. Bojarski, A. Choromanska, K. Choromanski, F. Fagan, C. Gouy-Pailler, A. Morvan, N. Sakr, T. Sarlos, and
J. Atif. Structured adaptive and random spinners for fast machine learning computations. In to appear in
AISTATS, 2017.
Y. Cho and L. K. Saul. Kernel methods for deep learning. In NIPS, 2009.
A. Choromanska, K. Choromanski, M. Bojarski, T. Jebara, S. Kumar, and Y. LeCun. Binary embeddings with
structured hashed projections. In ICML, 2016.
K. Choromanski and V. Sindhwani. Recycling randomness with structure for sublinear time kernel expansions.
In ICML, 2016.
A. Hinrichs and J. Vyb?ral. Johnson-Lindenstrauss lemma for circulant matrices. Random Structures &
Algorithms, 39(3):391?398, 2011.
H. J?gou, M. Douze, and C. Schmid. Product quantization for nearest neighbor search. IEEE Transactions on
Pattern Analysis and Machine Intelligence, 33(1):117?128, 2011.
Thorsten Joachims. Training linear svms in linear time. In Proceedings of the 12th ACM SIGKDD International
Conference on Knowledge Discovery and Data Mining, KDD ?06, pages 217?226, New York, NY, USA,
2006. ACM. ISBN 1-59593-339-5. doi: 10.1145/1150402.1150429. URL http://doi.acm.org/10.
1145/1150402.1150429.
W. Johnson and J. Lindenstrauss. Extensions of Lipschitz mappings into a Hilbert space. Contemporary
Mathematics, 26:189?206, 1984.
Q. Le, T. Sarl?s, and A. Smola. Fastfood - approximating kernel expansions in loglinear time. In ICML, 2013.
A. Rahimi and B. Recht. Random features for large-scale kernel machines. In NIPS, 2007.
Y.-L. K. Samo and S. Roberts. Generalized spectral kernels. CoRR, abs/1506.02236, 2015.
L. Schmidt, M. Sharifi, and I. Moreno. Large-scale speaker identification. In Acoustics, Speech and Signal
Processing (ICASSP), 2014 IEEE International Conference on, pages 1650?1654. IEEE, 2014.
G. Sidorov, A. Gelbukh, H. G?mez-Adorno, and D. Pinto. Soft similarity and soft cosine measure: Similarity of
features in vector space model. Computaci?n y Sistemas, 18(3), 2014.
N. Sundaram, A. Turmukhametova, N. Satish, T. Mostak, P. Indyk, S. Madden, and P. Dubey. Streaming
similarity search over one billion tweets using parallel locality-sensitive hashing. Proceedings of the VLDB
Endowment, 6(14):1930?1941, 2013.
J. Vyb?ral. A variant of the Johnson-Lindenstrauss lemma for circulant matrices. Journal of Functional Analysis,
260(4):1096?1105, 2011.
C. Williams. Computation with infinite neural networks. Neural Computation, 10(5):1203?1216, 1998.
F. Yu, A. Suresh, K. Choromanski, D. Holtmann-Rice, and S. Kumar. Orthogonal random features. In NIPS,
pages 1975?1983, 2016.
H. Zhang and L. Cheng. New bounds for circulant Johnson-Lindenstrauss embeddings. CoRR, abs/1308.6339,
2013.
Xu Zhang, Felix X. Yu, Ruiqi Guo, Sanjiv Kumar, Shengjin Wang, and Shih-Fu Chang. Fast orthogonal
projection based on kronecker product. In 2015 IEEE International Conference on Computer Vision, ICCV
2015, Santiago, Chile, December 7-13, 2015, pages 2929?2937, 2015. doi: 10.1109/ICCV.2015.335. URL
http://dx.doi.org/10.1109/ICCV.2015.335.
10
| 6626 |@word msr:4 middle:2 version:3 illustrating:1 norm:2 ruiqi:1 adrian:1 open:1 unif:6 km:3 cos2:1 vldb:1 recursively:1 reduction:12 initial:1 contains:1 interestingly:4 kx0:1 com:1 chazelle:3 si:2 yet:2 dx:2 sanjiv:1 razenshteyn:1 shape:1 analytic:1 kdd:1 moreno:1 plot:2 sundaram:2 stationary:1 half:3 fewer:1 selected:1 intelligence:1 kyk:1 complementing:1 isotropic:1 beginning:1 xk:3 ith:1 chile:1 colored:1 provides:2 contribute:1 node:1 gx:2 lx:10 org:2 zhang:5 jiri:1 prove:4 hermitian:1 introduce:1 theoretically:1 x0:3 pairwise:1 expected:2 behavior:2 examine:2 brain:2 jlt:6 decreasing:1 gou:2 becomes:1 provided:3 mass:1 what:1 kind:1 developed:1 whilst:1 finding:1 transformation:2 guarantee:8 exactly:4 rm:3 demonstrates:1 uk:3 brute:1 unit:1 grant:2 superiority:2 appear:1 before:1 felix:1 engineering:1 sd:22 might:3 doctoral:1 co:1 fastest:2 walsh:4 range:2 perpendicular:2 practical:1 lecun:1 practice:3 block:24 union:1 procedure:2 suresh:1 cfi:1 j0:2 empirical:6 projection:11 convenient:1 suggest:2 close:2 storage:1 context:1 applying:2 writing:1 equivalent:2 map:7 demonstrated:1 deterministic:1 quick:1 sarlos:2 go:1 attention:1 straightforward:1 independently:1 convex:1 williams:2 ergodic:1 unstructured:14 estimator:39 wie:1 importantly:1 orthonormal:1 spanned:1 datapoints:1 hd:8 embedding:3 ortho:4 coordinate:1 play:1 heavily:1 us:3 element:4 approximated:1 particularly:1 cayley:4 observed:3 role:1 bottom:2 epsrc:2 ep:2 solved:1 wang:1 worst:1 laarhoven:1 region:1 contemporary:1 intuition:1 complexity:2 cam:2 depend:2 deliver:2 efficiency:2 basis:1 usps:2 strikingly:1 easily:1 icassp:1 various:3 worsened:1 fast:11 effective:1 doi:4 sarl:1 larger:5 widely:1 g1:9 transform:12 indyk:2 advantage:1 isbn:1 douze:1 product:40 relevant:1 hadamard:18 combining:1 mixing:2 poorly:1 achieve:1 intuitive:2 hdk:2 frobenius:1 billion:1 r1:1 rademacher:17 help:5 illustrate:1 wider:2 depending:1 ac:2 derive:1 measured:1 nearest:2 odd:4 received:1 pendent:1 strong:2 implies:2 come:1 quantify:2 direction:1 sgn:2 enable:2 kchoro:1 require:2 hx:3 suffices:1 proposition:3 probable:3 strictly:2 extension:1 hold:2 practically:1 mapping:2 achieves:1 estimation:4 sensitive:2 council:1 agrees:1 repetition:1 tf:1 hope:1 gaussian:17 always:1 aim:1 rather:2 casting:1 broader:1 corollary:3 derived:1 focus:2 joachim:2 improvement:4 maria:1 ral:6 rank:1 mainly:1 indicates:1 sinn:1 contrast:2 sigkdd:1 attains:1 baseline:1 helpful:3 inference:1 dependent:1 lowercase:1 i0:2 streaming:1 typically:2 diminishing:3 choromanska:3 art:5 special:1 equal:1 construct:1 saving:2 having:2 beach:1 sampling:14 intriguingly:1 adversarially:1 yu:6 icml:3 bojarski:3 future:1 others:1 employ:1 randomly:2 preserve:1 replacement:11 recalling:1 ab:2 highly:1 mining:1 adjust:1 extreme:1 yielding:2 light:1 uppercase:1 superfluous:1 chain:9 fu:1 necessary:1 orthogonal:23 holtmann:1 walk:1 circle:1 re:1 theoretical:12 subfigure:2 instance:1 earlier:4 soft:2 stacking:1 cost:1 subset:1 entry:5 uniform:6 wonder:1 satish:1 johnson:10 conducted:1 weller:1 kn:1 mostak:1 aw:1 my:1 combined:1 cho:2 st:1 recht:3 international:3 explores:1 randomized:2 analogously:1 concrete:1 squared:2 central:2 nm:2 again:2 hn:1 choose:1 worse:2 leading:2 style:1 return:1 gy:2 bold:2 includes:2 santiago:1 explicitly:1 depends:2 multiplicative:1 h1:1 closed:1 observing:1 analyze:1 red:1 gouy:1 parallel:1 contribution:2 formed:2 square:1 accuracy:11 wiener:1 qk:2 efficiently:1 yield:16 ofthe:1 yellow:1 identification:1 iid:11 none:1 marginally:1 multiplying:1 dave:1 randomness:2 datapoint:2 explain:1 fagan:1 definition:7 against:1 gp1:2 johnsonlindenstrauss:1 proof:5 di:5 associated:2 gain:3 sampled:4 dataset:3 recall:1 knowledge:2 kyk22:1 dimensionality:12 conversation:1 hilbert:1 subtle:1 matej:1 appears:1 hashing:2 higher:3 reflected:1 improved:8 though:3 generality:1 furthermore:1 angular:24 just:6 smola:1 mez:1 sketch:1 hand:1 replacing:3 trust:1 nonlinear:3 google:4 quality:5 aj:6 indicated:1 believe:1 pailler:1 usa:2 effect:2 k22:1 normalized:6 unbiased:7 requiring:1 true:2 tamas:1 hence:3 attractive:1 indistinguishable:1 kyk2:2 maintained:1 speaker:2 noted:1 cosine:1 generalized:1 prominent:2 logd:1 image:1 wise:1 recently:2 functional:1 empirically:6 physical:1 conditioning:2 discussed:2 extend:1 belong:1 significant:2 refer:1 cambridge:4 ai:14 rd:1 mathematics:1 i6:2 neatly:1 centre:2 language:1 dot:11 lsh:1 moving:1 hashed:1 similarity:3 ort:7 base:11 gt:2 multivariate:1 showed:1 perspective:4 belongs:1 lomeli:1 store:1 inequality:2 binary:2 sdi:4 arbitrarily:1 success:1 outperforming:1 ojlt:7 yi:1 seen:1 additional:4 greater:1 mr:1 gp2:2 surely:2 period:2 signal:1 ii:1 full:3 rahimi:3 alan:2 calculation:2 long:1 sphere:1 a1:1 variant:8 involving:2 vision:1 expectation:3 metric:1 kernel:46 sometimes:1 robotics:2 background:2 addition:2 remarkably:1 whereas:1 crucial:1 parallelization:1 sr:1 markedly:1 comment:1 member:1 december:1 effectiveness:4 call:2 embeddings:6 easy:2 variety:3 affect:2 independence:1 gave:1 inner:4 regarding:1 computable:2 whether:1 motivated:1 expression:3 url:2 reuse:1 padding:1 effort:1 speech:1 proceed:1 york:1 deep:1 n510129:1 detailed:1 clear:3 dubey:1 transforms:2 dark:1 ang:16 inspiring:1 png:2 svms:1 diameter:1 http:2 outperform:5 notice:1 sign:5 leverhulme:1 per:1 blue:2 discrete:1 shall:2 group:1 key:2 salient:1 shih:1 demonstrating:3 choromanski:7 iof:1 krzysztof:1 asymptotically:1 graph:3 geometrically:1 tweet:1 cone:1 sum:1 turing:2 angle:4 letter:2 family:5 almost:2 throughout:1 appendix:9 scaling:2 ki:3 hi:3 bound:1 guaranteed:2 cheng:3 replaces:1 vyb:6 orthogonality:4 idf:1 kronecker:3 precisely:2 ri:4 speed:2 min:1 kumar:3 performing:2 aw665:1 tauberian:1 structured:19 ailon:3 smaller:6 y0:1 wi:2 projecting:1 iccv:3 thorsten:1 taken:2 equation:3 alluded:1 previously:2 visualization:1 discus:1 mechanism:5 needed:1 end:1 evalu:1 generalizes:1 available:1 rewritten:1 unreasonable:1 apply:1 observe:2 spectral:1 simulating:1 schmidt:3 alternative:1 jn:2 original:1 vikas:1 top:2 running:1 include:2 subsampling:1 recycling:1 build:2 establish:1 approximating:2 implied:1 question:1 realized:1 strategy:2 costly:1 diagonal:13 nr:1 loglinear:1 mx:1 distance:2 thank:2 samo:2 induction:1 rom:22 pointwise:6 illustration:1 providing:1 msh:5 ellipsoid:1 kk:2 setup:2 robert:2 statement:1 stoc:1 motivates:1 policy:2 perform:1 allowing:1 upper:1 observation:5 markov:7 datasets:2 beat:1 extended:2 sd1:3 rn:8 arbitrary:1 jebara:1 introduced:4 nonlinearly:1 required:1 connection:1 acoustic:1 established:2 nip:5 able:1 beyond:1 proceeds:1 below:4 pattern:1 built:1 including:4 gaining:1 deleting:1 power:1 event:1 natural:2 force:1 hybrid:5 mn:1 improve:1 gort:17 kxk22:2 carried:1 acknowledges:2 madden:1 schmid:1 text:1 geometric:2 understanding:3 l2:1 acknowledgement:1 multiplication:1 discovery:1 law:1 loss:1 expect:2 highlight:2 fully:1 inde:1 mixed:1 sublinear:1 storing:1 endowment:1 row:26 kkk2:1 copy:1 allow:1 understand:4 institute:2 circulant:4 saul:2 taking:5 neighbor:2 sparse:2 benefit:14 curve:1 dimension:11 lindenstrauss:10 stand:3 gram:6 computes:1 adaptive:1 bm:23 employing:1 rowland:1 transaction:1 approximate:2 search:2 timation:1 why:6 ca:1 expansion:2 mse:44 complex:8 constructing:2 hinrichs:3 did:1 aistats:1 spread:1 main:2 linearly:1 yi2:2 fastfood:1 n2:1 xu:1 representative:1 strengthened:1 ny:1 sub:23 theme:1 turmukhametova:1 deterministically:1 concatenating:1 kxk2:2 theorem:17 formula:3 specific:1 shengjin:1 explored:2 dk:5 exists:1 quantization:1 andoni:2 corr:2 magnitude:1 conditioned:2 nk:1 locality:2 cx:7 led:1 explore:3 desire:1 kxk:1 contained:1 g2:3 doubling:3 sindhwani:4 applies:1 pinto:1 sdk:3 chang:1 satisfies:5 acm:3 rice:1 conditional:1 goal:2 viewed:1 identity:1 replace:1 lipschitz:1 specifically:1 infinite:1 uniformly:1 lemma:6 hd1:1 experimental:1 select:1 support:3 mark:1 latter:1 guo:1 evaluate:1 d1:1 scratch:1 |
6,220 | 6,627 | From Parity to Preference-based Notions
of Fairness in Classification
Muhammad Bilal Zafar
MPI-SWS
[email protected]
Krishna P. Gummadi
MPI-SWS
[email protected]
Isabel Valera
MPI-IS
[email protected]
Manuel Gomez Rodriguez
MPI-SWS
[email protected]
Adrian Weller
University of Cambridge & Alan Turing Institute
[email protected]
Abstract
The adoption of automated, data-driven decision making in an ever expanding
range of applications has raised concerns about its potential unfairness towards
certain social groups. In this context, a number of recent studies have focused on
defining, detecting, and removing unfairness from data-driven decision systems.
However, the existing notions of fairness, based on parity (equality) in treatment
or outcomes for different social groups, tend to be quite stringent, limiting the
overall decision making accuracy. In this paper, we draw inspiration from the fairdivision and envy-freeness literature in economics and game theory and propose
preference-based notions of fairness?given the choice between various sets of
decision treatments or outcomes, any group of users would collectively prefer its
treatment or outcomes, regardless of the (dis)parity as compared to the other groups.
Then, we introduce tractable proxies to design margin-based classifiers that satisfy
these preference-based notions of fairness. Finally, we experiment with a variety
of synthetic and real-world datasets and show that preference-based fairness allows
for greater decision accuracy than parity-based fairness.
1
Introduction
As machine learning is increasingly being used to automate decision making in domains that affect
human lives (e.g., credit ratings, housing allocation, recidivism risk prediction), there are growing
concerns about the potential for unfairness in such algorithmic decisions [23, 25]. A flurry of recent
research on fair learning has focused on defining appropriate notions of fairness and then designing
mechanisms to ensure fairness in automated decision making [12, 14, 18, 19, 20, 21, 28, 32, 33, 34].
Existing notions of fairness in the machine learning literature are largely inspired by the concept of
discrimination in social sciences and law. These notions call for parity (i.e., equality) in treatment,
in impact, or both. To ensure parity in treatment (or treatment parity), decision making systems need
to avoid using users? sensitive attribute information, i.e., avoid using the membership information in
socially salient groups (e.g., gender, race), which are protected by anti-discrimination laws [4, 10]. As
a result, the use of group-conditional decision making systems is often prohibited. To ensure parity in
impact (or impact parity), decision making systems need to avoid disparity in the fraction of users
belonging to different sensitive attribute groups (e.g., men, women) that receive beneficial decision
outcomes. A number of learning mechanisms have been proposed to achieve parity in treatment [24],
An open-source code implementation of our scheme is available at: http://fate-computing.mpi-sws.org/
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
f2
-ve +ve
f2
M (100)
f2
M (100)
-ve +ve
M (100)
+ve
-ve
W (100)
M (200)
W (100)
W (200)
W (100)
W (200)
M (200)
+ve
M (200)
W (200)
-ve
f1
Acc: 0.83
Benefit: 0% (M), 67% (W)
f1
f1
Acc: 0.72
Benefit: 22% (M), 22% (W)
Acc: 1.00
Benefit: 33% (M), 67% (W)
Figure 1: A fictitious decision making scenario involving two groups: men (M) and women (W). Feature f1
(x-axis) is highly predictive for women whereas f2 (y-axis) is highly predictive for men. Green (red) quadrants
denote the positive (negative) class. Within each quadrant, the points are distributed uniformly and the numbers
in parenthesis denote the number of subjects in that quadrant. The left panel shows the optimal classifier
satisfying parity in treatment. This classifier leads to all the men getting classified as negative. The middle
panel shows the optimal classifier satisfying parity in impact (in addition to parity in treatment). This classifier
achieves impact parity by misclassifying women from positive class into negative class, and in the process,
incurs a significant cost in terms of accuracy. The right panel shows a classifier consisting of group-conditional
classifiers for men (purple) and women (blue). Both the classifiers satisfy the preferred treatment criterion since
for each group, adopting the other group?s classifier would lead to a smaller fraction of beneficial outcomes.
Additionally, this group-conditional classifier is also a preferred impact classifier since both groups get more
benefit as compared to the impact parity classifier. The overall accuracy is better than the parity classifiers.
parity in impact [7, 18, 21] or both [12, 14, 17, 20, 32, 33, 34]. However, these mechanisms pay a
significant cost in terms of the accuracy (or utility) of their predictions. In fact, there exist some
inherent tradeoffs (both theoretical and empirical) between achieving high prediction accuracy and
satisfying treatment and / or impact parity [9, 11, 15, 22].
In this work, we introduce, formalize and evaluate new notions of fairness that are inspired by the
concepts of fair division and envy-freeness in economics and game theory [5, 26, 31]. Our work
is motivated by the observation that, in certain decision making scenarios, the existing parity-based
fairness notions may be too stringent, precluding more accurate decisions, which may also be desired
by every sensitive attribute group. To relax these parity-based notions, we introduce the concept of a
user group?s preference for being assigned one set of decision outcomes over another. Given the
choice between various sets of decision outcomes, any group of users would collectively prefer the
set that contains the largest fraction (or the greatest number) of beneficial decision outcomes for that
group.1 More specifically, our new preference-based notions of fairness, which we formally define in
the next section, use the concept of user group?s preference as follows:
? From Parity Treatment to Preferred Treatment: To offer preferred treatment, a decision making
system should ensure that every sensitive attribute group (e.g., men and women) prefers the set of
decisions they receive over the set of decisions they would have received had they collectively
presented themselves to the system as members of a different sensitive group.
The preferred treatment criterion represents a relaxation of treatment parity. That is, every decision
making system that achieves treatment parity also satisfies the preferred treatment condition, which
implies (in theory) that the optimal decision accuracy that can be achieved under the preferred
treatment condition is at least as high as the one achieved under treatment parity. Additionally,
preferred treatment allows group-conditional decision making (not allowed by treatment parity),
which is necessary to achieve high decision accuracy in scenarios when the predictive power of
features varies greatly between different sensitive user groups [13], as shown in Figure 1.
While preferred treatment is a looser notion of fairness than treatment parity, it retains a core fairness
property embodied in treatment parity, namely, envy-freeness at the level of user groups. Under
preferred treatment, no group of users (e.g., men or women, blacks or whites) would feel that they
would be collectively better off by switching their group membership (e.g., gender, race). Thus,
1
Although it is quite possible that certain individuals from the group may not prefer the set that maximizes the benefit for the group as a
whole.
2
preferred treatment decision making, despite allowing group-conditional decision making, is not
vulnerable to being characterized as ?reverse discrimination? against, or "affirmative action? for
certain groups.
? From Parity Impact to Preferred Impact: To offer preferred impact, a decision making system
needs to ensure that every sensitive attribute group (e.g., men and women) prefers the set of decisions
they receive over the set of decisions they would have received under the criterion of impact parity.
The preferred impact criterion represents a relaxation of impact parity. That is, every decision making
system that achieves impact parity also satisfies the preferred impact condition, which implies (in
theory) that the optimal decision accuracy that can be achieved under the preferred impact condition
is at least as high as the one achieved under impact parity. Additionally, preferred impact allows
disparity in benefits received by different groups, which may be justified in scenarios where insisting
on impact parity would only lead to a reduction in the beneficial outcomes received by one or more
groups, without necessarily improving them for any other group. In such scenarios, insisting on
impact parity can additionally lead to a reduction in the decision accuracy, creating a case of tragedy
of impact parity with a worse decision making all round, as shown in Figure 1.
While preferred impact is a looser notion of fairness compared to impact parity, by guaranteeing
that every group receives at least as many beneficial outcomes as they would would have received
under impact parity, it retains the core fairness gains in beneficial outcomes that discriminated groups
would have achieved under the fairness criterion of impact parity.
Finally, we note that our preference-based fairness notions, while having many attractive properties,
are not the most suitable notions of fairness in all scenarios. In certain cases, parity fairness may well
be the eventual goal [3] and the more desirable notion.
In the remainder of this paper, we formalize our preference-based fairness notions in the context
of binary classification (Section 2), propose tractable and efficient proxies to include these notions
in the formulations of convex margin-based classifiers in the form of convex-concave constraints
(Section 3), and show on several real world datasets that our preference-based fairness notions can
provide significant gains in overall decision making accuracy as compared to parity-based fairness
(Section 4).
2
Defining preference-based fairness for classification
In this section, we will first introduce two useful quality metrics?utility and group benefit?in the
context of fairness in classification, then revisit parity-based fairness definitions in the light of these
quality metrics, and finally formalize the two preference-based notions of fairness introduced in
Section 1 from the perspective of the above metrics. For simplicity, we consider binary classification
tasks, however, the definitions can be easily extended to m-ary classification.
Quality metrics in fair classification. In a fair (binary) classification task, one needs to find a
mapping between the user feature vectors x 2 Rd and class labels y 2 { 1, 1}, where (x, y)
are drawn from an (unknown) distribution f (x, y). This is often achieved by finding a mapping
function ? : Rd ! R such that given a feature vector x with an unknown label y, the corresponding
classifier predicts y? = sign(?(x)). However, this mapping function also needs to be fair with respect
to the values of a user sensitive attribute z 2 Z ? Z 0 (e.g., sex, race), which are drawn from
an (unknown) distribution f (z) and may be dependent of the feature vectors and class labels, i.e.,
f (x, y, z) = f (x, y|z)f (z) 6= f (x, y)f (z).
Given the above problem setting, we introduce the following quality metrics, which we will use to
define and compare different fairness notions:
I. Utility (U ): overall profit obtained by the decision maker using the classifier. For example, in a
loan approval scenario, the decision maker is the bank that gives the loan and the utility can be
the overall accuracy of the classifier, i.e.:
U (?) = Ex,y [I{sign(?(x)) = y}],
where I(?) denotes the indicator function and the expectation is taken over the distribution
f (x, y). It is in the decision maker?s interest to use classifiers that maximize utility. Moreover,
depending on the scenario, one can attribute different profit to true positives and true negatives?
or conversely, different cost to false negatives and false positives?while computing utility. For
3
example, in the loan approval scenario, marking an eventual defaulter as non-defaulter may have
a higher cost than marking a non-defaulter as defaulter. For simplicity, in the remainder of the
paper, we will assume that the profit (cost) for true (false) positives and negatives is the same.
II. Group benefit (Bz ): the fraction of beneficial outcomes received by users sharing a certain value
of the sensitive attribute z (e.g., blacks, hispanics, whites). For example, in a loan approval
scenario, the beneficial outcome for a user may be receiving the loan and the group benefit for
each value of z can be defined as:
Bz (?) = Ex|z [I{sign(?(x)) = 1}],
where the expectation is taken over the conditional distribution f (x|z) and the bank offers a loan
to a user if sign(?(x)) = 1. Moreover, as suggested by some recent studies in fairness-aware
learning [18, 22, 32], the group benefits can also be defined as the fraction of beneficial outcomes
conditional on the true label of the user. For example, in a recidivism prediction scenario, the
group benefits can be defined as the fraction of eventually non-offending defendants sharing a
certain sensitive attribute value getting bail, that is:
Bz (?) = Ex|z,y=1 [I{sign(?(x)) = 1}],
where the expectation is taken over the conditional distribution f (x|z, y = 1), y = 1 indicates
that the defendant does not re-offend, and bail is granted if sign(?(x)) = 1.
Parity-based fairness. A number of recent studies [7, 14, 18, 21, 32, 33, 34] have considered a
classifier to be fair if it satisfies the impact parity criterion. That is, it ensures that the group benefits
for all the sensitive attribute values are equal, i.e.:
Bz (?) = Bz0 (?) for all z, z 0 2 Z.
(1)
In this context, different (or often same) definitions of group benefit (or beneficial outcome) have
lead to different terminology, e.g., disparate impact [14, 33], indirect discrimination [14, 21], redlining [7], statistical parity [12, 11, 22, 34], disparate mistreatment [32], or equality of opportunity [18].
However, all of these group benefit definitions invariably focus on achieving impact parity. We
point interested readers to Feldman et al. [14] and Zafar et al. [32] regarding the discussion on this
terminology.
Although not always explicitly sought, most of the above studies propose classifiers that also satisfy
treatment parity in addition to impact parity, i.e., they do not use the sensitive attribute z in the
decision making process. However, some of them [7, 18, 21] do not satisfy treatment parity since
they resort to group-conditional classifiers, i.e., ? = {?z }z2Z . In such case, we can rewrite the above
parity condition as:
Bz (?z ) = Bz0 (?z0 ) for all z, z 0 2 Z.
(2)
Fairness beyond parity. Given the above quality metrics, we can now formalize the two preferencebased fairness notions introduced in Section 1.
? Preferred treatment: if a classifier ? resorts to group-conditional classifiers, i.e., ? = {?z }z2Z ,
it is a preferred treatment classifier if each group sharing a sensitive attribute value z benefits
more from its corresponding group-conditional classifier ?z than it would benefit if it would be
classified by any of the other group-conditional classifiers ?z0 , i.e.,
Bz (?z ) Bz (?z0 ) for all z, z 0 2 Z.
(3)
Note that, if a classifier ? does not resort to group-conditional classifiers, i.e., ?z = ? for all
z 2 Z, it will be always be a preferred treatment classifier. If, in addition, such classifier ensures
impact parity, it is easy to show that its utility cannot be larger than a preferred treatment classifier
consisting of group-conditional classifiers.
? Preferred impact: a classifier ? offers preferred impact over a classifier ? 0 ensuring impact
parity if it achieves higher group benefit for each sensitive attribute value group, i.e.,
Bz (?) Bz (? 0 ) for all z 2 Z.
(4)
One can also rewrite the above condition for group-conditional classifiers, i.e., ? = {?z }z2Z
and ? 0 = {?z0 }z2Z , as follows:
Bz (?z ) Bz (?z0 ) for all z 2 Z.
(5)
0
Note that, given any classifier ? ensuring impact parity, it is easy to show that there will always
exist a preferred impact classifier ? with equal or higher utility.
4
Connection to the fair division literature. Our notion of preferred treatment is inspired by the
concept of envy-freeness [5, 31] in the fair division literature. Intuitively, an envy-free resource
division ensures that no user would prefer the resources allocated to another user over their own
allocation. Similarly, our notion of preferred treatment ensures envy-free decision making at the
level of sensitive attribute groups. Specifically, with preferred treatment classification, no sensitive
attribute group would prefer the outcomes from the classifier of another group.
Our notion of preferred impact draws inspiration from the two-person bargaining problem [26] in
the fair division literature. In a bargaining scenario, given a base resource allocation (also called the
disagreement point), two parties try to divide some additional resources between themselves. If the
parties cannot agree on a division, no party gets the additional resources, and both would only get the
allocation specified by the disagreement point. Taking the resources to be the beneficial outcomes,
and the disagreement point to be the allocation specified by the impact parity classifier, a preferred
impact classifier offers enhanced benefits to all the sensitive attribute groups. Put differently, the
group benefits provided by the preferred impact classifier Pareto-dominate the benefits provided by
the impact parity classifier.
On individual-level preferences. Notice that preferred treatment and preferred impact notions are
defined based on the group preferences, i.e., whether a group as a whole prefers (or, gets more
beneficial outcomes from) a given set of outcomes over another set. It is quite possible that a set
of outcomes preferred by the group collectively is not preferred by certain individuals in the group.
Consequently, one can extend our proposed notions to account for individual preferences as well,
i.e., a set of outcomes is preferred over another if all the individuals in the group prefer it. In the
remainder of the paper, we focus on preferred treatment and preferred impact in the context of group
preferences, and leave the case of individual preferences and its implications on the cost of achieving
fairness for future work.
3
Training preferred classifiers
In this section, our goal is training preferred treatment and preferred impact group-conditional
classifiers, i.e., ? = {?z }z2Z , that maximize utility given a training set D = {(xi , yi , zi )}N
i=1 , where
(xi , yi , zi ) ? f (x, y, z). In both cases, we will assume that:2
I. Each group-conditional classifier is a convex boundary-based classifier. For ease of exposition,
in this section, we additionally assume these classifiers to be linear, i.e., ?z (x) = ?zT x, where
?z is a parameter that defines the decision boundary in the feature space. We relax the linearity
assumption in Appendix A and extend our methodology to a non-linear SVM classifier.
II. The utility function U is defined as the overall accuracy of the group-conditional classifiers, i.e.,
X
U (?) = Ex,y [I{sign(?(x)) = y}] =
Ex,y|z [I{sign(?zT x) = y}]f (z).
(6)
z2Z
III. The group benefit Bz for users sharing the sensitive attribute value z is defined as their average
probability of being classified into the positive class, i.e.,
Bz (?) = Ex|z [I{sign(?(x)) = 1}] = Ex|z [I{sign(?zT x) = 1}].
(7)
Preferred impact classifiers. Given a impact parity classifier with decision boundary parameters
{?z0 }z2Z , one could think of finding the decision boundary parameters {?z }z2Z of a preferred impact
classifier that maximizes utility by solving the following optimization problem:
P
1
T
minimize
(x,y,z)2D I{sign(?z x) = y}
N
{?z }
(8)
P
P
T
0T
subject to
x2Dz I{sign(?z x) = 1}
x2Dz I{sign(?z x) = 1} for all z 2 Z,
where Dz = {(xi , yi , zi ) 2 D | zi = z} denotes the set of users in the training set sharing sensitive
attribute value z, the objective uses an empirical estimate of the utility, defined by Eq. 6, and the
preferred impact constraints, defined by Eq. 5, use empirical estimates of the group benefits, defined
by Eq. 7. Here, note that the right hand side of the inequalities does not contain any variables and can
be precomputed, i.e., the impact parity classifiers {?z0 }z2Z are given.
2
Exploring the relaxations of these assumptions is a very interesting avenue for future work.
5
Unfortunately, it is very challenging to solve the above optimization problem since both the objective
and constraints are nonconvex. To overcome this difficulty, we minimize instead a convex loss
function `? (x, y), which is classifier dependent [6], and approximate the group benefits using a ramp
(convex) function r(x) = max(0, x), i.e.,
P
P
1
minimize
(x,y,z)2D `?z (x, y) +
z2Z z ?(?z )
N
{?z }
(9)
P
P
T
0T
subject to
x2Dz max(0, ?z x)
x2Dz max(0, ?z x) for all z 2 Z,
which, for any convex regularizer ?(?), is a disciplined convex-concave program (DCCP) and thus
can be efficiently solved using well-known heuristics [30]. For example, if we particularize the above
formulation to group-conditional (standard) logistic regression classifiers ?z0 and ?z and L2 -norm
regularizer, then, Eq. 9 adopts the following form:
P
P
1
2
minimize
(x,y,z)2D log p(y|x, ?z ) +
z2Z z ||?z ||
N
{?z }
(10)
P
P
T
0T
subject to
x2Dz max(0, ?z x)
x2Dz max(0, ? z x) for all z 2 Z.
where p(y = 1|x, ?z ) =
1
1+e
Tx
?z
.
The constraints can similarly be added to other convex boundary-based classifiers like linear SVM.
We further expand on particularizing the constraints for non-linear SVM in Appendix A.
Preferred treatment classifiers. Similarly as in the case of preferred impact classifiers, one could
think of finding the decision boundary parameters {?z }z2Z of a preferred treatment classifier that
maximizes utility by solving the following optimization problem:
P
1
T
minimize
(x,y,z)2D I{sign(?z x) = y}
N
{?z }
(11)
P
P
T
T
0
subject to
x2Dz I{sign(?z x) = 1}
x2Dz I{sign(?z 0 x) = 1} for all z, z 2 Z,
where Dz = {(xi , yi , zi ) 2 D | zi = z} denotes the set of users in the training set sharing sensitive
attribute value z, the objective uses an empirical estimate of the utility, defined by Eq. 6, and the
preferred treatment constraints, defined by Eq. 3, use empirical estimates of the group benefits, defined
by Eq. 7. Here, note that both the left and right hand side of the inequalities contain optimization
variables.
However, the objective and constraints in the above problem are also nonconvex and thus we adopt a
similar strategy as in the case of preferred impact classifiers. More specifically, we solve instead the
following tractable problem:
P
P
1
minimize
(x,y,z)2D `?z (x, y) +
z2Z z ?(?z )
N
{?z }
(12)
P
P
T
T
0
subject to
x2Dz max(0, ?z x)
x2Dz max(0, ?z 0 x) for all z, z 2 Z,
which, for any convex regularizer ?(?), is also a disciplined convex-concave program (DCCP) and
can be efficiently solved.
4
Evaluation
In this section, we compare the performance of preferred treatment and preferred impact classifiers
against unconstrained, treatment parity and impact parity classifiers on a variety of synthetic and
real-world datasets. More specifically, we consider the following classifiers, which we train to
maximize utility subject to the corresponding constraints:
? Uncons: an unconstrained classifier that resorts to group-conditional classifiers. It violates
treatment parity?it trains a separate classifier per sensitive attribute value group?and potentially
violates impact parity?it may lead to different benefits for different groups.
? Parity: a parity classifier that does not use the sensitive attribute group information in the decision
making, but only during the training phase, and is constrained to satisfy both treatment parity?
its decisions do not change based on the users? sensitive attribute value as it does not resort to
group-conditional classifiers?and impact parity?it ensures that the benefits for all groups are
the same. We train this classifier using the methodology proposed by Zafar et al. [33].
? Preferred treatment: a classifier that resorts to group-conditional classifiers and is constrained
to satisfy preferred treatment?each group gets the highest benefit with its own classifier than
any other group?s classifier.
6
Acc : 0.87
Acc : 0.57
Acc : 0.76
Acc : 0.73
B0 : 0.16; B1 : 0.77
B0 : 0.20; B1 : 0.85
B0 : 0.51; B1 : 0.49
B0 : 0.58; B1 : 0.96
B0 : 0.21; B1 : 0.86
B0 : 0.58; B1 : 0.96
B0 : 0.43; B1 : 0.97
(a) Uncons
(b) Parity
(c) Preferred impact
(d) Preferred both
Figure 2: [Synthetic data] Crosses denote group-0 (points with z = 0) and circles denote group-1.
Green points belong to the positive class in the training data whereas red points belong to the negative
class. Each panel shows the accuracy of the decision making scenario along with group benefits (B0
and B1 ) provided by each of the classifiers involved. For group-conditional classifiers, cyan (blue)
line denotes the decision boundary for the classifier of group-0 (group-1). Parity case (panel (b))
consists of just one classifier for both groups in order to meet the treatment parity criterion.
? Preferred impact: a classifier that resorts to group-conditional classifiers and is constrained to
be preferred over the Parity classifier.
? Preferred both: a classifier that resort to group-conditional classifiers and is constrained to satisfy
both preferred treatment and preferred impact.
For the experiments in this section, we use logistic regression classifiers with L2 -norm regularization.
We randomly split the corresponding dataset into 70%-30% train-test folds 5 times, and report the
average accuracy and group benefits in the test folds. Appendix B describes the details for selecting
the optimal L2 -norm regularization parameters. Here, we compute utility (U ) as the overall accuracy
of a classifier and group benefits (Bz ) as the fraction of users sharing sensitive attribute z that are
classified into the positive class. Moreover, the sensitive attribute is always binary, i.e., z 2 {0, 1}.
4.1 Experiments on synthetic data
Experimental setup. Following Zafar et al. [33], we generate a synthetic dataset in which the unconstrained classifier (Uncons) offers different benefits to each sensitive attribute group. In particular, we
generate 20,000 binary class labels y 2 { 1, 1} uniformly at random along with their corresponding
two-dimensional feature vectors sampled from the following Gaussian distributions: p(x|y = 1) =
N ([2; 2], [5, 1; 1, 5]) and p(x|y = 1) = N ([ 2; 2], [10, 1; 1, 3]). Then, we generate each sensitive attribute from the Bernoulli distribution p(z = 1) = p(x0 |y = 1)/(p(x0 |y = 1)+p(x0 |y = 1)),
where x0 is a rotated version of x, i.e., x0 = [cos(?/8), sin(?/8); sin(?/8), cos(?/8)]. Finally,
we train the five classifiers described above and compute their overall (test) accuracy and (test) group
benefits.
Results. Figure 2 shows the trained classifiers, along with their overall accuracy and group benefits.
We can make several interesting observations:
The Uncons classifier leads to an accuracy of 0.87, however, the group-conditional boundaries and
high disparity in treatment for the two groups (0.16 vs. 0.85) mean that it satisfies neither treatment
parity nor impact parity. Moreover, it leads to only a small violation of preferred treatment?benefits
for group-0 would increase slightly from 0.16 to 0.20 by adopting the classifier of group-1. However,
this will not always be the case, as we will later show in the experiments on real data.
The Parity classifier satisfies both treatment and impact parity, however, it does so at a large cost in
terms of accuracy, which drops from 0.87 for Uncons to 0.57 for Parity.
The Preferred treatment classifier (not shown in the figure), leads to a minor change in decision
boundaries as compared to the Uncons classifier to achieve preferred treatment. Benefits for group-0
(group-1) with its own classifier are 0.20 (0.84) as compared to 0.17 (0.83) while using the classifier
of group-1 (group-0). The accuracy of this classifier is 0.87.
The Preferred impact classifier, by making use of a looser notion of fairness compared to impact
parity, provides higher benefits for both groups at a much smaller cost in terms of accuracy than the
Parity classifier (0.76 vs. 0.57). Note that, while the Parity classifier achieved equality in benefits by
misclassifying negative examples from group-0 into the positive class and misclassifying positive
7
B0(?0)
B0(?1)
B1(?1)
B1(?0)
Acc
0.7
0.8
0.6
0.6
0.5
0.4
Uncons.
Parity
Prf-treat.
Prf-imp.
Prf-both
Accuracy
Benefits
ProPublica COMPAS dataset
1
0.4
Adult dataset
0.85
0.84
0.2
0.83
0.82
0
Uncons.
Parity
Prf-treat.
Prf-imp.
Prf-both
Accuracy
Benefits
0.4
0.81
0.8
0.7
0.6
0.5
Uncons.
Parity
Prf-treat.
Prf-imp.
Accuracy
Benefits
NYPD SQF dataset
1
0.8
0.6
0.4
0.2
0
Prf-both
Figure 3: The figure shows the accuracy and benefits received by the two groups for various decision
making scenarios. ?Prf-treat.?, ?Prf-imp.?, and ?Prf-both? respectively correspond to the classifiers
satisfying preferred treatment, preferred impact, and both preferred treatment and impact criteria.
Sensitive attribute values 0 and 1 denote blacks and whites in ProPublica COMPAS dataset and
NYPD SQF datasets, and women and men in the Adult dataset. Bi (?j ) denotes the benefits obtained
by group i when using the classifier of group j. For the Parity case, we train just one classifier for
both the groups, so the benefits do not change by adopting other group?s classifier.
examples from group-1 into the negative class, the Preferred impact classifier only incurs the former
type of misclassifications. However, the outcomes of the Preferred impact classifier do not satisfy the
preferred treatment criterion: group-1 would attain higher benefit if it used the classifier of group-0
(0.96 as compared to 0.86).
Finally, the classifier that satisfies preferred treatment and preferred impact (Preferred both) achieves
an accuracy and benefits at par with the Preferred impact classifier.
We present the results of applying our fairness constraints on a non linearly-separable dataset with a
SVM classifier with a radial basis function (RBF) kernel in Appendix C.
4.2
Experiments on real data
Dataset description and experimental setup. We experiment with three real-world datasets: the
COMPAS recidivism prediction dataset compiled by ProPublica [23], the Adult income dataset from
UCI machine learning repository [2], and the New York Police Department (NYPD) Stop-questionand-frisk (SQF) dataset made publicly available by NYPD [1]. These datasets have been used by a
number of prior studies in the fairness-aware machine learning literature [14, 29, 32, 34, 33].
In the COMPAS dataset, the classification task is to predict whether a criminal defendant would
recidivate within two years (negative class) or not (positive class); in the Adult dataset, the task
is to predict whether a person earns more than 50K USD per year (positive class) or not; and, in
the SQF dataset, the task is to predict whether a pedestrian should be stopped on the suspicion
of having an illegal weapon or not (positive class). In all datasets, we assume being classified as
positive to be the beneficial outcome. Additionally, we divide the subjects in each dataset into two
sensitive attribute value groups: women (group-0) and men (group-1) in the Adult dataset and blacks
(group-0) and whites (group-1) in the COMPAS and SQF datasets. The supplementary material
8
(Appendix D) contains more information on the sensitive and the non-sensitive features as well as the
class distributions.3
Results. Figure 3 shows the accuracy achieved by the five classifiers described above along with the
benefits they provide for the three datasets. We can draw several interesting observations:4
In all cases, the Uncons classifier, in addition to violating treatment parity (a separate classifier for
each group) and impact parity (high disparity in group benefits), also violates the preferred treatment
criterion (in all cases, at least one of group-0 or group-1 would benefit more by adopting the other
group?s classifier). On the other hand, the Parity classifier satisfies the treatment parity and impact
parity but it does so at a large cost in terms of accuracy.
The Preferred treatment classifier provides a much higher accuracy than the Parity classifier?its
accuracy is at par with that of the Uncons classifier?while satisfying the preferred treatment criterion.
However, it does not meet the preferred impact criterion. The Preferred impact classifier meets the
preferred impact criterion but does not always satisfy preferred treatment. Moreover, it also leads to a
better accuracy then Parity classifier in all cases. However, the gain in accuracy is more substantial
for the SQF datasets as compared to the COMPAS and Adult dataset.
The classifier satisfying preferred treatment and preferred impact (Preferred both) has a somewhat
underwhelming performance in terms of accuracy for the Adult dataset. While the performance of
this classifier is better than the Parity classifier in the COMPAS dataset and NYPD SQF dataset, it is
slightly worse for the Adult dataset.
In summary, the above results show that ensuring either preferred treatment or preferred impact is
less costly in terms of accuracy loss than ensuring parity-based fairness, however, ensuring both
preferred treatment and preferred impact can lead to comparatively larger accuracy loss in certain
datasets. We hypothesize that this loss in accuracy may be partly due to splitting the number of
available samples into groups during training?each group-conditional classifier use only samples
from the corresponding sensitive attribute group?hence decreasing the effectiveness of empirical
risk minimization.
5
Conclusion
In this paper, we introduced two preference-based notions of fairness?preferred treatment and
preferred impact?establishing a previously unexplored connection between fairness-aware machine
learning and the economics and game theoretic concepts of envy-freeness and bargaining. Then,
we proposed tractable proxies to design boundary-based classifiers satisfying these fairness notions
and experimented with a variety of synthetic and real-world datasets, showing that preference-based
fairness often allows for greater decision accuracy than existing parity-based fairness notions.
Our work opens many promising venues for future work. For example, our methodology is limited
to convex boundary-based classifiers. A natural follow up would be to extend our methodology to
other types of classifiers, e.g., neural networks and decision trees. In this work, we defined preferred
treatment and preferred impact in the context of group preferences, however, it would be worth
revisiting the proposed definitions in the context of individual preferences. The fair division literature
establishes a variety of fairness axioms [26] such as Pareto-optimality and scale invariance. It would
be interesting to study such axioms in the context of fairness-aware machine learning.
Finally, we note that while moving from parity to preference-based fairness offers many attractive
properties, we acknowledge it may not always be the most appropriate notion, e.g., in some scenarios,
parity-based fairness may very well present the eventual goal and be more desirable [3].
Acknowledgments
AW acknowledges support by the Alan Turing Institute under EPSRC grant EP/N510129/1, and by
the Leverhulme Trust via the CFI.
3
Since the SQF dataset is highly skewed in terms of class distribution (?97% points in the positive class) resulting in a trained classifier
predicting all points as positive (yet having 97% accuracy), we subsample the dataset to have equal class distribution. Another option would be
using penalties proportional to the size of the class, but we observe that an unconstrained classifier with class penalties gives similar predictions
as compared to a balanced dataset. We decided to experiment with the balanced dataset since the accuracy drops in this dataset are easier to
interpret.
4
The unfairness in the SQF dataset is different from what one would expect [27]?an unconstrained classifier gives more benefits to blacks
as compared to whites. This is due to the fact that a larger fraction of stopped whites were found to be in possession on an illegal weapon
(Tables 3 and 4 in Appendix D).
9
References
[1] Stop, Question and Frisk Data. http://www1.nyc.gov/site/nypd/stats/reports-analysis/stopfrisk.page, 2017.
[2] Adult data. https://archive.ics.uci.edu/ml/datasets/adult, 1996.
[3] A. Altman. Discrimination. In The Stanford Encyclopedia of Philosophy. Metaphysics Research Lab,
Stanford University, 2016. https://plato.stanford.edu/archives/win2016/entries/discrimination/.
[4] S. Barocas and A. D. Selbst. Big Data?s Disparate Impact. California Law Review, 2016.
[5] M. Berliant and W. Thomson. On the Fair Division of a Heterogeneous Commodity. Journal of Mathematics
Economics , 1992.
[6] C. M. Bishop. Pattern Recognition and Machine Learning. Springer, 2006.
[7] T. Calders and S. Verwer. Three Naive Bayes Approaches for Discrimination-Free Classification. Data
Mining and Knowledge Discovery, 2010.
[8] O. Chapelle. Training a Support Vector Machine in the Primal. Neural Computation, 2007.
[9] A. Chouldechova. Fair Prediction with Disparate Impact:A Study of Bias in Recidivism Prediction
Instruments. arXiv preprint, arXiv:1610.07524, 2016.
[10] Civil Rights Act. Civil Rights Act of 1964, Title VII, Equal Employment Opportunities, 1964.
[11] S. Corbett-Davies, E. Pierson, A. Feller, S. Goel, and A. Huq. Algorithmic Decision Making and the Cost
of Fairness. In KDD, 2017.
[12] C. Dwork, M. Hardt, T. Pitassi, and O. Reingold. Fairness Through Awareness. In ITCSC, 2012.
[13] C. Dwork, N. Immorlica, A. T. Kalai, and M. Leiserson. Decoupled Classifiers for Fair and Efficient
Machine Learning. arXiv preprint arXiv:1707.06613, 2017.
[14] M. Feldman, S. A. Friedler, J. Moeller, C. Scheidegger, and S. Venkatasubramanian. Certifying and
Removing Disparate Impact. In KDD, 2015.
[15] S. A. Friedler, C. Scheidegger, and S. Venkatasubramanian. On the (im)possibility of Fairness. arXiv
preprint arXiv:1609.07236, 2016.
[16] J. E. Gentle, W. K. H?rdle, and Y. Mori. Handbook of Computational Statistics: Concepts and Methods.
Springer Science & Business Media, 2012.
[17] G. Goh, A. Cotter, M. Gupta, and M. Friedlander. Satisfying Real-world Goals with Dataset Constraints.
In NIPS, 2016.
[18] M. Hardt, E. Price, and N. Srebro. Equality of Opportunity in Supervised Learning. In NIPS, 2016.
[19] M. Joseph, M. Kearns, J. Morgenstern, and A. Roth. Fairness in Learning: Classic and Contextual Bandits.
In NIPS, 2016.
[20] F. Kamiran and T. Calders. Classification with No Discrimination by Preferential Sampling. In BENELEARN, 2010.
[21] T. Kamishima, S. Akaho, H. Asoh, and J. Sakuma. Fairness-aware Classifier with Prejudice Remover
Regularizer. In PADM, 2011.
[22] J. Kleinberg, S. Mullainathan, and M. Raghavan. Inherent Trade-Offs in the Fair Determination of Risk
Scores. In ITCS, 2017.
[23] J. Larson, S. Mattu, L. Kirchner, and J. Angwin. https://github.com/propublica/compas-analysis, 2016.
[24] B. T. Luong, S. Ruggieri, and F. Turini. kNN as an Implementation of Situation Testing for Discrimination
Discovery and Prevention. In KDD, 2011.
[25] C. Mu?oz, M. Smith, and D. Patil. Big Data: A Report on Algorithmic Systems, Opportunity, and Civil
Rights. Executive Office of the President. The White House., 2016.
[26] J. F. Nash Jr. The Bargaining Problem. Econometrica: Journal of the Econometric Society, 1950.
[27] NYCLU. Stop-and-Frisk Data. https://www.nyclu.org/en/stop-and-frisk-data, 2017.
10
[28] D. Pedreschi, S. Ruggieri, and F. Turini. Discrimination-aware Data Mining. In KDD, 2008.
[29] R. S. Sharad Goel, Justin M. Rao. Precinct or Prejudice? Understanding Racial Disparities in New York
City?s Stop-and-Frisk Policy. Annals of Applied Statistics, 2015.
[30] X. Shen, S. Diamond, Y. Gu, and S. Boyd. Disciplined Convex-Concave Programming. arXiv:1604.02639,
2016.
[31] H. R. Varian. Equity, Envy, and Efficiency. Journal of Economic Theory, 1974.
[32] M. B. Zafar, I. Valera, M. G. Rodriguez, and K. P. Gummadi. Fairness Beyond Disparate Treatment &
Disparate Impact: Learning Classification without Disparate Mistreatment. In WWW, 2017.
[33] M. B. Zafar, I. Valera, M. G. Rodriguez, and K. P. Gummadi. Fairness Constraints: Mechanisms for Fair
Classification. In AISTATS, 2017.
[34] R. Zemel, Y. Wu, K. Swersky, T. Pitassi, and C. Dwork. Learning Fair Representations. In ICML, 2013.
11
| 6627 |@word repository:1 version:1 middle:1 norm:3 sex:1 adrian:1 open:2 incurs:2 profit:3 offending:1 reduction:2 venkatasubramanian:2 contains:2 disparity:5 selecting:1 score:1 precluding:1 bilal:1 existing:4 contextual:1 com:1 manuel:1 yet:1 kdd:4 hypothesize:1 drop:2 propublica:4 discrimination:10 v:2 smith:1 core:2 detecting:1 provides:2 preference:22 org:5 five:2 along:4 consists:1 introduce:5 x0:5 mpg:1 themselves:2 growing:1 nor:1 metaphysics:1 inspired:3 socially:1 approval:3 decreasing:1 gov:1 provided:3 moreover:5 unfairness:4 panel:5 maximizes:3 linearity:1 medium:1 what:1 affirmative:1 morgenstern:1 finding:3 possession:1 every:6 unexplored:1 commodity:1 concave:4 act:2 classifier:130 uk:1 grant:1 positive:16 manuelgr:1 treat:4 switching:1 despite:1 uncons:11 establishing:1 meet:3 black:5 conversely:1 challenging:1 co:2 ease:1 limited:1 range:1 adoption:1 bi:1 decided:1 acknowledgment:1 testing:1 tragedy:1 cfi:1 empirical:6 axiom:2 attain:1 illegal:2 davy:1 boyd:1 radial:1 quadrant:3 get:5 cannot:2 put:1 context:8 risk:3 applying:1 www:2 dz:2 roth:1 economics:4 regardless:1 convex:12 focused:2 shen:1 simplicity:2 splitting:1 stats:1 dominate:1 classic:1 notion:32 altman:1 limiting:1 feel:1 enhanced:1 president:1 annals:1 user:22 programming:1 us:2 designing:1 satisfying:8 recognition:1 predicts:1 mistreatment:2 ep:1 epsrc:1 preprint:3 solved:2 revisiting:1 ensures:5 verwer:1 trade:1 highest:1 substantial:1 balanced:2 feller:1 mu:1 nash:1 econometrica:1 cam:1 flurry:1 employment:1 trained:2 rewrite:2 solving:2 pedreschi:1 predictive:3 division:8 f2:4 efficiency:1 basis:1 gu:1 easily:1 indirect:1 differently:1 isabel:2 various:3 tx:1 regularizer:4 train:6 zemel:1 outcome:23 quite:3 heuristic:1 larger:3 solve:2 supplementary:1 stanford:3 relax:2 ramp:1 remover:1 statistic:2 knn:1 think:2 housing:1 offend:1 propose:3 remainder:3 uci:2 frisk:5 moeller:1 achieve:3 oz:1 description:1 gentle:1 getting:2 guaranteeing:1 leave:1 rotated:1 depending:1 ac:1 minor:1 b0:10 received:7 eq:7 implies:2 attribute:29 human:1 raghavan:1 stringent:2 violates:3 material:1 muhammad:1 f1:4 im:1 z2z:13 exploring:1 credit:1 considered:1 ic:1 prohibited:1 algorithmic:3 mapping:3 predict:3 automate:1 hispanic:1 achieves:5 sought:1 adopt:1 friedler:2 label:5 maker:3 prf:12 title:1 sensitive:32 largest:1 establishes:1 city:1 cotter:1 minimization:1 offs:1 always:7 gaussian:1 kalai:1 avoid:3 asoh:1 office:1 focus:2 bernoulli:1 indicates:1 greatly:1 dependent:2 membership:2 bandit:1 expand:1 interested:1 overall:9 classification:14 prevention:1 raised:1 constrained:4 equal:4 aware:6 having:3 beach:1 defaulter:4 sampling:1 represents:2 icml:1 fairness:53 imp:4 future:3 report:3 inherent:2 barocas:1 randomly:1 ve:8 individual:7 phase:1 consisting:2 invariably:1 interest:1 highly:3 mining:2 dwork:3 leiserson:1 evaluation:1 possibility:1 violation:1 light:1 primal:1 implication:1 accurate:1 mullainathan:1 necessary:1 preferential:1 decoupled:1 tree:1 divide:2 goh:1 desired:1 re:1 circle:1 varian:1 theoretical:1 stopped:2 rao:1 retains:2 cost:10 entry:1 too:1 weller:1 varies:1 aw:1 synthetic:6 st:1 person:2 venue:1 off:1 receiving:1 kirchner:1 woman:10 usd:1 worse:2 creating:1 resort:8 luong:1 account:1 potential:2 de:1 pedestrian:1 satisfy:9 explicitly:1 race:3 later:1 try:1 lab:1 red:2 bayes:1 option:1 minimize:6 purple:1 publicly:1 accuracy:40 largely:1 efficiently:2 correspond:1 itcs:1 worth:1 ary:1 classified:5 acc:8 sharing:7 definition:5 against:2 bargaining:4 involved:1 gain:3 sampled:1 dataset:29 treatment:72 stop:5 hardt:2 rdle:1 ruggieri:2 knowledge:1 formalize:4 higher:6 violating:1 follow:1 methodology:4 supervised:1 disciplined:3 formulation:2 just:2 hand:3 receives:1 trust:1 rodriguez:3 defines:1 logistic:2 quality:5 usa:1 concept:7 true:4 contain:2 former:1 equality:5 inspiration:2 assigned:1 regularization:2 nypd:6 hence:1 white:7 attractive:2 round:1 sin:2 game:3 during:2 skewed:1 mpi:8 larson:1 criterion:13 theoretic:1 thomson:1 discriminated:1 extend:3 belong:2 interpret:1 significant:3 cambridge:1 feldman:2 rd:2 unconstrained:5 nyc:1 mathematics:1 similarly:3 akaho:1 had:1 moving:1 chapelle:1 compiled:1 pitassi:2 base:1 own:3 recent:4 freeness:5 perspective:1 driven:2 reverse:1 scenario:15 certain:9 nonconvex:2 inequality:2 binary:5 life:1 yi:4 krishna:1 greater:2 additional:2 somewhat:1 goel:2 maximize:3 ii:2 desirable:2 alan:2 characterized:1 determination:1 offer:7 long:1 cross:1 gummadi:4 parenthesis:1 impact:83 prediction:8 involving:1 ensuring:5 regression:2 heterogeneous:1 metric:6 expectation:3 bz:14 arxiv:7 kernel:1 adopting:4 achieved:8 receive:3 whereas:2 addition:4 justified:1 bz0:2 compas:8 scheidegger:2 source:1 allocated:1 weapon:2 archive:2 subject:8 tend:1 plato:1 member:1 reingold:1 fate:1 effectiveness:1 call:1 iii:1 easy:2 split:1 automated:2 variety:4 affect:1 zi:6 misclassifications:1 earns:1 economic:1 regarding:1 avenue:1 tradeoff:1 whether:4 motivated:1 utility:16 granted:1 penalty:2 york:2 prefers:3 action:1 useful:1 n510129:1 encyclopedia:1 kamiran:1 http:6 generate:3 exist:2 misclassifying:3 revisit:1 notice:1 sign:16 leverhulme:1 per:2 blue:2 group:127 salient:1 terminology:2 achieving:3 drawn:2 neither:1 econometric:1 relaxation:3 fraction:8 year:2 turing:2 defendant:3 selbst:1 sakuma:1 swersky:1 reader:1 wu:1 looser:3 draw:3 decision:53 prefer:6 appendix:6 cyan:1 pay:1 gomez:1 fold:2 constraint:11 certifying:1 kleinberg:1 optimality:1 separable:1 aw665:1 recidivism:4 department:1 marking:2 belonging:1 jr:1 beneficial:13 smaller:2 increasingly:1 describes:1 slightly:2 angwin:1 joseph:1 making:25 www1:1 intuitively:1 taken:3 mori:1 resource:6 agree:1 previously:1 calder:2 eventually:1 mechanism:4 precomputed:1 tractable:4 instrument:1 available:3 observe:1 appropriate:2 disagreement:3 denotes:5 ensure:5 include:1 opportunity:4 patil:1 sw:7 society:1 comparatively:1 objective:4 added:1 question:1 strategy:1 costly:1 separate:2 tue:1 code:1 racial:1 setup:2 unfortunately:1 potentially:1 negative:10 disparate:8 design:2 implementation:2 zt:3 policy:1 unknown:3 diamond:1 allowing:1 observation:3 datasets:13 acknowledge:1 anti:1 defining:3 extended:1 ever:1 situation:1 police:1 rating:1 introduced:3 criminal:1 namely:1 specified:2 connection:2 california:1 nip:4 adult:10 beyond:2 suggested:1 justin:1 pattern:1 program:2 green:2 max:7 greatest:1 power:1 suitable:1 difficulty:1 natural:1 business:1 predicting:1 valera:4 indicator:1 scheme:1 github:1 suspicion:1 axis:2 acknowledges:1 naive:1 embodied:1 prior:1 literature:7 l2:3 review:1 discovery:2 friedlander:1 understanding:1 law:3 loss:4 par:2 bail:2 expect:1 huq:1 men:10 interesting:4 allocation:5 fictitious:1 proportional:1 srebro:1 executive:1 awareness:1 proxy:3 bank:2 pareto:2 summary:1 parity:93 free:3 dis:1 side:2 bias:1 institute:2 taking:1 benefit:49 distributed:1 boundary:11 overcome:1 world:6 adopts:1 made:1 turini:2 party:3 income:1 social:3 approximate:1 preferred:92 ml:1 handbook:1 b1:10 pierson:1 xi:4 corbett:1 protected:1 table:1 additionally:6 promising:1 expanding:1 ca:1 improving:1 necessarily:1 zafar:6 domain:1 aistats:1 linearly:1 whole:2 subsample:1 big:2 fair:16 allowed:1 site:1 en:1 envy:8 house:1 removing:2 z0:8 bishop:1 showing:1 experimented:1 svm:4 gupta:1 concern:2 false:3 mattu:1 margin:2 easier:1 civil:3 vii:1 sharad:1 vulnerable:1 collectively:5 springer:2 gender:2 chouldechova:1 satisfies:7 kamishima:1 insisting:2 conditional:27 goal:4 consequently:1 exposition:1 towards:1 eventual:3 rbf:1 price:1 change:3 loan:6 specifically:4 uniformly:2 prejudice:2 kearns:1 called:1 partly:1 experimental:2 invariance:1 equity:1 formally:1 immorlica:1 support:2 philosophy:1 evaluate:1 ex:7 |
6,221 | 6,628 | Inferring Generative Model Structure
with Static Analysis
Paroma Varma1 , Bryan He2 , Payal Bajaj2 ,
Nishith Khandwala2 , Imon Banerjee3 , Daniel Rubin3,4 , Christopher R?2
1
Electrical Engineering, 2 Computer Science, 3 Biomedical Data Science, 4 Radiology
Stanford University
{paroma,bryanhe,pabajaj,nishith,imonb,rubin}@stanford.edu,
[email protected]
Abstract
Obtaining enough labeled data to robustly train complex discriminative models is a
major bottleneck in the machine learning pipeline. A popular solution is combining
multiple sources of weak supervision using generative models. The structure of these
models affects the quality of the training labels, but is difficult to learn without any
ground truth labels. We instead rely on weak supervision sources having some structure by virtue of being encoded programmatically. We present Coral, a paradigm
that infers generative model structure by statically analyzing the code for these
heuristics, thus significantly reducing the amount of data required to learn structure.
We prove that Coral?s sample complexity scales quasilinearly with the number of
heuristics and number of relations identified, improving over the standard sample
complexity, which is exponential in n for learning nth degree relations. Empirically,
Coral matches or outperforms traditional structure learning approaches by up to
3.81 F1 points. Using Coral to model dependencies instead of assuming independence results in better performance than a fully supervised model by 3.07 accuracy
points when heuristics are used to label radiology data without ground truth labels.
1
Introduction
Complex discriminative models like deep neural networks rely on a large amount of labeled training
data for their success. For many real-world applications, obtaining this magnitude of labeled
data is one of the most expensive and time consuming aspects of the machine learning pipeline.
Recently, generative models have been used to create training labels from various weak supervision
sources, such as heuristics or knowledge bases, by modeling the true class label as a latent variable
[1, 2, 27, 31, 36, 37]. After the necessary parameters for the generative models are learned using
unlabeled data, the distribution over the true labels can be inferred. Properly specifying the structure
of these generative models is essential in estimating the accuracy of the supervision sources. While
traditional structure learning approaches have focused on the supervised case [23, 28, 41], previous
works related to weak supervision assume that the structure is user-specified [1, 27, 31, 36]. Recently,
Bach et al. [2] showed that it is possible to learn the structure of these models with a sample complexity
that scales sublinearly with the number of possible binary dependencies. However, the sample
complexity scales exponentially for higher degree dependencies, limiting its ability to learn complex
dependency structures. Moreover, the time required to learn the dependencies also grows exponentially
with the degree of dependencies, hindering the development of user-defined heuristics.
This poses a problem in many domains where high degree dependencies are common among heuristics
that operate over a shared set of inputs. These inputs are interpretable characteristics extracted from the
data. For example, various approaches in computer vision use predicted bounding box or segmentation
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
attributes [18, 19, 29], like location and size, to weakly supervise more complex image-based learning
tasks [5, 7, 11, 26, 38]. Another example comes from the medical imaging domain, where attributes
include characteristics such as the area, intensity and perimeter of a tumor, as shown in Figure 1. Note
that these attributes are computationally represented, and the heuristics written over them are encoded
programmatically as well. There are typically a relatively small set of interpretable characteristics, so
the heuristics often share these attributes. This results in high order dependency structures among these
sources, which are crucial to model in the generative model that learns accuracies for these sources.
To address the issue of learning higher order dependencies efficiently, we present Coral, a paradigm
that statically analyzes the source code of the weak supervision sources to infer, rather than learn,
the complex relations among heuristics. Coral?s sample complexity scales quasilinearly with the
number of relevant dependencies and does not scale with the degree of the dependency, unlike the
sample complexity for Bach et al. [2], which scales exponentially with the degree of the dependency.
Moreover, the time to identify these relations is constant in the degree of dependencies, since it only
requires looking at the source code for each heuristic to find which heuristics share the same input.
This allows Coral to infer high degree dependencies more efficiently than techniques that rely only
on statistical methods to learn them, and thus generate a more accurate dependency structure for the
heuristics. Coral then uses a generative model to learn the proper weights for this dependency structure
to assign probabilistic labels to training data.
We experimentally validate the performance of Coral across various domains and show it outperforms
traditional structure learning under various conditions while being significantly more computationally
efficient. We show how modeling dependencies leads to an improvement of 3.81 F1 points compared
to standard structure learning approaches. Additionally, we show that Coral can assign labels to data
that have no ground truth labels, and this augmented training set results in improving the discriminative
model performance by 3.07 points. For a complex relation-based image classification task, 6 heuristic
functions written over only bounding box attributes as primitives are able to train a model that
performs within 0.74 points of the F1 score achieved by a fully-supervised model trained on the rich,
hand-labeled attribute and relation information in the Visual Genome database [21].
2
The Coral Paradigm
The Coral paradigm takes as input a set of domain-specific primitives and a set of programmatic
user-defined heuristic functions that operate over the primitives. We formally define these abstractions
in Section 2.1. Coral runs static analysis on the source code that defines the primitives and the heuristic
functions to identify which sets of heuristics are related by virtue of sharing primitives (Section 2.2).
Once Coral identifies these dependencies, it uses a factor graph to model the relationship between
the heuristics, primitives and the true class label. We describe the conditions under which Coral can
learn the structure of the generative model with significantly less data than traditional approaches
in Section 2.3 and demonstrate how this affects generative model accuracy via simulations. Finally,
we discuss how Coral learns the accuracies of the each heuristic and outputs probabilistic labels for
the training data in Section 2.4.
Raw Images
Segmentations
Heuristic
Functions
Noisy
Labels
p1: Area
?1(p1)
Aggressive
p2: Perimeter
?2(p2)
NonAggressive
p3: Intensity
?3(p2, p3)
Aggressive
Domain Specific
Primitives
Primitives and Heuristics are
User-Defined
Training
Labels
Generative
Model
75%
75%
Aggressive
Aggressive
Discriminative
Model
The Coral Pipeline
Figure 1: Running example for the Coral paradigm. Users apply standard algorithms to segment tumors
from the X-ray and extract the domain-specific primitives from the image and segmentation. They
write heuristic functions over the primitives that output a noisy label for each image. The generative
model takes these as inputs and provides probabilistic training labels for the discriminative model.
2
2.1
Coral Abstractions
Domain-Specific Primitives Domain-specific primitives (DSPs) in Coral are the simplest elements
that heuristic functions take as input and operate over. DSPs in Coral have semantic meaning, making
them interpretable for users. This is akin to the concept of language primitives in programming
languages, in which they are the smallest unit of processing with meaning. The motivation for making
the DSPs domain-specific instead of a general construct for the various data modalities is to allow users
to take advantage of existing work in their field to extract meaningful characteristics from the raw data.
Figure 1 shows an example of a pipeline for bone tumor classification as aggressive or non-aggressive,
inspired by one of our real experiments. First, an automated segmentation algorithm is used to generate
a binary mask for where the tumor is [20, 25, 34, 39]. Then, we define 3 DSPs based on the segmentation:
area (p1 ), perimeter (p2 ) and total intensity (p3 ) of the segmented area. More complex characteristics
such as those that capture texture, shape and edge features can also be used [4, 14, 22] (see Appendix).
We now define a formal construct for how DSPs are encoded programmatically. Users generate
DSPs in Coral through a primitive specifier function, such as create_primitives in Figure 2(a).
Specifically, this function takes as input a single unlabeled data point (and necessary intermediate
representations such as the segmentation) and returns an instance of PrimitiveSet, which maps
primitive names to primitive values, like integers (we refer to a specific instance of this class as P).
Note that P.ratio is composed of two other primitives, while the rest of the primitives are generated
independently from the image and segmentation.
def create_primitives(image,segmentation):
P = PrimitiveSet()
P.area = get_area(segmentation)
P.perimeter = get_perimeter(segmentation)
P.intensity = np.sum(segmentation*image)
P.perimeter
get_perimeter()
value
function
P.area
P.intensity
P.ratio
/
value
value
operator
P.perimeter
P.ratio = P.intensity/P.perimeter
return P
P.intensity
/
P.ratio
P.perimeter
value
(b)
(a)
(c)
Figure 2: (a) The create_primitives function that generates primitives. (b) Part of the AST for the
create_primitives function. (c) The composition structure that results from traversing the AST.
Heuristic Functions In Coral, heuristic functions (HFs) can be viewed as mapping a subset of the
DSPs to a noisy label for the training data, as shown in Figure 1. In our experience with user-defined
HFs, we observe that HFs are usually nested if-then statements in which each statement checks whether
the value of a single primitive or a combination of them are above or below a user-set threshold (see
Appendix). As shown in Figure 3(a), they take fields of the object P as input and return a label (or
abstain) based on the value of the input primitives. While our running example focuses on a single
data point for DSP generation and HF construction, both procedures are applied to the entire training
set to assign a set of noisy labels from each HF to each data point.
2.2
Static Dependency Analysis
Since the number of DSPs in some domains can be relatively small, multiple HFs can operate over
the same DSPs. HFs that share at least one primitive are trivially related to each other. Prior work
[2] learns these dependencies using the labels HFs assign to data points and its probability of success
scales with the amount of data available. However, only pairwise HF dependencies can be learned
efficiently, since the data required grows exponentially with the degree of the HF relation. This in
turn limits the complexity of the dependency structure this method can accurately learn and model.
Heuristic Function Inputs Coral takes advantage of the fact that users write HFs over a known,
finite set of primitives. It infers dependencies that exist among HFs by simply looking at the source
code of how the DSPs and HFs are constructed. This process requires no data to successfully learn
the dependencies, making it more computationally efficient than standard approaches. In order to
determine whether any set of HFs share at least one DSP, Coral looks at the input for each HF. Since
the HFs only take as input the DSP they operate over, simply grouping HFs by the primitives they
share is an efficient approach for recognizing these dependencies.
3
As shown in our running example, this would result in Coral not recognizing any dependencies among
the HFs since the input for all 3 HFs are different (Figure 3(a)). This, however, would be incorrect,
since the primitive P.ratio is composed of P.perimeter and P.intensity, which makes ?2
and ?3 related. Therefore, along with looking at the primitives that each HF takes as input, it is also
essential to model how these primitives are composed.
Primitive Compositions We use our running example in Figure 2 to explain how Coral gathers
information about DSP compositions. Coral builds an abstract syntax tree (AST) to represent the
computations the create_primitives function performs. An AST represents operations involving
the primitives as a tree, as shown in Figure 2(b). To find primitive compositions from the AST, Coral
first finds the expressions in the AST that add primitives to P (denoted in the AST as P.name). Then,
for each assignment expression, Coral traverses the subtree rooted at the assignment expression and
adds all other encountered primitives as a dependency for P.name. If no primitives are encountered
in the subtree, the primitive is registered as being independent of the rest. The composition structure
that results from traversing the AST is shown in Figure 2(c), where P.area, P.intensity, and
P.perimeter are independent while P.ratio is a composition.
Heuristic Function Dependency Structure With knowledge of how the DSPs are composed, we
return to our original method of looking at the inputs of the HFs. As before, we identify that ?1 and
?2 use P.area and P.perimeter, respectively. However, we now know that ?3 uses P.ratio, which
is a composition of P.intensity and P.perimeter. This implies that ?3 will be related to any HF
that takes either P.intensity, P.perimeter, or both as inputs. We proceed to build a relational
structure among the HFs and DSPs. As shown in Figure 3(b), this structure shows which independent
DSPs each HF operates over. The relational structure implicitly encodes dependency information
about the HFs ? if an edge points from one primitive to n HFs, those n HFs are in an n-way relation
by virtue of sharing that primitive. This dependency information can more formally be encoded in
a factor graph shown in Figure 3(c), which is discussed in Section 2.3. Note that we chose a particular
programmatic setup for creating DSPs and HFs to explain how static analysis can infer dependencies;
however, this process can be modified to work with other setups that encode DSPs and HFs as well.
def ?_1(P.area):
if P.area >= 2.0:
return 1
else:
return -1
def ?_2(P.perimeter):
if P.perimeter <= 12.0:
return 1
else:
return 0
def ?_3(P.ratio):
if P.ratio <= 5.0:
return 1
else:
return -1
?1
P.area
?1
???
1
p1
?2
???
2
p2
?3
???
3
p3
????
1
?2
P.perimeter
Y
????
2
????
????
3
?3
(a)
P.intensity
(b)
(c)
Figure 3: (a) shows the encoded HFs. (b) shows the HF dependency structure where DSP nodes have
an edge going to the HFs that use them as inputs (explicitly or implicitly). (c) shows the factor graph
Coral uses to model the relationship between HFs, DSPs, and latent class label Y.
2.3
Creating the Generative Model
We now describe the generative model used to predict the true class labels. The Coral model uses a
factor graph (Figure 3(c)) to model the relationship between the primitives (p ? R), heuristic functions
(? ? {?1,0,1}) and latent class label (Y ? {?1,1}). We show that by incorporating information about
how primitives are shared across HFs from static analysis, this factor graph infers all dependencies
between the heuristics that are guaranteed to be present. We also describe how Coral recovers additional
dependencies among the heuristics by studying empirical relationships between the primitives.
Modeling Heuristic Function Dependencies Now that dependencies have been inferred via static
analysis, the goal is to learn the accuracies for each HF and assign labels to training data accordingly.
4
The factor graph thus consists of two types of factors: accuracy factors ?Acc and HF factors from static
analysis ?HF .
The accuracy factors specify the accuracy of each heuristic function and are defined as
?Acc
i (Y,?i ) = Y ?i , i = 1,...,n
where n is the total number of heuristic functions.
The static analysis factors ensure that the heuristics are correctly evaluated based on the HF
dependencies found via static analysis. They ensure that a probability of zero is given to any
configuration where an HF does not have the correct value given the primitives it depends on. The
static analysis factors are defined as.
0
if ?i is valid given p1 ,...,pm
?HF
(?
,p
,...,p
)
=
, i = 1,...,n
i 1
m
i
?? otherwise
Since these factors are obtained directly from static analysis, they can be recovered with no data.
However, we note that static analysis is not sufficient to capture all dependencies required in the factor
graph to accurately model the process of generating training labels. Specifically, static analysis can
(i) pick up spurious dependencies among HFs that are not truly dependent on each other, or
(ii) miss key dependencies among HFs that exist due to dependencies among the DSPs in the HFs.
(i) can occur if some ?A takes as input DSPs pi ,pj and ?B takes as input DSPs pi ,pk , but pi always has
the same value. Although static analysis would pick up that ?A and ?B share a primitive and should
have a dependency, this may not be true if pj and pk are independent. (ii) can occur if two HFs depend
on different primitives, but these primitives happen to always have the same value. In this case, it is
impossible for static analysis to infer the dependency between the HFs if the primitives have different
names and are generated independently, as described in Section 2.2. A more realistic scenario comes
from our running example, where we would expect the area and perimeter of the tumor to be related.
To account for both cases, it is necessary to capture the possible dependencies that occur among the
DSPs to ensure that the dependencies from static analysis do not misspecify the factor graph. We
introduce a factor to account for additional dependencies among the primitives, ?DSP . There are many
possible choices for this dependency factor, but one simple choice is to model pairwise similarity
between the primitives. For binary and discrete primitives, the dependency factor with pairwise
similarity can be represented as
X
Sim
?DSP (p1 ,...,pm ) =
?Sim
ij (pi ,pj ), where ?ij (pi ,pj ) = I[pi = pj ].
i<j
The dependency factor can be generalized to continuous-valued primitives by binning the primitives
into discrete values before comparing for similarity.
Finally, with three types of factors, the probability distribution specified by the factor graph is
?
?
n
m X
m
n
X
X
X
Sim Sim ?
P (y,?1 ,...,?n ,p1 ,...,pm ) = exp? ?iAcc ?Acc
?HF
?ij
?ij
i +
i +
i=1
i=1
i=1 j=i+1
Sim
where ?Acc and ?ij
are weights that specify the strength of factors ?Acc and ?Sim
ij .
Inferring Dependencies without Data The HF factors capture all dependencies among the
heuristic functions that are not represented by the ?DSP factor. The dependencies represented by the
?DSP factor are precisely the dependencies that cannot be inferred via static analysis due to the fact
that this factor depends solely on the content of the primitives. It is therefore impossible to determine
what this factor is without data.
While assuming that we have the true ?DSP seems like a strong condition, we find that in real-world experiments, including the ?DSP factor rarely leads to improvements over the case when we only include the
?Acc and ?HF factors. In some of our experiments (see Section 3), we use bounding box location, size and
object labels as domain-specific primitives for image and video querying tasks. Since these primitives
are not correlated, modeling the primitive dependency does not lead to any improvement over just modeling HF dependencies from static analysis. Moreover, in other experiments where modeling the relation
among primitives helps, we observe relatively small benefits above what modeling HF dependencies
provides (Section 3). Therefore, even without data, it is possible to model the most important dependencies among HFs that lead to significant gains over the case in which no dependencies are modeled.
5
2.4
Generating Probabilistic Training Labels
Relative Improvement (%)
Given the probability distribution of the factor graph, our goal is to learn the proper weights ?iAcc
Sim
and ?ij
. Coral adopts structure learning approaches described in recent work [2], which learns
dependency structures in the weak supervision setting and maximizes the `1 -regularized marginal
pseudolikelihood of each primitive to learn the weights of the relevant factors.
50 Binary Dependencies
25
0
Relative Improvement over Structure Learning
50 3-ary Dependencies
50 4-ary Dependencies
25
0
250
N
500
0
n=6
n=8
n=10
25
0
250
N
500
0
0
250
N
500
Figure 4: Simulation demonstrating improved generative model accuracy with Coral compared to
structure learning [2] and Coral. Relative improvement of Coral over structure learning is plotted
against number of unlabeled data points (N ) and number of HFs (n).
To learn the weights of the generative model, we use contrastive divergence [15] as a maximum
likelihood estimation routine and maximize the marginal likelihood of the observed primitives. Gibbs
sampling is used to estimate the intractable gradients, which are then used in stochastic gradient descent.
Because the HFs are typically deterministic functions of the primitives (represented as the ?? value
of the correctness factors for invalid heuristic values), standard Gibbs sampling will not be able to
mix properly. As a result, we modify the Gibbs sampler to simultaneously sample one primitive along
with all heuristics that depend on it. Despite the fact that the true class label is latent, this process still
converges to the correct parameter values [27]. Additionally, the amount of data necessary to learn the
parameters scales quasilinearly with the number of parameters. In our case, the number of parameters
is simply the number of heuristics n and the number of relevant primitive similarity dependencies s.
We now formally state the conditions for this result, which match those of Ratner et al. [27], and give the
sample complexity of our method. First, we assume that there exists some feasible parameter set ? ? Rn
that is known to contain the parameter ?? = (?Acc , ?Sim ) that models the true distribution ? ? of the data:
??? ? ? s.t. ?? ? (p1 ,...,pm ,Y ) = ?? (p1 ,...,pm ,Y ).
(1)
?
Next, we must be able to accurately learn ? if we are provided with labeled samples of the true
distribution. Specifically, there must be an asymptotically unbiased estimator ?? that takes some set
of labeled data T independently sampled from ? ? such that for some c > 0,
? ) (2c|T |)?1 I.
Cov ?(T
(2)
Finally, we must have enough sufficiently accurate heuristics so that we have a reasonable estimate
of Y. For any two feasible models ?1 ,?2 ? ?,
h
i
c
E(p1 ,...,pm ,Y )???1 Var(p01 ,...,p0m ,Y 0 )???2 (Y 0 | p1 = p01 ,...,pm = p0m ) ?
(3)
n+s
Proposition 1. Suppose that we run stochastic gradient descent to produce estimates of the weights
??= (??Acc , ??Sim ) in a setup satisfying conditions (1), (2), and (3). Then, for any fixed error > 0, if the
number of unlabeled
N is at least ?[(n+s)log(n+s)], then our expected parameter error
h data points
i
? 2
2
?
is bounded by E k??? k ? .
The proof follows from the sample complexity of Ratner et al. [27] and appears in the Appendix. With
Sim
the weights ??iAcc and ??ij
maximizing the marginal likelihood of the observed primitives, we have
a fully specified factor graph and complete generative model, which can be used to predict the latent
class label. For each data point, we compute the label each heuristic function applies to it using the
6
values of the domain-specific primitives. Through the accuracy factors, we then estimate a distribution
for the latent class label and use these noisy labels to train a discriminative model.
We present a simulation to empirically compare our sample complexity with that of structure learning
[2]. In our simulation, we have n HFs, each with an accuracy of 75%, and explore settings in which
there exists one binary, 3-ary and 4-ary dependency among the HFs. The dependent HFs share exactly
one primitive, and the primitives themselves are independent (s = 0). We show our results in Figure
4. In the case with a binary dependency, structure learning recovers the necessary dependency with
few samples, and has similar performance to Coral. In contrast, in the second and third settings with
high-order dependencies, structure learning struggles to recover the relevant dependency, and performs
worse than Coral even as more training data is provided.
3
Experimental Results
We seek to experimentally validate the following claims about our approach. Our first claim is that HF
dependencies inferred via static analysis perform significantly better than a model that does not take
dependencies into account. Second, we compare to a structure learning approach for weak supervision
[2] and show how we outperform it over a variety of domains. Finally, we show that in case primitive
dependencies exist, Coral can learn and model those as well. We show that modeling the dependencies
between the heuristic functions and primitives can generate training sets that, in some cases, beat
fully supervised models by labeling additional unlabeled data. Our classification tasks range from
specialized medical domains to natural images and video, and we include details of the DSPs and
HFs in the Appendix. Note that while the number of HFs and DSPs is fairly low (Table 1), using static
analysis to automatically infer dependencies rather than ask users to identify them saves significant
effort since the number of possible dependencies grows exponentially with the number of HFs present.
We compare our approach to majority vote (MV), generative models that learn the accuracies of different
heuristics, specifically one that assumes the heuristics are independent (Indep) [27], and Bach et al. [2]
that learns the binary inter-heuristic dependencies (Learn Dep). We also compare to the fully supervised
(FS) case, and measure the performance of the discriminative model trained with labels generated using
the above methods. We split our approach into two parts: inferring HF dependencies using only static
analysis (HF Dep) and additionally learning primitive level dependencies (HF+DSP Dep).
Coral Performance
Figure 5: Discriminative model performance comparing HF Dep (HF dependencies from static
analysis) and HF+DSP Dep (HF and DSP dependencies) to other methods. Numbers in Appendix.
Visual Genome and ActivityNet Classification We explore how to extract complex relations in
images and videos given object labels and their bounding boxes. We used subsets of two datasets,
Visual Genome [21] and ActivityNet [9], and defined our task as finding images of ?a person biking
down a road? and finding basketball videos, respectively. For both tasks, a small set of DSPs were
shared heavily among HFs, and modeling the dependencies observed by static analysis led to a
significant improvement over the independent case. Since these dependencies involved groups of
3 or more heuristics, Coral improved significantly over structure learning as well, which was unable to
model these dependencies due to the lack of enough data. Moreover, modeling primitive dependencies
did not help since the primitives were indeed independent (Table 1). We report our results for these
tasks in terms of the F1 score (harmonic mean of the precision and recall) since there was significant
class imbalance which accuracy would not capture well.
Bone Tumor Classification We used a set of 802 labeled bone tumor X-ray images along with their
radiologist-drawn segmentations. Our task was to differentiate between aggressive and non-aggressive
7
Table 1: Heuristic Function (HF) and Domain-Specific Primitive (DSP) statistics. Discriminative
model improvement with HF+DSP Dep over other methods. *improvements shown in terms of F1
score, rest in terms of accuracy. ActivityNet model is LR using VGGNet embeddings as features.
Number of
Application
Improvement Over
Model
DSPs
HFs
Shared DSPs
MV
Indep
Learn Dep
FS
Visual Genome
ActivityNet
7
5
5
4
2
2
GoogLeNet
VGGNet+LR
7.49*
6.23*
2.90*
3.81*
2.90*
3.81*
-0.74*
-1.87*
Bone Tumor
Mammogram
17
6
7
6
0
0
LR
GoogLeNet
5.17
4.62
3.57
1.11
3.06
0
3.07
-0.64
tumors. We generated HFs that were a combination of hand-tuned rules and decision-tree generated
rules (tuned on a small held out subset of the dataset). The discriminative model utilized a set of 400
hand-tuned features (note that there is no overlap between these features and the DSPs) that encoded
various shape, texture, edge and intensity-based characteristics. Although there were no explicitly
shared primitives in this dataset, the generative model was still able to model the training labels more
accurately with knowledge of how heuristics used primitives, which affects the relative false positive
and false negative rates. Thus, the generative model significantly improved over the independent
model. Moreover, a small dataset size hindered structure learning, which gave a minimal boost over
the independent case (Table 1). When we used heuristics in Coral to label an additional 800 images
that had no ground truth labels, we beat the previous FS score by 3.07 points (Figure 5, Table 1).
Mammogram Tumor Classification We used the DDSM-CBIS [32] dataset, which consists of
1800 scanned film mammograms and associated segmentations for the tumors in the form of binary
masks. Our task was to identify whether a tumor is malignant or benign, and each heuristic only
operated over one primitive, resulting in no dependencies that static analysis could identify. In
this case, structure learning performed better than Coral when we only used static analysis to infer
dependencies (Figure 5). However, including primitive dependencies allowed us to match structure
learning, resulting in a 1.11 point improvement over the independent case (Figure 5, Table 1).
4
Related Work
As the need for labeled training data grows, a common alternative is to utilize weak supervision sources
such as distant supervision [10, 24], multi-instance learning [16, 30], and heuristics [8, 35]. Specifically
for images, weak supervision using object detection and segmentation or visual databases is a popular
technique as well (detailed discussion in Appendix). Estimating the accuracies of these sources without
access to ground truth labels is a classic problem [13]. Methods such as crowdsourcing [12, 17, 40],
boosting[3, 33], co-training [6], and learning from noisy labels are some of the popular approaches that
can combine various sources of weak supervision to assign noisy labels to data. However, Coral does
not require any labeled data to model the dependencies among the heuristics, which can be interpreted
as workers, classifiers or views for the above methods, and domain-specific primitives.
Recently, generative models have also been used to combine various sources of weak supervision
[1, 31, 36, 37]. One specific example, data programming [27], proposes using multiple sources of
weak supervision for text data in order to describe a generative model and subsequently learns the
accuracies of these sources. Coral also focuses on multiple programmatically encoded heuristics that
can weakly label data and learns their accuracies to assign labels to training data. However, Coral adds
an additional layer of domain-specific primitives in its generative model, which allows it to generalize
beyond text-based heuristics. It also infers the dependencies among the heuristics and the primitives,
rather than requiring users to specify them.
Other previous work also assume that this structure in generative models is user-specified [1, 31, 36].
However, Bach et al. [2] recently showed that it is possible to learn the dependency structure among
sources of weak supervision with a sample complexity that scales sublinearly with the number
of possible pairwise dependencies. Coral instead identifies the dependencies among the heuristic
functions by inspecting the content of the programmable functions, therefore relying on significantly
less data to learn the generative model structure. Moreover, Coral can also pick up higher-order
dependencies, for which Bach et al. [2] needs large amounts of data to detect.
8
5
Conclusion and Future Work
In this paper, we introduced Coral, a paradigm that models the dependency structure of weak
supervision heuristics and systematically combines their outputs to assign probabilistic labels to
training data. We described how Coral takes advantage of the programmatic nature of these heuristics
in order to infer dependencies among them via static analysis. Coral therefore requires a sample
complexity that is quasilinear in the number of heuristics and relations found. We showed how Coral
leads to significant improvements in discriminative model accuracy over traditional structure learning
approaches across various domains. Coral scratches the surface of the possible ways weak supervision
can borrow from the field of programming languages, especially as weak supervision sources are used to
label large magnitudes of data and need to be encoded programmatically. We look at a natural extension
of treating the process of encoding heuristics as writing functions and hope to explore the interactions
between systematic training set creation and concepts from the programming language field.
Acknowledgments We thank Shoumik Palkar, Stephen Bach, and Sen Wu for their helpful
conversations and feedback. We are grateful to Darvin Yi for his assistance with the DDSM dataset
based experiments and associated deep learning models. We acknowledge the use of the bone
tumor dataset annotated by Drs. Christopher Beaulieu and Bao Do and carefully collected over his
career by the late Henry H. Jones, M.D. (aka ?Bones Jones?). This material is based on research
sponsored by Defense Advanced Research Projects Agency (DARPA) under agreement number
FA8750-17-2-0095. We gratefully acknowledge the support of the DARPA SIMPLEX program
under No. N66001-15-C-4043, DARPA FA8750-12-2-0335 and FA8750-13-2-0039, DOE 108845,
the National Science Foundation (NSF) Graduate Research Fellowship under No. DGE-114747,
Joseph W. and Hon Mai Goodman Stanford Graduate Fellowship, National Institute of Health
(NIH) U54EB020405, the Office of Naval Research (ONR) under awards No. N000141210041
and No. N000141310129, the Moore Foundation, the Okawa Research Grant, American Family
Insurance, Accenture, Toshiba, and Intel. This research was supported in part by affiliate members
and other supporters of the Stanford DAWN project: Intel, Microsoft, Teradata, and VMware. The
U.S. Government is authorized to reproduce and distribute reprints for Governmental purposes
notwithstanding any copyright notation thereon. The views and conclusions contained herein are
those of the authors and should not be interpreted as necessarily representing the official policies or
endorsements, either expressed or implied, of DARPA or the U.S. Government. Any opinions, findings,
and conclusions or recommendations expressed in this material are those of the authors and do not
necessarily reflect the views of DARPA, AFRL, NSF, NIH, ONR, or the U.S. government.
9
References
[1] E. Alfonseca, K. Filippova, J.-Y. Delort, and G. Garrido. Pattern learning for relation extraction with a
hierarchical topic model. In Proceedings of the 50th Annual Meeting of the Association for Computational
Linguistics: Short Papers-Volume 2, pages 54?59. Association for Computational Linguistics, 2012.
[2] S. H. Bach, B. He, A. Ratner, and C. R?. Learning the structure of generative models without labeled data.
In ICML, 2017.
[3] A. Balsubramani and Y. Freund. Scalable semi-supervised aggregation of classifiers. In Advances in Neural
Information Processing Systems, pages 1351?1359, 2015.
[4] I. Banerjee, L. Hahn, G. Sonn, R. Fan, and D. L. Rubin. Computerized multiparametric mr image analysis
for prostate cancer aggressiveness-assessment. arXiv preprint arXiv:1612.00408, 2016.
[5] M. Blaschko, A. Vedaldi, and A. Zisserman. Simultaneous object detection and ranking with weak
supervision. In Advances in neural information processing systems, pages 235?243, 2010.
[6] A. Blum and T. Mitchell. Combining labeled and unlabeled data with co-training. In Proceedings of the
eleventh annual conference on Computational learning theory, pages 92?100. ACM, 1998.
[7] S. Branson, P. Perona, and S. Belongie. Strong supervision from weak annotation: Interactive training
of deformable part models. In Computer Vision (ICCV), 2011 IEEE International Conference on, pages
1832?1839. IEEE, 2011.
[8] R. Bunescu and R. Mooney. Learning to extract relations from the web using minimal supervision. In ACL,
2007.
[9] F. Caba Heilbron, V. Escorcia, B. Ghanem, and J. Carlos Niebles. Activitynet: A large-scale video
benchmark for human activity understanding. In Proceedings of the IEEE Conference on Computer Vision
and Pattern Recognition, pages 961?970, 2015.
[10] M. Craven, J. Kumlien, et al. Constructing biological knowledge bases by extracting information from
text sources. In ISMB, pages 77?86, 1999.
[11] J. Dai, K. He, and J. Sun. Boxsup: Exploiting bounding boxes to supervise convolutional networks for
semantic segmentation. In Proceedings of the IEEE International Conference on Computer Vision, pages
1635?1643, 2015.
[12] N. Dalvi, A. Dasgupta, R. Kumar, and V. Rastogi. Aggregating crowdsourced binary ratings. In Proceedings
of the 22nd international conference on World Wide Web, pages 285?294. ACM, 2013.
[13] A. P. Dawid and A. M. Skene. Maximum likelihood estimation of observer error-rates using the EM
algorithm. Applied statistics, pages 20?28, 1979.
[14] R. M. Haralick, K. Shanmugam, et al. Textural features for image classification. IEEE Transactions on
systems, man, and cybernetics, 3(6):610?621, 1973.
[15] G. E. Hinton. Training products of experts by minimizing contrastive divergence. Neural computation,
14(8):1771?1800, 2002.
[16] R. Hoffmann, C. Zhang, X. Ling, L. Zettlemoyer, and D. S. Weld. Knowledge-based weak supervision
for information extraction of overlapping relations. In Proceedings of the 49th Annual Meeting of the
Association for Computational Linguistics: Human Language Technologies-Volume 1, pages 541?550.
Association for Computational Linguistics, 2011.
[17] M. Joglekar, H. Garcia-Molina, and A. Parameswaran. Comprehensive and reliable crowd assessment
algorithms. In Data Engineering (ICDE), 2015 IEEE 31st International Conference on, pages 195?206.
IEEE, 2015.
[18] D. Kang, J. Emmons, F. Abuzaid, P. Bailis, and M. Zaharia. Optimizing deep cnn-based queries over video
streams at scale. CoRR, abs/1703.02529, 2017. URL http://arxiv.org/abs/1703.02529.
[19] A. Karpathy and L. Fei-Fei. Deep visual-semantic alignments for generating image descriptions. In
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 3128?3137, 2015.
[20] M. R. Kaus, S. K. Warfield, A. Nabavi, P. M. Black, F. A. Jolesz, and R. Kikinis. Automated segmentation
of mr images of brain tumors 1. Radiology, 218(2):586?591, 2001.
[21] R. Krishna, Y. Zhu, O. Groth, J. Johnson, K. Hata, J. Kravitz, S. Chen, Y. Kalantidis, L.-J. Li, D. A. Shamma,
et al. Visual genome: Connecting language and vision using crowdsourced dense image annotations. arXiv
preprint arXiv:1602.07332, 2016.
[22] C. Kurtz, A. Depeursinge, S. Napel, C. F. Beaulieu, and D. L. Rubin. On combining image-based and
ontological semantic dissimilarities for medical image retrieval applications. Medical image analysis, 18
(7):1082?1100, 2014.
10
[23] N. Meinshausen and P. B?hlmann. High-dimensional graphs and variable selection with the lasso. The
annals of statistics, pages 1436?1462, 2006.
[24] M. Mintz, S. Bills, R. Snow, and D. Jurafsky. Distant supervision for relation extraction without labeled data.
In Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL and the 4th International
Joint Conference on Natural Language Processing of the AFNLP: Volume 2-Volume 2, pages 1003?1011.
Association for Computational Linguistics, 2009.
[25] A. Oliver, J. Freixenet, J. Marti, E. P?rez, J. Pont, E. R. Denton, and R. Zwiggelaar. A review of automatic
mass detection and segmentation in mammographic images. Medical image analysis, 14(2):87?110, 2010.
[26] M. Oquab, L. Bottou, I. Laptev, and J. Sivic. Is object localization for free? - Weakly-supervised learning
with convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, pages 685?694, 2015.
[27] A. J. Ratner, C. M. De Sa, S. Wu, D. Selsam, and C. R?. Data programming: Creating large training sets,
quickly. In Advances in Neural Information Processing Systems, pages 3567?3575, 2016.
[28] P. Ravikumar, M. J. Wainwright, J. D. Lafferty, et al. High-dimensional ising model selection using
l1-regularized logistic regression. The Annals of Statistics, 38(3):1287?1319, 2010.
[29] J. Redmon, S. Divvala, R. Girshick, and A. Farhadi. You only look once: Unified, real-time object detection.
In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 779?788, 2016.
[30] S. Riedel, L. Yao, and A. McCallum. Modeling relations and their mentions without labeled text. In Joint
European Conference on Machine Learning and Knowledge Discovery in Databases, pages 148?163.
Springer, 2010.
[31] B. Roth and D. Klakow. Combining generative and discriminative model scores for distant supervision.
In EMNLP, pages 24?29, 2013.
[32] R. Sawyer-Lee, F. Gimenez, A. Hoogi, and D. Rubin. Curated breast imaging subset of ddsm, 2016.
[33] R. E. Schapire and Y. Freund. Boosting: Foundations and algorithms. MIT press, 2012.
[34] N. Sharma, L. M. Aggarwal, et al. Automated medical image segmentation techniques. Journal of medical
physics, 35(1):3, 2010.
[35] J. Shin, S. Wu, F. Wang, C. De Sa, C. Zhang, and C. R?. Incremental knowledge base construction using
DeepDive. Proceedings of the VLDB Endowment, 8(11):1310?1321, 2015.
[36] S. Takamatsu, I. Sato, and H. Nakagawa. Reducing wrong labels in distant supervision for relation
extraction. In Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics:
Long Papers-Volume 1, pages 721?729. Association for Computational Linguistics, 2012.
[37] P. Varma, B. He, D. Iter, P. Xu, R. Yu, C. De Sa, and C. R?. Socratic learning: Augmenting generative
models to incorporate latent subsets in training data. arXiv preprint arXiv:1610.08123, 2017.
[38] W. Xia, C. Domokos, J. Dong, L.-F. Cheong, and S. Yan. Semantic segmentation without annotating segments.
In Proceedings of the IEEE International Conference on Computer Vision, pages 2176?2183, 2013.
[39] D. Yi, M. Zhou, Z. Chen, and O. Gevaert. 3-d convolutional neural networks for glioblastoma segmentation.
arXiv preprint arXiv:1611.04534, 2016.
[40] Y. Zhang, X. Chen, D. Zhou, and M. I. Jordan. Spectral methods meet em: A provably optimal algorithm
for crowdsourcing. Journal of Machine Learning Research, 17(102):1?44, 2016.
[41] P. Zhao and B. Yu. On model selection consistency of lasso. Journal of Machine learning research, 7(Nov):
2541?2563, 2006.
11
| 6628 |@word cnn:1 seems:1 nd:1 vldb:1 simulation:4 seek:1 programmatically:5 contrastive:2 pick:3 mention:1 configuration:1 score:5 daniel:1 tuned:3 fa8750:3 outperforms:2 existing:1 recovered:1 comparing:2 written:2 must:3 realistic:1 happen:1 distant:4 benign:1 shape:2 treating:1 interpretable:3 sponsored:1 takamatsu:1 generative:28 accordingly:1 mccallum:1 short:1 ratner:4 lr:3 provides:2 boosting:2 node:1 location:2 traverse:1 org:1 zhang:3 along:3 constructed:1 incorrect:1 prove:1 consists:2 combine:3 eleventh:1 ray:2 dalvi:1 introduce:1 pairwise:4 inter:1 mask:2 expected:1 indeed:1 sublinearly:2 p1:11 themselves:1 multi:1 brain:1 inspired:1 relying:1 automatically:1 farhadi:1 provided:2 estimating:2 moreover:6 bounded:1 maximizes:1 project:2 notation:1 blaschko:1 what:2 mass:1 interpreted:2 klakow:1 unified:1 finding:3 teradata:1 interactive:1 exactly:1 classifier:2 wrong:1 unit:1 medical:7 grant:1 kurtz:1 before:2 positive:1 engineering:2 aggregating:1 modify:1 struggle:1 limit:1 specifier:1 textural:1 despite:1 encoding:1 analyzing:1 meet:1 solely:1 black:1 chose:1 acl:2 meinshausen:1 specifying:1 co:2 branson:1 jurafsky:1 shamma:1 range:1 graduate:2 ismb:1 acknowledgment:1 glioblastoma:1 procedure:1 shin:1 area:12 empirical:1 yan:1 significantly:7 vedaldi:1 road:1 cannot:1 unlabeled:6 selection:3 operator:1 ast:8 impossible:2 writing:1 bill:1 map:1 deterministic:1 roth:1 maximizing:1 primitive:76 independently:3 focused:1 chrismre:1 estimator:1 rule:2 borrow:1 varma:1 his:2 classic:1 biking:1 limiting:1 annals:2 construction:2 suppose:1 heavily:1 user:14 programming:5 us:5 agreement:1 element:1 dawid:1 expensive:1 satisfying:1 utilized:1 recognition:4 curated:1 ising:1 labeled:13 database:3 binning:1 observed:3 preprint:4 electrical:1 capture:5 wang:1 sun:1 indep:2 agency:1 complexity:12 bryanhe:1 trained:2 weakly:3 depend:2 segment:2 grateful:1 laptev:1 creation:1 localization:1 darpa:5 joint:3 various:9 represented:5 train:3 describe:4 query:1 labeling:1 crowd:1 encoded:8 stanford:5 heuristic:58 valued:1 film:1 otherwise:1 annotating:1 ability:1 cov:1 statistic:4 radiology:3 noisy:7 afnlp:1 differentiate:1 advantage:3 sen:1 hindering:1 interaction:1 product:1 relevant:4 combining:4 deformable:1 description:1 validate:2 bao:1 exploiting:1 produce:1 generating:3 incremental:1 converges:1 object:7 help:2 augmenting:1 pose:1 ij:8 dep:7 sa:3 strong:2 sonn:1 p2:5 c:1 predicted:1 come:2 implies:1 sim:10 snow:1 correct:2 attribute:6 annotated:1 stochastic:2 subsequently:1 aggressiveness:1 human:2 opinion:1 material:2 ddsm:3 require:1 government:3 assign:8 f1:5 niebles:1 proposition:1 biological:1 inspecting:1 extension:1 sufficiently:1 ground:5 exp:1 mapping:1 predict:2 claim:2 garrido:1 major:1 smallest:1 purpose:1 estimation:2 label:45 cheong:1 correctness:1 create:1 successfully:1 hope:1 joglekar:1 mit:1 always:2 modified:1 rather:3 zhou:2 office:1 encode:1 focus:2 dsp:16 naval:1 properly:2 improvement:12 likelihood:4 check:1 aka:1 haralick:1 contrast:1 detect:1 parameswaran:1 helpful:1 abstraction:2 dependent:2 typically:2 entire:1 spurious:1 relation:17 perona:1 reproduce:1 pont:1 going:1 provably:1 issue:1 among:22 classification:7 hon:1 denoted:1 development:1 proposes:1 fairly:1 mammographic:1 marginal:3 field:4 construct:2 once:2 having:1 beach:1 sampling:2 extraction:4 represents:1 look:3 denton:1 jones:2 icml:1 yu:2 future:1 simplex:1 np:1 prostate:1 report:1 few:1 composed:4 simultaneously:1 divergence:2 national:2 vmware:1 comprehensive:1 kalantidis:1 mintz:1 microsoft:1 ab:2 detection:4 insurance:1 alignment:1 truly:1 operated:1 copyright:1 held:1 radiologist:1 perimeter:17 accurate:2 oliver:1 edge:4 worker:1 necessary:5 experience:1 traversing:2 tree:3 plotted:1 girshick:1 minimal:2 instance:3 n000141310129:1 modeling:11 thereon:1 assignment:2 hlmann:1 subset:5 recognizing:2 johnson:1 dependency:94 st:2 person:1 international:6 probabilistic:5 systematic:1 lee:1 physic:1 dong:1 connecting:1 quickly:1 yao:1 reflect:1 accenture:1 emnlp:1 worse:1 creating:3 american:1 expert:1 zhao:1 return:10 li:1 aggressive:8 account:3 distribute:1 de:3 explicitly:2 mv:2 depends:2 ranking:1 stream:1 performed:1 bone:6 view:3 observer:1 hf:72 recover:1 aggregation:1 carlos:1 annotation:2 crowdsourced:2 accuracy:19 convolutional:3 characteristic:6 efficiently:3 rastogi:1 identify:6 generalize:1 weak:19 raw:2 accurately:4 computerized:1 cybernetics:1 mooney:1 ary:4 acc:8 explain:2 simultaneous:1 sharing:2 against:1 involved:1 proof:1 associated:2 recovers:2 static:26 gain:1 sampled:1 dataset:6 popular:3 ask:1 mitchell:1 recall:1 knowledge:7 conversation:1 infers:4 segmentation:20 routine:1 carefully:1 appears:1 afrl:1 higher:3 supervised:7 specify:3 improved:3 zisserman:1 evaluated:1 box:5 just:1 biomedical:1 hand:3 web:2 christopher:2 banerjee:1 lack:1 overlapping:1 assessment:2 defines:1 logistic:1 quality:1 grows:4 dge:1 usa:1 name:4 concept:2 true:9 contain:1 unbiased:1 requiring:1 moore:1 semantic:5 assistance:1 basketball:1 rooted:1 generalized:1 syntax:1 complete:1 demonstrate:1 performs:3 l1:1 image:26 meaning:2 harmonic:1 abstain:1 recently:4 dawn:1 nih:2 common:2 specialized:1 empirically:2 exponentially:5 volume:5 discussed:1 googlenet:2 association:7 he:3 refer:1 composition:7 significant:5 gibbs:3 automatic:1 trivially:1 pm:7 consistency:1 language:7 had:1 henry:1 gratefully:1 access:1 supervision:25 similarity:4 surface:1 base:3 add:3 showed:3 recent:1 optimizing:1 scenario:1 binary:9 success:2 onr:2 meeting:4 yi:2 molina:1 krishna:1 analyzes:1 additional:5 dai:1 mr:2 oquab:1 determine:2 paradigm:6 maximize:1 affiliate:1 sharma:1 semi:1 ii:2 multiple:4 mix:1 stephen:1 infer:7 aggarwal:1 segmented:1 match:3 hata:1 bach:7 long:2 retrieval:1 ravikumar:1 award:1 involving:1 scalable:1 regression:1 breast:1 vision:9 arxiv:9 represent:1 achieved:1 programmatic:3 zettlemoyer:1 fellowship:2 else:3 source:20 crucial:1 modality:1 goodman:1 operate:5 unlike:1 rest:3 p01:2 member:1 lafferty:1 jordan:1 integer:1 extracting:1 intermediate:1 split:1 enough:3 embeddings:1 automated:3 variety:1 affect:3 independence:1 gave:1 identified:1 lasso:2 p0m:2 hindered:1 okawa:1 selsam:1 supporter:1 bottleneck:1 whether:3 expression:3 defense:1 url:1 effort:1 akin:1 f:3 quasilinear:1 proceed:1 programmable:1 deep:4 detailed:1 karpathy:1 amount:5 bunescu:1 u54eb020405:1 simplest:1 generate:4 mai:1 outperform:1 exist:3 http:1 nsf:2 schapire:1 governmental:1 correctly:1 bryan:1 write:2 discrete:2 dasgupta:1 group:1 key:1 deepdive:1 iter:1 threshold:1 demonstrating:1 blum:1 drawn:1 pj:5 utilize:1 n66001:1 imaging:2 graph:12 asymptotically:1 icde:1 sum:1 run:2 palkar:1 you:1 family:1 reasonable:1 wu:3 p3:4 endorsement:1 decision:1 appendix:6 def:4 layer:1 guaranteed:1 fan:1 sawyer:1 encountered:2 annual:5 activity:1 sato:1 strength:1 occur:3 scanned:1 precisely:1 toshiba:1 fei:2 riedel:1 encodes:1 weld:1 generates:1 aspect:1 kumar:1 statically:2 relatively:3 skene:1 cbis:1 combination:2 craven:1 across:3 jolesz:1 em:2 joseph:1 making:3 supervise:2 iccv:1 pipeline:4 computationally:3 discus:1 turn:1 malignant:1 know:1 drs:1 studying:1 available:1 operation:1 apply:1 observe:2 hierarchical:1 balsubramani:1 spectral:1 robustly:1 save:1 alternative:1 original:1 assumes:1 running:5 include:3 ensure:3 linguistics:7 coral:54 build:2 especially:1 hahn:1 implied:1 hoffmann:1 traditional:5 gradient:3 unable:1 thank:1 majority:1 topic:1 collected:1 assuming:2 code:5 alfonseca:1 modeled:1 relationship:4 boxsup:1 ratio:9 minimizing:1 difficult:1 setup:3 statement:2 negative:1 proper:2 policy:1 perform:1 imbalance:1 datasets:1 benchmark:1 finite:1 acknowledge:2 descent:2 beat:2 relational:2 looking:4 hinton:1 rn:1 dsps:26 intensity:13 inferred:4 rating:1 introduced:1 required:4 specified:4 sivic:1 learned:2 registered:1 herein:1 kang:1 boost:1 nip:1 address:1 able:4 beyond:1 usually:1 below:1 pattern:5 program:1 including:2 reliable:1 video:6 wainwright:1 overlap:1 natural:3 rely:3 regularized:2 advanced:1 nth:1 representing:1 zhu:1 technology:1 identifies:2 vggnet:2 reprint:1 extract:4 health:1 text:4 prior:1 understanding:1 review:1 discovery:1 relative:4 freund:2 fully:5 expect:1 kikinis:1 generation:1 zaharia:1 querying:1 var:1 ghanem:1 foundation:3 degree:9 ontological:1 gather:1 sufficient:1 kravitz:1 rubin:4 systematically:1 share:7 pi:6 endowment:1 cancer:1 supported:1 free:1 formal:1 allow:1 pseudolikelihood:1 divvala:1 institute:1 wide:1 shanmugam:1 benefit:1 feedback:1 xia:1 world:3 valid:1 rich:1 genome:5 adopts:1 author:2 transaction:1 nov:1 implicitly:2 belongie:1 consuming:1 discriminative:12 continuous:1 latent:7 table:6 additionally:3 learn:23 nature:1 ca:1 correlated:1 career:1 obtaining:2 improving:2 bottou:1 european:1 complex:8 necessarily:2 constructing:1 domain:18 official:1 did:1 pk:2 dense:1 bounding:5 motivation:1 ling:1 allowed:1 he2:1 xu:1 augmented:1 intel:2 precision:1 inferring:3 exponential:1 marti:1 third:1 late:1 learns:7 rez:1 down:1 mammogram:3 beaulieu:2 specific:13 virtue:3 grouping:1 essential:2 incorporating:1 intractable:1 exists:2 false:2 n000141210041:1 corr:1 texture:2 magnitude:2 notwithstanding:1 subtree:2 dissimilarity:1 chen:3 authorized:1 led:1 garcia:1 simply:3 explore:3 visual:7 expressed:2 contained:1 recommendation:1 applies:1 springer:1 nested:1 truth:5 extracted:1 acm:2 groth:1 viewed:1 goal:2 invalid:1 shared:5 man:1 content:2 experimentally:2 feasible:2 specifically:5 nakagawa:1 reducing:2 operates:1 sampler:1 redmon:1 miss:1 tumor:14 total:2 experimental:1 vote:1 meaningful:1 rarely:1 formally:3 support:1 incorporate:1 scratch:1 crowdsourcing:2 |
6,222 | 6,629 | Structured Embedding Models for Grouped Data
Maja Rudolph
Columbia Univ.
[email protected]
Francisco Ruiz
Univ. of Cambridge
Columbia Univ.
Susan Athey
Stanford Univ.
David Blei
Columbia Univ.
Abstract
Word embeddings are a powerful approach for analyzing language, and exponential
family embeddings (EFE) extend them to other types of data. Here we develop
structured exponential family embeddings (S - EFE), a method for discovering
embeddings that vary across related groups of data. We study how the word
usage of U.S. Congressional speeches varies across states and party affiliation,
how words are used differently across sections of the ArXiv, and how the copurchase patterns of groceries can vary across seasons. Key to the success of our
method is that the groups share statistical information. We develop two sharing
strategies: hierarchical modeling and amortization. We demonstrate the benefits
of this approach in empirical studies of speeches, abstracts, and shopping baskets.
We show how S - EFE enables group-specific interpretation of word usage, and
outperforms EFE in predicting held-out data.
1
Introduction
Word embeddings (Bengio et al., 2003; Mikolov et al., 2013d,c,a; Pennington et al., 2014; Levy
and Goldberg, 2014; Arora et al., 2015) are unsupervised learning methods for capturing latent
semantic structure in language. Word embedding methods analyze text data to learn distributed
representations of the vocabulary that capture its co-occurrence statistics. These representations are
useful for reasoning about word usage and meaning (Harris, 1954; Rumelhart et al., 1986). Word
embeddings have also been extended to data beyond text (Barkan and Koenigstein, 2016; Rudolph
et al., 2016), such as items in a grocery store or neurons in the brain. Exponential family embeddings
(EFE) is a probabilistic perspective on embeddings that encompasses many existing methods and
opens the door to bringing expressive probabilistic modeling (Bishop, 2006; Murphy, 2012) to the
problem of learning distributed representations.
We develop structured exponential family embeddings (S - EFE), an extension of EFE for studying
how embeddings can vary across groups of related data. We will study several examples: in U.S.
Congressional speeches, word usage can vary across states or party affiliations; in scientific literature,
the usage patterns of technical terms can vary across fields; in supermarket shopping data, co-purchase
patterns of items can vary across seasons of the year. We will see that S - EFE discovers a per-group
embedding representation of objects. While the na?ve approach of fitting an individual embedding
model for each group would typically suffer from lack of data?especially in groups for which fewer
observations are available?we develop two methods that can share information across groups.
Figure 1a illustrates the kind of variation that we can capture. We fit an S - EFE to ArXiv abstracts
grouped into different sections, such as computer science (cs), quantitative finance (q-fin), and
nonlinear sciences (nlin). S- EFE results in a per-section embedding of each term in the vocabulary.
Using the fitted embeddings, we illustrate similar words to the word INTELLIGENCE. We can see
that how INTELLIGENCE is used varies by field: in computer science the most similar words include
ARTIFICIAL and AI ; in finance, similar words include ABILITIES and CONSCIOUSNESS .
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
(s)
?v
?v
V
X (s)
S
hierarchical:
amortized:
(a) S - EFE uncover variations in the usage of the word INTELLIGENCE.
(s)
(0)
?v ? N (?v , ??2 I)
(s)
(0)
?v = fs (?v )
(b) Graphical repres. of S - EFE.
Figure 1: (a) INTELLIGENCE is used differently across the ArXiv sections. Words with the closest
embedding to the query are listed for 5 sections. (The embeddings were obtained by fitting an
amortized S - EFE.) The method automatically orders the sections along the horizontal axis by their
similarity in the usage of INTELLIGENCE. See Section 3 additional for details. (b) Graphical
(s)
representation of S - EFE for data in S categories. The embedding vectors ?v are specific to each
group, and the context vectors ?v are shared across all categories.
In more detail, embedding methods posit two representation vectors for each term in the vocabulary;
an embedding vector and a context vector. (We use the language of text for concreteness; as we
mentioned, EFE extend to other types of data.) The idea is that the conditional probability of each
observed word depends on the interaction between the embedding vector and the context vectors of
the surrounding words. In S - EFE, we posit a separate set of embedding vectors for each group but a
shared set of context vectors; this ensures that the embedding vectors are in the same space.
We propose two methods to share statistical strength among the embedding vectors. The first
approach is based on hierarchical modeling (Gelman et al., 2003), which assumes that the groupspecific embedding representations are tied through a global embedding. The second approach is
based on amortization (Dayan et al., 1995; Gershman and Goodman, 2014), which considers that the
individual embeddings are the output of a deterministic function of a global embedding representation.
We use stochastic optimization to fit large data sets.
Our work relates closely to two threads of research in the embedding literature. One is embedding
methods that study how language evolves over time (Kim et al., 2014; Kulkarni et al., 2015; Hamilton
et al., 2016; Rudolph and Blei, 2017; Bamler and Mandt, 2017; Yao et al., 2017). Time can be thought
of as a type of ?group?, though with evolutionary structure that we do not consider. The second
thread is multilingual embeddings (Klementiev et al., 2012; Mikolov et al., 2013b; Ammar et al.,
2016; Zou et al., 2013); our approach is different in that most words appear in all groups and we are
interested in the variations of the embeddings across those groups.
Our contributions are thus as follows. We introduce the S - EFE model, extending EFE to grouped
data. We present two techniques to share statistical strength among the embedding vectors, one based
on hierarchical modeling and one based on amortization. We carry out a thorough experimental study
on two text databases, ArXiv papers by section and U.S. Congressional speeches by home state and
political party. Using Poisson embeddings, we study market basket data from a large grocery store,
grouped by season. On all three data sets, S - EFE outperforms EFE in terms of held-out log-likelihood.
Qualitatively, we demonstrate how S - EFE discovers which words are used most differently across U.S.
states and political parties, and show how word usage changes in different ArXiv disciplines.
2
Model Description
In this section, we develop structured exponential family embeddings (S - EFE), a model that builds
on exponential family embeddings (EFE) (Rudolph et al., 2016) to capture semantic variations across
groups of data. In embedding models, we represent each object (e.g., a word in text, or an item in
shopping data) using two sets of vectors, an embedding vector and a context vector. In this paper, we
2
are interested in how the embeddings vary across groups of data, and for each object we want to learn
a separate embedding vector for each group. Having a separate embedding for each group allows
us to study how the usage of a word like INTELLIGENCE varies across categories of the ArXiv, or
which words are used most differently by U.S. Senators depending on which state they are from and
whether they are Democrats or Republicans.
The S - EFE model extends EFE to grouped data, by having the embedding vectors be specific for each
group, while sharing the context vectors across all groups. We review the EFE model in Section 2.1.
We then formalize the idea of sharing the context vectors in Section 2.2, where we present two
approaches to build a hierarchical structure over the group-specific embeddings.
2.1
Background: Exponential Family Embeddings
In exponential family embeddings, we have a collection of objects, and our goal is to learn a vector
representation of these objects based on their co-occurrence patterns.
Let us consider a dataset represented as a (typically sparse) matrix X, where columns are datapoints
and rows are objects. For example, in text, each column corresponds to a location in the text, and
each entry xvi is a binary variable that indicates whether word v appears at location i.
In EFE, we represent each object v with two sets of vectors, embeddings vectors ?v [i] and context
vectors ?v [i], and we posit a probability distribution of data entries xvi in which these vectors interact.
The definition of the EFE model requires three ingredients: a context, a conditional exponential
family, and a parameter sharing structure. We next describe these three components.
Exponential family embeddings learn the vector representation of objects based on the conditional
probability of each observation, conditioned on the observations in its context. The context cvi gives
the indices of the observations that appear in the conditional probability distribution of xvi . The
definition of the context varies across applications. In text, it corresponds to the set of words in a
fixed-size window centered at location i.
Given the context cvi and the corresponding observations xcvi indexed by cvi , the distribution for xvi
is in the exponential family,
xvi | xcvi ? ExpFam (t(xvi ), ?v (xcvi )) ,
(1)
with sufficient statistics t(xvi ) and natural parameter ?v (xcvi ). The parameter vectors interact in the
conditional probability distributions of each observation xvi as follows. The embedding vectors ?v [i]
and the context vectors ?v [i] are combined to form the natural parameter,
?
?
X
?v (xcvi ) = g ??v [i]>
?v0 [i0 ]xv0 i0 ? ,
(2)
(v 0 ,i0 )?cvi
where g(?) is the link function. Exponential family embeddings can be understood as a bank of
generalized linear models (GLMs). The context vectors are combined to give the covariates, and the
?regression coefficients? are the embedding vectors. In Eq. 2, the link function g(?) plays the same
role as in GLMs and is a modeling choice. We use the identity link function.
The third ingredient of the EFE model is the parameter sharing structure, which indicates how the
embedding vectors are shared across observations. In the standard EFE model, we use ?v [i] ? ?v and
?v [i] ? ?v for all columns of X. That is, each unique object v has a shared representation across all
instances.
The objective function. In EFE, we maximize the objective function, which is given by the sum of
the log-conditional likelihoods in Eq. 1. In addition, we add an `2 -regularization term (we use the
notation of the log Gaussian pdf) over the embedding and context vectors, yielding
X
L = log p(?) + log p(?) +
log p xvi xcvi ; ?, ? ,
(3)
v,i
Note that maximizing the regularized conditional likelihood is not equivalent to maximum a posteriori.
Rather, it is similar to maximization of the pseudo-likelihood in conditionally specified models
(Arnold et al., 2001; Rudolph et al., 2016).
3
2.2
Structured Exponential Family Embeddings
Here, we describe the S - EFE model for grouped data. In text, some examples of grouped data are
Congressional speeches grouped into political parties or scientific documents grouped by discipline.
Our goal is to learn group-specific embeddings from data partitioned into S groups, i.e., each instance
i is associated with a group si ? {1, . . . , S}. The S - EFE model extends EFE to learn a separate set
of embedding vectors for each group.
To build the S - EFE model, we impose a particular parameter sharing structure over the set of
embedding and context vectors. We posit a structured model in which the context vectors are shared
across groups, i.e., ?v [i] ? ?v (as in the standard EFE model), but the embedding vectors are only
(s )
(s)
shared at the group level, i.e., for an observation i belonging to group si , ?v [i] ? ?v i . Here, ?v
denotes the embedding vector corresponding to group s. We show a graphical representation of the
S - EFE in Figure 1b.
Sharing the context vectors ?v has two advantages. First, the shared structure reduces the number of
parameters, while the resulting S - EFE model is still flexible to capture how differently words are
(s)
used across different groups, as ?v is allowed to vary.1 Second, it has the important effect of uniting
(s)
all embedding parameters in the same space, as the group-specific vectors ?v need to agree with the
components of ?v . While one could learn a separate embedding model for each group, as has been
done for text grouped into time slices (Kim et al., 2014; Kulkarni et al., 2015; Hamilton et al., 2016),
this approach would require ad-hoc postprocessing steps to align the embeddings.2
When there are S groups, the S - EFE model has S times as many embedding vectors than the standard
embedding model. This may complicate inferences about the group-specific vectors, especially for
groups with less data. Additionally, an object v may appear with very low frequency in a particular
group. Thus, the na?ve approach for building the S - EFE model without additional structure may be
detrimental for the quality of the embeddings, especially for small-sized groups. To address this
(s)
problem, we propose two different methods to tie the individual ?v together, sharing statistical
strength among them. The first approach consists in a hierarchical embedding structure. The second
approach is based on amortization. In both methods, we introduce a set of global embedding vectors
(0)
(s)
(0)
?v , and impose a particular structure to generate ?v from ?v .
Hierarchical embedding structure. Here, we impose a hierarchical structure that allows sharing
(s)
(0)
statistical strength among the per-group variables. For that, we assume that each ?v ? N (?v , ??2 I),
2
where ?? is a fixed hyperparameter. Thus, we replace the EFE objective function in Eq. 3 with
X
X
Lhier = log p(?) + log p(?(0) ) +
log p(?(s) | ?(0) ) +
log p xvi xcvi ; ?, ? .
(4)
s
v,i
(0)
where the `2 -regularization term now applies only on ?v and the global vectors ?v .
(0)
(s)
Fitting the hierarchical model involves maximizing Eq. 4 with respect to ?v , ?v , and ?v . We
note that we have not reduced the number of parameters to be inferred; rather, we tie them together
through a common prior distribution. We use stochastic gradient ascent to maximize Eq. 4.
Amortization. The idea of amortization has been applied in the literature to develop amortized
inference algorithms (Dayan et al., 1995; Gershman and Goodman, 2014). The main insight behind
amortization is to reuse inferences about past experiences when presented with a new task, leveraging
the accumulated knowledge to quickly solve the new problem. Here, we use amortization to control
(s)
the number of parameters of the S - EFE model. In particular, we set the per-group embeddings ?v to
(s)
(0)
be the output of a deterministic function of the global embedding vectors, ?v = fs (?v ). We use a
different function fs (?) for each group s, and we parameterize them using neural networks, similarly
to other works on amortized inference (Korattikara et al., 2015; Kingma and Welling, 2014; Rezende
et al., 2014; Mnih and Gregor, 2014). Unlike standard uses of amortized inference, in S - EFE the
(s)
1
Alternatively, we could share the embedding vectors ?v and have group-specific context vectors ?v . We
did not explore that avenue and leave it for future work.
2
Another potential advantage of the proposed parameter sharing structure is that, when the context vectors
are held fixed, the resulting objective function is convex, by the convexity properties of exponential families
(Wainwright and Jordan, 2008).
4
input to the functions fs (?) is unobserved and must be estimated together with the parameters of the
functions fs (?).
Depending on the architecture of the neural networks, the amortization can significantly reduce the
number of parameters in the model (as compared to the non-amortized model), while still having
the flexibility to model different embedding vectors for each group. The number of parameters in
the S - EFE model is KL(S + 1), where S is the number of groups, K is the dimensionality of the
embedding vectors, and L is the number of objects (e.g., the vocabulary size). With amortization,
we reduce the number of parameters to 2KL + SP , where P is the number of parameters of the
neural network. Since typically L P , this corresponds to a significant reduction in the number of
parameters, even when P scales linearly with K.
In the amortized S - EFE model, we need to introduce a new set of parameters ?(s) ? RP for each
group s, corresponding to the neural network parameters. Given these, the group-specific embedding
(s)
vectors ?v are obtained as
(0)
(0)
(s)
?(s)
).
v = fs (?v ) = f (?v ; ?
(5)
We compare two architectures for the function fs (?): fully connected feed-forward neural networks
and residual networks (He et al., 2016). For both, we consider one hidden layer with H units. Hence,
the network parameters ?(s) are two weight matrices,
(s)
?(s) = {W1
(s)
? RH?K , W2
? RK?H },
(6)
(0)
?v ,
i.e., P = 2KH parameters. The neural network takes as input the global embedding vector
(s)
(s)
and it outputs the group-specific embedding vectors ?v . The mathematical expression for ?v for a
feed-forward neural network and a residual network is respectively given by
(s)
(s) (0)
(0)
(s)
?(s)
=
f
(?
;
?
)
=
W
tanh
W
?
,
(7)
ffnet
v
v
v
2
1
(s)
(s) (0)
(0)
(s)
?(s)
) = ?(0)
,
(8)
v = fresnet (?v ; ?
v + W2 tanh W1 ?v
where we have considered the hyperbolic tangent nonlinearity. The main difference between both
network architectures is that the residual network focuses on modeling how the group-specific
(s)
(0)
embedding vectors ?v differ from the global vectors ?v . That is, if all weights were set to 0, the
(0)
feed-forward network would output 0, while the residual network would output the global vector ?v
for all groups.
The objective function under amortization is given by
X
Lamortiz = log p(?) + log p(?(0) ) +
log p xvi xcvi ; ?, ?(0) , ? .
(9)
v,i
(0)
We maximize this objective with respect to ?v , ?v , and ?(s) using stochastic gradient ascent. We
implement the hierarchical and amortized S - EFE models in TensorFlow (Abadi et al., 2015), which
allows us to leverage automatic differentiation.3
Example: structured Bernoulli embeddings for grouped text data. Here, we consider a set
of documents broken down into groups, such as political affiliations or scientific disciplines. We
can represent the data as a binary matrix X and a set of group indicators si . Since only one word
can appear in a certain position i, the matrix X contains one non-zero element per column. In
embedding models, we ignore this one-hot constraint for computational efficiency, and consider
that the observations are generated following a set of conditional Bernoulli distributions (Mikolov
et al., 2013c; Rudolph et al., 2016). Given that most of the entries in X are zero, embedding models
typically downweigh the contribution of the zeros to the objective function. Mikolov et al. (2013c)
use negative sampling, which consists in randomly choosing a subset of the zero observations. This
corresponds to a biased estimate of the gradient in a Bernoulli exponential family embedding model
(Rudolph et al., 2016).
The context cvi is given at each position i by the set of surrounding words in the document, according
to a fixed-size window.
3
Code is available at https://github.com/mariru/structured_embeddings
5
ArXiv abstracts
Senate speeches
Shopping data
data
text
text
counts
embedding of
15k terms
15k terms
5.5k items
groups
19
83
12
grouped by
subject areas
home state/party
months
size
15M words
20M words
0.5M trips
Table 1: Group structure and size of the three corpora analyzed in Section 3.
Example: structured Poisson embeddings for grouped shopping data. EFE and S - EFE extend to
applications beyond text and we use S - EFE to model supermarket purchases broken down by month.
For each market basket i, we have access to the month si in which that shopping trip happened. Now,
the rows of the data matrix X index items, while columns index shopping trips. Each element xvi
denotes the number of units of item v purchased at trip i. Unlike text, each column of X may contain
more than one non-zero element. The context cvi corresponds to the set of items purchased in trip i,
excluding v.
In this case, we use the Poisson conditional distribution, which is more appropriate for count data. In
Poisson S - EFE, we also downweigh the contribution of the zeros in the objective function, which
provides better results because it allows the inference to focus on the positive signal of the actual
purchases (Rudolph et al., 2016; Mikolov et al., 2013c).
3
Empirical Study
In this section, we describe the experimental study. We fit the S - EFE model on three datasets and
compare it against the EFE (Rudolph et al., 2016). Our quantitative results show that sharing the
context vectors provides better results, and that amortization and hierarchical structure give further
improvements.
Data. We apply the S - EFE on three datasets: ArXiv papers, U.S. Senate speeches, and purchases on
supermarket grocery shopping data. We describe these datasets below, and we provide a summary of
the datasets in Table 1.
ArXiv papers: This dataset contains the abstracts of papers published on the ArXiv under the 19
different tags between April 2007 and June 2015. We treat each tag as a group and fit S - EFE with the
goal of uncovering which words have the strongest shift in usage. We split the abstracts into training,
validation, and test sets, with proportions of 80%, 10%, and 10%, respectively.
Senate speeches: This dataset contains U.S. Senate speeches from 1994 to mid 2009. In contrast to
the ArXiv collection, it is a transcript of spoken language. We group the data into state of origin of
the speaker and his or her party affiliation. Only affiliations with the Republican and Democratic
Party are considered. As a result, there are 83 groups (Republicans from Alabama, Democrats from
Alabama, Republicans from Arkansas, etc.). Some of the state/party combinations are not available
in the data, as some of the 50 states have only had Senators with the same party affiliation. We split
the speeches into training (80%), validation (10%), and testing (10%).
Grocery shopping data: This dataset contains the purchases of 3, 206 customers. The data covers a
period of 97 weeks. After removing low-frequency items, the data contains 5, 590 unique items at the
UPC (Universal Product Code) level. We split the data into a training, test, and validation sets, with
proportions of 90%, 5%, and 5%, respectively. The training data contains 515, 867 shopping trips
and 5, 370, 623 purchases in total.
For the text corpora, we fix the vocabulary to the 15k most frequent terms and remove all words
that are not in the vocabulary.
Following Mikolov et al. (2013c), we additionally remove each word
p
with probability 1 ? 10?5 /fv , where fv is the word frequency. This downsamples especially the
frequent words and speeds up training. (Sizes reported in Table 1 are the number of words remaining
after preprocessing.)
Models. Our goal is to fit the S - EFE model on these datasets. For the text data, we use the Bernoulli
distribution as the conditional exponential family, while for the shopping data we use the Poisson
distribution, which is more appropriate for count data.
6
On each dataset, we compare four approaches based on S - EFE with two EFE (Rudolph et al., 2016)
baselines. All are fit using stochastic gradient descent (SGD) (Robbins and Monro, 1951). In
particular, we compare the following methods:
?
?
?
?
?
?
A global EFE model, which cannot capture group structure.
Separate EFE models, fitted independently on each group.
(this paper) S - EFE without hierarchical structure or amortization.
(this paper) S - EFE with hierarchical group structure.
(this paper) S - EFE, amortized with a feed-forward neural network (Eq. 7).
(this paper) S - EFE, amortized using a residual network (Eq. 8).
Experimental setup and hyperparameters. For text we set the dimension of the embeddings to
K = 100, the number of hidden units to H = 25, and we experiment with two context sizes, 2 and 8.4
In the shopping data, we use K = 50 and H = 20, and we randomly truncate the context of baskets
larger than 20 to reduce their size to 20. For both methods, we use 20 negative samples.
For all methods, we subsample minibatches of data in the same manner. Each minibatch contains
subsampled observations from all groups and each group is subsampled proportionally to its size.
For text, the words subsampled from within a group are consecutive, and for shopping data the
observations are sampled at the shopping trip level. This sampling scheme reduces the bias from
imbalanced group sizes. For text, we use a minibatch size of N/10000, where N is the size of the
corpus, and we run 5 passes over the data; for the shopping data we use N/100 and run 50 passes.
We use the default learning rate setting of TensorFlow for Adam5 (Kingma and Ba, 2015).
We use the standard initialization schemes for the
network parameters. The weights are
? neural
?
drawn from a uniform distribution bounded at ? 6/ K + H (Glorot and Bengio, 2010). For the
embeddings, we try 3 initialization schemes and choose the best one based on validation error. In
particular, these schemes are: (1) all embeddings are drawn from the Gaussian prior implied by the
regularizer; (2) the embeddings are initialized from a global embedding; (3) the context vectors are
initialized from a global embedding and held constant, while the embeddings vectors are drawn
randomly from the prior. Finally, for each method we choose the regularization variance from the set
{100, 10, 1, 0.1}, also based on validation error.
Runtime. We implemented all methods in Tensorflow. On the Senate speeches, the runtime of
S - EFE is 4.3 times slower than the runtime of global EFE, hierarchical EFE is 4.6 times slower than
the runtime of global EFE, and amortized S - EFE is 3.3 times slower than the runtime of global EFE.
(The Senate speeches have the most groups and hence the largest difference in runtime between
methods.)
Evaluation metric. We evaluate the fits by held-out pseudo (log-)likelihood. For each model,
we compute the test pseudo log-likelihood, according to the exponential family distribution used
(Bernoulli or Poisson). For each test entry, a better model will assign higher probability to the
observed word or item, and lower probability to the negative samples. This is a fair metric because the
competing methods all produce conditional likelihoods from the same exponential family.6 To make
results comparable, we train and evaluate all methods with the same number of negative samples (20).
The reported held out likelihoods give equal weight to the positive and negative samples.
Quantitative results. We show the test pseudo log-likelihood of all methods in Table 2 and report that
our method outperforms the baseline in all experiments. We find that S - EFE with either hierarchical
structure or amortization outperforms the competing methods based on standard EFE highlighted
in bold. This is because the global EFE ignores per-group variations, whereas the separate EFE
cannot share information across groups. The results of the global EFE baseline are better than fitting
separate EFE (the other baseline), but unlike the other methods the global EFE cannot be used for the
exploratory analysis of variations across groups. Our results show that using a hierarchical S - EFE
is always better than using the simple S - EFE model or fitting a separate EFE on each group. The
hierarchical structure helps, especially for the Senate speeches, where the data is divided into many
4
To save space we report results for context size 8 only. Context size 2 shows the same relative performance.
Adam needs to track a history of the gradients for each parameter that is being optimized. One advantage
from reducing the number of parameters with amortization is that it results in a reduced computational overhead
for Adam (as well as for other adaptive stepsize schedules).
6
Since we hold out chunks of consecutive words usually both a word and its context are held out. For all
methods we have to use the words in the context to compute the conditional likelihoods.
5
7
ArXiv papers
Senate speeches
Shopping data
?2.239 ? 0.002
?0.772 ? 0.000
?2.915 ? 0.004
?0.807 ? 0.002
S - EFE
?2.645 ? 0.002
?0.770 ? 0.001
S - EFE (hierarchical)
? 2.217 ? 0.001
?0.767 ? 0.000
S - EFE (amortiz+feedf)
?2.484 ? 0.002
?0.774 ? 0.000
S - EFE (amortiz+resnet)
?2.249 ? 0.002
?0.762 ? 0.000
Table 2: Test log-likelihood on the three considered datasets. S - EFE consistently achieves the
highest held-out likelihood. The competing methods are the global EFE, which can not capture group
variations, and the separate EFE, which cannot share information across groups.
Global EFE (Rudolph et al., 2016)
Separated EFE (Rudolph et al., 2016)
?2.176 ? 0.005
?2.500 ? 0.012
?2.287 ? 0.007
?2.170 ? 0.003
?2.153 ? 0.004
?2.120 ? 0.004
groups. Among the amortized S - EFE models we developed, at least amortization with residual
networks outperforms the base S - EFE. The advantage of residual networks over feed-forward neural
networks is consistent with the results reported by (He et al., 2016).
While both hierarchical S - EFE and amortized S - EFE share information about the embedding of a
(0)
particular word across groups (through the global embedding ?v ), amortization additionally ties
the embeddings of all words within a group (through learning the neural network of that group). We
hypothesize that for the Senate speeches, which are split into many groups, this is a strong modeling
constraint, while it helps for all other experiments.
Structured embeddings reveal a spectrum of word usage. We have motivated S - EFE with the
example that the usage of INTELLIGENCE varies by ArXiv category (Figure 1a). We now explain
how for each term the per-group embeddings place the groups on a spectrum. For a specific term
(s)
v we take its embeddings vectors {?v } for all groups s, and project them onto a one-dimensional
space using the first component of principal component analysis (PCA). This is a one-dimensional
summary of how close the embeddings of v are across groups. Such comparison is posible because
the S - EFE shares the context vectors, which grounds the embedding vectors in a joint space.
The spectrum for the word INTELLIGENCE along its first principal component is the horizontal
axis in Figure 1a. The dots are the projections of the group-specific embeddings for that word.
(The embeddings come from a fitted S - EFE with feed-forward amortization.) We can see that in an
unsupervised manner, the method has placed together groups related to physics on one end on the
spectrum, while computer science, statistics and math are on the other end of the spectrum.
To give additional intuition of what the usage of INTELLIGENCE is at different locations on the
spectrum, we have listed the 8 most similar words for the groups computer science (cs), quantitative
finance (q-fin), math (math), statistics (stat), and nonlinear sciences (nlin). Word similarities are
computed using cosine distance in the embedding space. Eventhough their embeddings are relatively
close to each other on the spectrum, the model has the flexibility to capture high variabilty in the lists
of similar words.
Exploring group variations with structured embeddings. The result of the S - EFE also allows us
to investigate which words have the highest deviation from their average usage for each group. For
example, in the Congressional speeches, there are many terms that we do not expect the Senators to
use differently (e.g., most stopwords). We might however want to ask a question like ?which words
do Republicans from Texas use most differently from other Senators?? By suggesting an answer,
our method can guide an exploratorydata analysis. For eachgroup s (state/party combination), we
PS
(s)
(t)
compute the top 3 words in argsortv ||?v ? S1 t=1 ?v || from within the top 1k words.
Table 3 shows a summary of our findings (the full table is in the Appendix). According to the S - EFE
(with residual network amortization), Republican Senators from Texas use BORDER and the phrase
OUR COUNTRY in different contexts than other Senators.
Some of these variations are probably influenced by term frequency, as we expect Democrats from
Washington to talk about WASHINGTON more frequently than other states. But we argue that our
method provides more insights than a frequency based analysis, as it is also sensitive to the context in
which a word appears. For example, WASHINGTON might in some groups be used more often in
8
TEXAS
border
our country
iraq
FLORIDA
medicaid
prescription
medicare
IOWA
bankruptcy
water
waste
agriculture
farmers
food
prescription
drug
drugs
WASHINGTON
washington
energy
oil
Table 3: List of the three most different words for different groups for the Congressional speeches.
S - EFE uncovers which words are used most differently by Republican Senators (red) and Democratic
Senators (blue) from different states. The complete table is in the Appendix.
the context of PRESIDENT and GEORGE, while in others it might appear in the context of DC and
CAPITAL, or it may refer to the state.
4
Discussion
We have presented several structured extensions of EFE for modeling grouped data. Hierarchical
S - EFE can capture variations in word usage across groups while sharing statistical strength between
them through a hierarchical prior. Amortization is an effective way to reduce the number of parameters
in the hierarchical model. The amortized S - EFE model leverages the expressive power of neural
networks to reduce the number of parameters, while still having the flexibility to capture variations
between the embeddings of each group. Below are practical guidelines for choosing a S - EFE.
How can I fit embeddings that vary across groups of data? To capture variations across groups,
never fit a separate embedding model for each group. We recommend at least sharing the context
vectors, as all the S - EFE models do. This ensures that the latent dimensions of the embeddings
are aligned across groups. In addition to sharing context vectors, we also recommend sharing
statistical strength between the embedding vectors. In this paper we have presented two ways to do
so, hierarchical modeling and amortization.
Should I use a hierarchical prior or amortization? The answer depends on how many groups the
data contain. In our experiments, the hierarchical S - EFE works better when there are many groups.
With less groups, the amortized S - EFE works better.
The advantage of the amortized S - EFE is that it has fewer parameters than the hierarchical model,
while still having the flexibility to capture across-group variations. The global embeddings in an
amortized S - EFE have two roles. They capture the semantic similarities of the words, and they also
serve as the input into the amortization networks. Thus, the global embeddings of words with similar
pattern of across-group variation need to be in regions of the embedding space that lead to similar
modifications by the amortization network. As the number of groups in the data increases, these two
roles become harder to balance. We hypothesize that this is why the amortized S - EFE has stronger
performance when there are fewer groups.
Should I use feed-forward or residual networks? To amortize a S - EFE we recommend residual
networks. They perform better than the feed-forward networks in all of our experiments. While the
feed-forward network has to output the entire meaning of a word in the group-specific embedding,
the residual network only needs the capacity to model how the group-specific embedding differs from
the global embedding.
Acknowledgements
We thank Elliott Ash and Suresh Naidu for the helpful discussions and for sharing the Senate speeches.
This work is supported by NSF IIS-1247664, ONR N00014-11-1-0651, DARPA PPAML FA8750-142-0009, DARPA SIMPLEX N66001-15-C-4032, the Alfred P. Sloan Foundation, and the John Simon
Guggenheim Foundation. Francisco J. R. Ruiz is supported by the EU H2020 programme (Marie
Sk?odowska-Curie grant agreement 706760).
9
References
Abadi, M., Agarwal, A., Barham, P., Brevdo, E., Chen, Z., Citro, C., Corrado, G., Davis, A., Dean, J., Devin, M.,
et al. (2015). TensorFlow: Large-scale machine learning on heterogeneous systems. Software available from
tensorflow.org.
Ammar, W., Mulcaire, G., Tsvetkov, Y., Lample, G., Dyer, C., and Smith, N. A. (2016). Massively multilingual
word embeddings. arXiv preprint arXiv:1602.01925.
Arnold, B. C., Castillo, E., Sarabia, J. M., et al. (2001). Conditionally specified distributions: an introduction
(with comments and a rejoinder by the authors). Statistical Science, 16(3):249?274.
Arora, S., Li, Y., Liang, Y., Ma, T., and Risteski, A. (2015). Rand-walk: A latent variable model approach to
word embeddings. arXiv preprint arXiv:1502.03520.
Bamler, R. and Mandt, S. (2017). Dynamic word embeddings. International Conference on Machine Learning.
Barkan, O. and Koenigstein, N. (2016). Item2vec: neural item embedding for collaborative filtering. In Machine
Learning for Signal Processing (MLSP), 2016 IEEE 26th International Workshop on, pages 1?6. IEEE.
Bengio, Y., Ducharme, R., Vincent, P., and Jauvin, C. (2003). A neural probabilistic language model. Journal of
machine learning research, 3(Feb):1137?1155.
Bishop, C. M. (2006). Machine learning and pattern recognition. Information Science and Statistics. Springer,
Heidelberg.
Dayan, P., Hinton, G. E., Neal, R. M., and Zemel, R. S. (1995). The Helmholtz machine. Neural Computation,
7(5):889?904.
Gelman, A., Carlin, J. B., Stern, H. S., and Rubin, D. B. (2003). Bayesian data analysis. Chapman and
Hall/CRC.
Gershman, S. J. and Goodman, N. D. (2014). Amortized inference in probabilistic reasoning. In Proceedings of
the Thirty-Sixth Annual Conference of the Cognitive Science Society.
Glorot, X. and Bengio, Y. (2010). Understanding the difficulty of training deep feedforward neural networks. In
Aistats, volume 9, pages 249?256.
Hamilton, W. L., Leskovec, J., and Jurafsky, D. (2016). Diachronic word embeddings reveal statistical laws of
semantic change. arXiv preprint arXiv:1605.09096.
Harris, Z. S. (1954). Distributional structure. Word, 10(2-3):146?162.
He, K., Zhang, X., Ren, S., and Sun, J. (2016). Deep residual learning for image recognition. In IEEE Conference
on Computer Vision and Pattern Recognition.
Kim, Y., Chiu, Y.-I., Hanaki, K., Hegde, D., and Petrov, S. (2014). Temporal analysis of language through neural
language models. arXiv preprint arXiv:1405.3515.
Kingma, D. P. and Ba, J. L. (2015). Adam: A method for stochastic optimization. In International Conference
on Learning Representations.
Kingma, D. P. and Welling, M. (2014). Auto-encoding variational Bayes. In International Conference on
Learning Representations.
Klementiev, A., Titov, I., and Bhattarai, B. (2012). Inducing crosslingual distributed representations of words.
Korattikara, A., Rathod, V., Murphy, K. P., and Welling, M. (2015). Bayesian dark knowledge. In Advances in
Neural Information Processing Systems.
Kulkarni, V., Al-Rfou, R., Perozzi, B., and Skiena, S. (2015). Statistically significant detection of linguistic
change. In Proceedings of the 24th International Conference on World Wide Web, pages 625?635. ACM.
Levy, O. and Goldberg, Y. (2014). Neural word embedding as implicit matrix factorization. In Neural Information
Processing Systems, pages 2177?2185.
Mikolov, T., Chen, K., Corrado, G., and Dean, J. (2013a). Efficient estimation of word representations in vector
space. ICLR Workshop Proceedings. arXiv:1301.3781.
Mikolov, T., Le, Q. V., and Sutskever, I. (2013b). Exploiting similarities among languages for machine translation.
arXiv preprint arXiv:1309.4168.
Mikolov, T., Sutskever, I., Chen, K., Corrado, G. S., and Dean, J. (2013c). Distributed representations of words
and phrases and their compositionality. In Neural Information Processing Systems, pages 3111?3119.
10
Mikolov, T., Yih, W.-T. a., and Zweig, G. (2013d). Linguistic regularities in continuous space word representations. In HLT-NAACL, pages 746?751.
Mnih, A. and Gregor, K. (2014). Neural variational inference and learning in belief networks. In International
Conference on Machine Learning.
Murphy, K. P. (2012). Machine learning: a probabilistic perspective. MIT press.
Pennington, J., Socher, R., and Manning, C. D. (2014). Glove: Global vectors for word representation. In
Conference on Empirical Methods on Natural Language Processing, volume 14, pages 1532?1543.
Rezende, D. J., Mohamed, S., and Wierstra, D. (2014). Stochastic backpropagation and approximate inference
in deep generative models. In International Conference on Machine Learning.
Robbins, H. and Monro, S. (1951). A stochastic approximation method. The annals of mathematical statistics,
pages 400?407.
Rudolph, M. and Blei, D. (2017). Dynamic bernoulli embeddings for language evolution. arXiv preprint at
arXiv:1703.08052.
Rudolph, M., Ruiz, F., Mandt, S., and Blei, D. (2016). Exponential family embeddings. In Advances in Neural
Information Processing Systems, pages 478?486.
Rumelhart, D. E., Hintont, G. E., and Williams, R. J. (1986). Learning representations by back-propagating
errors. Nature, 323:9.
Wainwright, M. J. and Jordan, M. I. (2008). Graphical models, exponential families, and variational inference.
Foundations and Trends in Machine Learning, 1(1?2):1?305.
Yao, Z., Sun, Y., Ding, W., Rao, N., and Xiong, H. (2017). Discovery of evolving semantics through dynamic
word embedding learning. arXiv preprint arXiv:1703.00607.
Zou, W. Y., Socher, R., Cer, D. M., and Manning, C. D. (2013). Bilingual word embeddings for phrase-based
machine translation. In EMNLP, pages 1393?1398.
11
| 6629 |@word proportion:2 stronger:1 open:1 uncovers:1 sgd:1 yih:1 harder:1 carry:1 reduction:1 contains:7 document:3 fa8750:1 outperforms:5 existing:1 past:1 com:1 si:4 must:1 john:1 devin:1 enables:1 remove:2 hypothesize:2 intelligence:9 discovering:1 fewer:3 item:11 generative:1 smith:1 blei:4 barkan:2 provides:3 math:3 location:4 org:1 zhang:1 stopwords:1 mathematical:2 along:2 wierstra:1 become:1 abadi:2 consists:2 fitting:5 overhead:1 manner:2 introduce:3 market:2 frequently:1 brain:1 automatically:1 food:1 actual:1 window:2 project:1 maja:2 notation:1 bounded:1 what:1 kind:1 developed:1 spoken:1 unobserved:1 finding:1 differentiation:1 pseudo:4 quantitative:4 thorough:1 temporal:1 finance:3 tie:3 runtime:6 klementiev:2 control:1 unit:3 farmer:1 grant:1 appear:5 hamilton:3 positive:2 understood:1 treat:1 encoding:1 analyzing:1 mandt:3 might:3 initialization:2 co:3 jurafsky:1 factorization:1 statistically:1 unique:2 practical:1 thirty:1 testing:1 implement:1 differs:1 backpropagation:1 suresh:1 area:1 empirical:3 universal:1 drug:2 thought:1 significantly:1 hyperbolic:1 projection:1 word:73 evolving:1 downsamples:1 cannot:4 onto:1 close:2 gelman:2 context:38 equivalent:1 deterministic:2 customer:1 dean:3 maximizing:2 hegde:1 williams:1 independently:1 convex:1 insight:2 datapoints:1 his:1 embedding:66 exploratory:2 variation:14 president:1 annals:1 play:1 goldberg:2 us:1 origin:1 agreement:1 amortized:19 rumelhart:2 element:3 recognition:3 helmholtz:1 trend:1 iraq:1 distributional:1 database:1 observed:2 role:3 preprint:7 ding:1 capture:12 parameterize:1 susan:1 region:1 ensures:2 connected:1 sun:2 eu:1 highest:2 mentioned:1 intuition:1 convexity:1 broken:2 covariates:1 dynamic:3 hintont:1 serve:1 efficiency:1 joint:1 darpa:2 differently:8 represented:1 ppaml:1 talk:1 regularizer:1 surrounding:2 train:1 univ:5 nlin:2 separated:1 describe:4 effective:1 artificial:1 query:1 zemel:1 choosing:2 stanford:1 solve:1 larger:1 ducharme:1 ability:1 statistic:6 highlighted:1 rudolph:14 hoc:1 advantage:5 propose:2 interaction:1 product:1 frequent:2 aligned:1 korattikara:2 flexibility:4 arkansas:1 description:1 kh:1 inducing:1 sutskever:2 exploiting:1 regularity:1 p:1 extending:1 produce:1 adam:3 leave:1 h2020:1 object:11 koenigstein:2 illustrate:1 develop:6 depending:2 propagating:1 stat:1 help:2 resnet:1 transcript:1 eq:7 strong:1 implemented:1 c:3 involves:1 come:1 differ:1 posit:4 closely:1 stochastic:7 centered:1 crc:1 require:1 assign:1 shopping:16 fix:1 extension:2 exploring:1 hold:1 cvi:6 considered:3 ground:1 hall:1 rfou:1 week:1 vary:9 consecutive:2 achieves:1 agriculture:1 estimation:1 tanh:2 sensitive:1 robbins:2 grouped:14 largest:1 mit:1 gaussian:2 always:1 rather:2 season:3 linguistic:2 rezende:2 focus:2 june:1 improvement:1 consistently:1 bernoulli:6 likelihood:12 indicates:2 political:4 contrast:1 kim:3 baseline:4 posteriori:1 inference:10 helpful:1 dayan:3 jauvin:1 i0:3 accumulated:1 typically:4 entire:1 hidden:2 her:1 interested:2 semantics:1 uncovering:1 among:6 flexible:1 grocery:5 field:2 equal:1 never:1 having:5 beach:1 sampling:2 washington:5 chapman:1 cer:1 unsupervised:2 athey:1 purchase:6 simplex:1 report:2 recommend:3 others:1 future:1 randomly:3 ve:2 individual:3 senator:8 murphy:3 subsampled:3 sarabia:1 skiena:1 detection:1 investigate:1 mnih:2 evaluation:1 analyzed:1 yielding:1 behind:1 held:8 expfam:1 experience:1 indexed:1 initialized:2 walk:1 leskovec:1 fitted:3 instance:2 column:6 modeling:9 rao:1 cover:1 maximization:1 phrase:3 deviation:1 entry:4 subset:1 uniform:1 reported:3 answer:2 varies:5 combined:2 chunk:1 st:1 international:7 probabilistic:5 physic:1 discipline:3 together:4 quickly:1 yao:2 na:2 w1:2 choose:2 emnlp:1 cognitive:1 li:1 suggesting:1 potential:1 bold:1 waste:1 coefficient:1 mlsp:1 sloan:1 depends:2 ad:1 bhattarai:1 try:1 analyze:1 red:1 bayes:1 simon:1 curie:1 monro:2 contribution:3 collaborative:1 odowska:1 variance:1 bayesian:2 vincent:1 ren:1 published:1 history:1 explain:1 strongest:1 influenced:1 sharing:16 basket:4 complicate:1 definition:2 sixth:1 against:1 petrov:1 energy:1 hlt:1 frequency:5 mohamed:1 associated:1 sampled:1 dataset:5 ask:1 knowledge:2 dimensionality:1 formalize:1 schedule:1 uncover:1 back:1 appears:2 feed:9 lample:1 higher:1 april:1 rand:1 done:1 though:1 implicit:1 eventhough:1 glms:2 horizontal:2 web:1 expressive:2 nonlinear:2 lack:1 minibatch:2 xv0:1 quality:1 reveal:2 scientific:3 oil:1 usa:1 usage:15 effect:1 building:1 contain:2 naacl:1 evolution:1 regularization:3 hence:2 consciousness:1 bankruptcy:1 semantic:4 neal:1 conditionally:2 davis:1 speaker:1 cosine:1 generalized:1 pdf:1 complete:1 demonstrate:2 reasoning:2 postprocessing:1 meaning:2 image:1 variational:3 discovers:2 common:1 volume:2 crosslingual:1 extend:3 interpretation:1 he:3 significant:2 refer:1 cambridge:1 ai:1 automatic:1 similarly:1 nonlinearity:1 language:11 had:1 dot:1 access:1 risteski:1 similarity:4 v0:1 etc:1 add:1 align:1 base:1 feb:1 closest:1 imbalanced:1 perspective:2 massively:1 store:2 certain:1 n00014:1 affiliation:6 success:1 binary:2 onr:1 additional:3 george:1 impose:3 maximize:3 period:1 corrado:3 signal:2 ii:1 relates:1 full:1 reduces:2 technical:1 long:1 zweig:1 divided:1 prescription:2 regression:1 heterogeneous:1 vision:1 metric:2 poisson:6 arxiv:28 represent:3 agarwal:1 background:1 want:2 addition:2 whereas:1 diachronic:1 country:2 goodman:3 w2:2 biased:1 unlike:3 bringing:1 perozzi:1 ascent:2 pass:2 subject:1 probably:1 comment:1 leveraging:1 jordan:2 leverage:2 door:1 feedforward:1 bengio:4 embeddings:58 congressional:6 split:4 fit:9 carlin:1 architecture:3 competing:3 reduce:5 idea:3 barham:1 avenue:1 texas:3 shift:1 thread:2 whether:2 expression:1 motivated:1 pca:1 reuse:1 suffer:1 f:7 speech:18 deep:3 useful:1 proportionally:1 listed:2 dark:1 mid:1 category:4 reduced:2 generate:1 http:1 nsf:1 happened:1 efe:109 estimated:1 per:7 track:1 blue:1 alfred:1 hyperparameter:1 group:100 key:1 four:1 drawn:3 capital:1 marie:1 n66001:1 concreteness:1 year:1 sum:1 run:2 powerful:1 extends:2 family:20 place:1 home:2 appendix:2 comparable:1 capturing:1 layer:1 annual:1 strength:6 constraint:2 software:1 tag:2 speed:1 mikolov:10 relatively:1 structured:11 according:3 truncate:1 combination:2 manning:2 guggenheim:1 belonging:1 across:33 partitioned:1 evolves:1 modification:1 s1:1 agree:1 count:3 dyer:1 end:2 studying:1 available:4 brevdo:1 apply:1 titov:1 hierarchical:26 appropriate:2 occurrence:2 stepsize:1 save:1 xiong:1 slower:3 rp:1 florida:1 assumes:1 denotes:2 include:2 remaining:1 top:2 graphical:4 especially:5 build:3 gregor:2 society:1 purchased:2 implied:1 objective:8 question:1 strategy:1 evolutionary:1 gradient:5 detrimental:1 iclr:1 distance:1 separate:11 link:3 thank:1 capacity:1 argue:1 considers:1 water:1 code:2 index:3 balance:1 liang:1 setup:1 negative:5 ba:2 guideline:1 stern:1 perform:1 neuron:1 observation:12 datasets:6 fin:2 descent:1 extended:1 excluding:1 hinton:1 dc:1 inferred:1 compositionality:1 david:1 specified:2 kl:2 trip:7 optimized:1 fv:2 alabama:2 xvi:12 tensorflow:5 kingma:4 nip:1 address:1 beyond:2 below:2 pattern:7 usually:1 democratic:2 encompasses:1 belief:1 wainwright:2 hot:1 power:1 natural:3 difficulty:1 regularized:1 predicting:1 indicator:1 residual:12 senate:10 medicare:1 scheme:4 github:1 republican:7 arora:2 axis:2 auto:1 columbia:4 supermarket:3 text:20 review:1 literature:3 rathod:1 prior:5 ammar:2 tangent:1 acknowledgement:1 relative:1 understanding:1 law:1 fully:1 expect:2 filtering:1 rejoinder:1 gershman:3 ingredient:2 validation:5 ash:1 iowa:1 foundation:3 sufficient:1 consistent:1 elliott:1 rubin:1 bank:1 share:9 amortization:24 translation:2 row:2 summary:3 placed:1 supported:2 bias:1 guide:1 arnold:2 wide:1 sparse:1 benefit:1 distributed:4 slice:1 dimension:2 vocabulary:6 default:1 world:1 ignores:1 forward:9 qualitatively:1 collection:2 preprocessing:1 adaptive:1 author:1 programme:1 party:11 welling:3 approximate:1 ignore:1 multilingual:2 global:24 corpus:3 francisco:2 alternatively:1 spectrum:7 discovery:1 continuous:1 latent:3 sk:1 why:1 table:9 additionally:3 learn:7 nature:1 ca:1 interact:2 heidelberg:1 bamler:2 zou:2 did:1 sp:1 main:2 aistats:1 linearly:1 rh:1 border:2 upc:1 hyperparameters:1 subsample:1 bilingual:1 allowed:1 fair:1 amortize:1 position:2 exponential:20 tied:1 levy:2 third:1 ruiz:3 uniting:1 rk:1 down:2 removing:1 specific:15 bishop:2 list:2 glorot:2 workshop:2 socher:2 pennington:2 illustrates:1 conditioned:1 chen:3 democrat:3 explore:1 applies:1 springer:1 corresponds:5 harris:2 minibatches:1 ma:1 acm:1 conditional:12 goal:4 identity:1 sized:1 month:3 shared:7 replace:1 change:3 glove:1 reducing:1 principal:2 total:1 castillo:1 experimental:3 citro:1 chiu:1 kulkarni:3 evaluate:2 |
6,223 | 663 | A Note on Learning Vector Quantization
Virginia R. de Sa
Dana H. Ballard
Department of Computer Science
University of Rochester
Rochester, NY 14627
Department of Computer Science
University of Rochester
Rochester, NY 14627
Abstract
Vector Quantization is useful for data compression. Competitive Learning which minimizes reconstruction error is an appropriate algorithm for
vector quantization of unlabelled data. Vector quantization of labelled
data for classification has a different objective, to minimize the number
of misclassifications, and a different algorithm is appropriate. We show
that a variant of Kohonen's LVQ2.1 algorithm can be seen as a multiclass extension of an algorithm which in a restricted 2 class case can
be proven to converge to the Bayes optimal classification boundary. We
compare the performance of the LVQ2.1 algorithm to that of a modified
version having a decreasing window and normalized step size, on a ten
class vowel classification problem.
1 Introduction
Vector quantization is a form of data compression that represents data vectors by a smaller
set of codebook vectors. Each data vector is then represented by its nearest codebook
vector. The goal of vector quantization is to represent the data with the fewest codebook
vectors while losing as little information as possible.
Vector quantization of unlabelled data seeks to minimize the reconstruction error. This can
be accomplished with Competitive learning[Grossberg, 1976; Kohonen, 1982], an iterative
learning algorithm for vector quantization that has been shown to perform gradient descent
on the following energy function [Kohonen, 1991]
J
/Ix -
220
ws?(x)
/l2p (x)dx.
A Note on Learning Vector Quantization
where p(x) is the probability distribution of the input patterns and Ws are the reference or
codebook vectors and s*(x) is defined by IIx - WSO(x) II ~ /Ix - will (for alIt). This minimizes the square reconstruction error of unlabelled data and may work reasonably well for
classification tasks if the patterns in the different classes are segregated.
In many classification tasks, however, the different member patterns may not be segregated
into separate clusters for each class. In these cases it is more important that members ofthe
same class be represented by the same codebook vector than that the reconstruction error
is minimized. To do this, the quantizer can m&ke use of the labelled data to encourage
appropriate quantization.
2
Previous approaches to Supervised Vector Quantization
The first use of labelled data (or a teaching signal) with Competitive Learning by Rumelhart
and Zipser [Rumelhart and Zipser, 1986] can be thought of as assigning a class to each
codebook vector and only allowing patterns from the appropriate class to influence each
reference vector.
This simple approach is far from optimal though as it fails to take into account interactions
between the classes. Kohonen addressed this in his LVQ( 1) algorithm[Kohonen, 1986]. He
argues that the reference vectors resulting from LVQ( 1) tend to approximate for a particular
class r,
P(xICr)P(Cr ) - ~#rP(xICs)P(Cs).
where P( Cj) is the a priori probability of Class i and P(xICj) is the conditional density of
Class i.
This approach is also not optimal for classification, as it addresses optimal places to put
the codebook vectors instead of optimal placement of the borders of the vector quantizer
which arise from the Voronoi tessellation induced by the codebook vectors. 1
3
Minimizing Misclassifications
In classification tasks the goal is to minimize the numbers of misclassifications of the
resultant quantizer. That is we want to minimize:
(1)
where, P(Classj) is the a priori probability of Classj and P(xIClassj) is the conditional
density of Classi and D.Rj is the decision region for class j (which in this case is all x such
that I~ - wkll < I~ - wjll (for all i) and Wk is a codebook vector for class j).
Consider a One-Dimensional problem of two classes and two codebook vectors wI and w2
defining a class boundary b = (wI + w2)/2 as shown in Figure 1. In this case Equation 1
reduces to:
1 Kohonen [1986] showed this by showing that the use of a "weighted" Voronoi tessellation (where
the relative distances of the borders from the reference vectors was changed) worked better. However
no principled way to calculate the relative weights was given and the application to real data used
the unweighted tessellation.
221
222
de Sa and Ballard
P(CIass i)P(xlClass i)
w2
b* b
wI
%
Figure 1: Codebook vectors Wl and'W2 define a border b. The optimal place for the border
is at b* where P(Cl)P(xICt} = P(C2)P(xIC2). The extra misclassification errors incurred by
placing the border at b is shown by the shaded region.
(2)
The derivative of Equation 2 with respect to b is
That is, the minimum number of misclassifications occurs at b* where
P(ClaSS1)P(b*IClasSl)
=P(Class2)P(b*IClass2).
If f(x) = (Classl)P(xIClassl) - P(Class2)P(xIClass2) was a regression function then we
could use stochastic approximation [Robbins and Monro, 1951] to estimate b* iteratively
as
ben + 1) = ben) + a(n)Z"
where Z" is a sample of the random variable Z whose expected value is
P(Classl)P(b(n)IClasst) - P(Class2)P(b(n)IClass2? and
lim a(n) = 0
,,-+co
l:ia(n) = 00
l:ia 2(n) < 00
However, we do not have immediate access to an appropriate random variable Z but can
express P( Classl )P(xIClassl)-P( Class2)P(xIClass2) as the limit of a sequence of regression
functions using the Parzen Window technique. In the Parzen window technique, probability
density functions are estimated as the sum of appropriately normalized pulses centered at
A Note on Learning Vector Quantization
the observed values. More formally, we can estimate P(xIClassi) as [Sklansky and Wassel,
1981]
Il
1~
Pi (x) = - L...J'?II(x-Xj,cll )
All
n )=1
.
where Xj is the sample data point at time j, and 'IIII(X- z, c(n)) is a Parzen window function
centred at Z with width parameter c(n) that satisfies the following conditions
'?II(X - z, c(n? ~ 0, Vx, Z
J~ '?II(X- Z, c(n?dx = 1
11-
lim n __ '?;(x- z, c(n))dx = 0
11-+-
lim '?1I(x-z,c(n? = c5(x-z)
II-+-
We can estimate f(x)
=P(Class1)P(xIClasst) - P(Class2)P(xIClass2) as
A
rex)
1
Il
= - LS(Xj)'?II(x-Xj,c(n?
n J=
.1
where S(Xj) is + 1 if Xj is from Class1 and -1 if Xj is from Class2.
Then
lim j"(X) = P(Class1)P(xIClass1) - P(Class2)P(xIClass2)
II-+-
and
lim E[S(X)'?ix - X, c(n)]
II-+-
=P(Class1)P(xIClassd -
P(Class2)P(xIClass2)
Wassel and Sklansky [1972] have extended the stochastic approximation method of Robbins and Monro [1951] to find the zero of a function that is the limit of a sequence of
regression functions and show rigourously that for the above case (where the distribution
of Class1 is to the left of that of Class2 and there is only one crossing point) the stochastic
approximation procedure
ben + 1) = ben) + a(n)ZII(xlI , Class(n), ben), c(n?
(3)
using
Z _ { 2c(n)'?(XII - ben), c(n?
II -
-2c(n)'?(XII - ben), c(n?
for XII
for XII
E Classl
E Class2
converges to the Bayes optimal border with probability one where '?(x - b, c) is a Parzen
window function. The following standard conditions for stochastic approximation convergence are needed in their proof
a(n), c(n)
> 0,
lim c(n)
II-+-
1:ia(n)c(n) = 00,
=0
lim a(n) = 0,
II-+-
223
224
de Sa and Ballard
as well as a condition that for rectangular Parzen functions reduces to a requirement that
P( Classl )P(xIClassl) - P( C lass2)P(xlClass2) be strictly positive to the left of b* and strictly
negative to the right of b* (for full details of the proof and conditions see [Wassel and
Sklansky, 1972]).
The above argument has only addressed the motion of the border. But b is defined as
b = (wI + w2)/2, thus we can move the codebook vectors according to
dE/dwl
=dEldw2 =.5dEldb.
We could now write Equation 3 as
wj(n + 1)
(X" - wj(n -
1?
=wj(n) + a2(n) IX" _ wj(n _ 1)1
if X" lies in window of width 2c(n) centred at ben), otherwise
Wi(n + 1)
=wi(n).
where we have used rectangular Parzen window functions and X" is from Classj. This
holds if Classl is to the right or left of Class2 as long as Wl and W2 are relatively ordered
appropriatel y.
Expanding the problem to more dimensions, and more classes with more codebook vectors per class, complicates the analysis as a change in two codebook vectors to better adjust
their border affects more than just the border between the two codebook vectors. However ignoring these effects for a first order approximation suggests the following update
procedure:
1?
*
(X" - wren 1) + a(n) IIX" _ wren _ 1)11
*.
(X" - w;(n - 1?
1) - a(n) IIX" _ wj(n _ 1)11
*
Wi
(n)
Wj (n)
=Wi (n =Wj (n -
where a(n) obeys the constraints above, X" is from Classj, and w;, wj are the two nearest
codebook vectors, one each from class i and j U* i) and x" lies within c(n) of the border
between them. (No changes are made if all the above conditions are not true). As above
this algorithm assumes that the initial positions of the codebook vectors are such that they
will not have to cross during the algorithm.
The above algorithm is similar to Kohonen's LVQ2.1 algorithm (which is performed after
appropriate initialization of the codebook vectors) except for the normalization of the step
size, the decreasing size of the window width c(n) and constraints on the learning rate a.
A Note on Learning Vector Quantization
4 Simulations
Motivated by the theory above, we decided to modify Kohonen's LVQ2.1 algorithm to
add normalization of the step size and a decreasing window. In order to allow closer
comparison with LVQ2.1, all other parts of the algorithm were kept the same. Thus a
decreased linearly. We used a linear decrease on the window size and defined it as in
LVQ2.1 for easier parameter matching. For a window size of w all input vectors satisfying
d;/dj> g:~ where di is the distance to the closest codebook vector and dj is the distance
to the next closest codebook vector, fall into the window between those two vectors (Note
however, that updates only occur if the two closest codebook vectors belong to different
classes).
The data used is a version of the Peterson and Barney vowel formant data 2. The dataset
consists of the first and second formants for ten vowels in a/hVdj context from 75 speakers
(32 males, 28 females, 15 children) who repeated each vowel twice 3. As we were not
testing generalization , the training set was used as the test set.
75.------.------.------.-----..-----~
~.
..
-fA
alpha-0.002
alpha-0.030
alpha-0.080
alpha-O .150
alpha-0.500
~A -:~~.::ra.?-..:::
.....
u
Q)
!\,.
70
.
....i
U
.....
r::Q)
Q)
'"
.
..
\.
.
\'
\\
\
".
\\
'B'"
... -
\..
\..~
\'"
-t--.
\
I""
\.
\
\\
\,:
,.,
\.
65
.......
~.~--;:::.-..--
; !
......
o
...u
;,oz:~,,; .. ~...
-+-
~
"
\
\\
??
\
\
\
\
~
I
it:
\
t
\,
\
\
\
\
60~-----L~----~--~-L--~~~----~
o
0.2
0.4
0.6
window size
0.8
Figure 2: The effect of different window sizes on the accuracy for different values of initial
a.
We ran three sets of experiments varying the number of codebook vectors and the number
of pattern presentations. For the first set of experiments there were 20 codebook vectors
and the algorithms ran for 40000 steps. Figure 2 shows the effect of varying the window
size for different initial learning rates a( 1) in the LVQ2.1 algorithm. The values plotted are
averaged over three runs (The order of presentation of patterns is different for the different
runs). The sensitivity of the algorithm to the window size as mentioned in [Kohonen, 1990]
is evident. In general we found that as the learning rate is increased the peak accuracy is
improved at the expense of the accuracy for other window widths. After a certain value
20 btained
from Steven Nowlan
33 speakers were missing one vowel and the raw data was linearly transfonned to have zero mean
and fall within the range [-3,3] in both components
225
226
de Sa and Ballard
85~----~----~------r------r----~
.
-,.
80
.....
u
.....
II'"
75
mod/100/40000
-?.?
~
-+-
... _._.-lI'-?-?-???-?- ? lIi~T2!t)lrotT(J?"''''':': ?
._.;:::~::::::... ---?--?~~~O..?.4.jlD.oJI._.~=-.
._~~. .- .
; .,
----~----+---------------------~
! ~=~.~'Il
r::
Qj
u
I-<
Qj
'"
?B???
~;.'
Qj
I-<
I-<
0
U
orig/20/40000
mod/20/40000
orig/20/4000
~-El'.'-
. . . . ". ~--- ~
?????????????????????????
70
65~----~----~------~----~~--~
o
0.1
0.2
0.3
window size
0.4
0.5
Figure 3: The performance of LVQ2.1 with and without the modifications (normalized step
size and decreasing window) for 3 different conditions. The legend gives in order [the alg
type/ the number of codebook vectors/ the number of pattern presentations]
the accuracy declines for further increases in learning rate.
Figure 3 shows the improvement achieved with normalization and a linearly decreasing
window size for three sets of experiments : (20 code book vectors/40000 pattern presentations), (20 code book vectors/4000 pattern presentations) and (100 code book vectors/40000 pattern presentations). For the decreasing window algorithm, the x-axis represents the window size in the middle of the run. As above, the values plotted were averaged
over three runs. The values of a(l) were the same within each algorithm over all three
conditions. A graph using the best a found for each condition separately is almost identical. The graph shows that the modifications provide a modest but consistent improvement
in accuracy across the conditions.
In summary the preliminary experiments indicate that a decreasing window and normalized
step size can be worthwhile additions to the LVQ2.1 algorithm and further experiments on
the generalization properties of the algorithm and with other data sets may be warranted.
For these tests we used a linear decrease of the window size and learning rate to allow for
easier comparison with the LVQ2.1 algorithm. Further modifications on the algorithm that
experiment with different functions (that obey the theoretical constraints) for the learning
rate and window size decrease may result in even better performance.
5 Summary
We have shown that Kohonen's LVQ2.1 algorithm can be considered as a variant on a
generalization of an algorithm which is optimal for a IDimensional/2 codebook vector
problem. We added a decreasing window and normalized step size, suggested from the
one dimensional algorithm. to the LVQ2.1 algorithm and found a small but consistent
improvement in accuracy.
A Note on Learning Vector Quantization
Acknowledgements
We would like to thank Steven Nowlan for his many helpful suggestions on an earlier draft
and for making the vowel formant data available to us. We are also grateful to Leonidas
Kontothanassis for his help in coding and discussion. This work was supported by a grant
from the Human Frontier Science Program and a Canadian NSERC 1967 Science and
Engineering Scholarship to the first author who also received A NIPS travel grant to attend
the conference.
References
[Grossberg, 1976] Stephen Grossberg, "Adaptive Pattern Classification and Universal Recoding: I. Parallel Development and Coding of Neural Feature Detectors," Biological
Cybernetics, 23:121-134,1976.
[Kohonen,1982] Teuvo Kohonen, "Self-Organized Formation of Topologically Correct
Feature Maps," Biological Cybernetics, 43:59--69, 1982.
[Kohonen,1986] Teuvo Kohonen, "Learning Vector Quantization for Pattern Recognition," Technical Report TKK-F-A601, Helsinki University of Technology, Department
of Technical Physics, Laboratory of Computer and Information Science, November
1986.
[Kohonen, 1990] Teuvo Kohonen, "Statistical Pattern Recognition Revisited," In R. Eckmiller, editor, Advanced Neural Computers, pages 137-144. Elsevier Science Publishers, 1990.
[Kohonen, 1991] Teuvo Kohonen, "Self-Organizing Maps: Optimization Approaches," In
T. Kohonen, K. Makisara, O. Simula, and J. Kangas, editors,Artijicial Neural Networks,
pages 981-990. Elsevier Science Publishers, 1991.
[Robbins and Monro, 1951J Herbert Robbins and Sutton Monro, "A Stochastic Approximation Method," Annals of Math. Stat., 22:400-407,1951.
[Rumelhart and Zipser, 1986] D. E. Rumelhart and D. Zipser, "Feature Discovery by
Competitive Learning," In David E. Rumelhart, James L. McClelland, and the PDP Research Group, editors, Parallel Distributed Processing: Explorations in the Microstructure of Cognition, volume 2, pages 151-193. MIT Press, 1986.
[Sklansky and Wassel, 1981] Jack Sklansky and Gustav N. Wassel, Pattern Classijiers
and Trainable Machines, Springer-Verlag, 1981.
[Wassel and Sklansky, 1972] Gustav N. Wassel and Jack Sklansky, "Training a OneDimensional Classifier to Minimize the Probability of Error," IEEE Transactions on
Systems, Man, and Cybernetics, SMC-2(4):533-541, 1972.
227
| 663 |@word middle:1 version:2 compression:2 pulse:1 seek:1 simulation:1 barney:1 initial:3 nowlan:2 assigning:1 dx:3 class1:6 xlclass:1 update:2 alit:1 draft:1 quantizer:3 revisited:1 math:1 codebook:25 zii:1 c2:1 consists:1 ra:1 expected:1 formants:1 decreasing:8 little:1 window:26 minimizes:2 classifier:1 grant:2 positive:1 engineering:1 attend:1 modify:1 limit:2 sutton:1 twice:1 initialization:1 suggests:1 shaded:1 co:1 smc:1 range:1 obeys:1 averaged:2 grossberg:3 decided:1 testing:1 procedure:2 universal:1 thought:1 matching:1 put:1 context:1 influence:1 map:2 missing:1 l:1 rectangular:2 ke:1 his:3 annals:1 losing:1 crossing:1 rumelhart:5 satisfying:1 recognition:2 simula:1 observed:1 steven:2 calculate:1 wj:8 region:2 decrease:3 ran:2 principled:1 mentioned:1 grateful:1 orig:2 represented:2 fewest:1 formation:1 whose:1 otherwise:1 formant:2 sequence:2 reconstruction:4 interaction:1 kohonen:19 organizing:1 oz:1 convergence:1 cluster:1 requirement:1 converges:1 ben:8 help:1 stat:1 nearest:2 received:1 sa:4 c:1 indicate:1 correct:1 stochastic:5 centered:1 vx:1 human:1 exploration:1 microstructure:1 generalization:3 preliminary:1 biological:2 extension:1 strictly:2 frontier:1 hold:1 considered:1 cognition:1 a2:1 wren:2 travel:1 robbins:4 wl:2 weighted:1 mit:1 modified:1 cr:1 varying:2 improvement:3 helpful:1 elsevier:2 voronoi:2 el:1 w:2 classification:8 priori:2 development:1 having:1 identical:1 represents:2 placing:1 makisara:1 minimized:1 t2:1 report:1 vowel:6 adjust:1 male:1 encourage:1 closer:1 modest:1 plotted:2 theoretical:1 complicates:1 increased:1 earlier:1 tessellation:3 teuvo:4 virginia:1 rex:1 density:3 peak:1 cll:1 sensitivity:1 physic:1 parzen:6 xict:1 lii:1 book:3 derivative:1 li:1 account:1 de:5 centred:2 coding:2 wk:1 leonidas:1 performed:1 competitive:4 bayes:2 parallel:2 rochester:4 monro:4 minimize:5 square:1 il:3 accuracy:6 who:2 ofthe:1 xli:1 raw:1 lvq2:12 cybernetics:3 detector:1 energy:1 james:1 resultant:1 proof:2 di:1 dataset:1 lim:7 cj:1 organized:1 supervised:1 improved:1 though:1 just:1 tkk:1 oji:1 effect:3 normalized:5 true:1 iteratively:1 laboratory:1 during:1 width:4 self:2 speaker:2 evident:1 argues:1 motion:1 jack:2 volume:1 belong:1 he:1 onedimensional:1 teaching:1 dj:2 access:1 add:1 closest:3 showed:1 female:1 certain:1 verlag:1 accomplished:1 seen:1 minimum:1 herbert:1 converge:1 signal:1 ii:12 stephen:1 full:1 rj:1 reduces:2 technical:2 unlabelled:3 cross:1 long:1 sklansky:7 wassel:7 variant:2 regression:3 represent:1 normalization:3 achieved:1 addition:1 want:1 separately:1 iiii:1 addressed:2 decreased:1 publisher:2 appropriately:1 w2:6 extra:1 induced:1 tend:1 member:2 legend:1 mod:2 zipser:4 gustav:2 canadian:1 xj:7 affect:1 misclassifications:4 decline:1 l2p:1 multiclass:1 qj:3 motivated:1 useful:1 ten:2 mcclelland:1 estimated:1 per:1 xii:4 write:1 express:1 eckmiller:1 group:1 kept:1 graph:2 sum:1 run:4 topologically:1 place:2 almost:1 decision:1 occur:1 placement:1 constraint:3 worked:1 helsinki:1 argument:1 relatively:1 department:3 according:1 smaller:1 across:1 wi:8 modification:3 making:1 restricted:1 equation:3 needed:1 available:1 obey:1 worthwhile:1 appropriate:6 cia:1 rp:1 artijicial:1 assumes:1 iix:3 scholarship:1 objective:1 move:1 added:1 occurs:1 fa:1 gradient:1 distance:3 separate:1 thank:1 transfonned:1 code:3 minimizing:1 expense:1 negative:1 perform:1 allowing:1 descent:1 november:1 immediate:1 defining:1 extended:1 pdp:1 kangas:1 david:1 class2:11 nip:1 address:1 suggested:1 pattern:14 program:1 classj:4 ia:3 misclassification:1 jld:1 advanced:1 technology:1 axis:1 acknowledgement:1 discovery:1 segregated:2 relative:2 suggestion:1 proven:1 dana:1 incurred:1 consistent:2 editor:3 pi:1 changed:1 summary:2 supported:1 allow:2 fall:2 peterson:1 recoding:1 distributed:1 boundary:2 dimension:1 unweighted:1 author:1 c5:1 made:1 adaptive:1 far:1 transaction:1 approximate:1 alpha:5 iterative:1 ballard:4 reasonably:1 expanding:1 ignoring:1 alg:1 warranted:1 cl:1 linearly:3 border:10 arise:1 child:1 repeated:1 ny:2 fails:1 position:1 lie:2 ix:4 showing:1 quantization:15 easier:2 ordered:1 nserc:1 springer:1 satisfies:1 conditional:2 goal:2 presentation:6 lvq:2 labelled:3 man:1 change:2 except:1 dwl:1 classi:1 a601:1 formally:1 trainable:1 |
6,224 | 6,630 | A Linear-Time Kernel Goodness-of-Fit Test
Wittawat Jitkrittum
Gatsby Unit, UCL
Wenkai Xu
Gatsby Unit, UCL
Zolt?n Szab??
CMAP, ?cole Polytechnique
[email protected]
[email protected]
[email protected]
Kenji Fukumizu
The Institute of Statistical Mathematics
[email protected]
Arthur Gretton?
Gatsby Unit, UCL
[email protected]
Abstract
We propose a novel adaptive test of goodness-of-fit, with computational cost
linear in the number of samples. We learn the test features that best indicate the
differences between observed samples and a reference model, by minimizing the
false negative rate. These features are constructed via Stein?s method, meaning that
it is not necessary to compute the normalising constant of the model. We analyse
the asymptotic Bahadur efficiency of the new test, and prove that under a mean-shift
alternative, our test always has greater relative efficiency than a previous linear-time
kernel test, regardless of the choice of parameters for that test. In experiments, the
performance of our method exceeds that of the earlier linear-time test, and matches
or exceeds the power of a quadratic-time kernel test. In high dimensions and where
model structure may be exploited, our goodness of fit test performs far better than
a quadratic-time two-sample test based on the Maximum Mean Discrepancy, with
samples drawn from the model.
1
Introduction
The goal of goodness of fit testing is to determine how well a model density p(x) fits an observed
sample D = {xi }ni=1 ? X ? Rd from an unknown distribution q(x). This goal may be achieved via
a hypothesis test, where the null hypothesis H0 : p = q is tested against H1 : p 6= q. The problem
of testing goodness of fit has a long history in statistics [11], with a number of tests proposed for
particular parametric models. Such tests can require space partitioning [18, 3], which works poorly in
high dimensions; or closed-form integrals under the model, which may be difficult to obtain, besides
in certain special cases [2, 5, 30, 26]. An alternative is to conduct a two-sample test using samples
drawn from both p and q. This approach was taken by [23], using a test based on the (quadratic-time)
Maximum Mean Discrepancy [16], however this does not take advantage of the known structure of p
(quite apart from the increased computational cost of dealing with samples from p).
More recently, measures of discrepancy with respect to a model have been proposed based on Stein?s
method [21]. A Stein operator for p may be applied to a class of test functions, yielding functions that
have zero expectation under p. Classes of test functions can include the W 2,? Sobolev space [14],
and reproducing kernel Hilbert spaces (RKHS) [25]. Statistical tests have been proposed by [9, 22]
based on classes of Stein transformed RKHS functions, where the test statistic is the norm of the
smoothness-constrained function with largest expectation under q . We will refer to this statistic as
the Kernel Stein Discrepancy (KSD). For consistent tests, it is sufficient to use C0 -universal kernels
[6, Definition 4.1], as shown by [9, Theorem 2.2], although inverse multiquadric kernels may be
preferred if uniform tightness is required [15].2
?
Zolt?n Szab??s ORCID ID: 0000-0001-6183-7603. Arthur Gretton?s ORCID ID: 0000-0003-3169-7624.
Briefly, [15] show that when an exponentiated quadratic kernel is used, a sequence of sets D may be
constructed that does not correspond to any q, but for which the KSD nonetheless approaches zero. In a statistical
testing setting, however, we assume identically distributed samples from q, and the issue does not arise.
2
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
The minimum variance unbiased estimate of the KSD is a U-statistic, with computational cost
quadratic in the number n of samples from q. It is desirable to reduce the cost of testing, however,
so that larger sample sizes may be addressed. A first approach is to replace the U-statistic with a
running average with linear cost, as proposed by [22] for the KSD, but this results in an increase in
variance and corresponding decrease in test power. An alternative approach is to construct explicit
features of the distributions, whose empirical expectations may be computed in linear time. In the
two-sample and independence settings, these features were initially chosen at random by [10, 8, 32].
More recently, features have been constructed explicitly to maximize test power in the two-sample
[19] and independence testing [20] settings, resulting in tests that are not only more interpretable, but
which can yield performance matching quadratic-time tests.
We propose to construct explicit linear-time features for testing goodness of fit, chosen so as to
maximize test power. These features further reveal where the model and data differ, in a readily interpretable way. Our first theoretical contribution is a derivation of the null and alternative distributions
for tests based on such features, and a corresponding power optimization criterion. Note that the
goodness-of-fit test requires somewhat different strategies to those employed for two-sample and
independence testing [19, 20], which become computationally prohibitive in high dimensions for
the Stein discrepancy (specifically, the normalization used in prior work to simplify the asymptotics
would incur a cost cubic in the dimension d and the number of features in the optimization). Details
may be found in Section 3.
Our second theoretical contribution, given in Section 4, is an analysis of the relative Bahadur
efficiency of our test vs the linear time test of [22]: this represents the relative rate at which the pvalue decreases under H1 as we observe more samples. We prove that our test has greater asymptotic
Bahadur efficiency relative to the test of [22], for Gaussian distributions under the mean-shift
alternative. This is shown to hold regardless of the bandwidth of the exponentiated quadratic kernel
used for the earlier test. The proof techniques developed are of independent interest, and we anticipate
that they may provide a foundation for the analysis of relative efficiency of linear-time tests in the
two-sample and independence testing domains. In experiments (Section 5), our new linear-time test
is able to detect subtle local differences between the density p(x), and the unknown q(x) as observed
through samples. We show that our linear-time test constructed based on optimized features has
comparable performance to the quadratic-time test of [9, 22], while uniquely providing an explicit
visual indication of where the model fails to fit the data.
2
Kernel Stein Discrepancy (KSD) Test
We begin by introducing the Kernel Stein Discrepancy (KSD) and associated statistical test, as
proposed independently by [9] and [22]. Assume that the data domain is a connected open set X ? Rd .
Consider a Stein operator Tp that takes in a multivariate function f (x) = (f1 (x), . . . , fd (x))> ? Rd
and constructs a function (Tp f ) (x) : Rd ? R. The constructed function has the key property that for
all f in an appropriate function class, Ex?q [(Tp f )(x)] = 0 if and only if q = p. Thus, one can use
this expectation as a statistic for testing goodness of fit.
The function class F d for the function f is chosen to be a unit-norm ball in a reproducing kernel Hilbert
space (RKHS) in [9, 22]. More precisely, let F be an RKHS associated with a positive definite kernel
k : X ? X ? R. Let ?(x) = k(x, ?) denote a feature map of k so that k(x, x0 ) = h?(x), ?(x0 )iF .
Assume that fi ? F for all i = 1, . . . , d so that f ? F ? ? ? ? ? F := F d where F d is equipped with
Pd
the standard inner product hf , giF d := i=1 hfi , gi iF . The kernelized Stein operator Tp studied
Pd
p(x)
?fi (x) (a)
in [9] is (Tp f ) (x) := i=1 ? log
f
(x)
+
= f , ? p (x, ?) F d , where at (a) we use the
i
?xi
?xi
reproducing property of F, i.e., fi (x) = hfi , k(x, ?)iF , and that
hence ? p (x, ?) := ? log?xp(x) k(x, ?)+ ?k(x,?)
is in F d .
?x
is defined such that (Tp f ) (x) ? Rd . This distinction
?k(x,?)
?xi
? F [28, Lemma 4.34],
We note that the Stein operator presented in [22]
is not crucial and leads to the same goodness-offit test. Under appropriate conditions, e.g. that limkxk?? p(x)fi (x) = 0 for all i = 1, . . . , d, it can
be shown using integration by parts that Ex?p (Tp f )(x) = 0 for any f ? F d [9, Lemma 5.1]. Based
on the Stein operator, [9, 22] define the kernelized Stein discrepancy as
Sp (q) :=
sup
kf kF d ?1
(a)
Ex?q f , ? p (x, ?) F d =
sup
kf kF d ?1
2
f , Ex?q ? p (x, ?) F d = kg(?)kF d ,
(1)
where at (a), ? p (x, ?) is Bochner integrable [28, Definition A.5.20] as long as Ex?q k? p (x, ?)kF d <
?, and g(y) := Ex?q ? p (x, y) is what we refer to as the Stein witness function. The Stein witness
function will play a crucial role in our new test statistic in Section 3. When a C0 -universal kernel is
used [6, Definition 4.1], and as long as Ex?q k?x log p(x) ? ?x log q(x)k2 < ?, it can be shown
that Sp (q) = 0 if and only if p = q [9, Theorem 2.2].
Ex?q Ex0 ?q hp (x, x0 ), where hp (x, y) :=
Pd ? 2 k(x,y)
>
>
s>
p (x)sp (y)k(x, y) + sp (y)?x k(x, y) + sp (x)?y k(x, y) +
i=1 ?xi ?yi , and sp (x) :=
c2 =
?x log p(x) is a column vector. An unbiased empirical estimator of Sp2 (q), denoted by S
P
2
i<j hp (xi , xj ) [22, Eq. 14], is a degenerate U-statistic under H0 . For the goodness-of-fit
n(n?1)
c2
test, the rejection threshold can be computed by a bootstrap procedure. All these properties make S
a very flexible criterion to detect the discrepancy of p and q: in particular, it can be computed even if
p is known only up to a normalization constant. Further studies on nonparametric Stein operators can
be found in [25, 14].
The KSD Sp (q) can be written as Sp2 (q)
=
c2 costs O(n2 ). To reduce this cost, a
Linear-Time Kernel Stein (LKS) Test Computation of S
linear-time (i.e., O(n)) estimator based on an incomplete U-statistic is proposed in [22, Eq. 17],
c2 := 2 Pn/2 h (x
given by S
i=1 p 2i?1 , x2i ), where we assume n is even for simplicity. Empirically
l
n
[22] observed that the linear-time estimator performs much worse (in terms of test power) than the
quadratic-time U-statistic estimator, agreeing with our findings presented in Section 5.
3
New Statistic: The Finite Set Stein Discrepancy (FSSD)
Although shown to be powerful, the main drawback of the KSD test is its high computational cost of
O(n2 ). The LKS test is one order of magnitude faster. Unfortunately, the decrease in the test power
outweighs the computational gain [22]. We therefore seek a variant of the KSD statistic that can be
computed in linear time, and whose test power is comparable to the KSD test.
Key Idea The fact that Sp (q) = 0 if and only if p = q implies that g(v) = 0 for all v ? X if and
only if p = q, where g is the Stein witness function in (1). One can see g as a function witnessing
the differences of p, q, in such a way that |gi (v)| is large when there is a discrepancy in the region
around v, as indicated by the ith output of g. The test statistic of [22, 9] is essentially given by the
degree of ?flatness? of g as measured by the RKHS norm k ? kF d . The core of our proposal is to use
a different measure of flatness of g which can be computed in linear time.
The idea is to use a real analytic kernel k which makes g1 , . . . , gd real analytic. If gi 6= 0 is an
analytic function, then the Lebesgue measure of the set of roots {x | gi (x) = 0} is zero [24]. This
property suggests that one can evaluate gi at a finite set of locations V = {v1 , . . . , vJ }, drawn
from a distribution with a density (w.r.t. the Lebesgue measure). If gi 6= 0, then almost surely
gi (v1 ), . . . , gi (vJ ) will not be zero. This idea was successfully exploited in recently proposed
linear-time tests of [8] and [19, 20]. Our new test statistic based on this idea is called the Finite Set
Stein Discrepancy (FSSD) and is given in Theorem 1. All proofs are given in the appendix.
Theorem 1 (The Finite Set Stein Discrepancy (FSSD)). Let V = {v1 , . . . , vJ } ? Rd be random
vectors drawn i.i.d. from a distribution ? which has a density. Let X be a connected open set
Pd PJ
1
2
in Rd . Define FSSD2p (q) := dJ
j=1 gi (vj ). Assume that 1) k : X ? X ? R is C0 i=1
universal [6, Definition 4.1] and real analytic i.e., for all v ? X , f (x) := k(x, v) is a real analytic
function on X . 2) Ex?q Ex0 ?q hp (x, x0 ) < ?. 3) Ex?q k?x log p(x) ? ?x log q(x)k2 < ?. 4)
limkxk?? p(x)g(x) = 0.
Then, for any J ? 1, ?-almost surely FSSD2p (q) = 0 if and only if p = q.
This measure depends on a set of J test locations (or features) {vi }Ji=1 used to evaluate the Stein
witness function, where J is fixed and is typically
small.2 A
kernel which is C0 -universal and real
kx?yk
analytic is the Gaussian kernel k(x, y) = exp ? 2?2 2 (see [20, Proposition 3] for the result
k
on analyticity). Throughout this work, we will assume all the conditions stated in Theorem 1, and
consider only the Gaussian kernel. Besides the requirement that the kernel be real and analytic,
the remaining conditions in Theorem 1 are the same as given in [9, Theorem 2.2]. Note that if the
3
FSSD is to be employed in a setting otherwise than testing, for instance to obtain pseudo-samples
converging to p, then stronger conditions may be needed [15].
3.1
Goodness-of-Fit Test with the FSSD Statistic
Given a significance level ? for the goodness-of-fit test, the test can be constructed so that H0 is
\2 is
\2 > T? , where T? is the rejection threshold (critical value), and FSSD
rejected when nFSSD
an empirical estimate of FSSD2p (q). The threshold which guarantees that the type-I error (i.e., the
probability of rejecting H0 when it is true) is bounded above by ? is given by the (1 ? ?)-quantile of
\2 under H0 . In the following, we start by giving
the null distribution i.e., the distribution of nFSSD
\2 , and summarize its asymptotic distributions in Proposition 2.
the expression for FSSD
?
Let ?(x) ? Rd?J such that [?(x)]i,j = ?p,i (x, vj )/ dJ. Define ? (x) := vec(?(x)) ? RdJ where
vec(M) concatenates columns of the matrix M into a column vector. We note that ? (x) depends
on the test locations V = {vj }Jj=1 . Let ?(x, y) := ? (x)> ? (y) = tr(?(x)> ?(y)). Given an i.i.d.
sample {xi }ni=1 ? q, a consistent, unbiased estimator of FSSD2p (q) is
d X
n X
J
X
X
X
2
1
\2 = 1
FSSD
?(xi , xj ),
?p,l (xi , vm )?p,l (xj , vm ) =
dJ
n(n ? 1) i=1
n(n ? 1) i<j
m=1
l=1
(2)
j6=i
which is a one-sample second-order U-statistic with ? as its U-statistic kernel [27, Section 5.1.1].
d
Being a U-statistic, its asymptotic distribution can easily be derived. We use ? to denote convergence
in distribution.
\2 ). Let Z1 , . . . , ZdJ i.i.d.
? N (0, 1). Let ? :=
Proposition 2 (Asymptotic distributions of FSSD
dJ?dJ
Ex?q [? (x)], ?r := covx?r [? (x)] ? R
for r ? {p, q}, and {?i }dJ
i=1 be the eigenvalues of
?p = Ex?p [? (x)? > (x)]. Assume that Ex?q Ey?q ?2 (x, y) < ?. Then, for any realization of
V = {vj }Jj=1 , the following statements hold.
d
\2 ?
1. Under H0 : p = q, nFSSD
PdJ
2
i=1 (Zi
? 1)?i .
2
2. Under H1 : p 6= q, if ?H
:= 4?> ?q ? > 0, then
1
?
d
2
\2 ? FSSD2 ) ?
n(FSSD
N (0, ?H
).
1
Proof. Recognizing that (2) is a degenerate U-statistic, the results follow directly from [27, Section
5.5.1, 5.5.2].
Claims 1 and 2 of Proposition 2 imply that under H1 , the test power (i.e., the probability of correctly
rejecting H1 ) goes to 1 asymptotically, if the threshold T? is defined as above. In practice, simulating
from the asymptotic null distribution in Claim 1 can be challenging, since the plug-in estimator of
?p requires a sample from p, which is not available. A straightforward solution is to draw sample
from p, either by assuming that p can be sampled easily or by using a Markov chain Monte Carlo
(MCMC) method, although this adds an additional computational burden to the test procedure. A
more subtle issue is that when dependent samples from p are used in obtaining the test threshold, the
test may become more conservative than required for i.i.d. data [7]. An alternative approach is to use
? q instead of ?p . The covariance matrix ?
? q can be directly computed from the
the plug-in estimate ?
data. This is the approach we take. Theorem 3 guarantees that the replacement of the covariance in
the computation of the asymptotic null distribution still yields a consistent test. We write PH1 for the
\2 under H1 .
distribution of nFSSD
? q := 1 Pn ? (xi )? > (xi ) ? [ 1 Pn ? (xi )][ 1 Pn ? (xj )]> with {xi }n ?
Theorem 3. Let ?
i=1
i=1
i=1
j=1
n
n
n
PdJ
q. Suppose that the test threshold T? is set to the (1??)-quantile of the distribution of i=1 (Zi2 ?1)??i
i.i.d.
? q . Then, under H0 , asymptotically
where {Zi }dJ
?1 , . . . , ??dJ are eigenvalues of ?
i=1 ? N (0, 1), and ?
J
the false positive rate is ?. Under H1 , for {vj }j=1 drawn from a distribution with a density, the test
\2 > T? ) ? 1 as n ? ?.
power PH (nFSSD
1
?q = ?
? p i.e., the plug-in
Remark 1. The proof of Theorem 3 relies on two facts. First, under H0 , ?
?
estimate of ?p . Thus, under H0 , the null distribution approximated with ?q is asymptotically
4
? p to ?p . Second, the rejection threshold obtained from the
correct, following the convergence of ?
approximated null distribution is asymptotically constant. Hence, under H1 , claim 2 of Proposition 2
d
\2 ?
\2 > T? ) ? 1.
implies that nFSSD
? as n ? ?, and consequently PH1 (nFSSD
3.2
Optimizing the Test Parameters
Theorem 1 guarantees that the population quantity FSSD2 = 0 if and only if p = q for any choice of
{vi }Ji=1 drawn from a distribution with a density. In practice, we are forced to rely on the empirical
\2 , and some test locations will give a higher detection rate (i.e., test power) than others for
FSSD
J
finite n. Following the approaches of [17, 20, 19, 29], we choose the test locations V = {vj }j=1
and kernel bandwidth ?k2 so as to maximize the test power i.e., the probability of rejecting H0 when
it is false. We first give an approximate expression for the test power when n is large.
\2 ). Under H1 , for large n and fixed r, the
Proposition 4 (Approximate test power of nFSSD
?
2
r
\2 > r) ? 1 ? ? ?
? n FSSD
test power PH1 (nFSSD
, where ? denotes the cumulative
?H1
n?H1
distribution function of the standard normal distribution, and ?H1 is defined in Proposition 2.
? r/n?FSSD2
\2 ?FSSD2
\2 > r) = PH (FSSD
\2 > r/n) = PH ?n FSSD
Proof. PH1 (nFSSD
n
.
>
1
1
?H1
?H1
For sufficiently large n, the alternative distribution
is approximatelynormal as given in Proposition 2.
\2 > r) ? 1 ? ? ? r ? ?n FSSD2 .
It follows that PH1 (nFSSD
?H
n?H
1
1
Let ? := {V, ?k2 } be the collection of all tuning parameters. Assume that n is sufficiently large.
?
2
r
Following the same argument as in [29], in ?n?
? n FSSD
?H1 , we observe that the first term
H1
?
2
? r
= O(n1/2 ), dominating
= O(n?1/2 ) going to 0 as n ? ?, while the second term n FSSD
?H
n?H
1
1
the first for large n. Thus, the best parameters that maximize the test power are given by ? ? =
\2 > T? ) ? arg max? FSSD2 . Since FSSD2 and ?H are unknown, we divide
arg max? PH1 (nFSSD
1
?H
1
\2
the sample {xi }ni=1 into two disjoint training and test sets, and use the training set to compute ??FSSD
,
H1 +?
where a small regularization parameter ? > 0 is added for numerical stability. The goodness-of-fit
test is performed on the test set to avoid overfitting. The idea of splitting the data into training and
test sets to learn good features for hypothesis testing was successfully used in [29, 20, 19, 17].
To find a local maximum of
{vi }Ji=1
\2
FSSD
?
?H1 +? ,
we use gradient ascent for its simplicity. The initial points of
are set to random draws from a normal distribution fitted to the training data, a heuristic we
found to perform well in practice. The objective is non-convex in general, reflecting many possible
ways to capture the differences of p and q. The regularization parameter ? is not tuned, and is
\2
fixed to a small constant. Assume that ?x log p(x) costs O(d2 ) to evaluate. Computing ?? ??FSSD
H +?
1
2
\2 and ?
costs O(d2 J 2 n). The computational complexity of nFSSD
?H
is O(d2 Jn). Thus, finding
1
a local optimum via gradient ascent is still linear-time, for a fixed maximum number of iterations.
? q costs O(d2 J 2 n), and obtaining all the eigenvalues of ?
? q costs O(d3 J 3 ) (required
Computing ?
only once). If the eigenvalues decay to zero sufficiently rapidly, one can approximate the asymptotic
null distribution with only a few eigenvalues. The cost to obtain the largest few eigenvalues alone can
be much smaller.
P
? := n1 ni=1 ? (xi ). It is possible to normalize the FSSD statistic to get a new
Remark 2. Let ?
? n := n?
? q + ?I)?1 ?
? > (?
? where ? ? 0 is a regularization parameter that goes to 0
statistic ?
as n ? ?. This was done in the case of the ME (mean embeddings) statistic of [8, 19]. The
asymptotic null distribution of this statistic takes the convenient form of ?2 (dJ) (independent of
? q . It turns out that the test power
p and q), eliminating the need to obtain the eigenvalues of ?
?
criterion for tuning the parameters in this case is the statistic ?n itself. However, the optimization
? q + ?I)?1 (costing O(d3 J 3 )) needs to be reevaluated in each
is computationally expensive as (?
gradient ascent iteration. This is not needed in our proposed FSSD statistic.
5
4
Relative Efficiency and Bahadur Slope
Both the linear-time kernel Stein (LKS) and FSSD tests have the same computational cost of O(d2 n),
and are consistent, achieving maximum power of 1 as n ? ? under H1 . It is thus of theoretical
interest to understand which test is more sensitive in detecting the differences of p and q. This can be
quantified by the Bahadur slope of the test [1]. Two given tests can then be compared by computing
the Bahadur efficiency (Theorem 7) which is given by the ratio of the slopes of the two tests. We
note that the constructions and techniques in this section may be of independent interest, and can be
generalised to other statistical testing settings.
We start by introducing the concept of Bahadur slope for a general test, following the presentation of
[12, 13]. Consider a hypothesis testing problem on a parameter ?. The test proposes a null hypothesis
H0 : ? ? ?0 against the alternative hypothesis H1 : ? ? ?\?0 , where ?, ?0 are arbitrary sets.
Let Tn be a test statistic computed from a sample of size n, such that large values of Tn provide
an evidence to reject H0 . We use plim to denote convergence in probability, and write Er for
Ex?r Ex0 ?r .
Approximate Bahadur Slope (ABS) For ?0 ? ?0 , let the asymptotic null distribution of Tn be
F (t) = limn?? P?0 (Tn < t), where we assume that the CDF (F ) is continuous and common to all
?0 ? ?0 . The continuity of F will be important later when Theorem 9 and 10 are used to compute
the slopes of LKS and FSSD tests. Assume that there exists a continuous strictly increasing function
(Tn ))
? : (0, ?) ? (0, ?) such that limn?? ?(n) = ?, and that ?2 plimn?? log(1?F
= c(?)
?(n)
where Tn ? P? , for some function c such that 0 < c(?A ) < ? for ?A ? ?\?0 , and c(?0 ) = 0 when
?0 ? ?0 . The function c(?) is known as the approximate Bahadur slope (ABS) of the sequence Tn .
The quantifier ?approximate? comes from the use of the asymptotic null distribution instead of the
exact one [1]. Intuitively the slope c(?A ), for ?A ? ?\?0 , is the rate of convergence of p-values (i.e.,
1 ? F (Tn )) to 0, as n increases. The higher the slope, the faster the p-value vanishes, and thus the
lower the sample size required to reject H0 under ?A .
(1)
(2)
Approximate Bahadur Efficiency Given two sequences of test statistics, Tn and Tn having the
(1)
(2)
same ?(n) (see Theorem 10), the approximate Bahadur efficiency of Tn relative to Tn is defined
(1)
as E(?A ) := c(1) (?A )/c(2) (?A ) for ?A ? ?\?0 . If E(?A ) > 1, then Tn is asymptotically more
(2)
efficient than Tn in the sense of Bahadur, for the particular problem specified by ?A ? ?\?0 . We
now give approximate Bahadur slopes for two sequences of linear time test statistics: the proposed
c2 discussed in Section 2.
\2 , and the LKS test statistic ?nS
nFSSD
l
\2 is c(FSSD) := FSSD2 /?1 , where ?1 is the
Theorem 5. The approximate Bahadur slope of nFSSD
>
maximum eigenvalue of ?p := Ex?p [? (x)? (x)] and ?(n) = n.
? c2
Theorem 6. The approximate Bahadur slope of the linear-time kernel Stein (LKS) test statistic nS
l
2
0
1 [Eq hp (x,x )]
(LKS)
is c
= 2 E h2 (x,x0 ) , where hp is the U-statistic kernel of the KSD statistic, and ?(n) = n.
]
p[ p
To make these results concrete, we consider the setting where p = N (0, 1) andq = N (?q , 1).
We assume that both tests use the Gaussian kernel k(x, y) = exp ?(x ? y)2 /2?k2 , possibly with
different bandwidths. We write ?k2 and ?2 for the FSSD and LKS bandwidths, respectively. Under
these assumptions, the slopes given in Theorem 5 and Theorem 6 can be derived explicitly. The
full expressions of the slopes are given in Proposition 12 and Proposition 13 (in the appendix). By
[12, 13] (recalled as Theorem 10 in the supplement), the approximate Bahadur efficiency can be
computed by taking the ratio of the two slopes. The efficiency is given in Theorem 7.
Theorem 7 (Efficiency in the Gaussian mean shift problem). Let E1 (?q , v, ?k2 , ?2 ) be the approxic2 for the case where p = N (0, 1), q = N (? , 1),
\2 relative to ?nS
mate Bahadur efficiency of nFSSD
l
q
\2 ). Fix ? 2 = 1 for nFSSD
\2 . Then, for any ?q 6= 0,
and J = 1 (i.e., one test location v for nFSSD
k
for some v ? R, and for any ?2 > 0, we have E1 (?q , v, ?k2 , ?2 ) > 2.
When p = N (0, 1) and q = N (?q , 1) for ?q 6= 0, Theorem 7 guarantees that our FSSD test is
asymptotically at least twice as efficient as the LKS test in the Bahadur sense. We note that the
6
efficiency is conservative in the sense that ?k2 = 1 regardless of ?q . Choosing ?k2 dependent on ?q
will likely improve the efficiency further.
5
Experiments
In this section, we demonstrate the performance of the proposed test on a number of problems. The
primary goal is to understand the conditions under which the test can perform well.
p
Sensitivity to Local Differences We start by demonstrating that
q
2
the test power objective FSSD /?H1 captures local differences
FSSD
?
of p and q, and that interpretable features v are found. Consider a one-dimensional
problem in which p = N (0, 1) and
?
?4
?2 v ? 0 v ? 2
4
q = Laplace(0,
1/
2),
a
zero-mean
Laplace
distribution
with
scale
?
parameter 1/ 2. These parameters are chosen so that p and q have Figure 1: The power criterion
the same mean and variance. Figure 1 plots the (rescaled) objective FSSD2 /?H1 as a function of
as a function of v. The objective illustrates that the best features test location v.
(indicated by v ? ) are at the most discriminative locations.
2
H1
Test Power We next investigate the power of different tests on two problems:
?
Qd
1. Gaussian vs. Laplace: p(x) = N (x|0, Id ) and q(x) = i=1 Laplace(xi |0, 1/ 2) where the
dimension d will be varied. The two distributions have the same mean and variance. The main
characteristic of this problem is local differences of p and q (see Figure 1). Set n = 1000.
2. Restricted Boltzmann Machine (RBM):
p(x) is the marginal distribution of p(x, h) =
1
1
>
>
>
2
exp
x
Bh
+
b
x
+
c
x
?
kxk
,
where
x ? Rd , h ? {?1}dh is a random vector of
Z
2
hidden
variables, and Z is the normalization constant. The exact marginal density p(x) =
P
dh
terms.
h?{?1,1}dh p(x, h) is intractable when dh is large, since it involves summing over 2
Recall that the proposed test only requires the score function ?x log p(x) (not the normalization
constant), which can be computed in closed form in this case. In this problem, q is another RBM
where entries of the matrix B are corrupted by Gaussian noise. This was the problem considered in
[22]. We set d = 50 and dh = 40, and generate samples by n independent chains (i.e., n independent
samples) of blocked Gibbs sampling with 2000 burn-in iterations.
We evaluate the following six kernel-based nonparametric tests with ? = 0.05, all using the Gaussian
kernel. 1. FSSD-rand: the proposed FSSD test where the test locations set to random draws from
a multivariate normal distribution fitted to the data. The kernel bandwidth is set by the commonly
used median heuristic i.e., ?k = median({kxi ? xj k, i < j}). 2. FSSD-opt: the proposed FSSD
test where both the test locations and the Gaussian bandwidth are optimized (Section 3.2). 3. KSD:
the quadratic-time Kernel Stein Discrepancy test with the median heuristic. 4. LKS: the linear-time
version of KSD with the median heuristic. 5. MMD-opt: the quadratic-time MMD two-sample
test of [16] where the kernel bandwidth is optimized by grid search to maximize a power criterion
as described in [29]. 6. ME-opt: the linear-time mean embeddings (ME) two-sample test of [19]
where parameters are optimized. We draw n samples from p to run the two-sample tests (MMD-opt,
ME-opt). For FSSD tests, we use J = 5 (see Section A for an investigation of test power as J varies).
All tests with optimization use 20% of the sample size n for parameter tuning. Code is available at
https://github.com/wittawatj/kernel-gof.
Figure 2 shows the rejection rates of the six tests for the two problems, where each problem is
repeated for 200 trials, resampling n points from q every time. In Figure 2a (Gaussian vs. Laplace),
high performance of FSSD-opt indicates that the test performs well when there are local differences
between p and q. Low performance of FSSD-rand emphasizes the importance of the optimization
of FSSD-opt to pinpoint regions where p and q differ. The power of KSD quickly drops as the
dimension increases, which can be understood since KSD is the RKHS norm of a function witnessing
differences in p and q across the entire domain, including where these differences are small.
We next consider the case of RBMs. Following [22], b, c are independently drawn from the standard
multivariate normal distribution, and entries of B ? R50?40 are drawn with equal probability from
{?1}, in each trial. The density q represents another RBM having the same b, c as in p, and with all
entries of B corrupted by independent zero-mean Gaussian noise with standard deviation ?per . Figure
7
0.0
1
5
10
dimension d
15
1.0
0.5
0.0
0.00
ME-opt
300
0.75
0.50
0.25
0.00
0.02
0.04
0.06
Perturbation SD ?per
MMD-opt
Time (s)
0.5
LKS
Rejection rate
1.0
KSD
2000
Sample size n
Rejection rate
FSSD-rand
Rejection rate
Rejection rate
FSSD-opt
1.0
4000
200
100
0
1000
2000
3000
Sample size n
4000
0.5 (d) Runtime (RBM)
(a) Gaussian vs. Laplace. (b) RBM. n = 1000. Per- (c) RBM. ?per = 0.1. Pern = 1000.
turb all entries of B.
turb B1,1 .
0.0
Figure 2: Rejection rates of the six tests. The proposed linear-time FSSD-opt
has0.02
a comparable
or
0.00
0.04
0.06
higher test power in some cases than the quadratic-time KSD test.
Perturbation SD ?per
2b shows the test powers as ?per increases, for a fixed sample size n = 1000. We observe that all the
tests have correct false positive rates (type-I errors) at roughly ? = 0.05 when there is no perturbation
noise. In particular, the optimization in FSSD-opt does not increase false positive rate when H0 holds.
We see that the performance of the proposed FSSD-opt matches that of the quadratic-time KSD at
all noise levels. MMD-opt and ME-opt perform far worse than the goodness-of-fit tests when the
difference in p and q is small (?per is low), since these tests simply represent p using samples, and do
not take advantage of its structure.
The advantage of having O(n) runtime can be clearly seen when the problem is much harder,
requiring larger sample sizes to tackle. Consider a similar problem on RBMs in which the parameter
B ? R50?40 in q is given by that of p, where only the first entry B1,1 is perturbed by random
N (0, 0.12 ) noise. The results are shown in Figure 2c where the sample size n is varied. We observe
that the two two-sample tests fail to detect this subtle difference even with large sample size. The
test powers of KSD and FSSD-opt are comparable when n is relatively small. It appears that KSD
has higher test power than FSSD-opt in this case for large n. However, this moderate gain in the test
power comes with an order of magnitude more computation. As shown in Figure 2d, the runtime
of the KSD is much larger than that of FSSD-opt, especially at large n. In these problems, the
performance of the new test (even without optimization) far exceeds that of the LKS test. Further
simulation results can be found in Section B.
0.20
Interpretable Features In the
0.16
final simulation, we demonstrate
0.12
that the learned test locations are
0.08
informative in visualising where
0.04
the model does not fit the data
0.00
well. We consider crime data
?0.04
from the Chicago Police Depart?0.08
ment, recording n = 11957
locations (latitude-longitude co- (a) p = 2-component GMM.
(b) p = 10-component GMM
ordinates) of robbery events in
optimization objective as a function of
Chicago in 2016.3 We address Figure 3: Plots of the
2
the situation in which a model p test location v ? R in the Gaussian mixture model (GMM)
for the robbery location density is evaluation task.
given, and we wish to visualise
where it fails to match the data. We fit a Gaussian mixture model (GMM) with the expectationmaximization algorithm to a subsample of 5500 points. We then test the model on a held-out test set
of the same size to obtain proposed locations of relevant features v. Figure 3a shows the test robbery
locations in purple, the model with two Gaussian components in wireframe, and the optimization
objective for v as a grayscale contour plot (a red star indicates the maximum). We observe that the
2-component model is a poor fit to the data, particularly in the right tail areas of the data, as indicated
in dark gray (i.e., the objective is high). Figure 3b shows a similar plot with a 10-component GMM.
The additional components appear to have eliminated some mismatch in the right tail, however a
discrepancy still exists in the left region. Here, the data have a sharp boundary on the right side
following the geography of Chicago, and do not exhibit exponentially decaying Gaussian-like tails.
We note that tests based on a learned feature located at the maximum both correctly reject H0 .
3
Data can be found at https://data.cityofchicago.org.
8
Acknowledgement
WJ, WX, and AG thank the Gatsby Charitable Foundation for the financial support. ZSz was
financially supported by the Data Science Initiative. KF has been supported by KAKENHI Innovative
Areas 25120012.
References
[1] R. R. Bahadur. Stochastic comparison of tests. The Annals of Mathematical Statistics, 31(2):
276?295, 1960.
[2] L. Baringhaus and N. Henze. A consistent test for multivariate normality based on the empirical
characteristic function. Metrika, 35:339?348, 1988.
[3] J. Beirlant, L. Gy?rfi, and G. Lugosi. On the asymptotic normality of the l1 - and l2 -errors in
histogram density estimation. Canadian Journal of Statistics, 22:309?318, 1994.
[4] R. Bhatia. Matrix analysis, volume 169. Springer Science & Business Media, 2013.
[5] A. Bowman and P. Foster. Adaptive smoothing and density based tests of multivariate normality.
Journal of the American Statistical Association, 88:529?537, 1993.
[6] C. Carmeli, E. De Vito, A. Toigo, and V. Umanit?. Vector valued reproducing kernel Hilbert
spaces and universality. Analysis and Applications, 08(01):19?61, Jan. 2010.
[7] K. Chwialkowski, D. Sejdinovic, and A. Gretton. A wild bootstrap for degenerate kernel tests. In
NIPS, pages 3608?3616, 2014.
[8] K. Chwialkowski, A. Ramdas, D. Sejdinovic, and A. Gretton. Fast two-sample testing with
analytic representations of probability measures. In NIPS, pages 1981?1989, 2015.
[9] K. Chwialkowski, H. Strathmann, and A. Gretton. A kernel test of goodness of fit. In ICML,
pages 2606?2615, 2016.
[10] T. Epps and K. Singleton. An omnibus test for the two-sample problem using the empirical
characteristic function. Journal of Statistical Computation and Simulation, 26(3?4):177?203,
1986.
[11] J. Frank J. Massey. The Kolmogorov-Smirnov test for goodness of fit. Journal of the American
Statistical Association, 46(253):68?78, 1951.
[12] L. J. Gleser. On a measure of test efficiency proposed by R. R. Bahadur. 35(4):1537?1544,
1964.
[13] L. J. Gleser. The comparison of multivariate tests of hypothesis by means of Bahadur efficiency.
28(2):157?174, 1966.
[14] J. Gorham and L. Mackey. Measuring sample quality with Stein?s method. In NIPS, pages
226?234, 2015.
[15] J. Gorham and L. Mackey. Measuring sample quality with kernels. In ICML, pages 1292?1301.
PMLR, 06?11 Aug 2017.
[16] A. Gretton, K. M. Borgwardt, M. J. Rasch, B. Sch?lkopf, and A. Smola. A kernel two-sample
test. JMLR, 13:723?773, 2012.
[17] A. Gretton, D. Sejdinovic, H. Strathmann, S. Balakrishnan, M. Pontil, K. Fukumizu, and
B. K. Sriperumbudur. Optimal kernel choice for large-scale two-sample tests. In NIPS, pages
1205?1213. 2012.
[18] L. Gy?rfi and E. C. van der Meulen. A consistent goodness of fit test based on the total variation
distance. In G. Roussas, editor, Nonparametric Functional Estimation and Related Topics, pages
631?645, 1990.
[19] W. Jitkrittum, Z. Szab?, K. P. Chwialkowski, and A. Gretton. Interpretable Distribution Features
with Maximum Testing Power. In NIPS, pages 181?189. 2016.
[20] W. Jitkrittum, Z. Szab?, and A. Gretton. An adaptive test of independence with analytic kernel
embeddings. In ICML, pages 1742?1751. PMLR, 2017.
[21] C. Ley, G. Reinert, and Y. Swan. Stein?s method for comparison of univariate distributions.
Probability Surveys, 14:1?52, 2017.
9
[22] Q. Liu, J. Lee, and M. Jordan. A kernelized Stein discrepancy for goodness-of-fit tests. In
ICML, pages 276?284, 2016.
[23] J. Lloyd and Z. Ghahramani. Statistical model criticism using kernel two sample tests. In NIPS,
pages 829?837, 2015.
[24] B. Mityagin. The Zero Set of a Real Analytic Function. Dec. 2015. arXiv: 1512.07276.
[25] C. J. Oates, M. Girolami, and N. Chopin. Control functionals for Monte Carlo integration.
Journal of the Royal Statistical Society: Series B (Statistical Methodology), 79(3):695?718, 2017.
[26] M. L. Rizzo. New goodness-of-fit tests for Pareto distributions. ASTIN Bulletin: Journal of the
International Association of Actuaries, 39(2):691?715, 2009.
[27] R. J. Serfling. Approximation Theorems of Mathematical Statistics. John Wiley & Sons, 2009.
[28] I. Steinwart and A. Christmann. Support Vector Machines. Springer, New York, 2008.
[29] D. J. Sutherland, H.-Y. Tung, H. Strathmann, S. De, A. Ramdas, A. Smola, and A. Gretton.
Generative models and model criticism via optimized Maximum Mean Discrepancy. In ICLR,
2016.
[30] G. J. Sz?kely and M. L. Rizzo. A new test for multivariate normality. Journal of Multivariate
Analysis, 93(1):58?80, 2005.
[31] A. W. van der Vaart. Asymptotic Statistics. Cambridge University Press, 2000.
[32] Q. Zhang, S. Filippi, A. Gretton, and D. Sejdinovic. Large-scale kernel methods for independence testing. Statistics and Computing, pages 1?18, 2017.
10
| 6630 |@word trial:2 version:1 briefly:1 eliminating:1 norm:4 stronger:1 smirnov:1 c0:4 open:2 d2:5 seek:1 simulation:3 covariance:2 zolt:2 tr:1 harder:1 initial:1 liu:1 series:1 score:1 tuned:1 rkhs:6 com:3 gmail:2 universality:1 written:1 readily:1 john:1 numerical:1 chicago:3 informative:1 wx:1 analytic:10 plot:4 interpretable:5 drop:1 v:4 alone:1 resampling:1 prohibitive:1 metrika:1 mackey:2 generative:1 ith:1 core:1 normalising:1 detecting:1 location:16 org:1 zhang:1 mathematical:2 bowman:1 constructed:6 c2:6 become:2 initiative:1 prove:2 wild:1 x0:5 roughly:1 equipped:1 increasing:1 begin:1 bounded:1 medium:1 null:12 what:1 kg:1 gif:1 developed:1 finding:2 ag:1 guarantee:4 pseudo:1 every:1 tackle:1 runtime:3 k2:10 uk:1 control:1 partitioning:1 unit:4 appear:1 positive:4 generalised:1 understood:1 local:7 sutherland:1 sd:2 id:3 approximately:1 lugosi:1 burn:1 twice:1 studied:1 quantified:1 suggests:1 challenging:1 co:1 testing:16 practice:3 definite:1 bootstrap:2 procedure:2 jan:1 pontil:1 asymptotics:1 area:2 universal:4 empirical:6 reject:3 matching:1 convenient:1 get:1 operator:6 bh:1 map:1 go:2 regardless:3 straightforward:1 independently:2 convex:1 survey:1 simplicity:2 splitting:1 estimator:6 financial:1 population:1 stability:1 variation:1 laplace:6 annals:1 construction:1 play:1 suppose:1 exact:2 hypothesis:7 approximated:2 expensive:1 particularly:1 located:1 observed:4 role:1 tung:1 visualising:1 capture:2 region:3 wj:1 connected:2 decrease:3 rescaled:1 yk:1 pd:4 vanishes:1 pdj:2 complexity:1 vito:1 incur:1 efficiency:17 easily:2 plimn:1 kolmogorov:1 derivation:1 forced:1 fast:1 monte:2 has0:1 bhatia:1 choosing:1 h0:15 gorham:2 quite:1 whose:2 larger:3 dominating:1 heuristic:4 valued:1 tightness:1 otherwise:1 statistic:37 gi:9 g1:1 vaart:1 analyse:1 itself:1 final:1 advantage:3 sequence:4 indication:1 eigenvalue:8 ucl:4 propose:2 ment:1 product:1 relevant:1 realization:1 rapidly:1 baringhaus:1 poorly:1 degenerate:3 normalize:1 convergence:4 requirement:1 optimum:1 strathmann:3 ac:2 measured:1 expectationmaximization:1 aug:1 eq:3 longitude:1 kenji:1 christmann:1 indicate:1 implies:2 qd:1 differ:2 rasch:1 come:2 involves:1 drawback:1 correct:2 ley:1 girolami:1 stochastic:1 require:1 f1:1 fix:1 geography:1 investigation:1 proposition:10 anticipate:1 zoltan:1 opt:18 strictly:1 hold:3 around:1 sufficiently:3 considered:1 normal:5 exp:3 henze:1 sp2:2 claim:3 estimation:2 cole:1 sensitive:1 largest:2 ex0:3 successfully:2 fukumizu:3 clearly:1 always:1 gaussian:16 pn:4 avoid:1 derived:2 kakenhi:1 indicates:2 criticism:2 detect:3 rdj:1 sense:3 dependent:2 typically:1 entire:1 initially:1 kernelized:3 hidden:1 transformed:1 going:1 chopin:1 issue:2 arg:2 flexible:1 denoted:1 proposes:1 constrained:1 special:1 integration:2 smoothing:1 marginal:2 equal:1 construct:3 once:1 having:3 beach:1 sampling:1 eliminated:1 represents:2 icml:4 discrepancy:17 others:1 simplify:1 few:2 ksd:21 szabo:1 lebesgue:2 replacement:1 n1:2 ab:2 detection:1 interest:3 fd:1 plim:1 investigate:1 robbery:3 evaluation:1 reinert:1 mixture:2 yielding:1 held:1 chain:2 integral:1 limkxk:2 arthur:3 necessary:1 conduct:1 incomplete:1 divide:1 theoretical:3 fitted:2 increased:1 column:3 earlier:2 instance:1 tp:7 goodness:19 measuring:2 cost:15 introducing:2 deviation:1 entry:5 uniform:1 recognizing:1 varies:1 corrupted:2 perturbed:1 kxi:1 gd:1 st:1 density:11 borgwardt:1 sensitivity:1 international:1 kely:1 lee:1 vm:2 quickly:1 concrete:1 choose:1 possibly:1 actuary:1 worse:2 american:2 filippi:1 de:2 singleton:1 gy:2 star:1 lloyd:1 explicitly:2 depends:2 vi:3 performed:1 h1:23 root:1 closed:2 later:1 sup:2 red:1 start:3 hf:1 decaying:1 astin:1 slope:15 contribution:2 purple:1 ni:4 variance:4 characteristic:3 correspond:1 yield:2 lkopf:1 rejecting:3 reevaluated:1 emphasizes:1 carlo:2 j6:1 history:1 definition:4 against:2 sriperumbudur:1 nonetheless:1 rbms:2 proof:5 associated:2 rbm:6 gain:2 sampled:1 recall:1 jitkrittum:3 hilbert:3 subtle:3 reflecting:1 appears:1 higher:4 follow:1 methodology:1 rand:3 done:1 rejected:1 smola:2 steinwart:1 continuity:1 quality:2 reveal:1 indicated:3 gray:1 roussas:1 usa:1 omnibus:1 concept:1 unbiased:3 true:1 requiring:1 hence:2 regularization:3 uniquely:1 criterion:5 polytechnique:2 demonstrate:2 tn:14 performs:3 l1:1 meaning:1 novel:1 recently:3 fi:4 common:1 functional:1 empirically:1 ji:3 jp:1 exponentially:1 visualise:1 discussed:1 tail:3 volume:1 association:3 refer:2 blocked:1 cambridge:1 vec:2 gibbs:1 smoothness:1 rd:9 tuning:3 grid:1 mathematics:1 hp:6 dj:9 add:1 multivariate:8 hfi:2 swan:1 optimizing:1 moderate:1 apart:1 certain:1 yi:1 exploited:2 der:2 integrable:1 seen:1 minimum:1 greater:2 somewhat:1 additional:2 employed:2 ey:1 surely:2 determine:1 maximize:5 bochner:1 full:1 desirable:1 flatness:2 gretton:12 exceeds:3 match:3 faster:2 plug:3 long:4 gof:1 e1:2 converging:1 variant:1 essentially:1 expectation:4 arxiv:1 iteration:3 kernel:41 normalization:4 mmd:5 represent:1 achieved:1 histogram:1 sejdinovic:4 proposal:1 dec:1 addressed:1 median:4 limn:2 crucial:2 sch:1 ascent:3 recording:1 rizzo:2 chwialkowski:4 balakrishnan:1 jordan:1 wittawat:1 canadian:1 identically:1 embeddings:3 independence:6 fit:23 xj:5 zi:2 bandwidth:7 reduce:2 inner:1 idea:5 shift:3 expression:3 six:3 york:1 jj:2 remark:2 rfi:2 nonparametric:3 stein:28 dark:1 ph:3 generate:1 http:2 bahadur:21 disjoint:1 correctly:2 per:7 write:3 wireframe:1 wittawatj:2 key:2 threshold:7 demonstrating:1 achieving:1 drawn:8 d3:2 costing:1 pj:1 gmm:5 v1:3 massey:1 asymptotically:6 run:1 inverse:1 powerful:1 almost:2 throughout:1 sobolev:1 draw:4 epps:1 appendix:2 comparable:4 quadratic:13 precisely:1 argument:1 innovative:1 relatively:1 ball:1 poor:1 carmeli:1 smaller:1 across:1 son:1 agreeing:1 serfling:1 intuitively:1 quantifier:1 restricted:1 taken:1 computationally:2 r50:2 turn:1 fail:1 needed:2 toigo:1 available:2 observe:5 appropriate:2 simulating:1 pmlr:2 zi2:1 alternative:8 jn:1 denotes:1 running:1 include:1 remaining:1 outweighs:1 ism:1 giving:1 quantile:2 especially:1 ghahramani:1 society:1 objective:7 added:1 quantity:1 depart:1 parametric:1 strategy:1 primary:1 exhibit:1 gradient:3 financially:1 iclr:1 distance:1 thank:1 me:6 topic:1 assuming:1 besides:2 code:1 providing:1 minimizing:1 ratio:2 difficult:1 unfortunately:1 statement:1 frank:1 negative:1 stated:1 boltzmann:1 unknown:3 perform:3 markov:1 finite:5 mate:1 situation:1 witness:4 varied:2 reproducing:4 perturbation:3 arbitrary:1 police:1 sharp:1 ordinate:1 required:4 specified:1 optimized:5 z1:1 crime:1 recalled:1 distinction:1 learned:2 nip:7 address:1 able:1 mismatch:1 latitude:1 summarize:1 max:2 including:1 oates:1 royal:1 power:31 critical:1 event:1 business:1 rely:1 analyticity:1 normality:4 improve:1 github:1 x2i:1 meulen:1 imply:1 pvalue:1 prior:1 acknowledgement:1 l2:1 kf:8 asymptotic:13 relative:8 foundation:2 h2:1 degree:1 sufficient:1 consistent:6 xp:1 foster:1 editor:1 charitable:1 pareto:1 supported:2 side:1 exponentiated:2 understand:2 institute:1 taking:1 bulletin:1 distributed:1 van:2 boundary:1 dimension:7 cumulative:1 contour:1 collection:1 adaptive:3 commonly:1 offit:1 far:3 functionals:1 approximate:12 preferred:1 dealing:1 sz:1 overfitting:1 summing:1 b1:2 xi:16 discriminative:1 grayscale:1 continuous:2 search:1 ph1:6 learn:2 lks:12 concatenates:1 ca:1 obtaining:2 beirlant:1 domain:3 vj:9 sp:8 significance:1 main:2 noise:5 arise:1 subsample:1 n2:2 ramdas:2 repeated:1 xu:1 cubic:1 gatsby:5 wiley:1 n:3 fails:2 explicit:3 wish:1 pinpoint:1 jmlr:1 theorem:23 er:1 decay:1 evidence:1 burden:1 exists:2 intractable:1 false:5 importance:1 cmap:1 supplement:1 magnitude:2 illustrates:1 kx:1 rejection:9 simply:1 likely:1 covx:1 univariate:1 visual:1 kxk:1 springer:2 relies:1 dh:5 cdf:1 goal:3 presentation:1 consequently:1 replace:1 specifically:1 szab:4 lemma:2 conservative:2 called:1 total:1 multiquadric:1 support:2 witnessing:2 evaluate:4 mcmc:1 tested:1 ex:15 |
6,225 | 6,631 | Cortical microcircuits as
gated-recurrent neural networks
Rui Ponte Costa?
Centre for Neural Circuits and Behaviour
Dept. of Physiology, Anatomy and Genetics
University of Oxford, Oxford, UK
[email protected]
Yannis M. Assael?
Dept. of Computer Science
University of Oxford, Oxford, UK
and DeepMind, London, UK
[email protected]
Brendan Shillingford?
Dept. of Computer Science
University of Oxford, Oxford, UK
and DeepMind, London, UK
[email protected]
Nando de Freitas
DeepMind
London, UK
[email protected]
Tim P. Vogels
Centre for Neural Circuits and Behaviour
Dept. of Physiology, Anatomy and Genetics
University of Oxford, Oxford, UK
[email protected]
Abstract
Cortical circuits exhibit intricate recurrent architectures that are remarkably similar
across different brain areas. Such stereotyped structure suggests the existence of
common computational principles. However, such principles have remained largely
elusive. Inspired by gated-memory networks, namely long short-term memory
networks (LSTMs), we introduce a recurrent neural network in which information
is gated through inhibitory cells that are subtractive (subLSTM). We propose a
natural mapping of subLSTMs onto known canonical excitatory-inhibitory cortical
microcircuits. Our empirical evaluation across sequential image classification
and language modelling tasks shows that subLSTM units can achieve similar
performance to LSTM units. These results suggest that cortical circuits can be
optimised to solve complex contextual problems and proposes a novel view on
their computational function. Overall our work provides a step towards unifying
recurrent networks as used in machine learning with their biological counterparts.
1
Introduction
Over the last decades neuroscience research has collected enormous amounts of data on the architecture and dynamics of cortical circuits, unveiling complex but stereotypical structures across
the neocortex (Markram et al., 2004; Harris and Mrsic-Flogel, 2013; Jiang et al., 2015). One of
the most prevalent features of cortical nets is their laminar organisation and their high degree of
recurrence, even at the level of local (micro-)circuits (Douglas et al., 1995; Song et al., 2005; Harris
and Mrsic-Flogel, 2013; Jiang et al., 2015) (Fig. 1a). Another key feature of cortical circuits is
the detailed and tight balance of excitation and inhibition, which has received growing support
* These authors contributed equally to this work.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
both at the experimental (Froemke et al., 2007; Xue et al., 2014; Froemke, 2015) and theoretical
level (van Vreeswijk and Sompolinsky, 1996; Brunel, 2000; Vogels and Abbott, 2009; Hennequin
et al., 2014, 2017). However, the computational processes that are facilitated by these architectures
and dynamics are still elusive. There remains a fundamental disconnect between the underlying
biophysical networks and the emergence of intelligent and complex behaviours.
Artificial recurrent neural networks (RNNs), on the other hand, are crafted to perform specific
computations. In fact, RNNs have recently proven very successful at solving complex tasks such
as language modelling, speech recognition, and other perceptual tasks (Graves, 2013; Graves et al.,
2013; Sutskever et al., 2014; van den Oord et al., 2016; Assael et al., 2016). In these tasks, the input
data contains information across multiple timescales that needs to be filtered and processed according
to its relevance. The ongoing presentation of stimuli makes it difficult to learn to separate meaningful
stimuli from background noise (Hochreiter et al., 2001; Pascanu et al., 2012). RNNs, and in particular
gated-RNNs, can solve this problem by maintaining a representation of relevant input sequences until
needed, without interference from new stimuli. In principle, such protected memories conserve past
inputs and thus allow back-propagation of errors further backwards in time (Pascanu et al., 2012).
Because of their memory properties, one of the first and most successful types of gated-RNNs was
named ?long short-term memory networks? (LSTMs, Hochreiter and Schmidhuber (1997), Fig. 1c).
Here we note that the architectural features of LSTMs overlap closely with known cortical structures,
but with a few important differences with regard to the mechanistic implementation of gates in a
cortical network and LSTMs (Fig. 1b). In LSTMs, the gates control the memory cell as a multiplicative
factor, but in biological networks, the gates, i.e. inhibitory neurons, act (to a first approximation)
subtractively ? excitatory and inhibitory (EI) currents cancel each other linearly at the level of
the postsynaptic membrane potential (Kandel et al., 2000; Gerstner et al., 2014). Moreover, such a
subtractive inhibitory mechanism must be well balanced (i.e. closely match the excitatory input) to act
as a gate to the inputs in the ?closed? state, without perturbing activity flow with too much inhibition.
Previous models have explored gating in subtractive excitatory and inhibitory balanced networks
(Vogels and Abbott, 2009; Kremkow et al., 2010), but without a clear computational role. On the other
hand, predictive coding RNNs with EI features have been studied (Bastos et al., 2012; Deneve and
Machens, 2016), but without a clear match to state-of-the-art machine learning networks. Regarding
previous neuroscientific interpretations of LSTMs, there have been suggestions of LSTMs as models
of working memory and different brain areas (e.g. prefrontal cortex, basal ganglia and hippocampus)
(O?Reilly and Frank, 2006; Krueger and Dayan, 2009; Cox and Dean, 2014; Marblestone et al., 2016;
Hassabis et al., 2017; Bhalla, 2017), but without a clear interpretation of the individual components
of LSTMs and a specific mapping to known circuits.
We propose to map the architecture and function of LSTMs directly onto cortical circuits, with
gating provided by lateral subtractive inhibition. Our networks have the potential to exhibit the
excitation-inhibition balance observed in experiments (Douglas et al., 1989; Bastos et al., 2012;
Harris and Mrsic-Flogel, 2013) and yield simpler gradient propagation than multiplicative gating.
We study these dynamics through our empirical evaluation showing that subLSTMs achieve similar
performance to LSTMs in the Penn Treebank and Wikitext-2 language modelling tasks, as well
as pixelwise sequential MNIST classification. By transferring the functionality of LSTMs into a
biologically more plausible network, our work provides testable hypotheses for the most recently
emerging, technologically advanced experiments on the functionality of entire cortical microcircuits.
2
Biological motivation
The architecture of LSTM units, with their general feedforward structure aided by additional recurrent
memory and controlled by lateral gates, is remarkably similar to the columnar architecture of cortical
circuits (Fig. 1). The central element in LSTMs and similar RNNs is the memory cell, which we
hypothesise to be implemented by local recurrent networks of pyramidal cells in layer-5. This is in
line with previous studies showing a relatively high level of recurrence and non-random connectivity
between pyramidal cells in layer-5 (Douglas et al., 1995; Thomson et al., 2002; Song et al., 2005).
Furthermore, layer-5 pyramidal networks display rich activity on (relatively) long time scales in
vivo (Barth? et al., 2009; Sakata and Harris, 2009; Harris and Mrsic-Flogel, 2013), consistent with
LSTM-like function. Slow memory decay in these networks can be controlled through short- (York
and van Rossum, 2009; Costa et al., 2013, 2017a) and long-term synaptic plasticity (Abbott and
2
Nelson, 2000; Senn et al., 2001; Pfister and Gerstner, 2006; Zenke et al., 2015; Costa et al., 2015,
2017a,b) at recurrent excitatory synapses.
The gates that protect a given memory in LSTMs can be mapped onto lateral inhibitory inputs in
cortical circuits. We propose that, similar to LSTMs, the input gate is implemented by inhibitory
neurons in layer-2/3 (or layer-4; Fig. 1a). Such lateral inhibition is consistent with the canonical view
of microcircuits (Douglas et al., 1989; Bastos et al., 2012; Harris and Mrsic-Flogel, 2013) and sparse
sensory-evoked responses in layer-2/3 (Sakata and Harris, 2009; Harris and Mrsic-Flogel, 2013).
In the brain, such inhibition is believed to originate from (parvalbumin) basket cells, providing a
near-exact balanced inhibitory counter signal to a given excitatory feedforward inputs (Froemke et al.,
2007; Xue et al., 2014; Froemke, 2015). Excitatory and inhibitory inputs thus cancel each other and
arriving signals are ignored by default. Consequently, any activity within the downstream memory
network remains largely unperturbed, unless it is altered through targeted modulation of the inhibitory
activity (Harris and Mrsic-Flogel, 2013; Vogels and Abbott, 2009; Letzkus et al., 2015). Similarly,
the memory cell itself can only affect the output of the LSTM when its activity is unaccompanied
by congruent inhibition (mapped onto layer-6 in the same microcircuit or 2/3 (or 4) in a separate
microcircuit), i.e. when lateral inhibition is turned down and the gate is open.
b
cortical circuit
IN
Layer 5
ct
memory
cell
f
PCs
L4 or 2/3
=
baseline
strong inh.
100
inh. = exc.
2
3
4
0
strong inh.
inh. = exc.
1
2
input
3
4
xt,ht-1
unit j
xt,ht-1 (input)
multiplicative gating
200
we
a
100
xt,ht-1
it
zt
unit j
baseline
e
at
kg
ea
w
100
0
xt,ht-1
ft
cell
xt,ht-1
it
subtractive gating
200
baseline
1
ct
xt,ht-1 (input)
weak inh.
200
0
=1
zt
unit j
ht (output)
ot
f
cell
PC
LSTM
xt,ht-1
=1
-
IN
input
output rate (Hz)
-
PC
ht (output)
ot
h.
L6, L4 or 2/3
output
=
c
subLSTM
k in
a
strong gate
closed gate
1
2
3
4
Figure 1: Biological and artificial gated recurrent neural networks. (a) Example unit of a simplified
cortical recurrent neural network. Sensory (or downstream) input arrives at pyramidal cells in
layer-2/3 (L2/3, or layer-4), which is then fed onto memory cells (recurrently connected pyramidal
cells in layer-5). The memory decays with a decay time constant f . Input onto layer-5 is balanced
out by inhibitory basket cells (BC). The balance is represented by the diagonal ?equal? connection.
The output of the memory cell is gated by basket cells at layer-6, 2/3 or 4 within the same area
(or at an upstream brain area). (b) Implementation of (a), following a similar notation to LSTM
units, but with it and ot as the input and output subtractive gates. Dashed connections represent the
potential to have a balance between excitatory and inhibitory input (weights are set to 1) (c) LSTM
recurrent neural network cell (see main text for details). The plots bellow illustrate the different
gating modes: (a) using a simple current-based noisy leaky-integrate-and-fire neuron (capped to
200Hz) with subtractive inhibition; (b) sigmoidal activation functions with subtractive gating; (c)
sigmoidal activation functions with multiplicative gating. Output rate represents the number of spikes
per second (Hz) as in biological circuits.
3
2.1
Why subtractive neural integration?
When a presynaptic cell fires, neurotransmitter is released by its synaptic terminals. The neurotransmitter is subsequently bound by postsynaptic receptors where it prompts a structural change of an ion
channel to allow the flow of electrically charged ions into or out of the postsynaptic cell. Depending
on the receptor type, the ion flux will either increase (depolarise) or decrease (hyperpolarise) the
postsynaptic membrane potential. If sufficiently depolarising ?excitatory? input is provided, the
postsynaptic potential will reach a threshold and fire a stereotyped action potential (?spike?, Kandel
et al. (2000)). This behaviour can be formalised as a RC?circuit (R = resistance, C = capacitance),
which follows Ohm?s laws u = RI and yields the standard leaky-integrate-and-fire neuron model
(Gerstner and Kistler, 2002), ?m u? = ?u + RIexc ? RIinh , where ?m = RC is the membrane time
constant, and Iexc and Iinh are the excitatory and inhibitory (hyperpolarizing) synaptic input currents,
respectively.
Action potentials are initiated in this standard model (Brette and Gerstner, 2005; Gerstner et al.,
2014)) when the membrane potential hits a hard threshold ?. They are modelled as a momentary
pulse and a subsequent reset to a resting potential. Neuronal excitation and inhibition have opposite
effects, such that inhibitory inputs acts linearly and subtractively on the membrane potential.
The leaky-integrate-and-fire model can be approximated at the level of firing rates as rate ?
?1
R(Iexc ?Iinh )
(see Fig. 1a for the input-output response; Gerstner and Kistler (2002)),
?m ln R(I
exc ?Iinh )??
which we used to demonstrate the impact of subtractive gating (Fig. 1b), and contrast it with multiplicative gating (Fig. 1c).
This firing-rate approximation forms the basis for our gated-RNN model which has a similar subtractive behaviour and input-output function (cf. Fig. 1b; bottom). Moreover, the rate formulation also
allows a cleaner comparison to LSTM units and the use of existing machine learning optimisation
methods.
It could be argued that a different form of inhibition (shunting inhibition), which counteracts excitatory
inputs by decreasing the over all membrane resistance, has a characteristic multiplicative gating effect
on the membrane potential. However, when analysed at the level of the output firing rate its effect
becomes subtractive (Holt and Koch, 1997; Prescott and De Koninck, 2003). This is consistent with
our approach in that our model is framed at the firing-rate level (rather than at the level of membrane
potentials).
3
Subtractive-gated long short-term memory
In an LSTM unit (Hochreiter and Schmidhuber, 1997; Greff et al., 2015) the access to the memory
cell ct is controlled by an input gate it (see Fig.1c). At the same time a forget gate ft controls
the decay of this memory1 , and the output gate ot controls whether the content of the memory
cell ct is transmitted to the rest of the network. A LSTM network consists of many LSTM units,
each containing its own memory cell ct , input it , forget ft and output ot gates. The LSTM state
is described as ht = f (xt , ht?1 , it , ft , ot ) and the unit follows the dynamics given in the middle
column below.
LSTM
subLSTM
[ft , ot , it ]T =
?(W xt + Rht?1 + b),
?(W xt + Rht?1 + b),
zt =
tanh(W xt + Rht?1 + b),
?(W xt + Rht?1 + b),
ct =
ct?1 ft + zt it ,
ct?1 ft + zt ? it ,
ht =
tanh(ct ) ot .
?(ct ) ? ot .
Here, ct is the memory cell (note the multiplicative control of the input gate), denotes element-wise
multiplication and zt is the new weighted input given with xt and ht?1 being the input vector and
recurrent input from other LSTM units, respectively. The overall output of the LSTM unit is then
1
Note that this leak is controlled by the input and recurrent units, which may be biologically unrealistic.
4
computed as ht . LSTM networks can have multiple layers with millions of parameters (weights and
biases), which are typically trained using stochastic gradient descent in a supervised setting. Above,
the parameters are W , R and b. The multiple gates allow the network to adapt the flow of information
depending on the task at hand. In particular, they enable writing to the memory cell (controlled by
input gate, it ), adjusting the timescale of the memory (controlled by forget gate, ft ) and exposing the
memory to the network (controlled by output gate, ot ). The combined effect of these gates makes it
possible for LSTM units to capture temporal (contextual) dependencies across multiple timescales.
Here, we introduce and study a new RNN unit, subLSTM. SubLSTM units are a mapping of LSTMs
onto known canonical excitatory-inhibitory cortical microcircuits (Douglas et al., 1995; Song et al.,
2005; Harris and Mrsic-Flogel, 2013). Similarly, subLSTMs are defined as ht = f (xt , ht?1 , it , ft , ot )
(Fig. 1b), however here the gating is subtractive rather than multiplicative. A subLSTM is defined by a
memory cell ct , the transformed input zt and the input gate it . In our model we use a simplified notion
of balance in the gating (?ztj ? ?ijt ) (for the jth unit), where ? = 1. 2 For the memory forgetting we
consider two options: (i) controlled by gates (as in an LSTM unit) as ft = ?(W xt + Rht?1 + b) or
(ii) a more biologically plausible learned simple decay [0, 1], referred to in the results as fix-subLSTM.
Similarly to its input, subLSTM?s output ht is also gated through a subtractive output gate ot (see
equations above). We evaluated different activation functions and sigmoidal transformations had the
highest performance.
The key differences to other gated-RNNs is in the subtractive inhibitory gating (it and ot ) that has
the potential to be balanced with the excitatory input (zt and ct , respectively; Fig. 1b). See below a
more detailed comparison of the different gating modes.
3.1
Subtractive versus multiplicative gating in RNNs
The key difference between subLSTMs and LSTMs lies in the implementation of the gating mechanism. LSTMs typically use a multiplicative factor to control the amplitude of the input signal.
SubLSTMs use a more biologically plausible interaction of excitation and inhibition. An important
consequence of subtractive gating is the potential for an improved gradient flow backwards towards
the input layers. To illustrate this we can compare the gradients for the subLSTMs and LSTMs in a
simple example.
First, we review the derivatives of the loss with respect to the various components of the subLSTM,
using notation based on (Greff et al., 2015). In this notation, ?a represents the derivative of the loss
def
with respect to a, and ?t = dloss
dht , the error from the layer above. Then by chain rule we have:
?ht = ?t
?ot = ??ht ? 0 (ot )
?ct = ?ht ? 0 (ct ) + ?ct+1 ft+1
?f t = ?ct ct?1 ? 0 (f t )
?it = ??ct ? 0 (it )
?zt = ?ct ? 0 (zt )
For comparison, the corresponding derivatives for an LSTM unit are given by:
?ht = ?t
?ot = ht tanh(ct ) ? 0 (ot )
?ct = ht ot tanh0 (ct ) + ?ct+1 ft+1
?f t = ?ct ct?1 ? 0 (f t )
?it = ?ct zt ? 0 (it )
?zt = ?ct it tanh0 (zt )
where ?(?) is the sigmoid activation function and the overlined variables ct , f t , etc. are the preactivation values of a gate or input transformation (e.g. ot = Wo xt + Ro ht?1 + bo for the output
2
These weights could also be optimised, but for this model we decided to keep the number of parameters to a
minimum for simplicity and ease of comparison with LSTMs.
5
gate of a subLSTM). Note that compared to the those of an LSTM, subLSTMs provide a simpler
gradient with fewer multiplicative factors.
Now, the LSTMs weights Wz of the input transformation z are updated according to
?Wz =
T X
T
X
t=0 t0 =t
?t0
?ht0
?ct ?zt
???
,
?ct0
?zt ?Wz
(1)
where T is the total number of temporal steps and the ellipsis abbreviates the recurrent gradient
paths through time, containing the path backwards through time via hs and cs for t ? s ? t0 . For
simplicity of analysis, we ignore these recurrent connections as they are the same in LSTM and
subLSTM, and only consider the depth-wise path through the network; we call this tth timestep
depth-only contribution to the derivative (?Wz )t . For an LSTM, by this slight abuse of notation, we
have
?ht ?ct ?zt ?zt
(?Wz )t = ?t
?ct ?zt ?zt ?Wz
(2)
= ?t ot tanh0 (ct ) it tanh0 (zt ) x>
t ,
|{z}
|{z}
output gate
input gate
where tanh0 (?) is the derivative of tanh. Notice that when either of the input or output gates are set to
zero (closed), the corresponding contributions to the gradient are zero. For a network with subtractive
gating, the depth-only derivative contribution becomes
(3)
(?Wz )t = ?t ? 0 (ct ) ? 0 (zt ) x>
t ,
where ? 0 (?) is the sigmoid derivative. In this case, the input and output gates, ot and it , are not
present. As a result, the subtractive gates in subLSTMs do not (directly) impair error propagation.
4
Results
The aims of our work were two-fold. First, inspired by cortical circuits we aimed to propose a
biological plausible implementation of an LSTM unit, which would allow us to better understand
cortical architectures and their dynamics. To compare the performance of subLSTM units to LSTMs,
we first compared the learning dynamics for subtractive and multiplicative networks mathematically.
In a second step, we empirically compared subLSTM and fix-subLSTM with LSTM networks in
two tasks: sequential MNIST classification and word-level language modelling on Penn Treebank
(Marcus et al., 1993) and Wikitext-2 (Merity et al., 2016). The network weights are initialised with
Glorot initialisation (Glorot and Bengio, 2010), and LSTM units have an initial forget gate bias of 1.
We selected the number of units for fix-subLSTM such that the number of parameters is held constant
across experiments to facilitate fair comparison with LSTMs and subLSTMs.
4.1
Sequential MNIST
In the ?sequential? MNIST digit classification task, each digit image from the MNIST dataset is
presented to the RNN as a sequence of pixels (Le et al. (2015); Fig. 2a) We decompose the MNIST
images of 28?28 pixels into sequences of 784 steps. The network was optimised using RMSProp
with momentum (Tieleman and Hinton, 2012), a learning rate of 10?4 , one hidden layer and 100
hidden units. Our results show that subLSTMs achieves similar results to LSTMs (Fig. 2b). Our
results are comparable to previous results using the same task (Le et al., 2015) and RNNs.
4.2
Language modelling
Language modelling represents a more challenging task for RNNs, with both short and long-term
dependencies. RNN language models (RNN LMs) models the probability of text by autoregressively
predicting a sequence of words. Each timestep is trained to predict the following word; in other
words, we model the word sequence as a product of conditional multinoulli distributions. We evaluate
the RNN LMs by measuring their perplexity, defined for a sequence of n words as
perplexity = P (w1 , . . . , wn )?1/n .
6
(4)
27
97
.
.2
9
.9
6
97
97.5
95.0
92.5
90.0
LS
TM
su
bL
ST
fix
M
-s
ub
LS
TM
28x28
...
100.0
97
b
seq.
testing accuracy (%)
a
Figure 2: Comparison of LSTM and subLSTM networks for sequential pixel-by-pixel MNIST, using
100 hidden units. (a) Samples from MNIST dataset. We converted each matrix of 28?28 pixels into
a temporal sequence of 784 timesteps. (b) Classification accuracy on the test set. fix-subLSTM has a
fixed but learned forget gate.
We first used the Penn Treebank (PTB) dataset to train our model on word-level language modelling
(929k training, 73k validation and 82k test words; with a vocabulary of 10k words).
All RNNs tested have 2 hidden layers; backpropagation is truncated to 35 steps, and a batch size of 20.
To optimise the networks we used RMSProp with momentum. We also performed a hyperparameter
search on the validation set over input, output, and update dropout rates, the learning rate, and
weight decay. The hyperparameter search was done with Google Vizier, which performs black-box
optimisation using Gaussian process bandits and transfer learning. Tables 2 and 3 show the resulting
hyperparameters. Table 1 reports perplexity on the test set (Golovin et al., 2017). To understand how
subLSTMs scale with network size we varied the number of hidden units between 10, 100, 200 and
650.
We also tested the Wikitext-2 language modelling dataset based on Wikipedia articles. This dataset
is twice as large as the PTB dataset (2000k training, 217k validation and 245k test words) and also
features a larger vocabulary (33k words). Therefore, it is well suited to evaluate model performance
on longer term dependencies and reduces the likelihood of overfitting.
On both datasets, our results show that subLSTMs achieve perplexity similar to LSTMs (Table 1a
and 1b). Interestingly, the more biological plausible version of subLSTM (with a simple decay as
forget gates) achieves performance similar to or better than subLSTMs.
(a) Penn Treebank (PTB) test perplexity
(b) Wikitext-2 test perplexity
size
subLSTM
fix-subLSTM
LSTM
size
subLSTM
fix-subLSTM
LSTM
10
100
200
650
222.80
91.46
79.59
76.17
213.86
91.84
81.97
70.58
215.93
88.39
74.60
64.34
10
100
200
650
268.33
103.36
89.00
78.92
259.89
105.06
94.33
79.49
271.44
102.77
86.15
74.27
Table 1: Language modelling (word-level) test set perplexities on (a) Penn Treebank and
(b) Wikitext-2. The models have two layers and fix-subLSTM uses a fixed but learned forget
gate f = [0, 1] for each unit. The number of units for fix-subLSTM was chosen such that the number
of parameters were the same as those of (sub)LSTM to facilitate fair comparison. Size indicates the
number of units.
The number of hidden units for fix-subLSTM were selected such that the number of parameters were
the same as LSTM and subLSTM, facilitating fair comparison.
7
Model
hidden
units
input
dropout
output
dropout
update
dropout
learning
rate
weight
decay
LSTM
subLSTM
fix-subLSTM
10
10
11
0.026
0.012
0.009
0.047
0.045
0.043
0.002
0.438
0
0.01186
0.01666
0.01006
0.000020
0.000009
0.000029
LSTM
subLSTM
fix-subLSTM
100
100
115
0.099
0.392
0.194
0.074
0.051
0.148
0.015
0.246
0.042
0.00906
0.01186
0.00400
0.000532
0.000157
0.000218
LSTM
subLSTM
fix-subLSTM
200
200
230
0.473
0.337
0.394
0.345
0.373
0.472
0.013
0.439
0.161
0.00496
0.01534
0.00382
0.000191
0.000076
0.000066
LSTM
subLSTM
fix-subLSTM
650
650
750
0.607
0.562
0.662
0.630
0.515
0.730
0.083
0.794
0.530
0.00568
0.00301
0.00347
0.000145
0.000227
0.000136
Table 2: Penn Treebank hyperparameters.
Model
hidden
units
input
dropout
output
dropout
update
dropout
learning
rate
weight
decay
LSTM
subLSTM
fix-subLSTM
10
10
11
0.015
0.002
0.033
0.039
0.030
0.070
0
0.390
0.013
0.01235
0.00859
0.00875
0
0.000013
0
LSTM
subLSTM
fix-subLSTM
100
100
115
0.198
0.172
0.130
0.154
0.150
0.187
0.002
0.009
0
0.01162
0.00635
0.00541
0.000123
0.000177
0.000172
LSTM
subLSTM
fix-subLSTM
200
200
230
0.379
0.342
0.256
0.351
0.269
0.273
0
0.018
0
0.00734
0.00722
0.00533
0.000076
0.000111
0.000160
LSTM
subLSTM
fix-subLSTM
650
650
750
0.572
0.633
0.656
0.566
0.567
0.590
0.071
0.257
0.711
0.00354
0.00300
0.00321
0.000112
0.000142
0.000122
Table 3: Wikitext-2 hyperparameters.
5
Conclusions & future work
Cortical microcircuits exhibit complex and stereotypical network architectures that support rich
dynamics, but their computational power and dynamics have yet to be properly understood. It is
known that excitatory and inhibitory neuron types interact closely to process sensory information
with great accuracy, but making sense of these interactions is beyond the scope of most contemporary
experimental approaches.
LSTMs, on the other hand, are a well-understood and powerful tool for contextual tasks, and their
structure maps intriguingly well onto the stereotyped connectivity of cortical circuits. Here, we
analysed if biologically constrained LSTMs (i.e. subLSTMs) could perform similarly well, and indeed,
such subtractively gated excitation-inhibition recurrent neural networks show promise compared
against LSTMs on benchmarks such as sequence classification and word-level language modelling.
While it is notable that subLSTMs could not outperform their traditional counterpart (yet), we hope
that our work will serve as a platform to discuss and develop ideas of cortical function and to establish
links to relevant experimental work on the role of excitatory and inhibitory neurons in contextual
learning (Froemke et al., 2007; Froemke, 2015; Poort et al., 2015; Pakan et al., 2016; Kuchibhotla
et al., 2017). In future work, it will be interesting to study how additional biological detail may affect
8
performance. Next steps should aim to include Dale?s principle (i.e. that a given neuron can only
make either excitatory or inhibitory connections, Strata and Harvey (1999)), and naturally focus
on the perplexing diversity of inhibitory cell types (Markram et al., 2004) and behaviour, such as
shunting inhibition and mixed subtractive and divisive control (Doiron et al., 2001; Mejias et al.,
2013; El Boustani and Sur, 2014; Seybold et al., 2015).
Overall, given the success of multiplicative gated LSTMs, it will be most insightful to understand if
some of the biological tricks of cortical networks may give LSTMs a further performance boost.
Acknowledgements
We would like to thank Everton Agnes, ?a?glar G?l?ehre, Gabor Melis and Jake Stroud for helpful
comments and discussion. R.P.C. and T.P.V. were supported by a Sir Henry Dale Fellowship by the
Wellcome Trust and the Royal Society (WT 100000). Y.M.A. was supported by the EPSRC and the
Research Council UK (RCUK). B.S. was supported by the Clarendon Fund.
References
Abbott, L. F. and Nelson, S. B. (2000). Synaptic plasticity: taming the beast. Nature Neuroscience, 3:1178.
Assael, Y. M., Shillingford, B., Whiteson, S., and de Freitas, N. (2016). Lipnet: Sentence-level lipreading. arXiv
preprint arXiv:1611.01599.
Barth?, P., Curto, C., Luczak, A., Marguet, S. L., and Harris, K. D. (2009). Population coding of tone stimuli in
auditory cortex: dynamic rate vector analysis. European Journal of Neuroscience, 30(9):1767?1778.
Bastos, A. M., Usrey, W. M., Adams, R. A., Mangun, G. R., Fries, P., and Friston, K. J. (2012). Canonical
microcircuits for predictive coding. Neuron, 76(4):695?711.
Bhalla, U. S. (2017). Dendrites, deep learning, and sequences in the hippocampus. Hippocampus.
Brette, R. and Gerstner, W. (2005). Adaptive exponential integrate-and-fire model as an effective description of
neuronal activity. Journal of Neurophysiology, 94(5):3637.
Brunel, N. (2000). Dynamics of Sparsely Connected Networks of Excitatory and Inhibitory Spiking Neurons.
Journal of Computational Neuroscience, 8(3):183?208.
Costa, R. P., Froemke, R. C., Sjostrom, P. J., and van Rossum, M. C. W. (2015). Unified pre- and postsynaptic
long-term plasticity enables reliable and flexible learning. eLife, 4:e09457.
Costa, R. P., Mizusaki, B. E. P., Sjostrom, P. J., and van Rossum, M. C. W. (2017a). Functional consequences of
pre- and postsynaptic expression of synaptic plasticity. Philosophical transactions of the Royal Society of
London. Series B, Biological sciences, 372(1715):20160153.
Costa, R. P., Padamsey, Z., D?amour, J. A., Emptage, N. J., Froemke, R. C., and Vogels, T. P. (2017b). Synaptic
Transmission Optimization Predicts Expression Loci of Long-Term Plasticity. Neuron, 96(1):177?189.e7.
Costa, R. P., Sjostrom, P. J., and van Rossum, M. C. W. (2013). Probabilistic inference of short-term synaptic
plasticity in neocortical microcircuits. Frontiers in Computational Neuroscience, 7:75.
Cox, D. D. and Dean, T. (2014). Neural Networks and Neuroscience-Inspired Computer Vision. Current Biology,
24(18):R921?R929.
Deneve, S. and Machens, C. K. (2016). Efficient codes and balanced networks. Nature Neuroscience, 19(3):375?
382.
Doiron, B., Longtin, A., Berman, N., and Maler, L. (2001). Subtractive and divisive inhibition: effect of
voltage-dependent inhibitory conductances and noise. Neural Computation, 13(1):227?248.
Douglas, R., Koch, C., Mahowald, M., Martin, K., and Suarez, H. (1995). Recurrent excitation in neocortical
circuits. Science, 269(5226):981?985.
Douglas, R. J., Martin, K. A. C., and Whitteridge, D. (1989). A Canonical Microcircuit for Neocortex. Neural
Computation, 1(4):480?488.
El Boustani, S. and Sur, M. (2014). Response-dependent dynamics of cell-specific inhibition in cortical networks
in vivo. Nature Communications, 5:5689.
9
Froemke, R. C. (2015). Plasticity of cortical excitatory-inhibitory balance. Annual Review of Neuroscience,
38(1):195?219.
Froemke, R. C., Merzenich, M. M., and Schreiner, C. E. (2007). A synaptic memory trace for cortical receptive
field plasticity. Nature.
Gerstner, W. and Kistler, W. M. (2002). Spiking Neuron Models. Single Neurons, Populations, Plasticity.
Cambridge University Press.
Gerstner, W., Kistler, W. M., Naud, R., and Paninski, L. (2014). Neuronal Dynamics. From Single Neurons to
Networks and Models of Cognition. Cambridge University Press.
Glorot, X. and Bengio, Y. (2010). Understanding the difficulty of training deep feedforward neural networks.
In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, pages
249?256.
Golovin, D., Solnik, B., Moitra, S., Kochanski, G., Karro, J., and Sculley, D. (2017). Google vizier: A service for
black-box optimization. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge
Discovery and Data Mining, pages 1487?1495. ACM.
Graves, A. (2013). Generating Sequences With Recurrent Neural Networks. arXiv.org.
Graves, A., Mohamed, A.-r., and Hinton, G. (2013). Speech recognition with deep recurrent neural networks.
arXiv preprint arXiv:1303.5778.
Greff, K., Srivastava, R. K., Koutn?k, J., Steunebrink, B. R., and Schmidhuber, J. (2015). LSTM: A Search
Space Odyssey. arXiv.org.
Harris, K. D. and Mrsic-Flogel, T. D. (2013). Cortical connectivity and sensory coding. Nature, 503(7474):51?58.
Hassabis, D., Kumaran, D., Summerfield, C., and Botvinick, M. (2017). Neuroscience-Inspired Artificial
Intelligence. Neuron, 95(2):245?258.
Hennequin, G., Agnes, E., and Vogels, T. P. (2017). Inhibitory Plasticity: Balance, Control, and Codependence.
Annual Review of Neuroscience, 40(1):557?579.
Hennequin, G., Vogels, T. P., and Gerstner, W. (2014). Optimal Control of Transient Dynamics in Balanced
Networks Supports Generation of Complex Movements. Neuron, 82(6):1394?1406.
Hochreiter, S., Bengio, Y., Frasconi, P., and Schmidhuber, J. (2001). Gradient flow in recurrent nets: the difficulty
of learning long-term dependencies.
Hochreiter, S. and Schmidhuber, J. (1997). Long short-term memory. Neural Computation, 9(8):1735?1780.
Holt, G. R. and Koch, C. (1997). Shunting inhibition does not have a divisive effect on firing rates. Neural
Computation, 9(5):1001?1013.
Jiang, X., Shen, S., Cadwell, C. R., Berens, P., Sinz, F., Ecker, A. S., Patel, S., and Tolias, A. S. (2015). Principles
of connectivity among morphologically defined cell types in adult neocortex. Science, 350(6264):aac9462?
aac9462.
Kandel, E. R., Schwartz, J. H., Jessell, T. M., and Siegelbaum, S. A. (2000). Principles of neural science.
Kremkow, J., Aertsen, A., and Kumar, A. (2010). Gating of signal propagation in spiking neural networks by
balanced and correlated excitation and inhibition. The Journal of neuroscience, 30(47):15760?15768.
Krueger, K. A. and Dayan, P. (2009). Flexible shaping: How learning in small steps helps. Cognition,
110(3):380?394.
Kuchibhotla, K. V., Gill, J. V., Lindsay, G. W., Papadoyannis, E. S., Field, R. E., Sten, T. A. H., Miller, K. D.,
and Froemke, R. C. (2017). Parallel processing by cortical inhibition enables context-dependent behavior.
Nature Neuroscience, 20(1):62?71.
Le, Q. V., Jaitly, N., and Hinton, G. E. (2015). A Simple Way to Initialize Recurrent Networks of Rectified
Linear Units. arXiv.org.
Letzkus, J. J., Wolff, S., and L?thi, A. (2015). Disinhibition, a Circuit Mechanism for Associative Learning and
Memory. Neuron.
10
Marblestone, A. H., Wayne, G., and Kording, K. P. (2016). Toward an Integration of Deep Learning and
Neuroscience. Frontiers in Computational Neuroscience, 10:94.
Marcus, M. P., Marcinkiewicz, M. A., and Santorini, B. (1993). Building a large annotated corpus of English:
the penn treebank. Computational Linguistics, 19(2):313?330.
Markram, H., Toledo-Rodriguez, M., Wang, Y., Gupta, A., Silberberg, G., and Wu, C. (2004). Interneurons of
the neocortical inhibitory system. Nature Reviews Neuroscience, 5(10):793?807.
Mejias, J. F., Kappen, H. J., Longtin, A., and Torres, J. J. (2013). Short-term synaptic plasticity and heterogeneity
in neural systems. 1510:185.
Merity, S., Xiong, C., Bradbury, J., and Socher, R. (2016). Pointer Sentinel Mixture Models. arXiv.org.
O?Reilly, R. C. and Frank, M. J. (2006). Making working memory work: a computational model of learning in
the prefrontal cortex and basal ganglia. Neural Computation, 18(2):283?328.
Pakan, J. M., Lowe, S. C., Dylda, E., Keemink, S. W., Currie, S. P., Coutts, C. A., Rochefort, N. L., and
Mrsic-Flogel, T. D. (2016). Behavioral-state modulation of inhibition is context-dependent and cell type
specific in mouse visual cortex. eLife, 5:e14985.
Pascanu, R., Mikolov, T., and Bengio, Y. (2012). On the difficulty of training Recurrent Neural Networks.
arXiv.org.
Pfister, J.-P. and Gerstner, W. (2006). Triplets of spikes in a model of spike timing-dependent plasticity. Journal
of Neuroscience, 26(38):9673?9682.
Poort, J., Khan, A. G., Pachitariu, M., Nemri, A., Orsolic, I., Krupic, J., Bauza, M., Sahani, M., Keller,
G. B., Mrsic-Flogel, T. D., and Hofer, S. B. (2015). Learning Enhances Sensory and Multiple Non-sensory
Representations in Primary Visual Cortex. Neuron, 86(6):1478?1490.
Prescott, S. A. and De Koninck, Y. (2003). Gain control of firing rate by shunting inhibition: roles of synaptic
noise and dendritic saturation. Proc. Natl. Acad. Sci. USA, 100(4):2076?2081.
Sakata, S. and Harris, K. D. (2009). Laminar structure of spontaneous and sensory-evoked population activity in
auditory cortex. Neuron, 64(3):404?418.
Senn, W., Markram, H., and Tsodyks, M. (2001). An algorithm for modifying neurotransmitter release probability
based on pre-and postsynaptic spike timing. Neural Computation, 13(1):35?67.
Seybold, B. A., Phillips, E. A. K., Schreiner, C. E., and Hasenstaub, A. R. (2015). Inhibitory Actions Unified by
Network Integration. Neuron, 87(6):1181?1192.
Song, S., Sj?str?m, P. J., Reigl, M., Nelson, S., and Chklovskii, D. B. (2005). Highly Nonrandom Features of
Synaptic Connectivity in Local Cortical Circuits. PLoS Biology, 3(3):e68.
Strata, P. and Harvey, R. (1999). Dale?s principle. Brain research bulletin, 50(5):349?350.
Sutskever, I., Vinyals, O., and Le, Q. V. (2014). Sequence to Sequence Learning with Neural Networks.
arXiv.org.
Thomson, A. M., West, D. C., Wang, Y., and Bannister, A. P. (2002). Synaptic connections and small circuits
involving excitatory and inhibitory neurons in layers 2-5 of adult rat and cat neocortex: triple intracellular
recordings and biocytin labelling in vitro. Cerebral cortex (New York, N.Y. : 1991), 12(9):936?953.
Tieleman, T. and Hinton, G. (2012). Lecture 6.5-rmsprop: Divide the gradient by a running average of its recent
magnitude. COURSERA: Neural networks for machine learning, 4(2):26?31.
van den Oord, A., Kalchbrenner, N., Vinyals, O., Espeholt, L., Graves, A., and Kavukcuoglu, K. (2016).
Conditional Image Generation with PixelCNN Decoders. arXiv.org.
van Vreeswijk, C. and Sompolinsky, H. (1996). Chaos in neuronal networks with balanced excitatory and
inhibitory activity. Science, 274(5293):1724?1726.
Vogels, T. P. and Abbott, L. F. (2009). Gating multiple signals through detailed balance of excitation and
inhibition in spiking networks. Nature Neuroscience, 12(4):483.
Xue, M., Atallah, B. V., and Scanziani, M. (2014). Equalizing excitation-inhibition ratios across visual cortical
neurons. Nature, 511(7511):596?600.
11
York, L. C. and van Rossum, M. C. W. (2009). Recurrent networks with short term synaptic depression. Journal
of Computational Neuroscience, 27(3):607?620.
Zenke, F., Agnes, E. J., and Gerstner, W. (2015). Diverse synaptic plasticity mechanisms orchestrated to form
and retrieve memories in spiking neural networks. Nature Communications, 6:6922.
12
| 6631 |@word h:1 neurophysiology:1 cox:2 version:1 middle:1 hippocampus:3 open:1 pulse:1 kappen:1 initial:1 contains:1 series:1 initialisation:1 bc:1 interestingly:1 past:1 freitas:2 existing:1 current:4 com:1 contextual:4 analysed:2 activation:4 yet:2 must:1 exposing:1 hyperpolarizing:1 subsequent:1 plasticity:13 enables:2 plot:1 update:3 fund:1 intelligence:2 fewer:1 selected:2 tone:1 short:9 filtered:1 pointer:1 provides:2 pascanu:3 nandodefreitas:1 simpler:2 sigmoidal:3 org:7 rc:2 consists:1 behavioral:1 introduce:2 forgetting:1 indeed:1 intricate:1 behavior:1 merity:2 growing:1 brain:5 terminal:1 ptb:3 inspired:4 decreasing:1 str:1 becomes:2 provided:2 underlying:1 moreover:2 circuit:20 notation:4 kg:1 deepmind:3 emerging:1 unified:2 transformation:3 sinz:1 nonrandom:1 temporal:3 act:3 ro:1 botvinick:1 hit:1 uk:12 control:9 unit:35 penn:7 schwartz:1 wayne:1 rossum:5 service:1 understood:2 local:3 timing:2 consequence:2 acad:1 receptor:2 oxford:8 jiang:3 initiated:1 optimised:3 firing:6 modulation:2 path:3 abuse:1 black:2 rnns:12 twice:1 studied:1 evoked:2 suggests:1 challenging:1 ease:1 maler:1 decided:1 testing:1 backpropagation:1 digit:2 area:4 thi:1 empirical:2 rnn:6 physiology:2 gabor:1 reilly:2 word:13 pre:3 holt:2 prescott:2 suggest:1 onto:8 context:2 writing:1 dean:2 map:2 charged:1 ecker:1 elusive:2 l:2 keller:1 shen:1 simplicity:2 schreiner:2 rule:1 stereotypical:2 hennequin:3 retrieve:1 population:3 notion:1 updated:1 spontaneous:1 lindsay:1 exact:1 us:1 machens:2 hypothesis:1 jaitly:1 trick:1 element:2 recognition:2 conserve:1 approximated:1 sparsely:1 predicts:1 observed:1 role:3 ft:12 bottom:1 epsrc:1 preprint:2 capture:1 suarez:1 wang:2 tsodyks:1 connected:2 coursera:1 sompolinsky:2 plo:1 counter:1 decrease:1 highest:1 contemporary:1 movement:1 balanced:9 leak:1 rmsprop:3 dynamic:13 trained:2 tight:1 solving:1 predictive:2 serve:1 basis:1 represented:1 neurotransmitter:3 various:1 cat:1 train:1 effective:1 london:4 artificial:4 kalchbrenner:1 larger:1 solve:2 plausible:5 statistic:1 sakata:3 timescale:1 emergence:1 itself:1 noisy:1 associative:1 sequence:12 biophysical:1 net:2 equalizing:1 propose:4 interaction:2 formalised:1 reset:1 product:1 relevant:2 turned:1 glar:1 achieve:3 description:1 sutskever:2 transmission:1 congruent:1 generating:1 adam:1 tim:2 illustrate:2 recurrent:23 ac:4 depending:2 develop:1 help:1 bradbury:1 received:1 strong:3 implemented:2 c:3 berman:1 anatomy:2 closely:3 mrsic:11 functionality:2 annotated:1 subsequently:1 stochastic:1 modifying:1 nando:1 enable:1 transient:1 kistler:4 argued:1 odyssey:1 behaviour:6 espeholt:1 fix:18 marcinkiewicz:1 decompose:1 koutn:1 biological:10 dendritic:1 mathematically:1 shillingford:3 frontier:2 sufficiently:1 koch:3 great:1 mapping:3 predict:1 scope:1 lm:2 cognition:2 achieves:2 released:1 proc:1 wikitext:6 tanh:4 council:1 tool:1 weighted:1 hope:1 gaussian:1 aim:2 e7:1 rather:2 voltage:1 release:1 focus:1 properly:1 modelling:10 prevalent:1 likelihood:1 indicates:1 contrast:1 sigkdd:1 brendan:2 baseline:3 sense:1 helpful:1 inference:1 dayan:2 el:2 dependent:5 entire:1 transferring:1 brette:2 typically:2 hidden:8 bandit:1 transformed:1 pixel:5 overall:3 classification:6 flexible:2 among:1 morphologically:1 proposes:1 art:1 integration:3 constrained:1 initialize:1 platform:1 equal:1 field:2 intriguingly:1 beach:1 frasconi:1 biology:2 represents:3 cancel:2 future:2 report:1 stimulus:4 intelligent:1 micro:1 few:1 individual:1 fire:6 assael:4 conductance:1 interneurons:1 mining:1 highly:1 evaluation:2 ztj:1 mixture:1 arrives:1 pc:3 natl:1 held:1 chain:1 unless:1 divide:1 theoretical:1 column:1 technologically:1 flogel:11 hasenstaub:1 measuring:1 hypothesise:1 mahowald:1 successful:2 too:1 ohm:1 pixelwise:1 dependency:4 e68:1 xue:3 combined:1 st:2 lstm:39 fundamental:1 stratum:2 oord:2 international:2 probabilistic:1 mouse:1 connectivity:5 w1:1 central:1 moitra:1 containing:2 prefrontal:2 agnes:3 derivative:7 potential:14 converted:1 de:4 diversity:1 coding:4 disconnect:1 notable:1 multiplicative:13 view:2 performed:1 closed:3 cadwell:1 lowe:1 option:1 parallel:1 vivo:2 contribution:3 accuracy:3 largely:2 characteristic:1 miller:1 yield:2 weak:1 modelled:1 kavukcuoglu:1 rectified:1 synapsis:1 reach:1 basket:3 synaptic:14 against:1 initialised:1 mohamed:1 naturally:1 costa:8 auditory:2 dataset:6 adjusting:1 gain:1 knowledge:1 shaping:1 amplitude:1 ea:1 back:1 barth:2 clarendon:1 supervised:1 response:3 improved:1 formulation:1 evaluated:1 microcircuit:11 ox:4 done:1 furthermore:1 box:2 until:1 hand:4 working:2 lstms:29 ei:2 su:1 trust:1 propagation:4 google:3 rodriguez:1 mode:2 vogels:9 usa:2 effect:6 facilitate:2 building:1 counterpart:2 merzenich:1 disinhibition:1 recurrence:2 excitation:9 rat:1 thomson:2 demonstrate:1 neocortical:3 performs:1 greff:3 image:4 wise:2 chaos:1 novel:1 recently:2 krueger:2 hofer:1 common:1 sigmoid:2 wikipedia:1 functional:1 spiking:5 empirically:1 perturbing:1 vitro:1 cerebral:1 million:1 interpretation:2 counteracts:1 resting:1 slight:1 seybold:2 cambridge:2 whitteridge:1 phillips:1 framed:1 rd:1 similarly:4 centre:2 language:11 had:1 henry:1 pixelcnn:1 access:1 cortex:7 longer:1 inhibition:24 etc:1 ct0:1 own:1 recent:1 perplexity:7 schmidhuber:5 harvey:2 success:1 emptage:1 lipreading:1 transmitted:1 minimum:1 additional:2 gill:1 signal:5 dashed:1 ii:1 multiple:6 reduces:1 match:2 adapt:1 x28:1 believed:1 long:11 dept:4 equally:1 shunting:4 ellipsis:1 controlled:8 impact:1 involving:1 optimisation:2 curto:1 vision:1 longtin:2 arxiv:11 represent:1 hochreiter:5 cell:29 ion:3 background:1 remarkably:2 fellowship:1 thirteenth:1 chklovskii:1 sjostrom:3 pyramidal:5 ot:21 rest:1 subtractively:3 comment:1 hz:3 recording:1 flow:5 call:1 structural:1 near:1 backwards:3 feedforward:3 bengio:4 wn:1 affect:2 timesteps:1 architecture:8 opposite:1 regarding:1 tm:2 idea:1 depolarise:1 t0:3 whether:1 expression:2 bastos:4 wo:1 song:4 resistance:2 speech:2 york:3 action:3 depression:1 deep:4 ignored:1 detailed:3 clear:3 cleaner:1 aimed:1 amount:1 neocortex:4 processed:1 tth:1 outperform:1 canonical:5 inhibitory:29 notice:1 senn:2 neuroscience:18 per:1 diverse:1 hyperparameter:2 promise:1 unveiling:1 bhalla:2 basal:2 key:3 threshold:2 enormous:1 douglas:7 abbott:6 ht:25 bannister:1 ht0:1 deneve:2 timestep:2 downstream:2 facilitated:1 powerful:1 named:1 architectural:1 wu:1 seq:1 comparable:1 dropout:7 layer:20 ct:34 atallah:1 bound:1 def:1 display:1 laminar:2 fold:1 annual:2 activity:8 ri:1 elife:2 kumar:1 mikolov:1 relatively:2 martin:2 reigl:1 according:2 electrically:1 membrane:8 across:7 postsynaptic:8 marblestone:2 beast:1 biologically:5 making:2 den:2 interference:1 wellcome:1 ln:1 equation:1 remains:2 discus:1 vreeswijk:2 mechanism:4 needed:1 locus:1 mechanistic:1 fed:1 ijt:1 pachitariu:1 fry:1 xiong:1 batch:1 hassabis:2 gate:35 existence:1 denotes:1 running:1 cf:1 include:1 linguistics:1 maintaining:1 l6:1 unifying:1 testable:1 establish:1 jake:1 society:2 iexc:2 dht:1 bl:1 capacitance:1 naud:1 spike:5 receptive:1 primary:1 diagonal:1 traditional:1 aertsen:1 exhibit:3 gradient:9 enhances:1 separate:2 mapped:2 lateral:5 link:1 thank:1 sci:1 decoder:1 nelson:3 originate:1 exc:3 presynaptic:1 collected:1 toward:1 marcus:2 code:1 sur:2 providing:1 balance:8 ratio:1 difficult:1 frank:2 trace:1 neuroscientific:1 implementation:4 zt:21 gated:13 contributed:1 perform:2 neuron:21 kumaran:1 datasets:1 benchmark:1 everton:1 descent:1 truncated:1 heterogeneity:1 hinton:4 communication:2 ponte:1 inh:5 santorini:1 varied:1 prompt:1 namely:1 khan:1 connection:5 sentence:1 philosophical:1 learned:3 protect:1 boost:1 toledo:1 nip:1 capped:1 impair:1 beyond:1 adult:2 below:2 biocytin:1 saturation:1 reliable:1 memory:32 wz:7 optimise:1 royal:2 unrealistic:1 overlap:1 power:1 natural:1 zenke:2 friston:1 predicting:1 difficulty:3 advanced:1 altered:1 marguet:1 sahani:1 summerfield:1 text:2 review:4 l2:1 acknowledgement:1 taming:1 multiplication:1 understanding:1 graf:5 law:1 sir:1 loss:2 discovery:1 lecture:1 mixed:1 suggestion:1 interesting:1 generation:2 proven:1 versus:1 jessell:1 triple:1 validation:3 integrate:4 degree:1 consistent:3 article:1 principle:7 treebank:7 silberberg:1 subtractive:23 ehre:1 genetics:2 excitatory:20 supported:3 last:1 arriving:1 jth:1 preactivation:1 english:1 bias:2 allow:4 understand:3 bulletin:1 markram:4 sparse:1 leaky:3 van:9 regard:1 default:1 cortical:29 depth:3 vocabulary:2 rich:2 sensory:7 author:1 dale:3 adaptive:1 simplified:2 flux:1 transaction:1 kording:1 sj:1 ignore:1 patel:1 keep:1 overfitting:1 corpus:1 tolias:1 sentinel:1 search:3 decade:1 protected:1 why:1 table:6 triplet:1 learn:1 channel:1 transfer:1 ca:1 golovin:2 nature:10 steunebrink:1 dendrite:1 kremkow:2 interact:1 whiteson:1 gerstner:12 complex:6 froemke:11 upstream:1 european:1 berens:1 timescales:2 stereotyped:3 linearly:2 main:1 motivation:1 noise:3 hyperparameters:3 intracellular:1 fair:3 facilitating:1 usrey:1 neuronal:4 fig:14 crafted:1 referred:1 west:1 torres:1 slow:1 sub:1 momentum:2 momentary:1 doiron:2 kandel:3 exponential:1 lie:1 perceptual:1 yannis:2 down:1 remained:1 specific:4 xt:16 gating:21 insightful:1 showing:2 unperturbed:1 explored:1 decay:9 recurrently:1 rht:5 gupta:1 organisation:1 glorot:3 socher:1 mnist:8 sequential:6 magnitude:1 labelling:1 rui:2 columnar:1 suited:1 forget:7 paninski:1 ganglion:2 visual:3 vinyals:2 bo:1 luczak:1 srivastava:1 brunel:2 tieleman:2 harris:13 acm:2 conditional:2 presentation:1 targeted:1 consequently:1 towards:2 sculley:1 content:1 change:1 aided:1 hard:1 abbreviates:1 wt:1 bellow:1 wolff:1 total:1 pfister:2 orchestrated:1 experimental:3 divisive:3 meaningful:1 iinh:3 l4:2 support:3 relevance:1 ub:1 ongoing:1 evaluate:2 tested:2 correlated:1 |
6,226 | 6,632 | k-Support and Ordered Weighted Sparsity for
Overlapping Groups: Hardness and Algorithms
Cong Han Lim
University of Wisconsin-Madison
[email protected]
Stephen J. Wright
University of Wisconsin-Madison
[email protected]
Abstract
The k-support and OWL norms generalize the `1 norm, providing better prediction
accuracy and better handling of correlated variables. We study the norms obtained
from extending the k-support norm and OWL norms to the setting in which there
are overlapping groups. The resulting norms are in general NP-hard to compute,
but they are tractable for certain collections of groups. To demonstrate this fact,
we develop a dynamic program for the problem of projecting onto the set of
vectors supported by a fixed number of groups. Our dynamic program utilizes tree
decompositions and its complexity scales with the treewidth. This program can
be converted to an extended formulation which, for the associated group structure,
models the k-group support norms and an overlapping group variant of the ordered
weighted `1 norm. Numerical results demonstrate the efficacy of the new penalties.
1
Introduction
The use of the `1 -norm to induce sparse solutions is ubiquitous in machine learning, statistics, and
signal processing. When the variables can be grouped into sets corresponding to different explanatory
factors, group variants of the `1 penalty can be used to recover solutions supported on a small number
of groups. When the collection of groups G forms a partition of the variables (that is, the groups do
not overlap), the group lasso penalty [19]
X
?GL (x) :=
kxG kp
(1)
G?G
is often used. In many cases, however, some variables may contribute to more than one explanatory
factor, which leads naturally to overlapping-group formulations. Such is the case in applications such
as finding relevant sets of genes in a biological process [10] or recovering coefficients in wavelet
trees [17]. In such contexts, the standard group lasso may introduce artifacts, since variables that are
contained in different numbers of groups are penalized differently. Another approach is to employ
the latent group lasso [10]:
X
X
?LGL (x) := min
kvG kp such that
vG = x,
(2)
x,v
G?G
G?G
where each vG is a separate vector of latent variables supported only on the group G. The latent
group lasso (2) can be written in terms of atomic norms, where the atomic set is
{x : kxkp ? 1, supp(x) ? G for some G ? G} .
This set allows vectors supported on any one group. The unit ball is the convex hull of this atomic set.
A different way of extending the `1 -norm involves explicit use of a sparsity parameter k. Argyriou
et al. [1] introduce the k-support norm ?k from the atomic norm perspective. The atoms are the set
of k-sparse vectors with unit norm, and the unit ball of the norm is thus
conv ({x : kxkp ? 1, |supp(x)|? k}) .
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
(3)
The k-support norm with p = 2 offers a tighter alternative to the elastic net, and like the elastic net, it
has better estimation performance than the `1 norm especially in the presence of correlated variables.
Another extension of the `1 norm is to the OSCAR/OWL/SLOPE norms [5, 20, 4], which order the
elements of x according to magnitude before weighing them:
X
(4)
?OWL (x) :=
wi |x?i |.
i?[n]
where the weights wi , i = 1, 2, . . . , n are nonnegative and decreasing and x? denotes the vector x
sorted by decreasing absolute value. This family of norms controls the false discovery rate and clusters
correlated variables. These norms correspond to applying the `? norm to a combinatorial penalty
function in the framework of Obozinski and Bach [11, 12], and can be generalized by considering
different `p -norms. For p = 2, we have the SOWL norm [18], whose variational form is
X
?SOWL (x) := 12 min??Rn+
x2i /?i + wi |?i? | .
i?[n]
We will refer to the generalized version of these norms as pOWL norms. The pOWL norms can be
viewed as extensions of the k-support norms from the atomic norm angle, which we will detail later.
`1
(Non-overlapping)
Group Lasso
(Non-overlapping)
k-Group
Support
k-support
OSCAR
OWL/SOWL
SLOPE
(Non-overlapping)
GrOWL
Latent
Group Lasso
Latent kGroup Support
(LG(k))
Latent Group
Smooth OWL
(LGS-OWL)
Figure 1: Some sparsity-inducing norms. Each arrow represents an extension of a previous norm. We study
the two shaded norms on the right.
In this paper, we study the norms obtained by combining the overlapping group formulations with
the k-sparse/OWL formulations, with the aim of obtaining the benefits of both worlds. When the
groups do not overlap, the combination is fairly straightforward; see the GrOWL norm introduced
by Oswal et al. [13]. We consider two classes of norms for overlapping groups. The latent k-group
support (LG(k)) norm, very recently introduced by Rao et al. [15], is defined by the unit ball
n
o
[
conv x : kxkp ? 1, supp(x) ?
G for some subset Gk ? G with k groups ,
(5)
G?Gk
directly extending the k-support norm definition to unions of groups. We introduce the latent group
smooth OWL (LGS-OWL) norm, which similarly extends OWL/SOWL/GrOWL. These norms can
be applied in the same settings where the latent group lasso has proven to be useful, while adapting
better to correlations. We explain how the norms are derived from a combinatorial penalty perspective
using the work of Obozinski and Bach [11, 12], and also provide explicit atomic-norm formulations.
The LGS-OWL norm can be seen as a combination of k-support norms across different k.
The rest of this focuses on computational aspects of these norms. Both the LG(k) norm and
the LGS-OWL norm are in general NP-hard to compute. Despite this hardness result, we devise
a computational approach that utilizes tree decompositions of the underlying group intersection
graph. The key parameter affecting the efficiency of our algorithms is the treewidth tw of the group
intersection graph, which is small for certain graph structures such as chains, trees, and cycles.
Certain problems with hierarchical groups like image recovery can have a tree structure [17, 3].
Our first main technical contribution is a dynamic program for the best k-group sparse approximation
problem, which has time complexity O(2O(tw) ? mk + n), where m is the total number of groups.
For group intersection graphs with a tree structure (tw = 2), this leads to a O(mk + n) algorithm,
significantly improving on the O(m2 k + n) algorithm presented in [3]. Next, we build on the
principles behind the dynamic program to construct extended formulations of O(2O(tw) ? mk 2 + n)
2
size for LG(k) and O(2O(tw) ? m3 + n) for LGS-OWL, improving by a factor of k or m respectively
in the special case in which the tree decomposition is a chain. This approach also yields extended
formulations of size O(nk) and O(n2 ) for the k-support and pOWL norms, respectively. (Previously,
only a O(n2 ) linear program was known for OWL [5].) We thus facilitate incorporation of these
norms into standard convex programming solvers.
Related Work. Obozinski and Bach [11, 12] develop a framework for penalties derived by convexifying the sum of a combinatorial function F and an `p term. They describe algorithms for computing
the proximal operators and norms for the case of submodular F . We use their framework, but note
that the algorithms they provide cannot be applied since our functions are not submodular.
Two other works focus directly on sparsity of unions of overlapping groups. Rao et al. [15] introduce
the LG(k) norm and approximates it via variable splitting. Baldassarre et al. [3] study the best
k-group sparse approximation problem, which they prove is NP-hard. For tree-structured intersection
graphs, they derive the aforementioned dynamic program with complexity O(m2 k + n).
For the case of p = ?, a linear programming relaxation for the unit ball of the latent k-group
support norm is provided by Halabi and Cevher [9, Section 5.4]. This linear program is tight if the
group-element incidence matrix augmented with an all-ones row is totally unimodular. This condition
can be violated by simple tree-structured intersection graphs with just four groups.
Notation and Preliminaries. Given A ? [n], the vector xA is the subvector of x ? Rn corresponding to the index set A. For collections
S of groups G, we use m to denote the number of groups in G,
that is, m = |G|. We assume that G?G G = [n], so that every index i ? [n] appears in at least one
group G ? G. The discrete function CG (A) denotes the minimum number of groups from G needed
to cover A (the smallest set cover).
2
Overlapping Group Norms with Group Sparsity-Related Parameters
We now describe the LG(k) and LGS-OWL norms from the combinatorial penalty perspective by
Obozinski and Bach [11, 12], providing an alternative theoretical motivation for the LG(k) norm
and formally motivating and defining LGS-OWL. Given a combinatorial function F : {A ? [n]} ?
R ? {+?} and an `p norm, a norm can be derived by taking the tightest positively homogeneous
convex lower bound of the combined penalty function F (supp(x)) + ?kxkpp . Defining q to satisfy
1/p + 1/q = 1 (so that `p and `q are dual), this procedure results in the norm ?F
p , which is given by
F
1/q
1/p
1/q
the convex envelope of the function ?p (x) := q (p?) F (supp(x)) kxkp , whose unit ball is
n
o
conv x ? Rn : kxkp ? F (supp(x))?1/q
.
(6)
The norms discussed in this paper can be cast in this framework. Recall that the definition of OWL (4)
includes nonnegative weights w1 ? w2 ? . . . wn ? 0. Defining h : [n] ? R to be the monotonically
Pk
increasing concave function h(k) = i=1 wi , we obtain
?
?
A = ?,
?0,
?0, A = ?,
k-support : F (A) = 1,
LG(k)
: F (A) = 1, CG (A) ? k,
|A|? k,
?
?
?, otherwise,
?, otherwise,
pOWL
: F (A) = h(|A|),
LGS-OWL : F (A) = h(CG (A)).
The definitions of the k-support and LG(k) balls from (3) and (5), respectively, match (6). As for the
OWL norms, we can express their unit ball by
!
m n
o
[
n
?1/q
conv
x ? R : kxkp ? h(i)
, CG (supp(x)) = i
.
(7)
i=1
This can be seen as taking all of the k-support or LG(k) atoms for each value of k, scaling them
according to the value of k, then taking the convex hull of the resulting set. Hence, the OWL norms
can be viewed as a way of interpolating the k-support norms across all values of k. We take advantage
of this interpretation in constructing extended formulations.
3
Hardness Results. Optimizing with the cardinality or non-overlapping group based penalties is
straightforward, since the well-known PAV algorithm [2] allows us to exactly compute the proximal
operator in O(n log n) time [12]. However, the picture is different when we allow overlapping groups.
There are no fast exact algorithms for overlapping group lasso, and iterative algorithms are typically
used. Introducing the group sparsity parameters makes the problem even harder.
Theorem 2.1. The following problems are NP-hard for both ?LG(k) and ?LGS-OWL when p > 1:
Compute ?(y),
arg minx?Rn
arg minx?Rn
1
2 kx
1
2 kx
?
?
(norm computation)
yk22 such that
yk22 +??(x).
?(x) ? ?,
(projection operator)
(proximal operator)
Therefore, other problems that incorporate these norm are also hard. Note that even if we only allow
each element to be in at most two groups, the problem is already hard. We will show in the next two
sections that these problems are tractable if the treewidth of the group intersection graph is small.
3
A Dynamic Program for Best k-Group Approximation
The best k-group approximation problem is the discrete optimization problem
arg min ky ? xk22 such that CG (supp(x)) ? k,
(8)
x
where the goal is to compute the projection of a vector y onto a union of subspaces each defined by a
subcollection of k groups. The solution to (8) has the form
yi i in chosen support,
x0i =
0 otherwise.
As mentioned above, Baldassarre et al. [3] show that this problem is NP-hard. They provide a
dynamic program that acts on the group intersection graph and focus specifically on the case where
this graph is a tree, obtaining a O(m2 k + n) dynamic programming algorithm. In this section, we
also start by using group intersection graphs, but instead focus on the tree decomposition of this
graph, which yields a more general approach.
3.1
Group Intersection Graphs and Tree Decompositions
We can represent the interactions between the different groups using an intersection graph,
which is an undirected graph IG = (G, EG ) in which each vertex denotes a group and two
groups are connected if and only if they overlap. For example, if the collection of groups is
{{1, 2, 3}, {3, 4, 5}, {5, 6, 7}, . . . }, then the intersection graph is simply a chain. If each group
corresponds to a parent and all its children in a rooted tree, the intersection graph is also a tree.
The group intersection graph highlights the dependencies between different groups. Algorithms for
this problem need to be aware of how picking one group may affect the choice of another, connnected
group. (If the groups do not overlap, then no groups are connected and a simple greedy approach
suffices.) A tree decomposition of IG is a more precise way of representing these dependencies.
We provide the definition of tree decompositions and treewidth below and illustrate the core ideas in
Figure 2. Tree decompositions are a fundamental tool in parametrized complexity, leading to efficient
algorithms if the parameter in question is small. See [7, 8] for a a more comprehensive overview
Figure 2: From groups to a group intersection graph to a tree decomposition of width 2.
4
A tree decomposition of (V, E) is a tree T with vertices X = {X1 , X2 , . . . , XN }, satisfying the
following conditions: (1) Each Xi is a subset of V , and the union of all sets Xi gives V . (2) For every
edge (v, w) in E, there is a vertex Xi that contains both v and w. (3) For each v ? V , the vertices
that contain v form a connected subtree of T . The width of a tree decomposition is maxi |Xi |?1, and
the treewidth of a graph (denoted by tw) is the smallest width among all tree decompositions. The
tree decomposition is not unique, and there is always a tree width with number of nodes |X |? |V |
(see for example Lemma 10.5.2 in [7]). Henceforth, we will assume that |X |? m.
The treewidth tw is modest for many types of graphs. For example, the treewidth is bounded
for the tree (tw = 1), the cycle (tw = 2), and series-parallel graphs (tw = 2). Computing tree
decompositions with optimal width for these
? graphs can be done in linear time. On the other hand,
the grid graph has large treewidth (tw ? n) and checking if a graph has tw ? k is NP-complete1 .
3.2
A Dynamic Program for Tree Decompositions
Given a collection of groups G, a corresponding tree decomposition T (G) of the group intersection
graph, and a vector y ? Rn , we provide a dynamic program for problem (8), the best k-group
approximation of y.
The tree decomposition has several features that we can exploit. The tree structure provides a natural
order for processing the vertices, which are subcollections of groups. Properties (1) and (2) yield a
natural way to map elements i ? [n] onto vertices in the tree, indicating when to include yi in the
process. Finally, the connected subtree corresponding to each group G as a result of property (3)
means that we only need to keep explicit information about G for that part of the computation.
The high-level view of our approach is described below. Details appear in the supplementary material.
Preprocessing: For each i ? [n], let G(i) denote the set of all groups that contain i. We have three
data structures: A and V, which are both indexed by (X, Y ), with X ? X and Y ? X; and T, which
is indexed by (X, Y, s), with s ? {0, 1 . . . , k}.
1. Root the tree decomposition and process the nodes from root to leaves: At each node X,
add an index i ? [n] to A(X, G(i) ) if i is unassigned and G(i) ? X.
P 2
2. Set V(X, Y ) ?
yi : i ? A(X, G(i) ), Y ? G(i) 6= 0 .
Main Process: At each vertex X in the tree decomposition, we are allowed to pick groups Y to
include in the support of the solution. The s term in T(X, Y, s) indicates the current group sparsity
?budget? that has been used. Proposition 3.2 below gives the semantic meaning behind each entry in
T . We process the nodes from the leaves to the root to fill T . At each step, the entries for node Xp
will be updated with information from its children.
The update for a leaf Xp is simply T(Xp , Yp , s) ? V(Xp , Yp ) if |Yp |= s. If |Yp |6= s, we mark
T(Xp , Yp , s) as invalid. For non-leaf Xp , we need to ensure that the groups chosen by the parent and
the child are compatible. We ensure this property via constraints of the form Yc ? Xp = Yp ? Xc .
For a single child Xc we have
T(Xp , Yp , s) ?
max
T(Xc , Y, s ? s0 ) : |Yp ? Xc |= s0 + V(Xp , Yp ),
(9)
Y :Y ?Xp =Yp ?Xc
and finally for Xp multiple children Xc(1) , . . . , Xc(d) of Xp , we set T(Xp , Yp , s) as
?
?
?
?
X
[
max
T(Xc(i) , Yi , si ) : Yp ?
Xc(i) = s0 + V(Xp , Yp ). (10)
Yi :Yi ?Xp =Yp ?Xc(i) ?P
?
d
for each i
i=1
i?[d]
si =s?s0
After making each update, we keep track of which Yi was used for each of the children for
T (Xp , Yp , s). This allows us to backtrack to recover the solution after T has been filled.
The next lemma and proposition prove the correctness of this dynamic program. The lemma follows
from the fact that every clique in a graph is contained in some node in any tree decomposition, while
the proposition from induction from the leaf nodes.
1
Nonetheless, there is significant research on developing exact and heuristic tree decomposition algorithms.
There are regular competitions for better implementations [6, pacechallenge.wordpress.com].
5
Lemma 3.1. Every index in [n] is assigned in the first preprocessing step.
Proposition 3.2. For a node X, let yX be the y vector restricted to just the indices i assigned to
nodes below and including X. Each entry T(X, Y, s) is the squared `2 -norm of the best projection of
yX , subject to the fact that besides the groups in Y , at most s ? |Y | are allowed to be used.
We now prove the time complexity of this algorithm. Proposition 3.4 describes the time complexity of
the update when there are many children. It uses the following simple lemma about max-convolutions.
Computing the other updates is straightforward.
Lemma 3.3. The max-convolution f between two concave functions g1 , g2 : {0, 1, . . . , k} ? R,
defined by f (i) := maxj {g1 (j) + g2 (i ? j)}, can be computed in O(k) time.
Proposition 3.4. The update (10) for a fixed Xp , Yp across all values s ? {0, 1, . . . , k} can be
implemented in O(2O(tw) ? dk) time.
Combining timing and correctness results gives us the desired algorithmic result. This approach
significantly improves on the results of Baldassarre et al. [3]. Their approach is specific to groups
whose intersection graph is a tree and uses O(m2 k + n) time.
Theorem 3.5. Given G and a corresponding tree decomposition TG with treewidth tw, projection
onto the corresponding k-group model can be done in O(2O(tw) ? (mk + n)) time. When the group
intersection graph is a tree, the projection takes O(mk + n) time.
4
Extended Formulations from Tree Decompositions
Here we model explicitly the unit ball of LG(k) (5) and LGS-OWL (7). The principles behind this
formulation are very similar to the dynamic program in the previous section.
We first consider the latent k-group support norm, whose atoms are
n
o
[
x : kxkp ? 1, supp(x) ?
G for some subset Gk ? G with k groups .
G?Gk
The following process describes a way of selecting an atom; our extended formulation encodes this
process mathematically. We introduce variables b, which represent the `p budget at a given node,
choice of groups, and group sparsity budget. We start at the root Xr , with `p budget of ? and group
sparsity budget of k:
X
b(Xr ,Y,k?|Y |) ? ?.
(11)
We then start moving towards the leaves, as follows.
1. Suppose we have picked some the groups at a node. Assign some of the `p budget to the xi
terms, where the index i is compatible with the node and the choice of groups.
2. Move on to the child and pick the groups we want to use, considering only groups that
are compatible with the parent. Debit the group budget accordingly. If there are multiple
children, spread the `p and group budgets among them before picking the groups.
The first step is represented by the following relations. Intermediate variables z and a are required to
ensure that we spread the `p budget correctly among the valid xi .
b(X,Y,s) ? k(z (X,Y,s) , u(X,Y,s) )kp ,
(X,Y,s)
(X,(Y,Y 0 ),s)
0
:Y ?Y =
6 ?}kp ,
? k{a
X
0
a(X,Y ,s) ?
a(X,(Y,Y ),s) ,
Y ?X
X
Xk
0
kxA(X,Y ) k2 ?
a(X,Y ,s) .
0
z
0
A(X,Y )=A(X,Y )
s=0
The second step is represented by the following inequality in the case of a single child.
o
Xn
u(Xp ,Yp ,s) ?
b(Xc ,(Yp ,Y ),s?s0 ) : Y ? Xp = Yp ? Xc , |Yp ? Xc |= s0 .
6
(12)
(13)
(14)
(15)
(16)
When there are multiple children, we need to introduce more intermediate variables to spread the
group budget correctly. The technique here is similar to the one used in the proof of Proposition 3.4;
we defer details to the supplementary material. In both cases, we need to collect the budgets that have
been sent from each Yp :
X
b(Xc ,Y,s) ?
b(Xc ,(Yp ,Y ),s) .
(17)
Yp
Those b variables unreachable by the budget transfer process are set to 0. Our main theorem about the
correctness of the construction in this section follows from the fact that when ? = 1, every extreme
point with nonzero x in our extended formulation is an atom of the corresponding LG(k).
Theorem 4.1. We can model the set ?LG(k) (x) ? ? using O(2O(tw) ? (mk 2 + n)) variables and
inequalities in general. When the tree decomposition is a chain, O(2O(tw) ? (mk + n)) suffices.
For the unit ball of ?LGS-OWL , we can exploit the fact that the atoms of ?LGS-OWL are obtained from
?LG(k) across different k at different scales. Instead of using the inequality (11) at the node, we have
X
h(k)1/q b(Xr ,Y,k?|Y |) ? ?,
Y ?Xr
O(tw)
which leads to a program of size O(2
5
? (m2 + n)) for chains and O(2O(tw) ? (m3 + n)) for trees.
Empirical Observations and Results
The extended formulations above can be implemented in modeling software such as CVX. This
may incur a large processing overhead, and it is often faster to implement these directly in a convex
optimization solver such as Gurobi or MOSEK. Use of the `? -norm leads to a linear program which
can be significantly faster than the second-order conic program that results from the `2 -norm.
We evaluated the performance of LG(k) and LGS-OWL on linear regression problems minx 12 ky ?
Axk2 +??(x). In the scenarios considered, we use the latent group lasso as a baseline. We test both
the `2 and `? variants of the various norms. Following [13] (which descrbes GrOWL), we consider
two different types of weights for LGS-OWL. The linear variant sets wi = 1 ? (i ? 1)/n for i ? [n],
whereas in the spike version, we set w1 = 1 and wi = 0.25 for i = 2, 3, . . . , n. The regularization
term ? was chosen by grid search over {10?2 , 10?1.95 , . . . , 104 } for each experiment.
The metrics we use are support recovery and estimation quality. For the support recovery experiments,
we count the number of times the correct support was identified. We also compute the root mean
square (RMSE) of kx ? x? k2 (estimation error).2
We had also tested the standard lasso, elastic net, and k-support and OWL norms, but these norms
performed poorly. In our experiments they were not able to recover the exact correct support in
any run. The estimation performance for the k-support norms and elastic net were worse than the
corresponding latent group lasso, and likewise for OWL vs. LGS-OWL.
Experiments. We used 20 groups of variables where each successive group overlaps by two elements with the next [10, 14]. The groups are given by {1, . . . , 10}, {9, . . . , 18}, . . . , {153, . . . , 162}.
For the first set of experiments, the support of the true input x? are a cluster of five groups in the
middle of x, with xi = 1 on the support. For the second set of experiments, the original x is supported
by the two disjoint clusters of five overlapping groups each, with xi = 2 on one cluster and xi = 3
on the other.
Each entry of the A matrix is chosen initially to be i.i.d. N (0, 1). We then introduce correlations
between groups in the same cluster in A. Within each cluster of groups, we replicate the same set of
columns for each group in the non-overlapping portions of the group (that is, every pair of groups in
a cluster shares at least 6 columns, and adjacent groups share 8 columns). We then introduce noise by
adding i.i.d. elements from N (0, 0.05) so that the replications are not exact. Finally, we generate y
by adding i.i.d. noise from N (0, 0.3) to each component of Ax? .
We present support recovery results in Figure 3 for the `2 variants of the norms which perform better
than the`? versions, though the relative results between the different norms hold. In the appendix we
provide the graphs for support recovery and estimation quality as well as other observations.
2
It is standard in the literature to compute the RMSE of the prediction or estimation quality. RMSE metrics
are not ideal in practice since we should ?debias? x to offset shrinkage due to the regularization term.
7
Figure 3: Support recovery performance as number of measurements (height of A) increases. The vertical
axis indicates the number of trials (out of 100) for which the correct support was identified. The two left graphs
correspond to the first configuration of group supports (five groups), while the others to the second configuration
(ten groups). Each line represents a different method. In the first and third graphs, we plot LG(k) for different
values of k, increasing from 1 to the ?ground truth? value. Note that k = 1 is exactly the latent group lasso. In
the second and fourth graphs, we plot LGS-OWL for the different choices of weights wi discussed in the text.
Our methods can significantly outperform latent group lasso in both support recovery and estimation
quality. We provide a summary below and more details are provided in the supplementary.
We first focus on support recovery. There is a significant jump in performance when k is the size
of the true support. Note that exceeding the ground-truth value makes recovery of the true support
impossible in the presence of noise. For smaller values of k, the results range from slight improvement
(especially when k = 4 or k = 8 in the first and second experiments respectively) to mixed results (for
large number of rows in A and small k). The LGS-OWL norms can provide performance almost as
good as the best settings of k for LG(k), and can be used when the number of groups is unknown. We
expect to see better performance for well-tuned OWL weights. We see similar results for estimation
performance. Smaller values of k provide little to no advantage, while larger values of k and the
LGS-OWL norms can offer significant improvement.
6
Discussion and Extensions
We introduce a variant of the OWL norm for overlapping groups and provide the first tractable
approaches for this and the latent k-group support norm (via extended formulations) under a bounded
treewidth condition. The projection algorithm for the best k-group sparse approximation problem
generalizes and improves on the algorithm by Baldassarre et al. [3]. Numerical results demonstrate
that the norms can provide significant improvement in support recovery and estimation.
A family of graphs with many applications and large treewidth is the set of grid graphs. Groups over
collections of adjacent pixels/voxels lead naturally to such group intersection graphs, and it remains
an open question whether polynomial time algorithms exist for this set of graphs. Another venue for
research is to derive and evaluate efficient approximations to these new norms.
It is tempting to apply recovery results on the latent group lasso here, since LG(k) can be cast as a
latent group lasso instance with groups {G0 : G0 is a union of up to k groups of G}. The consistency
results of [10] only applies under the strict condition that the target vector is supported exactly by a
unique set of k groups. The Gaussian width results of [16] do not give meaningful bounds even when
the groups are disjoint and k = 2. Developing theoretical guarantees on the performance of these
methods requires a much better understanding of the geometry of unions of overlapping groups.
We can easily extend the dynamic program to handle the case in which we want both k-group sparsity,
and overall sparsity of s. For tree-structured group intersection graphs, our dynamic program has
? 2 ks2 + mn) by [3]. This yields a variant of
time complexity O(mks + n log s) instead of the O(m
the above norms that again has a similar extended formulations. These variants could be employed as
an alternative to the sparse overlapping set LASSO by Rao et al. [14]. We leave this to future work.
Acknowledgements This work was supported by NSF award CMMI-1634597, ONR Award
N00014-13-1-0129, and AFOSR Award FA9550-13-1-0138.
8
References
[1] Argyriou, A., Foygel, R., and Srebro, N. (2012). Sparse prediction with the k-support norm. In Advances in
Neural Information Processing Systems, pages 1457?1465.
[2] Ayer, M., Brunk, H. D., Ewing, G. M., Reid, W. T., and Silverman, E. (1955). An empirical distribution
function for sampling with incomplete information. The Annals of Mathematical Statistics, 26(4):641?647.
[3] Baldassarre, L., Bhan, N., Cevher, V., Kyrillidis, A., and Satpathi, S. (2016). Group-Sparse Model Selection:
Hardness and Relaxations. IEEE Transactions on Information Theory, 62(11):6508?6534.
[4] Bogdan, M., van den Berg, E., Sabatti, C., Su, W., and Cand?s, E. J. (2015). Slope?adaptive variable
selection via convex optimization. Ann. Appl. Stat., 9(3):1103?1140.
[5] Bondell, H. D. and Reich, B. J. (2008). Simultaneous Regression Shrinkage, Variable Selection, and
Supervised Clustering of Predictors with OSCAR. Biometrics, 64(1):115?123.
[6] Dell, H., Husfeldt, T., Jansen, B. M. P., Kaski, P., Komusiewicz, C., and Rosamond, F. A. (2016). The first
parameterized algorithms and computational experiments challenge. In 11th International Symposium on
Parameterized and Exact Computation, IPEC 2016, August 24-26, 2016, Aarhus, Denmark, pages 30:1?30:9.
[7] Downey, R. G. and Fellows, M. R. (1999). Parameterized Complexity. Monographs in Computer Science.
Springer New York, New York, NY.
[8] Downey, R. G. and Fellows, M. R. (2013). Fundamentals of Parameterized Complexity. Texts in Computer
Science. Springer London, London.
[9] Halabi, M. E. and Cevher, V. (2015). A totally unimodular view of structured sparsity. In Lebanon, G. and
Vishwanathan, S., editors, Proceedings of the 18th International Conference on Artificial Intelligence and
Statistics (AISTATS 2015), pages 223?231.
[10] Jacob, L., Obozinski, G., and Vert, J.-P. (2009). Group lasso with overlap and graph lasso. In Proceedings
of the 26th Annual International Conference on Machine Learning, ICML ?09, pages 433?440, New York,
NY, USA. ACM.
[11] Obozinski, G. and Bach, F. (2012). Convex Relaxation for Combinatorial Penalties. Technical report.
[12] Obozinski, G. and Bach, F. (2016). A unified perspective on convex structured sparsity: Hierarchical,
symmetric, submodular norms and beyond. Technical report.
[13] Oswal, U., Cox, C., Lambon-Ralph, M., Rogers, T., and Nowak, R. (2016). Representational similarity
learning with application to brain networks. In Balcan, M. F. and Weinberger, K. Q., editors, Proceedings of
The 33rd International Conference on Machine Learning, pages 1041?1049, New York, NY, USA. PMLR.
[14] Rao, N., Cox, C., Nowak, R., and Rogers, T. T. (2013). Sparse overlapping sets lasso for multitask learning
and its application to fmri analysis. In Burges, C., Bottou, L., Welling, M., Ghahramani, Z., and Weinberger,
K., editors, Advances in Neural Information Processing Systems 26, pages 2202?2210.
[15] Rao, N., Dud?k, M., and Harchaoui, Z. (2017). The group k-support norm for learning with structured
sparsity. In 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP),
pages 2402?2406.
[16] Rao, N., Recht, B., and Nowak, R. (2012). Universal measurement bounds for structured sparse signal
recovery. In Lawrence, N. D. and Girolami, M., editors, Proceedings of the Fifteenth International Conference
on Artificial Intelligence and Statistics, pages 942?950, La Palma, Canary Islands. PMLR.
[17] Rao, N. S., Nowak, R. D., Wright, S. J., and Kingsbury, N. G. (2011). Convex approaches to model wavelet
sparsity patterns. In 2011 18th IEEE International Conference on Image Processing, pages 1917?1920. IEEE.
[18] Sankaran, R., Bach, F., and Bhattacharya, C. (2017). Identifying Groups of Strongly Correlated Variables
through Smoothed Ordered Weighted L1 -norms. In Singh, A. and Zhu, J., editors, Proceedings of the 20th
International Conference on Artificial Intelligence and Statistics, pages 1123?1131, Fort Lauderdale, FL,
USA. PMLR.
[19] Yuan, M. and Lin, Y. (2006). Model selection and estimation in regression with grouped variables. Journal
of the Royal Statistical Society: Series B (Statistical Methodology), 68(1):49?67.
[20] Zeng, X. and Figueiredo, M. A. T. (2014). The Ordered Weighted `1 Norm: Atomic Formulation,
Projections, and Algorithms. arXiv:1409.4271.
9
| 6632 |@word multitask:1 trial:1 cox:2 middle:1 version:3 polynomial:1 norm:85 replicate:1 open:1 palma:1 decomposition:24 jacob:1 pick:2 harder:1 configuration:2 contains:1 efficacy:1 series:2 selecting:1 tuned:1 current:1 com:1 incidence:1 si:2 written:1 numerical:2 partition:1 plot:2 update:5 v:1 greedy:1 leaf:6 weighing:1 intelligence:3 accordingly:1 xk:1 core:1 fa9550:1 provides:1 contribute:1 node:13 successive:1 five:3 height:1 mathematical:1 dell:1 kingsbury:1 symposium:1 replication:1 yuan:1 prove:3 overhead:1 introduce:9 hardness:4 cand:1 brain:1 decreasing:2 little:1 considering:2 solver:2 conv:4 provided:2 totally:2 underlying:1 notation:1 increasing:2 bounded:2 cardinality:1 unified:1 finding:1 guarantee:1 fellow:2 every:6 act:1 concave:2 unimodular:2 exactly:3 k2:2 control:1 unit:9 appear:1 reid:1 before:2 timing:1 despite:1 collect:1 shaded:1 appl:1 range:1 unique:2 atomic:7 union:6 practice:1 implement:1 silverman:1 xr:4 procedure:1 empirical:2 universal:1 adapting:1 significantly:4 projection:7 vert:1 induce:1 regular:1 onto:4 cannot:1 selection:4 operator:4 context:1 applying:1 impossible:1 map:1 straightforward:3 convex:10 recovery:12 splitting:1 identifying:1 m2:5 argyriou:2 fill:1 handle:1 updated:1 annals:1 construction:1 suppose:1 target:1 exact:5 programming:3 homogeneous:1 us:2 element:6 satisfying:1 ks2:1 cong:1 cycle:2 connected:4 mentioned:1 monograph:1 complexity:9 dynamic:14 singh:1 tight:1 incur:1 debias:1 efficiency:1 easily:1 icassp:1 differently:1 represented:2 various:1 kaski:1 fast:1 describe:2 london:2 kp:4 artificial:3 whose:4 heuristic:1 supplementary:3 larger:1 otherwise:3 statistic:5 g1:2 advantage:2 net:4 interaction:1 mks:1 relevant:1 combining:2 poorly:1 representational:1 inducing:1 competition:1 ky:2 parent:3 cluster:7 extending:3 leave:1 bogdan:1 derive:2 develop:2 illustrate:1 stat:1 x0i:1 recovering:1 c:1 involves:1 treewidth:11 implemented:2 girolami:1 correct:3 hull:2 material:2 owl:35 rogers:2 assign:1 suffices:2 preliminary:1 proposition:7 biological:1 tighter:1 mathematically:1 extension:4 hold:1 considered:1 wright:2 ground:2 lawrence:1 algorithmic:1 kxkpp:1 smallest:2 estimation:10 baldassarre:5 combinatorial:6 grouped:2 correctness:3 wordpress:1 tool:1 weighted:4 always:1 gaussian:1 aim:1 unassigned:1 shrinkage:2 derived:3 focus:5 ax:1 improvement:3 indicates:2 cg:5 baseline:1 typically:1 explanatory:2 initially:1 relation:1 ralph:1 pixel:1 arg:3 aforementioned:1 dual:1 among:3 denoted:1 unreachable:1 overall:1 jansen:1 special:1 fairly:1 ewing:1 construct:1 aware:1 beach:1 atom:6 sampling:1 represents:2 icml:1 mosek:1 future:1 fmri:1 np:6 others:1 report:2 employ:1 comprehensive:1 maxj:1 geometry:1 extreme:1 behind:3 chain:5 edge:1 nowak:4 modest:1 biometrics:1 tree:42 indexed:2 filled:1 incomplete:1 desired:1 theoretical:2 mk:7 cevher:3 instance:1 column:3 modeling:1 rao:7 cover:2 tg:1 introducing:1 vertex:7 subset:3 entry:4 predictor:1 motivating:1 dependency:2 proximal:3 combined:1 st:1 venue:1 fundamental:2 international:8 recht:1 lauderdale:1 picking:2 w1:2 squared:1 again:1 henceforth:1 worse:1 leading:1 yp:23 supp:9 converted:1 includes:1 coefficient:1 satisfy:1 explicitly:1 later:1 view:2 root:5 picked:1 performed:1 portion:1 start:3 recover:3 parallel:1 slope:3 defer:1 rmse:3 kxg:1 contribution:1 square:1 accuracy:1 likewise:1 correspond:2 yield:4 generalize:1 backtrack:1 explain:1 simultaneous:1 definition:4 nonetheless:1 naturally:2 associated:1 proof:1 recall:1 lim:1 improves:2 ubiquitous:1 appears:1 supervised:1 ayer:1 brunk:1 methodology:1 formulation:16 done:2 evaluated:1 though:1 strongly:1 just:2 xa:1 correlation:2 hand:1 su:1 zeng:1 overlapping:20 artifact:1 quality:4 usa:4 facilitate:1 contain:2 true:3 hence:1 assigned:2 regularization:2 dud:1 symmetric:1 nonzero:1 semantic:1 eg:1 pav:1 adjacent:2 width:6 rooted:1 generalized:2 demonstrate:3 l1:1 balcan:1 image:2 variational:1 meaning:1 recently:1 overview:1 discussed:2 interpretation:1 approximates:1 slight:1 extend:1 refer:1 significant:4 measurement:2 rd:1 grid:3 consistency:1 similarly:1 submodular:3 had:1 moving:1 reich:1 han:1 similarity:1 add:1 perspective:4 optimizing:1 scenario:1 certain:3 kxa:1 n00014:1 inequality:3 onr:1 yi:7 devise:1 seen:2 minimum:1 employed:1 monotonically:1 signal:3 stephen:1 tempting:1 multiple:3 harchaoui:1 smooth:2 technical:3 match:1 faster:2 offer:2 long:1 bach:7 lin:1 award:3 prediction:3 variant:8 regression:3 metric:2 fifteenth:1 arxiv:1 represent:2 affecting:1 want:2 whereas:1 kvg:1 subcollection:1 envelope:1 rest:1 w2:1 strict:1 subject:1 undirected:1 sent:1 presence:2 yk22:2 intermediate:2 ideal:1 wn:1 affect:1 lasso:19 identified:2 idea:1 kyrillidis:1 whether:1 downey:2 penalty:10 speech:1 york:4 useful:1 ten:1 generate:1 outperform:1 exist:1 nsf:1 disjoint:2 track:1 correctly:2 discrete:2 express:1 group:136 key:1 four:1 wisc:2 graph:38 relaxation:3 sum:1 run:1 angle:1 parameterized:4 oscar:3 fourth:1 extends:1 family:2 almost:1 utilizes:2 cvx:1 appendix:1 scaling:1 bound:3 fl:1 nonnegative:2 annual:1 incorporation:1 constraint:1 vishwanathan:1 x2:1 software:1 encodes:1 aspect:1 lambon:1 min:3 structured:7 developing:2 according:2 ball:9 combination:2 across:4 describes:2 smaller:2 wi:7 island:1 tw:19 making:1 axk2:1 projecting:1 restricted:1 den:1 handling:1 xk22:1 bondell:1 previously:1 remains:1 foygel:1 count:1 needed:1 halabi:2 tractable:3 generalizes:1 tightest:1 apply:1 hierarchical:2 pmlr:3 bhattacharya:1 alternative:3 weinberger:2 original:1 denotes:3 clustering:1 include:2 ensure:3 madison:2 xc:15 yx:2 exploit:2 ghahramani:1 especially:2 build:1 society:1 move:1 g0:2 already:1 question:2 spike:1 cmmi:1 minx:3 subspace:1 separate:1 parametrized:1 induction:1 denmark:1 besides:1 index:6 providing:2 lg:19 gk:4 implementation:1 unknown:1 perform:1 vertical:1 convolution:2 observation:2 defining:3 extended:10 precise:1 rn:6 smoothed:1 august:1 introduced:2 cast:2 subvector:1 required:1 gurobi:1 pair:1 fort:1 acoustic:1 nip:1 able:1 sabatti:1 beyond:1 below:5 pattern:1 yc:1 sparsity:15 challenge:1 program:19 max:4 including:1 royal:1 convexifying:1 overlap:6 natural:2 mn:1 representing:1 zhu:1 x2i:1 picture:1 conic:1 axis:1 canary:1 text:2 literature:1 discovery:1 checking:1 voxels:1 understanding:1 acknowledgement:1 relative:1 wisconsin:2 afosr:1 aarhus:1 expect:1 highlight:1 mixed:1 proven:1 srebro:1 vg:2 xp:19 s0:6 principle:2 editor:5 kxkp:7 share:2 row:2 compatible:3 penalized:1 summary:1 supported:7 gl:1 figueiredo:1 allow:2 burges:1 taking:3 absolute:1 sparse:11 benefit:1 van:1 xn:2 world:1 valid:1 collection:6 jump:1 preprocessing:2 ig:2 adaptive:1 welling:1 transaction:1 lebanon:1 gene:1 keep:2 clique:1 xi:9 search:1 latent:18 iterative:1 transfer:1 ca:1 elastic:4 obtaining:2 improving:2 bottou:1 interpolating:1 constructing:1 aistats:1 pk:1 main:3 spread:3 arrow:1 motivation:1 noise:3 n2:2 child:11 allowed:2 positively:1 augmented:1 x1:1 ny:3 explicit:3 exceeding:1 third:1 wavelet:2 theorem:4 specific:1 maxi:1 offset:1 dk:1 false:1 adding:2 magnitude:1 subtree:2 budget:12 kx:3 nk:1 intersection:19 simply:2 ordered:4 contained:2 g2:2 applies:1 springer:2 corresponds:1 truth:2 acm:1 obozinski:7 sorted:1 viewed:2 goal:1 ann:1 invalid:1 towards:1 hard:7 specifically:1 lemma:6 total:1 la:1 m3:2 meaningful:1 indicating:1 formally:1 swright:1 berg:1 support:43 mark:1 violated:1 incorporate:1 evaluate:1 tested:1 correlated:4 |
6,227 | 6,633 | A simple model of recognition and recall memory
Nisheeth Srivastava
Computer Science, IIT Kanpur
Kanpur, UP 208016
[email protected]
Edward Vul
Dept of Psychology, UCSD
9500 Gilman Drive La Jolla CA 92093
[email protected]
Abstract
We show that several striking differences in memory performance between recognition and recall tasks are explained by an ecological bias endemic in classic memory
experiments - that such experiments universally involve more stimuli than retrieval
cues. We show that while it is sensible to think of recall as simply retrieving
items when probed with a cue - typically the item list itself - it is better to think
of recognition as retrieving cues when probed with items. To test this theory, by
manipulating the number of items and cues in a memory experiment, we show
a crossover effect in memory performance within subjects such that recognition
performance is superior to recall performance when the number of items is greater
than the number of cues and recall performance is better than recognition when
the converse holds. We build a simple computational model around this theory,
using sampling to approximate an ideal Bayesian observer encoding and retrieving
situational co-occurrence frequencies of stimuli and retrieval cues. This model
robustly reproduces a number of dissociations in recognition and recall previously
used to argue for dual-process accounts of declarative memory.
1
Introduction
Over half a century, differences in memory performance in recognition and recall-based experiments
have been a prominent nexus of controversy and confusion. There is broad agreement among
memory researchers, following Mandler?s influential lead, that there are at least two different types
of memory activities - recollection, wherein we simply remember something we want to remember,
and familiarity, wherein we remember having seen something before, but nothing more beyond
it [8]. Recall-based experiments are obvious representatives of recollection. Mandler suggested that
recognition was a good example of familiarity activity.
Dual-process accounts of memory question Mandler?s premise that recognition is exclusively a
familiarity operation. They argue, phenomenologically, that recognition could also succeed successful
recollection, making the process a dual composition of recollection and familiarity [20]. Experimental
procedures and analysis methods have been designed to test for the relative presence of both processes
in recognition experiments, with variable success. These endeavors contrast with strength-based
single-process models of memory that treat recognition as the retrieval of a weak trace of item
memory, and recall as retrieval of a stronger trace of the same item [19].
The single/dual process dispute also spills over into the computational modeling of memory. Gillund
and Shiffrin?s influential SAM model is a single-process account of both recognition and recall [4]. In
SAM and other strength-based models of declarative memory, recognition is modeled as item-relevant
associative activation of memory breaching a threshold, while recall is modeled as sampling items
from memory using the relative magnitudes of these associative activations. In contrast, McClelland?s
equally influential CLS model is explicitly a dual-process model, where a fast learning hippocampal
component primarily responsible for recollection sits atop a slow learning neocortical component
responsible for familiarity [9]. Wixted?s signal detection model tries to bridge the gap between
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
these accounts by allowing dual process contributions to combine additively into a unidimensional
strength variable [19]. While such pragmatic syntheses are useful, the field is still looking for a more
satisfactory theoretical unification.
The depth of the difference between the postulated dual processes of recollection and familiarity
depends inevitably on the strength of the quantitative and qualitative dissociations that previous
research has documented in memory tasks, prominent among which are recognition and recall.
Mandler, for instance, postulated a one-to-one mapping between recognition and familarity on one
hand and recall and recollection on the other [8], although other authors hold more nuanced views [20].
Notwithstanding such differences of opinion, the road to discovering useful single-process accounts
of declarative memory has to go through explaining the multiple performance dissociations between
recognition and recall memory tasks. To the extent that single process accounts of both tasks can
explain such dissociations, differences between recollection and familarity will not seem nearly as
fundamental.
Improved strength-based models have competently modeled a large array of recognition-recall
dissociations [13], but fail, or have to make intricate assumptions, in the face of others [20]. More
importantly, the SAM model and its descendants are not purely single-process models. They model
recognition as a threshold event and recall as a sampling event, with the unification coming from
the fact that both events occur using the same information base of associative activation magnitude.
We present a much simpler single process model that capably reproduces many critical qualitative
recognition-recall dissociations. In the process, we rationalize the erstwhile abstract associative
activation of strength-based memory models as statistically efficient monitoring of environmental
co-occurrence frequencies. Finally, we show using simulations and a behavioral experiment, that the
large differences between recognition and recall in the literature can be explained by the responses of
an approximately Bayesian observer tracking these frequencies to two different questions.
2
Model
We use a very simple model, specified completely by heavily stylized encoding and retrieval processes.
The encoding component of our model simply learns the relative frequencies with which specific
conjunctions of objects are attended to in the world. We consider objects x of only two types: items
xi and lists xl . We model each timestep as as a Bernoulli trial between the propensity to attend to
any of the set of items or to the item-list itself, with a uniform prior probability of sampling any
of the objects. Observers update the probability of co-occurrence, defined in our case rigidly as
1-back occurrence, inductively as the items on the list are presented. We model this as the observer?s
sequential Bayesian updates of the probability p(x), stored at every time step as a discrete memory
engram m.
Thus, in this encoding model, information about the displayed list of items is available in distributed
form in memory as p(xi , xl |m), with each engram m storing one instance of co-occurrence. The true
joint distribution of observed items,to the extent that it is encoded within the set of all task-relevant
memory engrams M is then expressible as a simple probabilistic marginalization,
X
p(xi , xl ) =
p(xi , xl |m)p(m),
(1)
m?M
where we assume that p(m) is flat over M, i.e. we assume that within the set of memory engrams
relevant for the retrieval cue, memory access is random.
Our retrieval model is approximately Bayesian. It assumes that people sample a small subset of all
relevant engrams M0 ? M when making memory judgments. Thus, the joint distribution accessible
to the observer during retrieval becomes a function of the set of engrams actually retrieved,
X
pMk (xi , xl ) =
p(xi , xl |m)p(m),
(2)
m?Mk
where Mk denotes the set of first k engrams retrieved.
Following a common approach to sampling termination in strength-based sequential sampling memory
models, we use a novelty threshold that allows the memory retrieval process to self-terminate when
incoming engrams no longer convey significantly novel information [4, 13]. We treat the arrival of the
2
Recall
Recognition
Retrieval
Encoding
Retrieval
Encoding
A
A
B
B
C
C
Sample from p( |A)
p(x| )
p( |x)
Sample from p(x| )
A
A
B
X
C
Figure 1: Illustrating the ecological difference in retrieval during recognition and recall memory
experiments. We model recall retrieval as a probabilistic query about items conditioned on the item
list and recognition retrieval as a probabilistic query about the item list conditioned on the item
presented during retrieval. Since there are almost always more items than lists in classic memory
experiments, the second conditional distribution tends to be formed on a smaller discrete support set
than the former.
k th successive engram into working memory as a signal for the observer to probabilistically sample
from pMk . The stopping rule for memory retrieval in our model is for n consecutive identical samples
being drawn in succession during this internal sampling, n remaining a free parameter in the model.
This rule is designed to capture the fact that memory search is frugal and self-terminating [15]. The
sample drawn at the instant the novelty threshold is breached is overtly recalled. Since this sample is
drawn from a distribution constructed by approximately reconstructing the true encoded distribution
of situational co-occurrences, the retrieval model is approximately Bayesian. Finally, since our
encoding model ensures that the observer knows the joint distribution of event co-occurrences, which
contains all the information needed to compute marginals and conditionals also, we further assume
that these derivative distributions can also be sampled, using the same retrieval model, when required.
We show in this paper that this simple memory model yields both recognition and recall behavior.
The difference between recognition and recall is simply that these two retrieval modalities ask two
different questions of the same base of encoded memory - the joint distribution p(xi , xl ). We illustrate
this difference in Figure 1. During recall-based retrieval, experimenters ask participants to remember
all the items that were on a previously studied list. In this case, the probabilistic question being asked
is ?given xl , find xi ?, which our model would answer by sampling p(xi |xl ). In item-recognition
experiments, experimenters ask participants to determine whether each of several items was on a
previously shown list or not. We assert that in this case the probabilistic question being asked is
?given xi . find xl ?, which our model would answer by sampling p(xl |xi ).
Our operationalization of recognition as a question about the list rather than the item runs contrary to
previous formalizations, which have tended to model it as the associative activation engendered in
the brain by observing a previously seen stimulus - models of recognition memory assume that the
activation for previously seen stimuli is greater, for all sorts of reasons. In contrast, recall is modeled
in classical memory accounts much the same way as in ours - as a conditional activation of items
associated with retrieval cues, including both the item list and temporally contiguous items. Our
approach assumes that the same mechanism of conditional activation occurs in recognition as well the difference is that we condition on the item itself.
3
3
Basic prediction: fast recognition and slow recall
The sample-based threshold used to terminate memory retrieval in our model does not depend on
the size of the support of the probability distribution being sampled from. This immediately implies
that, for the same threshold sample value, the model will take longer to approach it when sampling
from a distribution with larger support than when sampling from distributions with smaller support.
In classical memory experiments, observers are typically asked to memorize multiple items associated
with one, or a few, lists. Thus, there is an ecological bias built into classic memory experiments such
that |items| |lists|. Making this assumption immediately rationalizes the apparent difference in
speed and effort between recognition and recall in our model. Because the recognition task samples
p(list|item), its sample complexity is lower than recall, which involves sampling p(item|list) from
memory.
To verify this numerically, starting from identical memory encodings in both cases, we ran 1000
simulations of recognition and recall respectively using our retrieval model, using a fixed n = 5 1 .
The results, measured in terms of the number of retrieval samples k drawn before termination in
each of the 1000 trials, are shown in the left panel of Figure 2. The sample complexity of recall is
evidently higher than for recognition2 . Thus, we suggest that the fundamental difference between
recognition and recall - that recognition is easier and recall is harder - is explicable simply by virtue
of the ecological bias of memory experiments that use fewer cues than stimuli.
The difference in speed between recollection and familiarity processes, as measured in recall and
recognition experiments, has been one of the fundamental motivations for proposing that two memory
processes are involved in declarative memory. Dual-process accounts have invoked priority arguments
instead, e.g. that information has to pass through semantic memory, which is responsible for
recognition, before accessing episodic memory which is responsible for recall [17].Single process
accounts following in the lineage of SAM [4] have explained the difference by arguing that recognition
involves a single comparison of activation values to a threshold, whereas recall involves competition
between multiple activations for sampling. Our model rationalizes this distinction made in SAM-style
sequential sampling models by arguing that recognition memory retrieval is identical to recall memory
retrieval; only the support of the distribution from which the memory trace is to be probabilistically
retrieved changes. Thus, instead of using a race to threshold for recognition and a sampling process
in recall, this model uses self-terminating sampling in both cases, explaining the main difference
between the two tasks - easy recognition and hard recall - as a function of typical ecological parameter
choices. This observation also explains the relative indifference of recognition tasks to divided
attention conditions, in contrast with recall which is heavily affected [2]. Because of the lower sample
complexity of recognition, fewer useful samples are needed to arrive at the correct conclusion.
4
An empirical test
The explanation our model offers is simple, but untested. To directly test it, we constructed a simple
behavioral experiment, where we would manipulate the number of items and cues keeping the total
number of presentations constant, and see how this affected memory performance in both recognition
and recall retrieval modalities. Our model predicts that memory performance difficulty scales up with
the size of the support of the conditional probability distribution relevant to the retrieval modality.
Thus recall, which samples from p(item|list), should become easier as the number of items to recall
per cue reduces. Similarly recognition, which samples from p(listlitem), should become harder as
the number of cues per item increases. Because classic memory experiments have tended to use more
items than cues (lists), our model predicts that such experiments would consistently find recognition
to be easier than recall. By inverting this pattern, having more cues than items, for instance, we would
expect to see the opposite pattern hold. We tested for this performance crossover using the following
experiment.
1
Our results are relatively independent of the choice of n, since for any value of n, recognition stays easier
than recall so long as the cue-item fan out remains large and vice versa.
2
Recall trials that timed out by not returning a sample beyond the maximum time limit (100 samples) are not
plotted. These corresponded to 55% of the trials, resulting in a recall hit rate of 45%. In contrast, the average
recognition hit rate was 82% for this simulation.
4
Recognition
400
Recall
2.5
Recognition
Recall
400
300
300
200
200
100
100
d'
Trials
2
1.5
1
0.5
0
0
50
100
0
0
50
100
0
{2,7}
Sample count
{3,5}
{4,4}
{5,3}
{7,2}
Condition
Figure 2: (Left) Simulation results show easier recognition and harder recall given typical ecological
choices for stimuli and cue set sizes. (Right) Results from experiment manipulating the stimuli and
cue set size ratio. By manipulating the number of stimuli and cues, we predicted that we would
be able to make recall harder than recognition for experiment participants. The results support our
prediction unambiguously.Error bars show s.e.m.
We used a 2?2 within subject factorial design for this experiment, testing for the effect of the retrieval
mode - recognition/recall and either a stimulus heavy, or cue heavy selection of task materials. In
addition, we ran two conditions between subjects, using different parameterization of the stimuli/cue
ratios. In the stimulus heavy condition, for instance, participants were exposed to 5 stimuli associated
with 3 cues, while for the cue heavy condition, they saw 3 stimuli associated with 5 cues. The semantic
identity of the stimuli and cue sets were varied across all four conditions randomly, and the order of
presentation of conditions to participants was counterbalanced. All participants worked on all four
of the memory tasks, with interference avoided with the use of semantically distinct category pairs
across the four conditions. Specifically, we used number-letter, vegetable-occupation, fruit-adjective
and animal-place category pairs for the four conditions. Within each category, stimuli/cues for a
particular presentation were sampled from a 16 item master list, such that a stimulus could not occur
twice in conjunction with the same cue, but could occur in conjunction with multiple cues.
120 undergraduates participated in the experiment for course credit. Voluntary consent was obtained
from all participants, and the experimental protocol was approved by an institutional IRB. We told
experiment participants that they would be participating in a memory experiment, and their goal was
to remember as many of the items we showed them as possible. We also told them that the experiment
would have four parts, and that once they started working on a part, there would be no opportunity to
take a break until it ended. 80 participants performed the experiment with 3/5 and 5/3 stimulus-to-cue
mappings, 40 did it with 2/7 and 7/2 stimulus-to-cue mappings. Note that in all cases, participants
saw approximately the same number of total stimulus-cue bindings (3x5 = 15 or 2x7 = 14), thus
undergoing equivalent cognitive load during encoding.
Stimuli and cues were presented onscreen, with each pair appearing on the screen for 3 seconds,
followed by an ITI of equal duration. To prevent mnemonic strategy use at the time of encoding, the
horizontal orientation of the stimulus-cue pair was randomly selected on each trial, and participants
were not told beforehand which item category would be the cue; they could only discover this at
the time of retrieval3 . Participants were permitted to begin retrieval at their own discretion once
the encoding segment of the trial had concluded within each condition. All participants chose to
commence retrieval without delay. Participants were also permitted to take breaks of between 2-5
minutes between working on the different conditions, with several choosing to do so.
Once participants had seen all item-pairs for one of the conditions, the experiment prompted them to,
when ready, click on a button to proceed to the testing phase. In the recall condition, they saw a text
box and a sentence asking them to recall all the items that occurred alongside item X, where X was
randomly chosen from the set of possible cues for that condition; they responded by typing in the
words they remembered. For recognition, participants saw a sentence asking them to identify if X had
occurred alongside Y, where Y was randomly chosen from the set of possible cues for that condition.
3
An active weblink to the actual experiment is available online at [anonymized weblink].
5
After each forced yes/no response, a new X was shown. Half the X?s shown in the recognition test
were ?lures? , they had not been originally displayed alongside Y.
Memory performance was measured using d?, which is simply the difference between the z-normed
hit rate and false alarm rate, as is conventional in recognition experiments. d? is generally not used
to measure recall performance, since the number of true negatives is undefined in classic recall
experiments, which leaves the false alarm rate undefined as well. In our setup, the number of true
negatives is obviously the number of stimuli the participant saw that were not on the specific list
being probed, which is what we used to calculate d-prime for recall as well.
The right panel in Figure 2 illustrates the results of our experiment. The predicted crossover is
unambiguously observed. Further, changes in memory performance across the stimulus-cue set size
manipulation is symmetric across recognition and recall. This is precisely what we?d expect if set
size dependence was symmetrically affecting memory performance across both tasks as occurs in
our model. While not wishing to read too much into the symmetry of the quantitative result, we note
that such symmetry under a simple manipulation of the retrieval conditions appears to suggest that
the manipulation does in fact affect memory performance very strongly. Overall, the data strongly
supports our thesis - that quantitative differences in memory performance in recognition and recall
tasks are driven by differences in the set size of the underlying memory distribution being sampled.
The set size of the distribution being sampled, in turn, is determined by task constraints - and ends up
being symmetric when comparing single-item recognition with cued recall.
5
Predicting more recognition-recall dissociations
The fact that recognition is usually easier than recall - more accurate and quicker for the same stimuli
sets - is simply the most prominent difference between the two paradigms. Experimentalists have
uncovered a number of interesting manipulations in memory experiments that affect performance
on these tasks differentially. These are called recognition-recall dissociations, and are prominent
challenges to single-process accounts of the two tasks. Why should a manipulation affect only one
task and not the other if they are both outcomes of the same underlying process? [20] Previous
single-process accounts have had success in explaining some such dissociations. We focus here
on some that have proved relatively hard to explain without making inelegant dissociation-specific
assumptions in earlier accounts [13].
5.1
List strength effects and part set cuing
Unidimensional strength-based models of memory like SAM and REM fail to predict the list strength
effect [12] where participants? memory performance in free recall is lower than a controlled baseline
for weaker items on mixed lists (lists containing both strongly and weakly encoded items). Such
behavior is predicted easily by strength-based models. What they find difficult to explain is that
performance does not deviate from baseline in recognition tasks. The classical explanation for this
discrepancy is the use of a differentiation assumption. It is assumed that stronger items are associated
more strongly to the encoding context, however differences between the item itself as shown, and
its encoded image are also stronger. In free recall, this second interaction does not have an effect,
since the item itself is not presented, so a positive list strength effect is seen. In recognition, it is
conjectured that the two influences cancel each other out, resulting in a null list strength effect [13].
A lot of intricate assumptions have to hold for the differentiation account to hold. Our model has
a much simpler explanation for the null list-strength effect in recognition. Recognition involves
sampling based on the strength of the associative activation of the list given a specific item and so
is independent of the encoding strength of other items. On the other hand, recall involves sampling
from p(item|list) across all items, in which case, having a distribution favoring other items will
reduce the probability that the unstrengthened items will be sampled. Thus, the difference in which
variable the retrieval operation conditions on explains the respective presence and absence of a list
strength effect in recall and recognition.
The left panel in Figure 3 presents simulation results from our model reproducing this effect, where we
implement mixed lists by presenting certain stimuli more frequently during encoding and retrieve in
the usual manner. Hit rates are calculated for less frequently presented stimuli. For either elicitation
modality, the actual outcome of the retrieval itself is sampled from the appropriate conditional
6
distribution as a specific item/cue. In this particular experiment, which manipulates how much
training observers have on some of the items on the list, the histories entering the simulation are
generated such that some items co-occur with the future retrieval cue more frequently than others, i.e.
two items occur with a probability of 0.4 and 0.3 respectively, and three items occur with a probability
of 0.1 each alongside the cue. The simulation shows a positive list strength effect for recall (weaker
hit rates for less studied items) and a null list strength effect for recognition, congruent with data.
Our model also reconciles the results of [1] who demonstrated that the list strength effect does not
occur if we examine only items that are the first in their category to be retrieved. For categoryinsensitive strength-based accounts, this is a serious problem. For our account, which is explicitly
concerned with how observers co-encode stimuli and retrieval cues, this result is no great mystery.
For multi-category memory tests, the presence of each semantic category instantiates a novel list
during encoding, such that the strength-dependent updates during retrieval apply to each individual
p(item|list) and do not apply across the other category lists.
More generally, the dynamic nature of the sampled distribution in our Bayesian theory accommodates
the theoretical views of both champions of strength-dependent activation and retrieval-dependent
suppression [1]. Strength-dependent activation is present in our model in the form of the Bayesian
posterior over cue-relevant targets at the time when cued recall commences; retrieval-dependent
suppression of competitors is present in the form of normalization of the distribution during further
sequential Bayesian updates as the retrieval process continues. Assigning credit differentially to
individual categories predicts an attenuation (though not removal) of the list strength effect, due
to the absence of learning-induced changes for the first-tested items, as well diminishing memory
performance with testing position seen in [1].
Mixed
0.8
0.6
0.4
Recognition FAR
0.8
Recall HR
Recall FAR
0.6
0.4
0.2
0.2
0
Recognition HR
Baseline
1.0
Fraction
Weak item hit rate
1
Recognition
Recall
0
0
0.2
0.4
0.6
0.8
1
Prepend ratio
Figure 3: Reproducing (left) list strength effects and (right) the word frequency mirror effect using
our model.
The part set cueing effect is the observation that showing participants a subset of the items to be
recalled during retrieval reduces their recall performance for non-shown items [11]. This effect does
not appear in recognition experiments, which is again problematic for unidimensional strength-based
memory models. Our model has a simple explanation. The presented items during retrieval are simply
treated as further encoding opportunities for the seen items, resulting in a list strength imbalance as
above. This affects recall, but not recognition for the same reasons the list strength effect does.
5.2
Mirror effect
Another interesting effect that strength-based memory models have found hard to explain is the
word-frequency mirror effect [5]. This is seen when participants see two different classes of items
in recognition experiments. It is found, for instance, that unique items are both recognized more
accurately as previously seen and unseen in such experiments than common items. Such a pattern of
memory performance is contrary to the predictions of nearly all accounts of memory that depend
on unidimensional measures of memory strength, who can only model adaptive changes in memory
performance via shifts in the response criterion [19] that do not permit both the hit rate and the false
alarm rate to improve simultaneously.
7
The essential insight of the mirror effect is that some types of stimuli are intrinsically more memorable
than others, a common-sense observation that has proved surprisingly difficult for strength-based
memory models to assimilate.This difficulty extends to our own model also, but our inductive framework allows us to express the assumptions about information that the stimuli base frequency adds
to the picture in a clean way. Specifically, in our model observers use p(list|item) for recognition, which is high for unique items and low for common items by Bayesian inversion because
p(item|list)/p(item) ? 1 for unique items, because they are unlikely to have been encountered
outside the experimental context, and 1 for common items. In contrast, observers sample from
p(item|list) during recall, removing the effect of the frequency base rate p(item), so that the pattern
of results is inverted: performance is equivalent or better than baseline for common stimuli than for
rare ones [6], since they are more likely to be retrieved in general.
The right panel in Figure 3 shows simulation results using our model wherein we used two possible
cues during encoding, one to test performance during retrieval and one to modify the non-retrieval
frequency of stimuli encounters. For this experiment, which manipulates where we have to influence
how often the relative frequency with which the observers have seen the items in task-irrelevant
contexts other than the retrieval task, we prepended the base case history (of size 50 time steps)
with differently sized prior history samples (between 10 and 50 time steps long, in steps of 5),
wherein items co-occurred with cues that were not used during retrieval. The simulation results show
that, in recognition, hit rates drop and false alarm rates rise with more exposure to items outside
the experimental list context (high frequency items). Since our model assumes unambiguous cue
conditioning, it predicts unchanged performance from baseline for recall. More intricate models that
permit cue-cue associations may reproduce the advantage for common items documented empirically.
6
Discussion
We have made a very simple proposal in this paper. We join multiple previous authors in arguing that
memory retrieval in cued recall tasks can be interpreted as a question about the likelihood of retrieving
an item given the retrieval cue, typically the list of items given at the time of encoding [17, 8, 4].
We depart from previous authors in arguing that memory retrieval in item recognition tasks asks the
precisely opposite question: what is the likelihood of a given item having been associated with the
list? We integrated this insight into a simple inference-based model of memory encoding, which
shares its formal motivations with recent inference-based models of conditioning [3, 14], and an
approximately Bayesian model of memory retrieval, which samples memory frugally along lines
motivated on information-theoretic [18] and ecological grounds [16] by recent work.
Our model is meant to be expository and ignores several large issues that other richer models typically
engage with. For instance, it is silent about the time decay of memory particles, the partitioning of
the world into items and cues, and how it would go about explaining other more intricate memory
tasks like plurality discrimination and remember-know judgments. These omissions are deliberate, in
the sense that we wanted to present a minimal model to deliver the core intuition behind our approach
- that differences in memory performance in recognition and recall are attributable to no deeper
issue than an ecological preference to test memory using more items than lists. This observation
can now subsequently guide and constrain the construction of more realistic models of declarative
memory [3]. To the extent that differences traditionally used to posit dual-process accounts of memory
can be accounted for using simpler models like ours, the need to proliferate neuroanatomical and
process-level distinctions for various memory operations can be concomitantly reduced.
The distinction between recall and recognition memory also has important implications for the
presumed architecture of machine learning systems. Modern ML systems increasingly rely on a
combination of distributed representation of sensory information (using deep nets) and state-centric
representation of utility information (using reinforcement learning) to achieve human-like learning
and transfer capabilities, for example in simple Atari games [10]. The elicitation of class or category
membership in neural networks is quintessentially a recognition task, while the elicitation of state
value functions, as well as other intermediate computations in RL are clearly recall tasks. Partly
in realization of the large differences in the sort of memory required to support these two classes
of learning models, researchers have taken to postulating dual-process artificial memories [7]. Our
demonstration of the fundamental unitarity of the two modes of memory performance can and should
constrain the design of deep RL models in simpler ways.
8
References
[1] Karl-heinz B?uml. The list-strength effect: Strength-dependent competition or suppression? Psychonomic
Bulletin & Review, 4(2):260?264, 1997.
[2] Fergus IM Craik, Richard Govoni, Moshe Naveh-Benjamin, and Nicole D Anderson. The effects of divided
attention on encoding and retrieval processes in human memory. Journal of Experimental Psychology:
General, 125(2):159, 1996.
[3] Samuel J Gershman, David M Blei, and Yael Niv. Context, learning, and extinction. Psychological review,
117(1):197, 2010.
[4] Gary Gillund and Richard M Shiffrin. A retrieval model for both recognition and recall. Psychological
review, 91(1):1, 1984.
[5] Murray Glanzer and John K Adams. The mirror effect in recognition memory: data and theory. Journal of
Experimental Psychology: Learning, Memory, and Cognition, 16(1):5, 1990.
[6] Vernon Gregg. Word frequency, recognition and recall. John Wiley & Sons, 1976.
[7] Dharshan Kumaran, Demis Hassabis, and James L McClelland. What learning systems do intelligent agents
need? complementary learning systems theory updated. Trends in Cognitive Sciences, 20(7):512?534,
2016.
[8] George Mandler. Recognizing: The judgment of previous occurrence. Psychological review, 87(3):252,
1980.
[9] James L McClelland, Bruce L McNaughton, and Randall C O?reilly. Why there are complementary learning
systems in the hippocampus and neocortex: insights from the successes and failures of connectionist
models of learning and memory. Psychological review, 102(3):419, 1995.
[10] Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Andrei A Rusu, Joel Veness, Marc G Bellemare,
Alex Graves, Martin Riedmiller, Andreas K Fidjeland, Georg Ostrovski, et al. Human-level control through
deep reinforcement learning. Nature, 518(7540):529?533, 2015.
[11] Raymond S Nickerson. Retrieval inhibition from part-set cuing: A persisting enigma in memory research.
Memory & cognition, 12(6):531?552, 1984.
[12] Roger Ratcliff, Steven E Clark, and Richard M Shiffrin. List-strength effect: I. data and discussion. Journal
of Experimental Psychology: Learning, Memory, and Cognition, 16(2):163, 1990.
[13] Richard M Shiffrin and Mark Steyvers. A model for recognition memory: Rem?retrieving effectively
from memory. Psychonomic bulletin & review, 4(2):145?166, 1997.
[14] Nisheeth Srivastava and Paul R Schrater. Classical conditioning via inference over observable situation
contexts. In Proceedings of the Annual Meeting of the Cognitive Science Society, 2014.
[15] Saul Sternberg. Memory-scanning: Mental processes revealed by reaction-time experiments. American
scientist, 57(4):421?457, 1969.
[16] Peter M Todd and Gerd Gigerenzer. Environments that make us smart: Ecological rationality. Current
Directions in Psychological Science, 16(3):167?171, 2007.
[17] Endel Tulving and Donald M Thomson. Retrieval processes in recognition memory: Effects of associative
context. Journal of Experimental Psychology, 87(1):116, 1971.
[18] Edward Vul, Noah Goodman, Thomas L Griffiths, and Joshua B Tenenbaum. One and done? optimal
decisions from very few samples. Cognitive science, 38(4):599?637, 2014.
[19] John T Wixted. Dual-process theory and signal-detection theory of recognition memory. Psychological
review, 114(1):152, 2007.
[20] Andrew P Yonelinas. The nature of recollection and familiarity: A review of 30 years of research. Journal
of memory and language, 46(3):441?517, 2002.
9
| 6633 |@word trial:7 illustrating:1 inversion:1 stronger:3 approved:1 hippocampus:1 extinction:1 termination:2 additively:1 simulation:9 irb:1 attended:1 asks:1 harder:4 contains:1 exclusively:1 uncovered:1 ours:2 reaction:1 current:1 comparing:1 activation:13 assigning:1 atop:1 john:3 realistic:1 engendered:1 wanted:1 designed:2 drop:1 update:4 discrimination:1 cue:48 half:2 discovering:1 item:94 fewer:2 parameterization:1 selected:1 leaf:1 prepended:1 core:1 blei:1 mental:1 cse:1 sits:1 successive:1 preference:1 simpler:4 gillund:2 along:1 constructed:2 become:2 retrieving:5 qualitative:2 descendant:1 combine:1 behavioral:2 manner:1 dispute:1 presumed:1 intricate:4 behavior:2 frequently:3 examine:1 multi:1 brain:1 heinz:1 rem:2 actual:2 becomes:1 begin:1 discover:1 underlying:2 competently:1 panel:4 null:3 what:5 atari:1 interpreted:1 dharshan:1 proposing:1 differentiation:2 ended:1 assert:1 remember:6 quantitative:3 every:1 attenuation:1 returning:1 hit:8 partitioning:1 control:1 converse:1 appear:1 before:3 positive:2 attend:1 scientist:1 treat:2 tends:1 limit:1 modify:1 todd:1 encoding:20 rigidly:1 discretion:1 approximately:6 lure:1 chose:1 twice:1 studied:2 co:9 statistically:1 unique:3 responsible:4 arguing:4 testing:3 implement:1 procedure:1 demis:1 episodic:1 riedmiller:1 empirical:1 crossover:3 significantly:1 reilly:1 word:4 road:1 donald:1 griffith:1 suggest:2 selection:1 context:7 influence:2 bellemare:1 equivalent:2 conventional:1 demonstrated:1 nicole:1 go:2 attention:2 starting:1 duration:1 commence:1 normed:1 exposure:1 immediately:2 lineage:1 manipulates:2 rule:2 insight:3 array:1 importantly:1 inelegant:1 retrieve:1 steyvers:1 classic:5 century:1 traditionally:1 mcnaughton:1 updated:1 target:1 construction:1 heavily:2 engage:1 rationality:1 us:1 agreement:1 trend:1 gilman:1 recognition:87 nisheeth:2 continues:1 predicts:4 observed:2 steven:1 quicker:1 capture:1 calculate:1 ensures:1 ran:2 accessing:1 intuition:1 benjamin:1 complexity:3 environment:1 asked:3 inductively:1 dynamic:1 controversy:1 terminating:2 depend:2 weakly:1 segment:1 gigerenzer:1 exposed:1 smart:1 purely:1 deliver:1 completely:1 easily:1 stylized:1 joint:4 differently:1 iit:1 various:1 distinct:1 fast:2 forced:1 query:2 artificial:1 corresponded:1 choosing:1 outcome:2 outside:2 apparent:1 encoded:5 larger:1 richer:1 breached:1 unseen:1 think:2 itself:6 associative:7 online:1 obviously:1 advantage:1 evidently:1 net:1 interaction:1 coming:1 relevant:6 realization:1 gregg:1 consent:1 shiffrin:4 achieve:1 participating:1 competition:2 differentially:2 congruent:1 adam:1 silver:1 object:3 cued:3 illustrate:1 andrew:1 ac:1 measured:3 naveh:1 edward:2 predicted:3 involves:5 implies:1 memorize:1 direction:1 untested:1 posit:1 correct:1 subsequently:1 human:3 opinion:1 material:1 explains:2 premise:1 niv:1 plurality:1 im:1 hold:5 around:1 credit:2 ground:1 great:1 mapping:3 predict:1 cognition:3 m0:1 consecutive:1 institutional:1 propensity:1 bridge:1 saw:5 vice:1 champion:1 clearly:1 always:1 unitarity:1 rather:1 rusu:1 probabilistically:2 conjunction:3 encode:1 focus:1 consistently:1 bernoulli:1 likelihood:2 ratcliff:1 contrast:6 suppression:3 wishing:1 sense:2 baseline:5 inference:3 enigma:1 dependent:6 stopping:1 membership:1 typically:4 unlikely:1 integrated:1 diminishing:1 favoring:1 manipulating:3 reproduce:1 expressible:1 issue:2 overall:1 dual:11 among:2 orientation:1 animal:1 field:1 once:3 equal:1 having:4 beach:1 sampling:18 koray:1 identical:3 veness:1 broad:1 cancel:1 nearly:2 discrepancy:1 future:1 others:3 stimulus:31 intelligent:1 serious:1 primarily:1 few:2 modern:1 randomly:4 richard:4 connectionist:1 simultaneously:1 individual:2 phase:1 detection:2 prepend:1 assimilate:1 ostrovski:1 mnih:1 joel:1 undefined:2 behind:1 implication:1 accurate:1 beforehand:1 unification:2 respective:1 concomitantly:1 timed:1 plotted:1 theoretical:2 minimal:1 mk:2 psychological:6 instance:6 modeling:1 rationalizes:2 asking:2 earlier:1 contiguous:1 subset:2 rare:1 uniform:1 delay:1 successful:1 recognizing:1 too:1 stored:1 answer:2 scanning:1 st:1 fundamental:4 accessible:1 stay:1 probabilistic:5 told:3 synthesis:1 thesis:1 again:1 containing:1 priority:1 cognitive:4 american:1 derivative:1 style:1 account:17 volodymyr:1 postulated:2 explicitly:2 race:1 depends:1 performed:1 try:1 observer:13 view:2 break:2 observing:1 lot:1 sort:2 participant:20 capability:1 bruce:1 contribution:1 formed:1 gerd:1 responded:1 who:2 dissociation:10 judgment:3 succession:1 yield:1 identify:1 yes:1 weak:2 bayesian:10 kavukcuoglu:1 accurately:1 monitoring:1 drive:1 researcher:2 history:3 explain:4 tended:2 competitor:1 failure:1 frequency:12 involved:1 james:2 obvious:1 associated:6 sampled:8 cueing:1 experimenter:2 proved:2 intrinsically:1 ask:3 recall:81 actually:1 back:1 appears:1 centric:1 higher:1 originally:1 unambiguously:2 wherein:4 improved:1 rationalize:1 response:3 permitted:2 done:1 box:1 strongly:4 though:1 anderson:1 roger:1 glanzer:1 until:1 hand:2 working:3 horizontal:1 mode:2 nuanced:1 usa:1 effect:29 verify:1 true:4 former:1 inductive:1 read:1 symmetric:2 entering:1 satisfactory:1 semantic:3 x5:1 during:16 self:3 game:1 unambiguous:1 samuel:1 criterion:1 prominent:4 hippocampal:1 presenting:1 neocortical:1 theoretic:1 confusion:1 thomson:1 image:1 invoked:1 novel:2 superior:1 common:7 psychonomic:2 empirically:1 rl:2 conditioning:3 association:1 occurred:3 marginals:1 numerically:1 schrater:1 composition:1 versa:1 similarly:1 particle:1 language:1 had:5 overtly:1 persisting:1 access:1 longer:2 inhibition:1 base:5 add:1 something:2 posterior:1 own:2 showed:1 recent:2 retrieved:5 conjectured:1 jolla:1 driven:1 prime:1 manipulation:5 irrelevant:1 certain:1 ecological:9 success:3 remembered:1 meeting:1 vul:2 joshua:1 yonelinas:1 inverted:1 seen:10 greater:2 george:1 recognized:1 novelty:2 determine:1 paradigm:1 signal:3 mandler:5 multiple:5 reduces:2 offer:1 long:3 retrieval:54 mnemonic:1 divided:2 equally:1 manipulate:1 controlled:1 prediction:3 basic:1 experimentalists:1 craik:1 normalization:1 proposal:1 whereas:1 want:1 conditionals:1 addition:1 participated:1 affecting:1 situational:2 concluded:1 modality:4 goodman:1 subject:3 induced:1 contrary:2 seem:1 presence:3 ideal:1 symmetrically:1 intermediate:1 easy:1 concerned:1 revealed:1 marginalization:1 affect:4 psychology:5 counterbalanced:1 architecture:1 opposite:2 click:1 reduce:1 silent:1 andreas:1 unidimensional:4 shift:1 whether:1 motivated:1 utility:1 effort:1 peter:1 proceed:1 deep:3 useful:3 generally:2 vegetable:1 involve:1 factorial:1 neocortex:1 tenenbaum:1 mcclelland:3 category:10 documented:2 reduced:1 deliberate:1 nickerson:1 problematic:1 per:2 discrete:2 probed:3 affected:2 express:1 georg:1 four:5 threshold:8 drawn:4 prevent:1 clean:1 timestep:1 button:1 fraction:1 year:1 run:1 mystery:1 letter:1 master:1 striking:1 arrive:1 almost:1 place:1 extends:1 decision:1 followed:1 fan:1 encountered:1 annual:1 activity:2 strength:35 occur:7 noah:1 precisely:2 worked:1 constraint:1 constrain:2 alex:1 flat:1 sternberg:1 x7:1 speed:2 argument:1 relatively:2 martin:1 uml:1 influential:3 expository:1 combination:1 instantiates:1 smaller:2 across:7 reconstructing:1 sam:6 increasingly:1 son:1 making:4 vernon:1 randall:1 explained:3 interference:1 taken:1 previously:6 remains:1 turn:1 count:1 fail:2 mechanism:1 needed:2 know:2 nexus:1 tulving:1 end:1 available:2 operation:3 yael:1 permit:2 apply:2 appropriate:1 occurrence:8 robustly:1 appearing:1 encounter:1 hassabis:1 thomas:1 neuroanatomical:1 assumes:3 denotes:1 remaining:1 spill:1 opportunity:2 instant:1 cuing:2 build:1 murray:1 classical:4 society:1 unchanged:1 question:8 moshe:1 occurs:2 depart:1 strategy:1 dependence:1 usual:1 onscreen:1 fidjeland:1 recollection:10 accommodates:1 sensible:1 argue:2 extent:3 declarative:5 reason:2 modeled:4 prompted:1 ratio:3 demonstration:1 setup:1 difficult:2 trace:3 negative:2 rise:1 design:2 allowing:1 pmk:2 imbalance:1 observation:4 kumaran:1 iti:1 inevitably:1 displayed:2 endemic:1 voluntary:1 situation:1 looking:1 ucsd:2 frugal:1 varied:1 reproducing:2 omission:1 david:2 inverting:1 pair:5 required:2 specified:1 sentence:2 recalled:2 distinction:3 nip:1 beyond:2 suggested:1 able:1 bar:1 pattern:4 alongside:4 usually:1 elicitation:3 challenge:1 adjective:1 built:1 including:1 memory:102 explanation:4 phenomenologically:1 event:4 critical:1 difficulty:2 typing:1 predicting:1 treated:1 hr:2 rely:1 improve:1 temporally:1 picture:1 started:1 ready:1 raymond:1 text:1 prior:2 literature:1 deviate:1 removal:1 review:8 relative:5 engram:9 occupation:1 graf:1 expect:2 mixed:3 interesting:2 gershman:1 clark:1 agent:1 anonymized:1 fruit:1 storing:1 share:1 heavy:4 karl:1 course:1 accounted:1 surprisingly:1 free:3 keeping:1 bias:3 weaker:2 formal:1 deeper:1 guide:1 explaining:4 saul:1 face:1 bulletin:2 distributed:2 depth:1 calculated:1 world:2 ignores:1 author:3 made:2 adaptive:1 universally:1 avoided:1 sensory:1 reinforcement:2 far:2 approximate:1 observable:1 iitk:1 ml:1 reproduces:2 active:1 incoming:1 assumed:1 xi:11 fergus:1 search:1 why:2 terminate:2 nature:3 transfer:1 ca:2 symmetry:2 cl:1 protocol:1 wixted:2 reconciles:1 did:1 marc:1 main:1 motivation:2 proliferate:1 alarm:4 arrival:1 paul:1 nothing:1 evul:1 convey:1 complementary:2 memorable:1 representative:1 join:1 screen:1 andrei:1 postulating:1 slow:2 attributable:1 wiley:1 formalization:1 position:1 xl:11 learns:1 kanpur:2 minute:1 removing:1 familiarity:8 specific:5 load:1 showing:1 undergoing:1 list:51 decay:1 virtue:1 operationalization:1 essential:1 undergraduate:1 false:4 sequential:4 effectively:1 mirror:5 magnitude:2 notwithstanding:1 conditioned:2 illustrates:1 gap:1 easier:6 simply:8 likely:1 indifference:1 tracking:1 binding:1 gary:1 environmental:1 succeed:1 conditional:5 identity:1 endeavor:1 presentation:3 goal:1 sized:1 absence:2 change:4 hard:3 typical:2 specifically:2 determined:1 semantically:1 total:2 called:1 pas:1 partly:1 experimental:8 la:1 pragmatic:1 internal:1 people:1 support:9 mark:1 commences:1 meant:1 dept:1 tested:2 srivastava:2 |
6,228 | 6,634 | On Structured Prediction Theory with Calibrated
Convex Surrogate Losses
Anton Osokin
INRIA/ENS?, Paris, France
HSE?, Moscow, Russia
Francis Bach
INRIA/ENS?, Paris, France
Simon Lacoste-Julien
MILA and DIRO
Universit? de Montr?al, Canada
Abstract
We provide novel theoretical insights on structured prediction in the context of
efficient convex surrogate loss minimization with consistency guarantees. For any
task loss, we construct a convex surrogate that can be optimized via stochastic
gradient descent and we prove tight bounds on the so-called ?calibration function?
relating the excess surrogate risk to the actual risk. In contrast to prior related
work, we carefully monitor the effect of the exponential number of classes in the
learning guarantees as well as on the optimization complexity. As an interesting
consequence, we formalize the intuition that some task losses make learning harder
than others, and that the classical 0-1 loss is ill-suited for structured prediction.
1
Introduction
Structured prediction is a subfield of machine learning aiming at making multiple interrelated
predictions simultaneously. The desired outputs (labels) are typically organized in some structured
object such as a sequence, a graph, an image, etc. Tasks of this type appear in many practical domains
such as computer vision [34], natural language processing [42] and bioinformatics [19].
The structured prediction setup has at least two typical properties differentiating it from the classical
binary classification problems extensively studied in learning theory:
1. Exponential number of classes: this brings both additional computational and statistical challenges.
By exponential, we mean exponentially large in the size of the natural dimension of output, e.g., the
number of all possible sequences is exponential w.r.t. the sequence length.
2. Cost-sensitive learning: in typical applications, prediction mistakes are not all equally costly.
The prediction error is usually measured with a highly-structured task-specific loss function, e.g.,
Hamming distance between sequences of multi-label variables or mean average precision for ranking.
Despite many algorithmic advances to tackle structured prediction problems [4, 35], there have been
relatively few papers devoted to its theoretical understanding. Notable recent exceptions that made
significant progress include Cortes et al. [13] and London et al. [28] (see references therein) which
proposed data-dependent generalization error bounds in terms of popular empirical convex surrogate
losses such as the structured hinge loss [44, 45, 47]. A question not addressed by these works is
whether their algorithms are consistent: does minimizing their convex bounds with infinite data lead
to the minimization of the task loss as well? Alternatively, the structured probit and ramp losses are
consistent [31, 30], but non-convex and thus it is hard to obtain computational guarantees for them.
In this paper, we aim at getting the property of consistency for surrogate losses that can be efficiently
minimized with guarantees, and thus we consider convex surrogate losses.
The consistency of convex surrogates is well understood in the case of binary classification [50, 5, 43]
and there is significant progress in the case of multi-class 0-1 loss [49, 46] and general multi?
?
DI ?cole normale sup?rieure, CNRS, PSL Research University
National Research University Higher School of Economics
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
class loss functions [3, 39, 48]. A large body of work specifically focuses on the related tasks of
ranking [18, 9, 40] and ordinal regression [37].
Contributions. In this paper, we study consistent convex surrogate losses specifically in the context
of an exponential number of classes. We argue that even while being consistent, a convex surrogate
might not allow efficient learning. As a concrete example, Ciliberto et al. [10] recently proposed a
consistent approach to structured prediction, but the constant in their generalization error bound can
be exponentially large as we explain in Section 5. There are two possible sources of difficulties from
the optimization perspective: to reach adequate accuracy on the task loss, one might need to optimize
a surrogate loss to exponentially small accuracy; or to reach adequate accuracy on the surrogate loss,
one might need an exponential number of algorithm steps because of exponentially large constants
in the convergence rate. We propose a theoretical framework that jointly tackles these two aspects
and allows to judge the feasibility of efficient learning. In particular, we construct a calibration
function [43], i.e., a function setting the relationship between accuracy on the surrogate and task
losses, and normalize it by the means of convergence rate of an optimization algorithm.
Aiming for the simplest possible application of our framework, we propose a family of convex
surrogates that are consistent for any given task loss and can be optimized using stochastic gradient
descent. For a special case of our family (quadratic surrogate), we provide a complete analysis
including general lower and upper bounds on the calibration function for any task loss, with exact
values for the 0-1, block 0-1 and Hamming losses. We observe that to have a tractable learning
algorithm, one needs both a structured loss (not the 0-1 loss) and appropriate constraints on the
predictor, e.g., in the form of linear constraints for the score vector functions. Our framework also
indicates that in some cases it might be beneficial to use non-consistent surrogates. In particular, a
non-consistent surrogate might allow optimization only up to specific accuracy, but exponentially
faster than a consistent one.
We introduce the structured prediction setting suitable for studying consistency in Sections 2 and 3.
We analyze the calibration function for the quadratic surrogate loss in Section 4. We review the
related works in Section 5 and conclude in Section 6.
2
Structured prediction setup
In structured prediction, the goal is to predict a structured output y ? Y (such as a sequence, a graph,
an image) given an input x ? X . The quality of prediction is measured by a task-dependent loss
? y | x) ? 0 specifying the cost for predicting y? when the correct output is y. In this
function L(y,
paper, we consider the case when the number of possible predictions and the number of possible
labels are both finite. For simplicity,1 we also assume that the sets of possible predictions and correct
outputs always coincide and do not depend on x. We refer to this set as the set of labels Y, denote its
cardinality by k, and map its elements to 1, . . . , k. In this setting, assuming that the loss function
depends only on y? and y, but not on x directly, the loss is defined by a loss matrix L ? Rk?k . We
assume that all the elements of the matrix L are non-negative and will use Lmax to denote the maximal
element. Compared to multi-class classification, k is typically exponentially large in the size of the
natural dimension of y, e.g., contains all possible sequences of symbols from a finite alphabet.
Following standard practices in structured prediction [12, 44], we define the prediction model by
a score function f : X ? Rk specifying a score fy (x) for each possible output y ? Y. The final
prediction is done by selecting a label with the maximal value of the score
pred(f(x)) := argmax fy? (x),
(1)
?
y?Y
with some fixed strategy to resolve ties. To simplify the analysis, we assume that among the labels
with maximal scores, the predictor always picks the one with the smallest index.
The goal of prediction-based machine learning consists in finding a predictor that works well on
the unseen test set, i.e., data points coming from the same distribution D as the one generating the
training data. One way to formalize this is to minimize the generalization error, often referred to as
the actual (or population) risk based on the loss L,
RL (f) := IE(x,y)?D L pred(f(x)), y .
(2)
Minimizing the actual risk (2) is usually hard. The standard approach is to minimize a surrogate risk,
which is a different objective easier to optimize, e.g., convex. We define a surrogate loss as a function
1
Our analysis is generalizable to rectangular losses, e.g., ranking losses studied by Ramaswamy et al. [40].
2
? : Rk ? Y ? R depending on a score vector f = f(x) ? Rk and a target label y ? Y as input
arguments. We denote the y-th component of f with fy . The surrogate risk (the ?-risk) is defined as
R? (f) := IE(x,y)?D ?(f(x), y),
(3)
where the expectation is taken w.r.t. the data-generating distribution D. To make the minimization
of (3) well-defined, we always assume that the surrogate loss ? is bounded from below and continuous.
Examples of common surrogate
losses include the structured hinge-loss [44, 47] ?SSVM (f , y) :=
? y) ? fy , the logP
maxy?Y
fy? + L(y,
loss (maximum likelihood learning) used, e.g., in conditional
?
random fields [25], ?log (f , y) := log( y?Y
exp fy? ) ? fy , and their hybrids [38, 21, 22, 41].
?
? y) := [y? 6= y],2 and the
In terms of task losses, we consider the unstructured 0-1 loss L01 (y,
? y) :=
two following structured losses: block 0-1 loss with b equal blocks of labels L01,b (y,
[y? and y are not in the same block]; and (normalized) Hamming loss between tuples of T binary
P
? y) := T1 Tt=1 [?
yt 6= yt ]. To illustrate some aspects of our analysis, we also
variables yt : LHam,T (y,
look at the mixed loss L01,b,? : a convex combination of the 0-1 and block 0-1 losses, defined as
L01,b,? := ?L01 + (1 ? ?)L01,b for some ? ? [0, 1].
3
Consistency for structured prediction
3.1
Calibration function
We now formalize the connection between the actual risk RL and the surrogate ?-risk R? via the
so-called calibration function, see Definition 1 below [5, 49, 43, 18, 3]. As it is standard for this
kind of analysis, the setup is non-parametric, i.e. it does not take into account the dependency of
scores on input variables x. For now, we assume that a family of score functions FF consists of all
vector-valued Borel measurable functions f : X ? F where F ? Rk is a subspace of allowed score
vectors, which will play an important role in our analysis. This setting is equivalent to a pointwise
analysis, i.e, looking at the different input x independently. We bring the dependency on the input
back into the analysis in Section 3.3 where we assume a specific family of score functions.
Let DX represent the marginal distribution for D on x and IP(? | x) denote its conditional given x.
We can now rewrite the risk RL and ?-risk R? as
RL (f) = IEx?DX `(f(x), IP(? | x)), R? (f) = IEx?DX ?(f(x), IP(? | x)),
where the conditional risk ` and the conditional ?-risk ? depend on a vector of scores f and a
conditional distribution on the set of output labels q as
Xk
Xk
`(f , q) :=
qc L(pred(f ), c), ?(f , q) :=
qc ?(f , c).
c=1
c=1
The calibration function H?,L,F between the surrogate loss ? and the task loss L relates the excess
surrogate risk with the actual excess risk via the excess risk bound:
H?,L,F (?`(f , q)) ? ??(f , q), ?f ? F, ?q ? ?k ,
(4)
?
?
where ??(f , q) = ?(f , q) ? inf ? ?(f , q), ?`(f , q) = `(f , q) ? inf ? `(f , q) are the excess
f ?F
f ?F
risks and ?k denotes the probability simplex on k elements.
In other words, to find a vector f that yields an excess risk smaller than ?, we need to optimize the
?-risk up to H?,L,F (?) accuracy (in the worst case). We make this statement precise in Theorem 2
below, and now proceed to the formal definition of the calibration function.
Definition 1 (Calibration function). For a task loss L, a surrogate loss ?, a set of feasible scores F,
the calibration function H?,L,F (?) (defined for ? ? 0) equals the infimum excess of the conditional
surrogate risk when the excess of the conditional actual risk is at least ?:
H?,L,F (?) := inf ??(f , q)
(5)
f ?F , q??k
s.t. ?`(f , q) ? ?.
We set H?,L,F (?) to +? when the feasible set is empty.
(6)
By construction, H?,L,F is non-decreasing on [0, +?), H?,L,F (?) ? 0, the inequality (4) holds,
and H?,L,F (0) = 0. Note that H?,L,F can be non-convex and even non-continuous (see examples
in Figure 1). Also, note that large values of H?,L,F (?) are better.
2
Here we use the Iverson bracket notation, i.e., [A] := 1 if a logical expression A is true, and zero otherwise.
3
no constraints
tight constraints
H(")
H(")
no constraints
tight constraints
0
0.5
1
0
"
(a): Hamming loss LHam,T
0.2
0.4
"
(b): Mixed loss L01,b,0.4
Figure 1: Calibration functions for the quadratic surrogate ?quad (12) defined in Section 4 and two
different task losses. (a) ? the calibration functions for the Hamming loss LHam,T when used without
constraints on the scores, F = Rk (in red), and with the tight constraints implying consistency,
F = span(LHam,T ) (in blue). The red curve can grow exponentially slower than the blue one. (b) ?
the calibration functions for the mixed loss L01,b,? with ? = 0.4 (see Section 2 for the definition) when
used without constraints on the scores (red) and with tight constraints for the block 0-1 loss (blue).
The blue curve represents level-0.2 consistency. The calibration function equals zero for ? ? ?/2,
but grows exponentially faster than the red curve representing a consistent approach and thus could
be better for small ?. More details on the calibration functions in this figure are given in Section 4.
3.2
Notion of consistency
We use the calibration function H?,L,F to set a connection between optimizing the surrogate and
task losses by Theorem 2, which is similar to Theorem 3 of Zhang [49].
Theorem 2 (Calibration connection). Let H?,L,F be the calibration function between the surrogate
? ?,L,F be a convex non-decreasing
loss ? and the task loss L with feasible set of scores F ? Rk . Let H
lower bound of the calibration function. Assume that ? is continuous and bounded from below. Then,
? ?,L,F (?) and any f ? FF , we have
for any ? > 0 with finite H
? ?,L,F (?) ? RL (f) < R? + ?,
R? (f) < R? + H
(7)
?,F
L,F
where R??,F := inf f?FF R? (f) and R?L,F := inf f?FF RL (f).
Proof. We take the expectation of (4) w.r.t. x, where the second argument of ` is set to the conditional
? ?,L,F is convex) to get
distribution IP(? | x). Then, we apply Jensen?s inequality (since H
? ?,L,F (RL (f) ? R?L,F ) ? R? (f) ? R??,F < H
? ?,L,F (?),
H
(8)
? ?,L,F .
which implies (7) by monotonicity of H
? ?,L,F (?) required by Theorem 2 always exists, e.g.,
A suitable convex non-decreasing lower bound H
the zero constant. However, in this case Theorem 2 is not informative, because the l.h.s. of (7) is
? ?,L,F defined as the lower convex envelope of
never true. Zhang [49, Proposition 25] claims that H
? ?,L,F (?) > 0, ?? > 0, if H?,L,F (?) > 0, ?? > 0, and,
the calibration function H?,L,F satisfies H
? ?,L,F always exists and
e.g., the set of labels is finite. This statement implies that an informative H
allows to characterize consistency through properties of the calibration function H?,L,F .
We now define a notion of level-? consistency, which is more general than consistency.
Definition 3 (level-? consistency). A surrogate loss ? is consistent up to level ? ? 0 w.r.t. a task
loss L and a set of scores F if and only if the calibration function satisfies H?,L,F (?) > 0 for all
? > ? and there exists ?? > ? such that H?,L,F (?
?) is finite.
Looking solely at (standard level-0) consistency vs. inconsistency might be too coarse to capture
practical properties related to optimization accuracy (see, e.g., [29]). For example, if H?,L,F (?) = 0
only for very small values of ?, then the method can still optimize the actual risk up to a certain
level which might be good enough in practice, especially if it means that it can be optimized faster.
Examples of calibration functions for consistent and inconsistent surrogate losses are shown in
Figure 1.
Other notions of consistency. Definition 3 with ? = 0 and F = Rk results in the standard setting
often appearing in the literature. In particular, in this case Theorem 2 implies Fisher consistency as
4
formulated, e.g., by Pedregosa et al. [37] for general losses and Lin [27] for binary classification.
This setting is also closely related to many definitions of consistency used in the literature. For
example, for a bounded from below and continuous surrogate, it is equivalent to infinite-sample
consistency [49], classification calibration [46], edge-consistency [18], (L, Rk )-calibration [39],
prediction calibration [48]. See [49, Appendix A] for the detailed discussion.
Role of F. Let the approximation error for the restricted set of scores F be defined as R?L,F ?R?L :=
inf f?FF RL (f) ? inf f RL (f). For any conditional distribution q, the score vector f := ?Lq will
yield an optimal prediction. Thus the condition span(L) ? F is sufficient for F to have zero
approximation error for any distribution D, and for our 0-consistency condition to imply the standard
Fisher consistency with respect to L. In the following, we will see that a restricted F can both play a
role for computational efficiency as well as statistical efficiency (thus losses with smaller span(L)
might be easier to work with).
3.3
Connection to optimization accuracy and statistical efficiency
The scale of a calibration function is not intrinsically well-defined: we could multiply the surrogate
function by a scalar and it would multiply the calibration function by the same scalar, without
changing the optimization problem. Intuitively, we would like the surrogate loss to be of order 1. If
with this scale the calibration function is exponentially small (has a 1/k factor), then we have strong
evidence that the stochastic optimization will be difficult (and thus learning will be slow).
To formalize this intuition, we add to the picture the complexity of optimizing the surrogate loss with
a stochastic approximation algorithm. By using a scale-invariant convergence rate, we provide a
natural normalization of the calibration function. The following two observations are central to the
theoretical insights provided in our work:
1. Scale. For a properly scaled surrogate loss, the scale of the calibration function is a good indication
of whether a stochastic approximation algorithm will take a large number of iterations (in the worst
case) to obtain guarantees of small excess of the actual risk (and vice-versa, a large coefficient
indicates a small number of iterations). The actual verification requires computing the normalization
quantities given in Theorem 6 below.
2. Statistics. The bound on the number of iterations directly relates to the number of training
examples that would be needed to learn, if we see each iteration of the stochastic approximation
algorithm as using one training example to optimize the expected surrogate.
To analyze the statistical convergence of surrogate risk optimization, we have to specify the set of
score functions that we work with. We assume that the structure on input x ? X is defined by a
positive definite kernel K : X ? X ? R. We denote the corresponding reproducing kernel Hilbert
space (RKHS) by H and its explicit feature map by ?(x) ? H. By the reproducing property, we
have hf, ?(x)iH = f (x) for all x ? X , f ? H, where h?, ?iH is the inner product in the RKHS. We
define the subspace of allowed scores F ? Rk via the span of the columns of a matrix F ? Rk?r .
The matrix F explicitly defines the structure of the score function. With this notation, we will assume
that the score function is of the form f(x) = F W ?(x), where W : H ? Rr is a linear operator
to be learned (a matrix if H is of finite dimension) that represents a collection of r elements in H,
transforming ?(x) to a vector in Rr by applying the RKHS inner product r times.3 Note that for
structured losses, we usually have r k. The set of all score functions is thus obtained by varying W
in this definition and is denoted by FF,H . As a concrete example of a score family FF,H for structured
prediction, consider the standard sequence model with unary and pairwise potentials. In this case, the
dimension r equals T s + (T ? 1)s2 , where T is the sequence length and s is the number of labels
of each variable. The columns of the matrix F consist of 2T ? 1 groups (one for each unary and
pairwise potential). Each row of F has exactly one entry equal to one in each column group (with
zeros elsewhere).
In this setting, we use the online projected averaged stochastic subgradient descent ASGD4 (stochastic
w.r.t. data (x(n) , y (n) ) ? D) to minimize the surrogate risk directly [6]. The n-th update consists in
W (n) := PD W (n?1) ? ? (n) F T ???(x(n) )T ,
(9)
3
Note that if rank(F ) = r, our setup is equivalent to assuming a joint kernel [47] in the product form:
Kjoint ((x, c), (x0 , c0 )) := K(x, x0 )F (c, :)F (c0 , :)T , where F (c, :) is the row c for matrix F .
4
See, e.g., [36] for the formal setup of kernel ASGD.
5
where F T ???(x(n) )T : H ? Rr is the stochastic functional gradient, ? (n) is the step size
and PD is the projection on the ball of radius D w.r.t. the Hilbert?Schmidt norm5. The vector
?? ? Rk is a regular gradient of the sampled surrogate ?(f(x(n) ), y (n) ) w.r.t. the scores, ?? =
?f ?(f , y (n) )|f =f(x(n) ) . We wrote the above update using an explicit feature map ? for notational
simplicity, but kernel ASGD can also be implemented without it by using the kernel trick. The
convergence properties of ASGD in RKHS are analogous to the finite-dimensional ASGD because
they rely on dimension-free quantities. To use a simple convergence analysis, we follow Ciliberto
et al. [10] and make the following simplifying assumption:
Assumption 4 (Well-specified optimization w.r.t. the function class FF,H ). The distribution D is
such that R??,F := inf f?FF R? (f) has some global minimum f? that also belongs to FF,H .
Assumption 4 simply means that each row of W ? defining f? belongs to the RKHS H implying
a finite norm kW ? kHS . Assumption 4 can be relaxed if the kernel K is universal, but then the
convergence analysis becomes much more complicated [36].
Theorem 5 (Convergence rate). Under Assumption 4 and assuming that (i) the functions ?(f , y)
are bounded from below and convex w.r.t. f ? Rk for all y ? Y; (ii) the expected square of the norm
of the stochastic gradient is bounded, IE(x,y)?D kF T ???(x)T k2HS ? M 2 and (iii) kW ? kHS ? D,
?
then running the ASGD algorithm (9) with the constant step-size ? := M2D
for N steps admits the
N
(N )
?
following expected suboptimality for the averaged iterate f :
XN
?f(N ) := 1
?
IE[R? (?f(N ) )] ? R??,F ? 2DM
where
F W (n) ?(x(n) )T .
(10)
N
N
n=1
Theorem 5 is a straight-forward extension of classical results [33, 36].
By combining the convergence rate of Theorem 5 with Theorem 2 that connects the surrogate and
actual risks, we get Theorem 6 which explicitly gives the number of iterations required to achieve
? accuracy on the expected population risk (see App. A for the proof). Note that since ASGD is
applied in an online fashion, Theorem 6 also serves as the sample complexity bound, i.e., says how
many samples are needed to achieve ? target accuracy (compared to the best prediction rule if F has
zero approximation error).
Theorem 6 (Learning complexity). Under the assumptions of Theorem 5, for any ? > 0, the random
(w.r.t. the observed training set) output ?f(N ) ? FF,H of the ASGD algorithm after
N > N ? :=
4D 2 M 2
?2
H
?,L,F (?)
(11)
iterations has the expected excess risk bounded with ?, i.e., IE[RL (?f(N ) )] < R?L,F + ?.
4
Calibration function analysis for quadratic surrogate
A major challenge to applying Theorem 6 is the computation of the calibration function H?,L,F . In
App. C, we present a generalization to arbitrary multi-class losses of a surrogate loss class from Zhang
[49, Section 4.4.2] that is consistent for any task loss L. Here, we consider the simplest example of
this family, called the quadratic surrogate ?quad , which has the advantage that we can bound or even
compute exactly its calibration function. We define the quadratic surrogate as
?quad (f , y) :=
1
2k kf
+ L(:, y)k22 =
1
2k
k
X
(fc2 + 2fc L(c, y) + L(c, y)2 ).
(12)
c=1
One simple sufficient condition for the surrogate (12) to be consistent and also to have zero approximation error is that F fully contains span(L). To make the dependence on the score subspace explicit,
we parameterize it with a matrix F ? Rk?r with the number of columns r typically being much
smaller than the number of labels k. With this notation, we have F = span(F ) = {F ? | ? ? Rr },
and the dimensionality of F equals the rank of F , which is at most r.6
?
The Hilbert?Schmidt norm of a linear operator A is defined as kAkHS = trA? A where A? is the adjoint
operator. In the case of finite dimension, the Hilbert?Schmidt norm coincides with the Frobenius matrix norm.
6
Evaluating ?quad requires computing F T F and F T L(:, y) for which direct computation is intractable when
k is exponential, but which can be done in closed form for the structured losses we consider (the Hamming and
block 0-1 loss). More generally, these operations require suitable inference algorithms. See also App. F.
5
6
For the quadratic surrogate (12), the excess of the expected surrogate takes a simple form:
??quad (F ?, q) =
1
2k kF ?
+ Lqk22 .
(13)
Equation (13) holds under the assumption that the subspace F contains the column space of the
loss matrix span(L), which also means that the set F contains the optimal prediction for any q (see
Lemma 9 in App. B for the proof). Importantly, the function ??quad (F ?, q) is jointly convex in the
conditional probability q and parameters ?, which simplifies its analysis.
Lower bound on the calibration function. We now present our main technical result: a lower
bound on the calibration function for the surrogate loss ?quad (12). This lower bound characterizes
the easiness of learning with this surrogate given the scaling intuition mentioned in Section 3.3. The
proof of Theorem 7 is given in App. D.1.
Theorem 7 (Lower bound on H?quad ). For any task loss L, its quadratic surrogate ?quad , and a score
subspace F containing the column space of L, the calibration function can be lower bounded:
H?quad ,L,F (?) ?
?2
2k maxi6=j kPF ?ij k22
?
?2
4k ,
(14)
where PF is the orthogonal projection on the subspace F and ?ij = ei ? ej ? Rk with ec being
the c-th basis vector of the standard basis in Rk .
Lower bound for specific losses. We now discuss the meaning of the bound (14) for some specific
losses (the detailed derivations are given in App. D.3). For the 0-1, block 0-1 and Hamming losses
(L01 , L01,b and LHam,T , respectively) with the smallest possible score subspaces F, the bound (14)
?2 ?2
?2
gives 4k
, 4b and 8T
, respectively. All these bounds are tight (see App. E). However, if F = Rk
the bound (14) is not tight for the block 0-1 and mixed losses (see also App. E). In particular, the
bound (14) cannot detect level-? consistency for ? > 0 (see Def. 3) and does not change when the
loss changes, but the score subspace stays the same.
Upper bound on the calibration function. Theorem 8 below gives an upper bound on the calibration
function holding for unconstrained scores, i.e, F = Rk (see the proof in App. D.2). This result shows
that without some appropriate constraints on the scores, efficient learning is not guaranteed (in the
worst case) because of the 1/k scaling of the calibration function.
Theorem 8 (Upper bound on H?quad ). If a loss matrix L with Lmax > 0 defines a pseudometric7 on
labels and there are no constraints on the scores, i.e., F = Rk , then the calibration function for the
?2
quadratic surrogate ?quad can be upper bounded: H?quad ,L,Rk (?) ? 2k
, 0 ? ? ? Lmax .
From our lower bound in Theorem 7 (which guarantees consistency), the natural constraint on
the score is F = span(L), with the dimension of this space giving an indication of the intrinsic
?difficulty? of a loss. Computations for the lower bounds in some specific cases (see App. D.3 for
details) show that the 0-1 loss is ?hard? while the block 0-1 loss and the Hamming loss are ?easy?.
Note that in all these cases the lower bound (14) is tight, see the discussion below.
Exact calibration functions. Note that the bounds proven in Theorems 7 and 8 imply that, in the
case of no constraints on the scores F = Rk , for the 0-1, block 0-1 and Hamming losses, we have
?2
4k
? H?quad ,L,Rk (?) ?
?2
2k ,
(15)
where L is the matrix defining a loss. For completeness, in App. E, we compute the exact calibration
functions for the 0-1 and block 0-1 losses. Note that the calibration function for the 0-1 loss equals the
lower bound, illustrating the worst-case scenario. To get some intuition, an example of a conditional
distribution q that gives the (worst case) value to the calibration function (for several losses) is
qi = 21 + 2? , qj = 21 ? 2? and qc = 0 for c ?6= {i, j}. See the proof of Proposition 12 in App. E.1.
In what follows, we provide the calibration functions in the cases with constraints on the scores. For
the block 0-1 loss with b equal blocks and under constraints that the scores within blocks are equal,
the calibration function equals (see Proposition 14 of App. E.2)
H?quad ,L01,b ,F01,b (?) =
7
?2
4b ,
0 ? ? ? 1.
(16)
A pseudometric is a function d(a, b) satisfying the following axioms: d(x, y) ? 0, d(x, x) = 0 (but
possibly d(x, y) = 0 for some x 6= y), d(x, y) = d(y, x), d(x, z) ? d(x, y) + d(y, z).
7
For the Hamming loss defined over T binary variables and under constraints implying separable
scores, the calibration function equals (see Proposition 15 in App. E.3)
H?quad ,LHam,T ,FHam,T (?) =
?2
8T
, 0 ? ? ? 1.
(17)
The calibration functions (16) and (17) depend on the quantities representing the actual complexities
of the loss (the number of blocks b and the length of the sequence T ) and can be exponentially larger
than the upper bound for the unconstrained case.
In the case of mixed 0-1 and block 0-1 loss, if the scores f are constrained to be equal inside the
blocks, i.e., belong to the subspace F01,b = span(L01,b ) ( Rk , then the calibration function is equal
to 0 for ? ? ?2 , implying inconsistency (and also note that the approximation error can be as big as ?
for F01,b ). However, for ? > ?2 , the calibration function is of the order 1b (? ? ?2 )2 . See Figure 1b for
the illustration of this calibration function and Proposition 17 of App. E.4 for the exact formulation
and the proof. Note that while the calibration function for the constrained case is inconsistent, its
value can be exponentially larger than the one for the unconstrained case for ? big enough and when
the blocks are exponentially large (see Proposition 16 of App. E.4).
Computation of the SGD constants. Applying the learning complexity Theorem 6 requires to
compute the quantity DM where D bounds the norm of the optimal solution and M bounds the
expected square of the norm of the stochastic gradient. In App. F, we provide a way to bound this
quantity for our quadratic surrogate (12) under the simplifying assumption that each conditional qc (x)
(seen as function of x) belongs to the RKHS H (which implies Assumption 4). In particular, we get
?
DM = L2max ?(?(F ) rRQmax ), ?(z) = z 2 + z,
(18)
where ?(F ) is the condition number of the matrix F , R is an upper bound on the RKHS norm of
Pk
object feature maps k?(x)kH . We define Qmax as an upper bound on c=1 kqc kH (can be seen as
Pk
the generalization of the inequality c=1 qc ? 1 for probabilities). The constants R and Qmax depend
on the data, the constant Lmax depends on the loss, r and ?(F ) depend on the choice of matrix F .
We compute the constant DM for the specific losses that we considered in App. F.1. For the 0-1, block
0-1 and Hamming losses, we have DM = O(k), DM = O(b) and DM = O(log32 k), respectively.
These computations indicate that the quadratic surrogate allows efficient learning for structured block
0-1 and Hamming losses, but that the convergence could be slow in the worst case for the 0-1 loss.
5
Related works
Consistency for multi-class problems. Building on significant progress for the case of binary
classification, see, e.g. [5], there has been a lot of interest in the multi-class case. Zhang [49] and
Tewari & Bartlett [46] analyze the consistency of many existing surrogates for the 0-1 loss. Gao &
Zhou [20] focus on multi-label classification. Narasimhan et al. [32] provide a consistent algorithm
for arbitrary multi-class loss defined by a function of the confusion matrix. Recently, Ramaswamy &
Agarwal [39] introduce the notion of convex calibrated dimension, as the minimal dimensionality of
the score vector that is required for consistency. In particular, they showed that for the Hamming loss
on T binary variables, this dimension is at most T . In our analysis, we use scores of rank (T + 1),
see (35) in App. D.3, yielding a similar result.
The task of ranking has attracted a lot of attention and [18, 8, 9, 40] analyze different families of
surrogate and task losses proving their (in-)consistency. In this line of work, Ramaswamy et al.
[40] propose a quadratic surrogate for an arbitrary low rank loss which is related to our quadratic
surrogate (12). They also prove that several important ranking losses, i.e., precision@q, expected
rank utility, mean average precision and pairwise disagreement, are of low-rank. We conjecture that
our approach is compatible with these losses and leave precise connections as future work.
Structured SVM (SSVM) and friends. SSVM [44, 45, 47] is one of the most used convex surrogates
for tasks with structured outputs, thus, its consistency has been a question of great interest. It is
known that Crammer-Singer multi-class SVM [15], which SSVM is built on, is not consistent for
0-1 loss unless there is a majority class with probability at least 12 [49, 31]. However, it is consistent
for the ?abstain? and ordinal losses in the case of 3 classes [39]. Structured ramp loss and probit
surrogates are closely related to SSVM and are consistent [31, 16, 30, 23], but not convex.
8
Recently, Do?gan et al. [17] categorized different versions of multi-class SVM and analyzed them
from Fisher and universal consistency point of views. In particular, they highlight differences between
Fisher and universal consistency and give examples of surrogates that are Fisher consistent, but not
universally consistent and vice versa. They also highlight that the Crammer-Singer SVM is neither
Fisher, not universally consistent even with a careful choice of regularizer.
Quadratic surrogates
for structured prediction. Ciliberto et al. [10] and Brouard et al. [7] consider
Pn
minimizing i=1 kg(xi ) ? ?o (yi )k2H aiming to match the RKHS embedding of inputs g : X ? H
to the feature maps of outputs ?o : Y ? H. In their frameworks, the task loss is not considered at
the learning stage, but only at the prediction stage. Our quadratic surrogate (12) depends on the loss
directly. The empirical risk defined by both their and our objectives can be minimized analytically with
the help of the kernel trick and, moreover, the resulting predictors are identical. However, performing
such computation in the case of large dataset can be intractable and the generalization properties have
to be taken care of, e.g., by the means of regularization. In the large-scale scenario, it is more natural
to apply stochastic optimization (e.g., kernel ASGD) that directly minimizes the population risk and
has better dependency on the dataset size. When combined with stochastic optimization, the two
approaches lead to different behavior. In our framework, we need to estimate r = rank(L) scalar
functions, but the alternative needs to estimate k functions (if, e.g., ?o (y) = ey ? Rk ), which results
in significant differences for low-rank losses, such as block 0-1 and Hamming.
Calibration functions. Bartlett et al. [5] and Steinwart [43] provide calibration functions for most
existing surrogates for binary classification. All these functions differ in term of shape, but are
roughly similar in terms of constants. Pedregosa et al. [37] generalize these results to the case of
ordinal regression. However, their calibration functions have at best a 1/k factor if the surrogate is
normalized w.r.t. the number of classes. The task of ranking has been of significant interest. However,
most of the literature [e.g., 11, 14, 24, 1], only focuses on calibration functions (in the form of regret
bounds) for bipartite ranking, which is more akin to cost-sensitive binary classification.
?vila Pires et al. [3] generalize the theoretical framework developed by Steinwart [43] and present
results for the multi-class SVM of Lee et al. [26] (the score vectors are constrained
P to sum to zero)
that can
Pbe built for any task loss of interest. Their surrogate ? is of the form c?Y L(c, y)a(fc )
where c?Y fc = 0 and a(f ) is some convex function with all subgradients at zero being positive.
The recent work by ?vila Pires & Szepesv?ri [2] refines the results, but specifically for the case of
0-1 loss. In this line of work, the surrogate is typically not normalized by k, and if normalized the
calibration functions have the constant 1/k appearing.
Finally, Ciliberto et al. [10] provide the calibration function for their quadratic surrogate. Assuming
? y) = hV ?o (y),
? ?o (y)iHY , y,
? y ? Y (this assumption can
that the loss can be represented as L(y,
always be satisfied in the case of a finite number of labels, by taking V as the loss matrix L and
?o (y) := ey ? Rk where ey is the y-th vector of the standard basis in Rk ). In their Theorem 2, they
provide an excess risk bound leading to a lower bound on the corresponding calibration function
2
H?,L,Rk (?) ? c?2 where a constant c? = kV k2 maxy?Y k?o (y)k simply equals the spectral norm
?
of the loss matrix for the finite-dimensional construction provided above. However, the spectral
norm of the loss matrix is exponentially large even for highly structured losses such as the block 0-1
and Hamming losses, i.e., kL01,b k2 = k ? kb , kLHam,T k2 = k2 . This conclusion puts the objective
of Ciliberto et al. [10] in line with ours when no constraints are put on the scores.
6
Conclusion
In this paper, we studied the consistency of convex surrogate losses specifically in the context
of structured prediction. We analyzed calibration functions and proposed an optimization-based
normalization aiming to connect consistency with the existence of efficient learning algorithms.
Finally, we instantiated all components of our framework for several losses by computing the
calibration functions and the constants coming from the normalization. By carefully monitoring
exponential constants, we highlighted the difference between tractable and intractable task losses.
These were first steps in advancing our theoretical understanding of consistent structured prediction.
Further steps include analyzing more losses such as the low-rank ranking losses studied by Ramaswamy et al. [40] and, instead of considering constraints on the scores, one could instead put
constraints on the set of distributions to investigate the effect on the calibration function.
9
Acknowledgements
We would like to thank Pascal Germain for useful discussions. This work was partly supported
by the ERC grant Activia (no. 307574), the NSERC Discovery Grant RGPIN-2017-06936 and the
MSR-INRIA Joint Center.
References
[1] Agarwal, Shivani. Surrogate regret bounds for bipartite ranking via strongly proper losses.
Journal of Machine Learning Research (JMLR), 15(1):1653?1674, 2014.
[2] ?vila Pires, Bernardo and Szepesv?ri, Csaba. Multiclass classification calibration functions.
arXiv, 1609.06385v1, 2016.
[3] ?vila Pires, Bernardo, Ghavamzadeh, Mohammad, and Szepesv?ri, Csaba. Cost-sensitive
multiclass classification risk bounds. In ICML, 2013.
[4] Bakir, G?khan, Hofmann, Thomas, Sch?lkopf, Bernhard, Smola, Alexander J., Taskar, Ben,
and Vishwanathan, S.V.N. Predicting Structured Data. MIT press, 2007.
[5] Bartlett, Peter L., Jordan, Michael I., and McAuliffe, Jon D. Convexity, classification, and risk
bounds. Journal of the American Statistical Association, 101(473):138?156, 2006.
[6] Bousquet, Olivier and Bottou, L?on. The tradeoffs of large scale learning. In NIPS, 2008.
[7] Brouard, C?line, Szafranski, Marie, and d?Alch?-Buc, Florence. Input output kernel regression:
Supervised and semi-supervised structured output prediction with operator-valued kernels.
Journal of Machine Learning Research (JMLR), 17(176):1?48, 2016.
[8] Buffoni, David, Gallinari, Patrick, Usunier, Nicolas, and Calauz?nes, Cl?ment. Learning scoring
functions with order-preserving losses and standardized supervision. In ICML, 2011.
[9] Calauz?nes, Cl?ment, Usunier, Nicolas, and Gallinari, Patrick. On the (non-)existence of convex,
calibrated surrogate losses for ranking. In NIPS, 2012.
[10] Ciliberto, Carlo, Rosasco, Lorenzo, and Rudi, Alessandro. A consistent regularization approach
for structured prediction. In NIPS, 2016.
[11] Cl?men?on, St?phan, Lugosi, G?bor, and Vayatis, Nicolas. Ranking and empirical minimization
of U-statistics. The Annals of Statistics, pp. 844?874, 2008.
[12] Collins, Michael. Discriminative training methods for hidden Markov models: Theory and
experiments with perceptron algorithms. In EMNLP, 2002.
[13] Cortes, Corinna, Kuznetsov, Vitaly, Mohri, Mehryar, and Yang, Scott. Structured prediction
theory based on factor graph complexity. In NIPS, 2016.
[14] Cossock, David and Zhang, Tong. Statistical analysis of bayes optimal subset ranking. IEEE
Transactions on Information Theory, 54(11):5140?5154, 2008.
[15] Crammer, Koby and Singer, Yoram. On the algorithmic implementation of multiclass kernelbased vector machines. Journal of Machine Learning Research (JMLR), 2:265?292, 2001.
[16] Do, Chuong B., Le, Quoc, Teo, Choon Hui, Chapelle, Olivier, and Smola, Alex. Tighter bounds
for structured estimation. In NIPS, 2009.
[17] Do?gan, ?r?n, Glasmachers, Tobias, and Igel, Christian. A unified view on multi-class support
vector classification. Journal of Machine Learning Research (JMLR), 17(45):1?32, 2016.
[18] Duchi, John C., Mackey, Lester W., and Jordan, Michael I. On the consistency of ranking
algorithms. In ICML, 2010.
[19] Durbin, Richard, Eddy, Sean, Krogh, Anders, and Mitchison, Graeme. Biological sequence
analysis: probabilistic models of proteins and nucleic acids. Cambridge university press, 1998.
10
[20] Gao, Wei and Zhou, Zhi-Hua. On the consistency of multi-label learning. In COLT, 2011.
[21] Gimpel, Kevin and Smith, Noah A. Softmax-margin CRFs: Training loglinear models with cost
functions. In NAACL, 2010.
[22] Hazan, Tamir and Urtasun, Raquel. A primal-dual message-passing algorithm for approximated
large scale structured prediction. In NIPS, 2010.
[23] Keshet, Joseph. Optimizing the measure of performance in structured prediction. In Advanced
Structured Prediction. MIT Press, 2014.
[24] Kotlowski, Wojciech, Dembczynski, Krzysztof, and Huellermeier, Eyke. Bipartite ranking
through minimization of univariate loss. In ICML, 2011.
[25] Lafferty, John, McCallum, Andrew, and Pereira, Fernando. Conditional random fields: Probabilistic models for segmenting and labeling sequence data. In ICML, 2001.
[26] Lee, Yoonkyung, Lin, Yi, and Wahba, Grace. Multicategory support vector machines: Theory
and application to the classification of microarray data and satellite radiance data. Journal of
the American Statistical Association, 99(465):67?81, 2004.
[27] Lin, Yi. A note on margin-based loss functions in classification. Statistics & Probability Letters,
68(1):73?82, 2004.
[28] London, Ben, Huang, Bert, and Getoor, Lise. Stability and generalization in structured prediction.
Journal of Machine Learning Research (JMLR), 17(222):1?52, 2016.
[29] Long, Phil and Servedio, Rocco. Consistency versus realizable H-consistency for multiclass
classification. In ICML, 2013.
[30] McAllester, D. A. and Keshet, J. Generalization bounds and consistency for latent structural
probit and ramp loss. In NIPS, 2011.
[31] McAllester, David. Generalization bounds and consistency for structured labeling. In Predicting
Structured Data. MIT Press, 2007.
[32] Narasimhan, Harikrishna, Ramaswamy, Harish G., Saha, Aadirupa, and Agarwal, Shivani.
Consistent multiclass algorithms for complex performance measures. In ICML, 2015.
[33] Nemirovski, A., Juditsky, A., Lan, G., and Shapiro, A. Robust stochastic approximation
approach to stochastic programming. SIAM Journal on Optimization, 19(4):1574?1609, 2009.
[34] Nowozin, Sebastian and Lampert, Christoph H. Structured learning and prediction in computer
vision. Foundations and Trends in Computer Graphics and Vision, 6(3?4):185?365, 2011.
[35] Nowozin, Sebastian, Gehler, Peter V., Jancsary, Jeremy, and Lampert, Christoph H. Advanced
Structured Prediction. MIT Press, 2014.
[36] Orabona, Francesco. Simultaneous model selection and optimization through parameter-free
stochastic learning. In NIPS, 2014.
[37] Pedregosa, Fabian, Bach, Francis, and Gramfort, Alexandre. On the consistency of ordinal
regression methods. Journal of Machine Learning Research (JMLR), 18(55):1?35, 2017.
[38] Pletscher, Patrick, Ong, Cheng Soon, and Buhmann, Joachim M. Entropy and margin maximization for structured output learning. In ECML PKDD, 2010.
[39] Ramaswamy, Harish G. and Agarwal, Shivani. Convex calibration dimension for multiclass
loss matrices. Journal of Machine Learning Research (JMLR), 17(14):1?45, 2016.
[40] Ramaswamy, Harish G., Agarwal, Shivani, and Tewari, Ambuj. Convex calibrated surrogates
for low-rank loss matrices with applications to subset ranking losses. In NIPS, 2013.
[41] Shi, Qinfeng, Reid, Mark, Caetano, Tiberio, van den Hengel, Anton, and Wang, Zhenhua. A
hybrid loss for multiclass and structured prediction. IEEE transactions on pattern analysis and
machine intelligence (TPAMI), 37(1):2?12, 2015.
11
[42] Smith, Noah A. Linguistic structure prediction. Synthesis lectures on human language technologies, 4(2):1?274, 2011.
[43] Steinwart, Ingo. How to compare different loss functions and their risks. Constructive Approximation, 26(2):225?287, 2007.
[44] Taskar, Ben, Guestrin, Carlos, and Koller, Daphne. Max-margin markov networks. In NIPS,
2003.
[45] Taskar, Ben, Chatalbashev, Vassil, Koller, Daphne, and Guestrin, Carlos. Learning structured
prediction models: a large margin approach. In ICML, 2005.
[46] Tewari, Ambuj and Bartlett, Peter L. On the consistency of multiclass classification methods.
Journal of Machine Learning Research (JMLR), 8:1007?1025, 2007.
[47] Tsochantaridis, I., Joachims, T., Hofmann, T., and Altun, Y. Large margin methods for
structured and interdependent output variables. Journal of Machine Learning Research (JMLR),
6:1453?1484, 2005.
[48] Williamson, Robert C., Vernet, Elodie, and Reid, Mark D. Composite multiclass losses. Journal
of Machine Learning Research (JMLR), 17(223):1?52, 2016.
[49] Zhang, Tong. Statistical analysis of some multi-category large margin classification methods.
Journal of Machine Learning Research (JMLR), 5:1225?1251, 2004.
[50] Zhang, Tong. Statistical behavior and consistency of classification methods based on convex
risk minimization. Annals of Statistics, 32(1):56?134, 2004.
12
| 6634 |@word msr:1 illustrating:1 version:1 norm:10 c0:2 simplifying:2 pick:1 sgd:1 harder:1 contains:4 score:43 selecting:1 rkhs:8 ours:1 existing:2 dx:3 attracted:1 john:2 refines:1 kpf:1 informative:2 shape:1 hofmann:2 christian:1 update:2 juditsky:1 v:1 implying:4 mackey:1 intelligence:1 xk:2 mccallum:1 smith:2 coarse:1 completeness:1 zhang:7 daphne:2 iverson:1 direct:1 m2d:1 prove:2 consists:3 inside:1 introduce:2 x0:2 pairwise:3 expected:8 roughly:1 pkdd:1 behavior:2 multi:15 decreasing:3 resolve:1 actual:11 quad:16 pf:1 cardinality:1 considering:1 becomes:1 provided:2 zhi:1 bounded:8 notation:3 moreover:1 what:1 kg:1 kind:1 minimizes:1 generalizable:1 narasimhan:2 developed:1 unified:1 finding:1 csaba:2 guarantee:6 bernardo:2 tackle:2 tie:1 exactly:2 universit:1 scaled:1 k2:4 gallinari:2 lester:1 grant:2 appear:1 mcauliffe:1 segmenting:1 t1:1 positive:2 understood:1 reid:2 mistake:1 consequence:1 aiming:4 despite:1 analyzing:1 solely:1 lugosi:1 inria:3 might:8 therein:1 studied:4 specifying:2 christoph:2 nemirovski:1 averaged:2 igel:1 practical:2 practice:2 block:23 definite:1 regret:2 empirical:3 universal:3 axiom:1 composite:1 projection:2 word:1 regular:1 protein:1 altun:1 get:4 cannot:1 selection:1 hse:1 operator:4 put:3 context:3 risk:35 applying:3 tsochantaridis:1 optimize:5 szafranski:1 map:5 measurable:1 shi:1 yt:3 equivalent:3 center:1 economics:1 attention:1 independently:1 convex:29 rectangular:1 phil:1 qc:5 simplicity:2 unstructured:1 insight:2 rule:1 importantly:1 population:3 proving:1 notion:4 embedding:1 stability:1 analogous:1 annals:2 target:2 play:2 construction:2 exact:4 olivier:2 programming:1 trick:2 element:5 trend:1 satisfying:1 approximated:1 gehler:1 observed:1 role:3 taskar:3 wang:1 capture:1 worst:6 parameterize:1 hv:1 caetano:1 mentioned:1 intuition:4 transforming:1 pd:2 complexity:7 convexity:1 alessandro:1 tobias:1 ong:1 ghavamzadeh:1 depend:5 tight:8 rewrite:1 bipartite:3 efficiency:3 basis:3 alch:1 joint:2 represented:1 regularizer:1 alphabet:1 derivation:1 instantiated:1 london:2 labeling:2 kevin:1 larger:2 valued:2 say:1 ramp:3 otherwise:1 statistic:5 lham:6 unseen:1 rgpin:1 jointly:2 highlighted:1 final:1 ip:4 online:2 sequence:11 indication:2 rr:4 advantage:1 tpami:1 propose:3 ment:2 maximal:3 coming:2 product:3 combining:1 achieve:2 adjoint:1 graeme:1 frobenius:1 kh:2 normalize:1 ihy:1 getting:1 kv:1 convergence:10 empty:1 satellite:1 generating:2 maxi6:1 leave:1 ben:4 object:2 help:1 depending:1 illustrate:1 friend:1 andrew:1 measured:2 ij:2 school:1 progress:3 krogh:1 strong:1 implemented:1 judge:1 implies:4 indicate:1 differ:1 radius:1 closely:2 correct:2 stochastic:16 kb:1 human:1 vila:4 mcallester:2 glasmachers:1 require:1 generalization:9 tiberio:1 proposition:6 tighter:1 biological:1 extension:1 hold:2 considered:2 exp:1 great:1 k2h:1 algorithmic:2 predict:1 claim:1 major:1 radiance:1 smallest:2 estimation:1 label:16 sensitive:3 cole:1 teo:1 vice:2 minimization:6 mit:4 always:6 aim:1 normale:1 zhou:2 ej:1 pn:1 varying:1 iex:2 linguistic:1 lise:1 focus:3 joachim:2 properly:1 notational:1 rank:10 indicates:2 likelihood:1 contrast:1 detect:1 realizable:1 inference:1 dependent:2 anders:1 cnrs:1 unary:2 chatalbashev:1 typically:4 hidden:1 koller:2 france:2 classification:19 ill:1 among:1 denoted:1 pascal:1 dual:1 colt:1 constrained:3 special:1 softmax:1 gramfort:1 marginal:1 field:2 construct:2 equal:14 never:1 beach:1 identical:1 represents:2 kw:2 look:1 icml:8 koby:1 jon:1 future:1 minimized:2 others:1 simplex:1 simplify:1 richard:1 few:1 saha:1 simultaneously:1 national:1 choon:1 argmax:1 connects:1 ciliberto:6 montr:1 interest:4 message:1 highly:2 investigate:1 multiply:2 analyzed:2 bracket:1 yielding:1 primal:1 devoted:1 edge:1 orthogonal:1 unless:1 desired:1 theoretical:6 minimal:1 column:6 logp:1 maximization:1 cost:5 entry:1 subset:2 predictor:4 too:1 graphic:1 characterize:1 dependency:3 connect:1 elodie:1 calibrated:4 combined:1 st:2 siam:1 ie:5 stay:1 lee:2 probabilistic:2 pbe:1 michael:3 synthesis:1 concrete:2 central:1 satisfied:1 containing:1 rosasco:1 russia:1 possibly:1 emnlp:1 huang:1 american:2 leading:1 wojciech:1 account:1 potential:2 jeremy:1 de:1 coefficient:1 tra:1 notable:1 explicitly:2 ranking:15 depends:3 asgd:8 view:2 ramaswamy:7 closed:1 lot:2 analyze:4 francis:2 sup:1 red:4 hf:1 characterizes:1 complicated:1 bayes:1 hazan:1 dembczynski:1 simon:1 carlos:2 florence:1 contribution:1 minimize:3 square:2 accuracy:10 acid:1 efficiently:1 yield:2 generalize:2 anton:2 lkopf:1 bor:1 carlo:1 monitoring:1 straight:1 app:19 explain:1 simultaneous:1 reach:2 sebastian:2 definition:8 servedio:1 pp:1 dm:7 proof:7 di:1 hamming:15 sampled:1 calauz:2 dataset:2 popular:1 intrinsically:1 logical:1 dimensionality:2 bakir:1 organized:1 formalize:4 hilbert:4 eddy:1 carefully:2 sean:1 back:1 harikrishna:1 alexandre:1 higher:1 supervised:2 follow:1 aadirupa:1 specify:1 wei:1 formulation:1 done:2 strongly:1 stage:2 smola:2 steinwart:3 ei:1 defines:2 brings:1 quality:1 infimum:1 grows:1 usa:1 effect:2 k22:2 normalized:4 true:2 building:1 naacl:1 analytically:1 regularization:2 eyke:1 coincides:1 suboptimality:1 complete:1 tt:1 confusion:1 mohammad:1 duchi:1 bring:1 image:2 meaning:1 novel:1 recently:3 abstain:1 common:1 functional:1 rl:10 exponentially:13 cossock:1 belong:1 association:2 relating:1 significant:5 refer:1 versa:2 cambridge:1 unconstrained:3 consistency:40 erc:1 language:2 brouard:2 chapelle:1 calibration:65 supervision:1 etc:1 add:1 patrick:3 recent:2 showed:1 perspective:1 optimizing:3 inf:8 belongs:3 rieure:1 scenario:2 certain:1 inequality:3 binary:9 inconsistency:2 yi:3 scoring:1 seen:2 minimum:1 additional:1 relaxed:1 care:1 preserving:1 ey:3 guestrin:2 fernando:1 ii:1 relates:2 multiple:1 semi:1 technical:1 faster:3 match:1 bach:2 long:2 lin:3 equally:1 feasibility:1 qi:1 prediction:42 regression:4 vision:3 expectation:2 arxiv:1 iteration:6 represent:1 normalization:4 kernel:11 agarwal:5 buffoni:1 vayatis:1 szepesv:3 addressed:1 grow:1 source:1 microarray:1 sch:1 envelope:1 kotlowski:1 vitaly:1 inconsistent:2 lafferty:1 jordan:2 structural:1 yang:1 iii:1 enough:2 easy:1 iterate:1 wahba:1 inner:2 simplifies:1 multiclass:9 tradeoff:1 psl:1 qj:1 whether:2 expression:1 bartlett:4 utility:1 akin:1 peter:3 proceed:1 passing:1 adequate:2 ssvm:5 generally:1 tewari:3 detailed:2 useful:1 extensively:1 shivani:4 category:1 simplest:2 shapiro:1 blue:4 group:2 easiness:1 lan:1 monitor:1 changing:1 neither:1 marie:1 lacoste:1 advancing:1 v1:1 krzysztof:1 graph:3 subgradient:1 sum:1 letter:1 qmax:2 raquel:1 family:7 k2hs:1 appendix:1 scaling:2 bound:44 def:1 guaranteed:1 rudi:1 cheng:1 quadratic:16 durbin:1 noah:2 constraint:20 vishwanathan:1 alex:1 ri:3 bousquet:1 aspect:2 argument:2 span:9 pseudometric:1 f01:3 separable:1 performing:1 relatively:1 conjecture:1 subgradients:1 structured:48 combination:1 ball:1 beneficial:1 smaller:3 joseph:1 making:1 quoc:1 maxy:2 intuitively:1 restricted:2 invariant:1 den:1 taken:2 equation:1 discus:1 needed:2 ordinal:4 singer:3 tractable:2 serf:1 studying:1 usunier:2 operation:1 apply:2 observe:1 vernet:1 appropriate:2 disagreement:1 spectral:2 appearing:2 schmidt:3 alternative:1 corinna:1 slower:1 existence:2 thomas:1 moscow:1 denotes:1 include:3 running:1 gan:2 standardized:1 harish:3 hinge:2 yoram:1 giving:1 multicategory:1 especially:1 classical:3 objective:3 question:2 quantity:5 strategy:1 costly:1 parametric:1 dependence:1 grace:1 surrogate:74 loglinear:1 rocco:1 gradient:6 subspace:9 distance:1 thank:1 fc2:1 majority:1 argue:1 fy:7 urtasun:1 assuming:4 length:3 index:1 relationship:1 pointwise:1 illustration:1 minimizing:3 setup:5 difficult:1 robert:1 statement:2 holding:1 negative:1 chuong:1 implementation:1 proper:1 l01:12 upper:8 observation:1 nucleic:1 markov:2 francesco:1 ingo:1 finite:11 fabian:1 descent:3 ecml:1 defining:2 looking:2 precise:2 reproducing:2 bert:1 arbitrary:3 canada:1 pred:3 david:3 germain:1 paris:2 required:3 specified:1 optimized:3 connection:5 khan:1 learned:1 pires:4 nip:11 usually:3 below:9 scott:1 pattern:1 challenge:2 ambuj:2 built:2 including:1 max:1 suitable:3 getoor:1 natural:6 difficulty:2 hybrid:2 predicting:3 rely:1 buhmann:1 advanced:2 pletscher:1 representing:2 technology:1 lorenzo:1 imply:2 ne:2 julien:1 picture:1 gimpel:1 prior:1 understanding:2 review:1 literature:3 kf:3 acknowledgement:1 discovery:1 interdependent:1 loss:136 lecture:1 subfield:1 probit:3 mixed:5 interesting:1 fully:1 highlight:2 men:1 proven:1 versus:1 foundation:1 sufficient:2 consistent:24 verification:1 huellermeier:1 nowozin:2 vassil:1 row:3 elsewhere:1 lmax:4 compatible:1 supported:1 mohri:1 free:2 soon:1 formal:2 allow:2 perceptron:1 taking:1 differentiating:1 van:1 curve:3 dimension:10 xn:1 evaluating:1 tamir:1 hengel:1 forward:1 made:1 collection:1 coincide:1 projected:1 universally:2 osokin:1 ec:1 transaction:2 excess:12 bernhard:1 buc:1 wrote:1 monotonicity:1 global:1 conclude:1 tuples:1 xi:1 discriminative:1 alternatively:1 mitchison:1 continuous:4 latent:1 learn:1 robust:1 ca:1 nicolas:3 mehryar:1 bottou:1 cl:3 complex:1 williamson:1 domain:1 pk:2 main:1 s2:1 big:2 lampert:2 allowed:2 categorized:1 body:1 crfs:1 referred:1 mila:1 en:2 ff:11 borel:1 fashion:1 slow:2 tong:3 precision:3 pereira:1 explicit:3 exponential:8 lq:1 jmlr:11 rk:27 theorem:25 specific:7 jensen:1 symbol:1 cortes:2 admits:1 svm:5 evidence:1 exists:3 consist:1 ih:2 intractable:3 intrinsic:1 hui:1 keshet:2 margin:7 easier:2 phan:1 suited:1 entropy:1 interrelated:1 simply:2 fc:3 univariate:1 gao:2 nserc:1 scalar:3 kuznetsov:1 hua:1 khs:2 satisfies:2 conditional:13 goal:2 formulated:1 careful:1 orabona:1 fisher:6 feasible:3 hard:3 change:2 jancsary:1 typical:2 infinite:2 specifically:4 qinfeng:1 lemma:1 called:3 partly:1 exception:1 pedregosa:3 kernelbased:1 support:2 mark:2 crammer:3 collins:1 alexander:1 bioinformatics:1 constructive:1 |
6,229 | 6,635 | Best of Both Worlds: Transferring Knowledge from
Discriminative Learning to a Generative Visual
Dialog Model
Jiasen Lu1?, Anitha Kannan2?, Jianwei Yang1 , Devi Parikh3,1 , Dhruv Batra3,1
1
Georgia Institute of Technology, 2 Curai, 3 Facebook AI Research
{jiasenlu, jw2yang, parikh, dbatra}@gatech.edu
Abstract
We present a novel training framework for neural sequence models, particularly
for grounded dialog generation. The standard training paradigm for these models
is maximum likelihood estimation (MLE), or minimizing the cross-entropy of the
human responses. Across a variety of domains, a recurring problem with MLE
trained generative neural dialog models (G) is that they tend to produce ?safe?
and generic responses (?I don?t know?, ?I can?t tell?). In contrast, discriminative
dialog models (D) that are trained to rank a list of candidate human responses
outperform their generative counterparts; in terms of automatic metrics, diversity,
and informativeness of the responses. However, D is not useful in practice since it
can not be deployed to have real conversations with users.
Our work aims to achieve the best of both worlds ? the practical usefulness of
G and the strong performance of D ? via knowledge transfer from D to G. Our
primary contribution is an end-to-end trainable generative visual dialog model,
where G receives gradients from D as a perceptual (not adversarial) loss of the
sequence sampled from G. We leverage the recently proposed Gumbel-Softmax
(GS) approximation to the discrete distribution ? specifically, a RNN augmented
with a sequence of GS samplers, coupled with the straight-through gradient estimator to enable end-to-end differentiability. We also introduce a stronger encoder for
visual dialog, and employ a self-attention mechanism for answer encoding along
with a metric learning loss to aid D in better capturing semantic similarities in
answer responses. Overall, our proposed model outperforms state-of-the-art on the
VisDial dataset by a significant margin (2.67% on recall@10). The source code
can be downloaded from https://github.com/jiasenlu/visDial.pytorch
1
Introduction
One fundamental goal of artificial intelligence (AI) is the development of perceptually-grounded
dialog agents ? specifically, agents that can perceive or understand their environment (through vision,
audio, or other sensors), and communicate their understanding with humans or other agents in natural
language. Over the last few years, neural sequence models (e.g. [47, 44, 46]) have emerged as the
dominant paradigm across a variety of setting and datasets ? from text-only dialog [44, 40, 23, 3] to
more recently, visual dialog [7, 9, 8, 33, 45], where an agent must answer a sequence of questions
grounded in an image, requiring it to reason about both visual content and the dialog history.
The standard training paradigm for neural dialog models is maximum likelihood estimation (MLE)
or equivalently, minimizing the cross-entropy (under the model) of a ?ground-truth? human response.
Across a variety of domains, a recurring problem with MLE trained neural dialog models is that they
tend to produce ?safe? generic responses, such as ?Not sure? or ?I don?t know? in text-only dialog [23],
and ?I can?t see? or ?I can?t tell? in visual dialog [7, 8]. One reason for this emergent behavior is that
?
Work was done while at Facebook AI Research.
the space of possible next utterances in a dialog is highly multi-modal (there are many possible paths
a dialog may take in the future). In the face of such highly multi-modal output distributions, models
?game? MLE by latching on to the head of the distribution or the frequent responses, which by nature
tend to be generic and widely applicable. Such safe generic responses break the flow of a dialog and
tend to disengage the human conversing with the agent, ultimately rendering the agent useless. It is
clear that novel training paradigms are needed; that is the focus of this paper.
One promising alternative to MLE training proposed by recent work [36, 27] is sequence-level
training of neural sequence models, specifically, using reinforcement learning to optimize taskspecific sequence metrics such as BLEU [34], ROUGE [24], CIDEr [48]. Unfortunately, in the case
of dialog, all existing automatic metrics correlate poorly with human judgment [26], which renders
this alternative infeasible for dialog models.
In this paper, inspired by the success of adversarial training [16], we propose to train a generative
visual dialog model (G) to produce sequences that score highly under a discriminative visual dialog
model (D). A discriminative dialog model receives as input a candidate list of possible responses and
learns to sort this list from the training dataset. The generative dialog model (G) aims to produce a
sequence that D will rank the highest in the list, as shown in Fig. 1.
Note that while our proposed approach is inspired by adversarial training, there are a number of
subtle but crucial differences over generative adversarial networks (GANs). Unlike traditional GANs,
one novelty in our setup is that our discriminator receives a list of candidate responses and explicitly
learns to reason about similarities and differences across candidates. In this process, D learns a
task-dependent perceptual similarity [12, 19, 15] and learns to recognize multiple correct responses
in the feature space. For example, as shown in Fig. 1 right, given the image, dialog history, and
question ?Do you see any bird??, besides the ground-truth answer ?No, I do not?, D can also assign
high scores to other options that are valid responses to the question, including the one generated by
G: ?Not that I can see?. The interaction between responses is captured via the similarity between the
learned embeddings. This similarity gives an additional signal that G can leverage in addition to the
MLE loss. In that sense, our proposed approach may be viewed as an instance of ?knowledge transfer?
[17, 5] from D to G. We employ a metric-learning loss function and a self-attention answer encoding
mechanism for D that makes it particularly conducive to this knowledge transfer by encouraging
perceptually meaningful similarities to emerge. This is especially fruitful since prior work has
demonstrated that discriminative dialog models significantly outperform their generative counterparts,
but are not as useful since they necessarily need a list of candidate responses to rank, which is only
available in a dialog dataset, not in real conversations with a user. In that context, our work aims to
achieve the best of both worlds ? the practical usefulness of G and the strong performance of D ? via
this knowledge transfer.
Our primary technical contribution is an end-to-end trainable generative visual dialog model, where
the generator receives gradients from the discriminator loss of the sequence sampled from G. Note
that this is challenging because the output of G is a sequence of discrete symbols, which na?vely is not
amenable to gradient-based training. We propose to leverage the recently proposed Gumbel-Softmax
(GS) approximation to the discrete distribution [18, 30] ? specifically, a Recurrent Neural Network
(RNN) augmented with a sequence of GS samplers, which when coupled with the straight-through
gradient estimator [2, 18] enables end-to-end differentiability.
Our results show that our ?knowledge transfer? approach is indeed successful. Specifically, our
discriminator-trained G outperforms the MLE-trained G by 1.7% on recall@5 on the VisDial dataset,
essentially improving over state-of-the-art [7] by 2.43% recall@5 and 2.67% recall@10. Moreover,
our generative model produces more diverse and informative responses (see Table 3).
As a side contribution specific to this application, we introduce a novel encoder for neural visual
dialog models, which maintains two separate memory banks ? one for visual memory (where do
we look in the image?) and another for textual memory (what facts do we know from the dialog
history?), and outperforms the encoders used in prior work.
2
Related Work
GANs for sequence generation. Generative Adversarial Networks (GANs) [16] have shown to be
effective models for a wide range of applications involving continuous variables (e.g. images) c.f
[10, 35, 22, 55]. More recently, they have also been used for discrete output spaces such as language
generation ? e.g. image captioning [6, 41], dialog generation [23], or text generation [53] ? by either
viewing the generative model as a stochastic parametrized policy that is updated using REINFORCE
2
Generator
Question Qt
A gray tiger cat sitting underneath
a metal bench.
Is it in color? Yes it is.
Is it day time? Yes.
Is the tiger big? No, it?s a regular
cat.
?"
HCIAE
Encoder
I do not see any birds
No , I do not
Nope
?
yes
Mangoes
HCIAE
Encoder
White
Score
No
Answer
Encoder
?"
No ,
Nope
Not that I can see
yes
?
?
Mangoes
I see small shops
HQI
HQI
No , I do not
White
I see small shops
t rounds of history
HQI
No bird
No
Answer
Decoder
Option answers (D)
No bird
I do not see any
birds
Image I
Do you see any birds?
Discriminator
Not that I can see
Deep Metric Loss
Gumbel
Sampler
Figure 1: Model architecture of the proposed model. Given the image, history, and question, the discriminator
receives as additional input a candidate list of possible responses and learns to sort this list. The generator aims
to produce a sequence that discriminator will rank the highest in the list. The right most block is D?s score for
different candidate answers. Note that the multiple plausible responses all score high. Image from the COCO
dataset [25].
with the discriminator providing the reward [53, 6, 41, 23], or (closer to our approach) through
continuous relaxation of discrete variables through Gumbel-Softmax to enable backpropagating the
response from the discriminator [21, 41].
There are a few subtle but significant differences w.r.t. to our application, motivation, and approach.
In these prior works, both the discriminator and the generator are trained in tandem, and from scratch.
The goal of the discriminator in those settings has primarily been to discriminate ?fake? samples
(i.e. generator?s outputs) from ?real? samples (i.e. from training data). In contrast, we would like
to transfer knowledge from the discriminator to the generator. We start with pre-trained D and G
models suited for the task, and then transfer knowledge from D to G to further improve G, while
keeping D fixed. As we show in our experiments, this procedure results in G producing diverse
samples that are close in the embedding space to the ground truth, due to perceptual similarity learned
in D. One can also draw connections between our work and Energy Based GAN (EBGAN) [54] ?
without the adversarial training aspect. The ?energy? in our case is a deep metric-learning based
scoring mechanism, instantiated in the visual dialog application.
Modeling image and text attention. Models for tasks at the intersection of vision and language
? e.g., image captioning [11, 13, 20, 49], visual question answering [1, 14, 31, 37], visual dialog
[7, 9, 8, 45, 33] ? typically involve attention mechanisms. For image captioning, this may be attending
to relevant regions in the image [49, 51, 28]. For VQA, this may be attending to relevant image
regions alone [4, 50, 52] or co-attending to image regions and question words/phrases [29].
In the context of visual dialog, [7] uses attention to identify utterances in the dialog history that may
be useful for answering the current question. However, when modeling the image, the entire image
embedding is used to obtain the answer. In contrast, our proposed encoder HCIAE (Section 4.1)
localizes the region in the image that can help reliably answer the question. In particular, in addition
to the history and the question guiding the image attention, our visual dialog encoder also reasons
about the history when identifying relevant regions of the image. This allows the model to implicitly
resolve co-references in the text and ground them back in the image.
3
Preliminaries: Visual Dialog
We begin by formally describing the visual dialog task setup as introduced by Das et al. [7]. The
machine learning task is as follows. A visual dialog model is given as input an image I, caption c
describing the image, a dialog history till round t ? 1, H = ( c , (q1 , a1 ), . . . , (qt?1 , at?1 )), and
|{z} | {z }
|
{z
}
H0
H1
Ht?1
the followup question qt at round t. The visual dialog agent needs to return a valid response to the
question.
Given the problem setup, there are two broad classes of methods ? generative and discriminative
models. Generative models for visual dialog are trained by maximizing the log-likelihood of the
ground truth answer sequence agt
t ? At given the encoded representation of the input (I, H, qt ).
3
On the other hand, discriminative models receive both an encoding of the input (I, H, qt ) and as
(1)
(100)
additional input a list of 100 candidate answers At = {at , . . . , at
}. These models effectively
learn to sort the list. Thus, by design, they cannot be used at test time without a list of candidates
available.
4
Approach: Backprop Through Discriminative Losses for Generative
Training
In this section, we describe our approach to transfer knowledge from a discriminative visual dialog
model (D) to generative visual dialog model (G). Fig. 1 (a) shows the overview of our approach.
Given the input image I, dialog history H, and question qt , the encoder converts the inputs into a
joint representation et . The generator G takes et as input, and produces a distribution over answer
sequences via a recurrent neural network (specifically an LSTM). At each word in the answer
sequence, we use a Gumbel-Softmax sampler S to sample the answer token from that distribution.
The discriminator D in it?s standard form takes et , ground-truth answer agt
t and N ? 1 ?negative?
gt
N ?1
answers {a?
}
as
input,
and
learns
an
embedding
space
such
that
similarity(e
t , f (at )) >
t,i i=1
?
similarity(et , f (at,? )), where f (?) is the embedding function. When we enable the communication
between D and G, we feed the sampled answer a
?t into discriminator, and optimize the generator G
to produce samples that get higher scores in D?s metric space.
We now describe each component of our approach in detail.
4.1
History-Conditioned Image Attentive Encoder (HCIAE)
?
An important characteristic in dialogs is the use of co-reference to avoid repeating entities that can
be contextually resolved. In fact, in the VisDial dataset [7] nearly all (98%) dialogs involve at least
one pronoun. This means that for a model to correctly answer a question, it would require a reliable
mechanism for co-reference resolution.
A common approach is to use an encoder architecture with an attention mechanism that implicitly
performs co-reference resolution by identifying the portion of the dialog history that can help in
answering the current question [7, 38, 39, 32]. while using a holistic representation for the image.
Intuitively, one would also expect that the answer is also localized to regions in the image, and be
consistent with the attended history.
With this motivation, we propose a novel encoder architecture (called HCIAE) shown in Fig. 2. Our
encoder first uses the current question to attend to the exchanges in the history, and then use the
question and attended history to attend to the image, so as to obtain the final encoding.
Specifically, we use the spatial image features V ?
Rd?k from a convolution layer of a CNN. qt is
encoded with an LSTM to get a vector mqt ?
CNN
?
Rd . Simultaneously, each previous round of history
(H0 , . . . , Ht?1 ) is encoded separately with another
LSTM
?"
LSTM as Mth ? Rd?t . Conditioned on the question
embedding, the model attends to the history. The at?"
tended representation of the history and the question
LSTM
?$
embedding are concatenated, and used as input to
attend to the image:
LSTM
?"%&
?ht
= softmax(zth )
(2)
?
= waT tanh(Wh Mth + (Wq mqt )1T ) (1)
?
?
zth
Encoder
Figure 2: Structure of the proposed encoder.
where 1 ? R is a vector with all elements set to 1.
Wh , Wq ? Rt?d and wa ? Rk are parameters to be learned. ? ? Rk is the attention weight over
? ht is a convex combination of columns of Mt , weighted
history. The attended history feature m
h
? ht as the query vector and get
appropriately by the elements of ?t . We further concatenate mqt and m
the attended image feature v?t in the similar manner. Subsequently, all three components are used to
obtain the final embedding et :
t
? ht , v?t ])
et = tanh(We [mqt , m
where We ? Rd?3d is weight parameters and [?] is the concatenation operation.
4
(3)
4.2
Discriminator Loss
Discriminative visual dialog models produce a distribution over the candidate answer list At and
maximize the log-likelihood of the correct option agt
t . The loss function for D needs to be conducive
for knowledge transfer. In particular, it needs to encourage perceptually meaningful similarities.
Therefore, we use a metric-learning multi-class N-pair loss [43] defined as:
logistic loss
z
? N ?1
LD = Ln?pair {et , agt
t , {at,i }i=1 }, f = log 1 +
N
X
}|
gt
?
>
exp e>
t f (at,i ) ? et f (at )
|
{z
}
i=1
!{
(4)
score margin
where f is an attention based LSTM encoder for the answer. This attention can help the discriminator
better deal with paraphrases across answers. The attention weight is learnt through a 1-layer MLP
over LSTM output at each time step. The N-pair loss objective encourages learning a space in which
the ground truth answer is scored higher than other options, and at the same time, encourages options
similar to ground truth answers to score better than dissimilar ones. This means that, unlike the
multiclass logistic loss, the options that are correct but different from the correct option may not be
overly penalized, and thus can be useful in providing a reliable signal to the generator. See Fig. 1 for
an example. Follwing [43], we regularize the L2 norm of the embedding vectors to be small.
4.3
Discriminant Perceptual Loss and Knowledge Transfer from D to G
At a high-level, our approach for transferring knowledge from D to G is as follows: G repeatedly
queries D with answers a
?t that it generates for an input embedding et to get feedback and update
itself. In each such update, G?s goal is to update its parameters to try and have a
?t score higher than
the correct answer, agt
,
under
D?s
learned
embedding
and
scoring
function.
Formally,
the perceptual
t
loss that G aims to optimize is given by:
gt
>
>
LG = L1?pair {et , a
?t , agt
},
f
=
log
1
+
exp
e
f
(a
)
?
e
f
(?
a
)
(5)
t
t
t
t
t
where f is the embedding function learned by the discriminator as in (4). Intuitively, updating
generator parameters to minimize LG can be interpreted as learning to produce an answer sequence
a
?t that ?fools? the discriminator into believing that this answer should score higher than the human
response agt
t under the discriminator?s learned embedding f (?) and scoring function.
While it is straightforward to sample an answer a
?t from the generator and perform a forward pass
through the discriminator, na?vely, it is not possible to backpropagate the gradients to the generator
parameters since sampling discrete symbols results in zero gradients w.r.t. the generator parameters.
To overcome this, we leverage the recently introduced continuous relaxation of the categorical
distribution ? the Gumbel-softmax distribution or the Concrete distribution [18, 30].
At an intuitive level, the Gumbel-Softmax (GS) approximation uses the so called ?Gumbel-Max trick?
to reparametrize sampling from a categorical distribution and replaces argmax with softmax to obtain
a continuous relaxation of the discrete random variable. Formally, let x denote a K-ary categorical
K
random variable with parameters denoted by (p1 , . . . pK ), or x ? Cat(p). Let gi 1 denote K
?g
IID samples from the standard Gumbel distribution, gi ? F (g) = e?e . Now, a sample from the
Concrete distribution can be produced via the following transformation:
e(log pi +gi )/?
yi = PK (log p +g )
j
j /?
j=1 e
?i ? {1, . . . , K}
(6)
where ? is a temperature parameter that control how close samples y from this Concrete distribution
approximate the one-hot encoding of the categorical variable x.
As illustrated in Fig. 1, we augment the LSTM in G with a sequence of GS samplers. Specifically,
at each position in the answer sequence, we use a GS sampler to sample an answer token from
that conditional distribution. When coupled with the straight-through gradient estimator [2, 18]
this enables end-to-end differentiability. Specifically, during the forward pass we discretize the GS
samples into discrete samples, and in the backward pass use the continuous relaxation to compute
gradients. In our experiments, we held the temperature parameter fixed at 0.5.
5
5
Experiments
Dataset and Setup. We evaluate our proposed approach on the VisDial dataset [7], which was
collected by Das et al. by pairing two subjects on Amazon Mechanical Turk to chat about an image.
One person was assigned the role of a ?questioner? and the other of ?answerer?. One worker (the
questioner) sees only a single line of text describing an image (caption from COCO [25]); the image
remains hidden to the questioner. Their task is to ask questions about this hidden image to ?imagine
the scene better?. The second worker (the answerer) sees the image and caption and answers the
questions. The two workers take turns asking and answering questions for 10 rounds. We perform
experiments on VisDial v0.9 (the latest available release) containing 83k dialogs on COCO-train and
40k on COCO-val images, for a total of 1.2M dialog question-answer pairs. We split the 83k into 82k
for train, 1k for val, and use the 40k as test, in a manner consistent with [7]. The caption is
considered to be the first round in the dialog history.
Evaluation Protocol. Following the evaluation protocol established in [7], we use a retrieval setting
to evaluate the responses at each round in the dialog. Specifically, every question in VisDial is
coupled with a list of 100 candidate answer options, which the models are asked to sort for evaluation
purposes. D uses its score to rank these answer options, and G uses the log-likelihood of these
options for ranking. Models are evaluated on standard retrieval metrics ? (1) mean rank, (2) recall
@k, and (3) mean reciprocal rank (MRR) ? of the human response in the returned sorted list.
Pre-processing. We truncate captions/questions/answers longer than 24/16/8 words respectively. We
then build a vocabulary of words that occur at least 5 times in train, resulting in 8964 words.
Training Details In our experiments, all 3 LSTMs are single layer with 512d hidden state. We use
VGG-19 [42] to get the representation of image. We first rescale the images to be 224 ? 224 pixels,
and take the output of last pooling layer (512 ? 7 ? 7) as image feature. We use the Adam optimizer
with a base learning rate of 4e-4. We pre-train G using standard MLE for 20 epochs, and D with
supervised training based on Eq (4) for 30 epochs. Following [43], we regularize the L2 norm of the
embedding vectors to be small. Subsequently, we train G with LG + ?LM LE , which is a combination
of discriminative perceptual loss and MLE loss. We set ? to be 0.5. We found that including LM LE
(with teacher-forcing) is important for encouraging G to generate grammatically correct responses.
5.1
Results and Analysis
Baselines. We compare our proposed techniques to the current state-of-art generative and discriminative models developed in [7]. Specifically, [7] introduced 3 encoding architectures ? Late Fusion (LF),
Hierarchical Recurrent Encoder (HRE), Memory Network (MN) ? each trained with a generative
(-G) and discriminative (-D) decoder. We compare to all 6 models.
Our approaches. We present a few variants of our approach to systematically study the individual
contributions of our training procedure, novel encoder (HCIAE), self-attentive answer encoding
(ATT), and metric-loss (NP).
? HCIAE-G-MLE is a generative model with our proposed encoder trained under the MLE objective. Comparing this variant to the generative baselines from [7] establishes the improvement
due to our encoder (HCIAE).
? HCIAE-G-DIS is a generative model with our proposed encoder trained under the mixed MLE
and discriminator loss (knowledge transfer). This forms our best generative model. Comparing
this model to HCIAE-G-MLE establishes the improvement due to our discriminative training.
? HCIAE-D-MLE is a discriminative model with our proposed encoder, trained under the standard discriminative cross-entropy loss. The answer candidates are encoded using an LSTM
(no attention). Comparing this variant to the discriminative baselines from [7] establishes the
improvement due to our encoder (HCIAE) in the discriminative setting.
? HCIAE-D-NP is a discriminative model with our proposed encoder, trained under the n-pair
discriminative loss (as described in Section 4.2). The answer candidates are encoded using an
LSTM (no attention). Comparing this variant to HCIAE-D-MLE establishes the improvement
due to the n-pair loss.
? HCIAE-D-NP-ATT is a discriminative model with our proposed encoder, trained under the
n-pair discriminative loss (as described in Section 4.2), and using the self-attentive answer
encoding. Comparing this variant to HCIAE-D-NP establishes the improvement due to the
self-attention mechanism while encoding the answers.
6
Table 1: Results (generative) on VisDial dataset. ?MRR? Table 2: Results (discriminative) on VisDial dataset.
is mean reciprocal rank and ?Mean? is mean rank.
Model
MRR R@1 R@5 R@10 Mean
Model
MRR R@1 R@5 R@10 Mean
LF-D [7]
0.5807 43.82 74.68 84.07 5.78
0.5868 44.82 74.81 84.36 5.66
LF-G [7]
0.5199 41.83 61.78 67.59 17.07 HREA-D [7]
0.5965 45.55 76.22 85.37 5.46
HREA-G [7]
0.5242 42.28 62.33 68.17 16.79 MN-D [7]
MN-G [7]
0.5259 42.29 62.85 68.88 17.06
HCIAE-D-MLE
0.6140 47.73 77.50 86.35 5.15
0.6182 47.98 78.35 87.16 4.92
HCIAE-G-MLE 0.5386 44.06 63.55 69.24 16.01 HCIAE-D-NP
HCIAE-G-DIS 0.5467 44.35 65.28 71.55 14.23 HCIAE-D-NP-ATT 0.6222 48.48 78.75 87.59 4.81
Results. Tables 1, 2 present results for all our models and baselines in generative and discriminative
settings. The key observations are:
1. Main Results for HCIAE-G-DIS: Our final generative model with all ?bells and whistles?,
HCIAE-G-DIS, uniformly performs the best under all the metrics, outperforming the
previous state-of-art model MN-G by 2.43% on R@5. This shows the importance of the
knowledge transfer from the discriminator and the benefit from our encoder architecture.
2. Knowledge transfer vs. encoder for G: To understand the relative importance of the proposed history conditioned image attentive encoder (HCIAE) and the knowledge transfer, we
compared the performance of HCIAE-G-DIS with HCIAE-G-MLE, which uses our proposed encoder but without any feedback from the discriminator. This comparison highlights
two points: first, HCIAE-G-MLE improves R@5 by 0.7% over the current state-of-art
method (MN-D) confirming the benefits of our encoder. Secondly, and importantly, its
performance is lower than HCIAE-G-DIS by 1.7% on R@5, confirming that the modifications to encoder alone will not be sufficient to gain improvements in answer generation;
knowledge transfer from D greatly improves G.
3. Metric loss vs. self-attentive answer encoding: In the purely discriminative setting, our
final discriminative model (HCIAE-D-NP-ATT) also beats the performance of the corresponding state-of-art models [7] by 2.53% on R@5. The n-pair loss used in the discriminator
is not only helpful for knowledge transfer but it also improves the performance of the
discriminator by 0.85% on R@5 (compare HCIAE-D-NP to HCIAE-D-MLE). The improvements obtained by using the answer attention mechanism leads to an additional, albeit
small, gains of 0.4% on R@5 to the discriminator performance (compare HCIAE-D-NP to
HCIAE-D-NP-ATT).
5.2
Does updating discriminator help?
Recall that our model training happens as follows: we independently train the generative model
HCIAE-G-MLE and the discriminative model HCIAE-D-NP-ATT. With HCIAE-G-MLE as the
initialization, the generative model is updated based on the feedback from HCIAE-D-NP-ATT and
this results in our final HCIAE-G-DIS.
We performed two further experiments to answer the following questions:
? What happens if we continue training HCIAE-D-NP-ATT in an adversarial setting? In particular, we continue training by maximizing the score of the ground truth answer agt
t and
minimizing the score of the generated answer a
?t , effectively setting up an adversarial training regime LD = ?LG . The resulting discriminator HCIAE-GAN1 has significant drop in
performance, as can be seen in Table. 4 (32.97% R@5). This is perhaps expected because
HCIAE-GAN1 updates its parameters based on only two answers, the ground truth and the
generated sample (which is likely to be similar to ground truth). This wrecks the structure that
HCIAE-D-NP-ATT had previously learned by leveraging additional incorrect options.
? What happens if we continue structure-preserving training of HCIAE-D-NP-ATT? In addition
to providing HCIAE-D-NP-ATT samples from G as fake answers, we also include incorrect
options as negative answers so that the structure learned by the discriminator is preserved.
HCIAE-D-NP-ATT continues to train under loss LD . In this case (HCIAE-GAN2 in Table. 4),
we find that there is a small improvement in the performance of G. The additional computational
overhead to training the discriminator supersedes the performance improvement. Also note that
HCIAE-D-NP-ATT itself gets worse at the dialog task.
7
Table 3: Qualitative comparison. ?Ours? are samples from G-DIS model with different gumbel noise z. Images
from the COCO dataset [25].
Q: Is it a home or restaurant?
A: I think restaurant.
G-MLE: I can?t tell.
Ours (z1): Hard to say.
Ours (z2): It looks like a restaurant.
Ours (z3): I can?t tell because it
is too close.
Q: Can you see his face?
A: I am not sure.
G-MLE: Yes.
Ours (z1): I can only see the
back of his body.
Ours (z2): No.
Ours (z3): No , he?s too far
away.
Q: How old does the man seem to
be?
A: 20?s.
G-MLE: Late teens.
Ours (z1): On his 20s.
Ours (z2): In his twenties.
Ours (z3): Mid 20?s.
Q: Can you see broccoli?
A: Yes, 3 larger pieces and some
small.
G-MLE: No, just the broccoli.
Ours (z1): I can see broccoli,
slightly butter.
Ours (z2): Yes, there is broccoli.
Ours (z3): Yes, broccoli is
green.
Table 4: Adversarial training results on VisDial dataset.
Discriminative
Generative
Model
MRR
R@1
R@5
R@10
Mean
MRR
R@1
R@5
R@10
Mean
HCIAE-D-NP-ATT
HCIAE-G-DIS
0.6222
-
48.48
-
78.75
-
87.59
-
4.81
-
0.5467
44.35
65.28
71.55
14.23
HCIAE-GAN1
HCIAE-GAN2
0.2177
0.6050
8.82
46.20
32.97
77.92
52.14
87.20
18.53
4.97
0.5298
0.5459
43.12
44.33
62.74
65.05
68.58
71.40
16.25
14.34
One might wonder, why not train a GAN for visual dialog? Formulating the task in a GAN setting
would involve G and D training in tandem with D providing feedback as to whether a response that
G generates is real or fake. We found this to be a particularly unstable setting, for two main reasons:
First, consider the case when the ground truth answer and the generated answers are the same. This
happens for answers that are typically short or ?cryptic? (e.g. ?yes?). In this case, D can not train itself
or provide feedback, as the answer is labeled both positive and negative. Second, in cases where the
ground truth answer is descriptive but the generator provides a short answer, D can quickly become
powerful enough to discard generated samples as fake. In this case, D is not able to provide any
information to G to get better at the task. Our experience suggests that the discriminator, if one were
to consider a ?GANs for visual dialog? setting, can not merely be focused on differentiating fake
from real. It needs to be able to score similarity between the ground truth and other answers. Such a
scoring mechanism provides a more reliable feedback to G. In fact, as we show in the previous two
results, a pre-trained D that captures this structure is the key ingredient in sharing knowledge with G.
The adversarial training of D is not central.
5.3
Qualitative Comparison
In Table 3 we present a couple of qualitative examples that compares the responses generated by
G-MLE and G-DIS. G-MLE predominantly produces ?safe? and less informative answers, such as
?Yes? and or ?I can?t tell?. In contrast, our proposed model G-DIS does so less frequently, and often
generates more diverse yet informative responses.
6
Conclusion
Generative models for (visual) dialog are typically trained with an MLE objective. As a result, they
tend to latch on to safe and generic responses. Discriminative (or retrieval) models on the other hand
have been shown to significantly outperform their generative counterparts. However, discriminative
models can not be deployed as dialog agents with a real user where canned candidate responses
are not available. In this work, we propose transferring knowledge from a powerful discriminative
visual dialog model to a generative model. We leverage the Gumbel-Softmax (GS) approximation to
the discrete distribution ?specifically, a RNN augmented with a sequence of GS samplers, coupled
with a ST gradient estimator for end-to-end differentiability. We also propose a novel visual dialog
encoder that reasons about image-attention informed by the history of the dialog; and employ a
metric learning loss along with a self-attentive answer encoding to enable the discriminator to
learn meaningful structure in dialog responses. The result is a generative visual dialog model that
significantly outperforms state-of-the-art.
8
References
[1] Stanislaw Antol, Aishwarya Agrawal, Jiasen Lu, Margaret Mitchell, Dhruv Batra, C Lawrence Zitnick,
and Devi Parikh. Vqa: Visual question answering. In Proceedings of the IEEE International Conference
on Computer Vision, pages 2425?2433, 2015.
[2] Yoshua Bengio, Nicholas L?onard, and Aaron C. Courville. Estimating or propagating gradients through
stochastic neurons for conditional computation. CoRR, abs/1308.3432, 2013.
[3] Antoine Bordes and Jason Weston.
arXiv:1605.07683, 2016.
Learning end-to-end goal-oriented dialog.
arXiv preprint
[4] Kan Chen, Jiang Wang, Liang-Chieh Chen, Haoyuan Gao, Wei Xu, and Ram Nevatia. Abc-cnn: An attention based convolutional neural network for visual question answering. arXiv preprint arXiv:1511.05960,
2015.
[5] Tianqi Chen, Ian Goodfellow, and Jonathon Shlens. Net2net: Accelerating learning via knowledge transfer.
arXiv preprint arXiv:1511.05641, 2015.
[6] Bo Dai, Dahua Lin, Raquel Urtasun, and Sanja Fidler. Towards diverse and natural image descriptions via
a conditional gan. arXiv preprint arXiv:1703.06029, 2017.
[7] Abhishek Das, Satwik Kottur, Khushi Gupta, Avi Singh, Deshraj Yadav, Jos? M.F. Moura, Devi Parikh,
and Dhruv Batra. Visual Dialog. In Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, 2017.
[8] Abhishek Das, Satwik Kottur, Jos? MF Moura, Stefan Lee, and Dhruv Batra. Learning cooperative visual
dialog agents with deep reinforcement learning. arXiv preprint arXiv:1703.06585, 2017.
[9] Harm de Vries, Florian Strub, Sarath Chandar, Olivier Pietquin, Hugo Larochelle, and Aaron Courville.
Guesswhat?! visual object discovery through multi-modal dialogue. arXiv preprint arXiv:1611.08481,
2016.
[10] Emily L. Denton, Soumith Chintala, Arthur Szlam, and Robert Fergus. Deep generative image models
using a laplacian pyramid of adversarial networks. Neural Information Processing Systems, 2015.
[11] Jeff Donahue, Lisa Anne Hendricks, Sergio Guadarrama, Marcus Rohrbach, Subhashini Venugopalan,
Kate Saenko, and Trevor Darrell. Long-term Recurrent Convolutional Networks for Visual Recognition
and Description. In CVPR, 2015.
[12] Alexey Dosovitskiy and Thomas Brox. Generating images with perceptual similarity metrics based on
deep networks. In Advances in Neural Information Processing Systems, pages 658?666, 2016.
[13] Hao Fang, Saurabh Gupta, Forrest N. Iandola, Rupesh Kumar Srivastava, Li Deng, Piotr Doll?r, Jianfeng
Gao, Xiaodong He, Margaret Mitchell, John C. Platt, C. Lawrence Zitnick, and Geoffrey Zweig. From
Captions to Visual Concepts and Back. In CVPR, 2015.
[14] Haoyuan Gao, Junhua Mao, Jie Zhou, Zhiheng Huang, Lei Wang, and Wei Xu. Are you talking to a
machine? dataset and methods for multilingual image question answering. In NIPS, 2015.
[15] Leon A Gatys, Alexander S Ecker, and Matthias Bethge. A neural algorithm of artistic style. arXiv preprint
arXiv:1508.06576, 2015.
[16] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron
Courville, and Yoshua Bengio. Generative adversarial nets. In Advances in neural information processing
systems, pages 2672?2680, 2014.
[17] Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. Distilling the knowledge in a neural network. arXiv
preprint arXiv:1503.02531, 2015.
[18] Eric Jang, Shixiang Gu, and Ben Poole. Categorical reparameterization with gumbel-softmax. arXiv
preprint arXiv:1611.01144, 2016.
[19] Justin Johnson, Alexandre Alahi, and Li Fei-Fei. Perceptual losses for real-time style transfer and
super-resolution. In European Conference on Computer Vision, pages 694?711. Springer, 2016.
[20] Andrej Karpathy and Li Fei-Fei. Deep visual-semantic alignments for generating image descriptions. In
CVPR, 2015.
[21] Matt J. Kusner and Jos? Miguel Hern?ndez-Lobato. Gans for sequences of discrete elements with the
gumbel-softmax distribution. CoRR, abs/1611.04051, 2016.
[22] Christian Ledig, Lucas Theis, Ferenc Huszar, Jose Caballero, Andrew P. Aitken, Alykhan Tejani, Johannes
Totz, Zehan Wang, and Wenzhe Shi. Photo-realistic single image super-resolution using a generative
adversarial network. CoRR, abs/1609.04802, 2016.
[23] Jiwei Li, Will Monroe, Tianlin Shi, Alan Ritter, and Dan Jurafsky. Adversarial learning for neural dialogue
generation. arXiv preprint arXiv:1701.06547, 2017.
[24] Chin-Yew Lin. Rouge: A package for automatic evaluation of summaries. In ACL 2004 Workshop, 2004.
9
[25] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Doll?r,
and C Lawrence Zitnick. Microsoft coco: Common objects in context. In European Conference on
Computer Vision, pages 740?755. Springer, 2014.
[26] Chia-Wei Liu, Ryan Lowe, Iulian V Serban, Michael Noseworthy, Laurent Charlin, and Joelle Pineau.
How not to evaluate your dialogue system: An empirical study of unsupervised evaluation metrics for
dialogue response generation. arXiv preprint arXiv:1603.08023, 2016.
[27] Siqi Liu, Zhenhai Zhu, Ning Ye, Sergio Guadarrama, and Kevin Murphy. Optimization of image description
metrics using policy gradient methods. arXiv preprint arXiv:1612.00370, 2016.
[28] Jiasen Lu, Caiming Xiong, Devi Parikh, and Richard Socher. Knowing when to look: Adaptive attention
via A visual sentinel for image captioning. CoRR, abs/1612.01887, 2016.
[29] Jiasen Lu, Jianwei Yang, Dhruv Batra, and Devi Parikh. Hierarchical question-image co-attention for
visual question answering. In Advances In Neural Information Processing Systems, pages 289?297, 2016.
[30] Chris J Maddison, Andriy Mnih, and Yee Whye Teh. The concrete distribution: A continuous relaxation of
discrete random variables. arXiv preprint arXiv:1611.00712, 2016.
[31] Mateusz Malinowski, Marcus Rohrbach, and Mario Fritz. Ask your neurons: A neural-based approach to
answering questions about images. In ICCV, 2015.
[32] Hongyuan Mei, Mohit Bansal, and Matthew R Walter. Coherent dialogue with attention-based language
models. arXiv preprint arXiv:1611.06997, 2016.
[33] Nasrin Mostafazadeh, Chris Brockett, Bill Dolan, Michel Galley, Jianfeng Gao, Georgios P Spithourakis,
and Lucy Vanderwende. Image-grounded conversations: Multimodal context for natural question and
response generation. arXiv preprint arXiv:1701.08251, 2017.
[34] Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. Bleu: a method for automatic evaluation
of machine translation. In ACL, 2002.
[35] Alec Radford, Luke Metz, and Soumith Chintala. Unsupervised representation learning with deep
convolutional generative adversarial networks. CoRR, abs/1511.06434, 2015.
[36] Marc?Aurelio Ranzato, Sumit Chopra, Michael Auli, and Wojciech Zaremba. Sequence level training with
recurrent neural networks. arXiv preprint arXiv:1511.06732, 2015.
[37] Mengye Ren, Ryan Kiros, and Richard Zemel. Exploring models and data for image question answering.
In NIPS, 2015.
[38] Iulian V Serban, Alessandro Sordoni, Yoshua Bengio, Aaron Courville, and Joelle Pineau. Building end-to-end dialogue systems using generative hierarchical neural network models. arXiv preprint
arXiv:1507.04808, 2015.
[39] Iulian Vlad Serban, Alessandro Sordoni, Ryan Lowe, Laurent Charlin, Joelle Pineau, Aaron Courville,
and Yoshua Bengio. A hierarchical latent variable encoder-decoder model for generating dialogues. arXiv
preprint arXiv:1605.06069, 2016.
[40] Iulian Vlad Serban, Alessandro Sordoni, Ryan Lowe, Laurent Charlin, Joelle Pineau, Aaron Courville,
and Yoshua Bengio. A hierarchical latent variable encoder-decoder model for generating dialogues. In
Thirty-First AAAI Conference on Artificial Intelligence, 2017.
[41] Rakshith Shetty, Marcus Rohrbach, Lisa Anne Hendricks, Mario Fritz, and Bernt Schiele. Speaking the
same language: Matching machine to human captions by adversarial training. CoRR, abs/1703.10476,
2017.
[42] Karen Simonyan and Andrew Zisserman. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556, 2014.
[43] Kihyuk Sohn. Improved deep metric learning with multi-class n-pair loss objective. In Advances in Neural
Information Processing Systems, pages 1849?1857, 2016.
[44] Alessandro Sordoni, Michel Galley, Michael Auli, Chris Brockett, Yangfeng Ji, Margaret Mitchell, JianYun Nie, Jianfeng Gao, and Bill Dolan. A neural network approach to context-sensitive generation of
conversational responses. arXiv preprint arXiv:1506.06714, 2015.
[45] Florian Strub, Harm de Vries, Jeremie Mary, Bilal Piot, Aaron Courville, and Olivier Pietquin. End-to-end
optimization of goal-driven and visually grounded dialogue systems. arXiv preprint arXiv:1703.05423,
2017.
[46] Sainbayar Sukhbaatar, Jason Weston, Rob Fergus, et al. End-to-end memory networks. In Advances in
neural information processing systems, pages 2440?2448, 2015.
[47] Ilya Sutskever, Oriol Vinyals, and Quoc V Le. Sequence to sequence learning with neural networks. In
Advances in neural information processing systems, pages 3104?3112, 2014.
[48] Ramakrishna Vedantam, C Lawrence Zitnick, and Devi Parikh. Cider: Consensus-based image description
evaluation. In CVPR, 2015.
10
[49] Oriol Vinyals, Alexander Toshev, Samy Bengio, and Dumitru Erhan. Show and tell: A neural image
caption generator. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,
pages 3156?3164, 2015.
[50] Huijuan Xu and Kate Saenko. Ask, attend and answer: Exploring question-guided spatial attention for
visual question answering. In European Conference on Computer Vision, pages 451?466. Springer, 2016.
[51] Kelvin Xu, Jimmy Ba, Ryan Kiros, Kyunghyun Cho, Aaron C. Courville, Ruslan Salakhutdinov, Richard S.
Zemel, and Yoshua Bengio. Show, attend and tell: Neural image caption generation with visual attention.
CoRR, abs/1502.03044, 2015.
[52] Zichao Yang, Xiaodong He, Jianfeng Gao, Li Deng, and Alex Smola. Stacked attention networks for image
question answering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,
pages 21?29, 2016.
[53] Lantao Yu, Weinan Zhang, Jun Wang, and Yong Yu. Seqgan: Sequence generative adversarial nets with
policy gradient. AAAI Conference on Artificial Intelligence, 2017.
[54] Junbo Jake Zhao, Micha?l Mathieu, and Yann LeCun. Energy-based generative adversarial network. CoRR,
abs/1609.03126, 2016.
[55] Jun-Yan Zhu, Taesung Park, Phillip Isola, and Alexei A Efros. Unpaired image-to-image translation using
cycle-consistent adversarial networks. arXiv preprint arXiv:1703.10593, 2017.
11
| 6635 |@word cnn:3 stronger:1 norm:2 q1:1 attended:4 mengye:1 ld:3 lantao:1 liu:2 ndez:1 score:14 att:14 ours:13 bilal:1 outperforms:4 existing:1 guadarrama:2 current:5 comparing:5 anne:2 haoyuan:2 com:1 yet:1 z2:4 must:1 john:1 concatenate:1 realistic:1 informative:3 confirming:2 christian:1 enables:2 drop:1 update:4 v:2 alone:2 gan2:2 sukhbaatar:1 intelligence:3 generative:39 alec:1 reciprocal:2 short:2 shi:2 provides:2 zhang:1 along:2 become:1 junbo:1 pairing:1 incorrect:2 qualitative:3 overhead:1 dan:1 manner:2 introduce:2 aitken:1 mohit:1 expected:1 indeed:1 behavior:1 p1:1 dialog:65 kiros:2 multi:5 frequently:1 whistle:1 gatys:1 salakhutdinov:1 inspired:2 resolve:1 encouraging:2 soumith:2 tandem:2 begin:1 estimating:1 moreover:1 what:3 interpreted:1 developed:1 informed:1 transformation:1 every:1 alahi:1 zaremba:1 platt:1 control:1 ramanan:1 szlam:1 sherjil:1 producing:1 kelvin:1 positive:1 attend:5 todd:1 rouge:2 encoding:11 jiang:1 laurent:3 path:1 might:1 alexey:1 bird:6 initialization:1 acl:2 suggests:1 challenging:1 luke:1 co:6 micha:1 contextually:1 jurafsky:1 range:1 practical:2 thirty:1 lecun:1 practice:1 block:1 lf:3 maire:1 procedure:2 mei:1 rnn:3 yan:1 bell:1 significantly:3 onard:1 matching:1 empirical:1 pre:4 word:5 regular:1 get:7 cannot:1 close:3 zehan:1 andrej:1 salim:1 context:5 yee:1 optimize:3 fruitful:1 dean:1 demonstrated:1 ecker:1 lobato:1 bill:2 straightforward:1 attention:23 latest:1 independently:1 emily:1 iulian:4 jimmy:1 focused:1 resolution:4 amazon:1 identifying:2 convex:1 pouget:1 perceive:1 attending:3 estimator:4 sarath:1 importantly:1 shlens:1 regularize:2 fang:1 his:4 reparameterization:1 embedding:13 updated:2 imagine:1 user:3 caption:9 olivier:2 us:6 samy:1 goodfellow:2 trick:1 element:3 recognition:5 particularly:3 updating:2 continues:1 cooperative:1 labeled:1 role:1 preprint:22 wang:4 capture:1 yadav:1 region:6 cycle:1 ranzato:1 highest:2 alessandro:4 environment:1 schiele:1 nie:1 reward:1 asked:1 warde:1 ultimately:1 trained:16 singh:1 deva:1 ferenc:1 purely:1 eric:1 gu:1 resolved:1 multimodal:1 joint:1 emergent:1 cat:3 reparametrize:1 train:10 stacked:1 walter:1 instantiated:1 effective:1 describe:2 query:2 artificial:3 zemel:2 mqt:4 avi:1 jianfeng:4 h0:2 kevin:1 tell:7 jean:1 bernt:1 larger:1 plausible:1 cvpr:4 say:1 emerged:1 widely:1 encoded:5 encoder:33 simonyan:1 gi:3 ward:1 think:1 itself:3 final:5 sequence:27 agrawal:1 descriptive:1 matthias:1 net:2 propose:5 interaction:1 zth:2 frequent:1 relevant:3 pronoun:1 holistic:1 till:1 poorly:1 achieve:2 papineni:1 margaret:3 jianyun:1 intuitive:1 description:5 sutskever:1 darrell:1 jing:1 captioning:4 generating:4 adam:1 tianqi:1 ben:1 object:2 help:4 attends:1 andrew:2 recurrent:5 propagating:1 miguel:1 rescale:1 qt:7 eq:1 strong:2 pietquin:2 taskspecific:1 larochelle:1 distilling:1 safe:5 guided:1 ning:1 correct:6 stochastic:2 subsequently:2 human:9 jonathon:1 enable:4 viewing:1 backprop:1 require:1 exchange:1 assign:1 preliminary:1 broccoli:5 ryan:5 sainbayar:1 secondly:1 exploring:2 pytorch:1 sordoni:4 considered:1 ground:14 dhruv:5 visually:1 exp:2 lawrence:4 caballero:1 lm:2 matthew:1 efros:1 optimizer:1 purpose:1 ruslan:1 estimation:2 applicable:1 tanh:2 sensitive:1 establishes:5 weighted:1 stefan:1 sensor:1 super:2 cider:2 aim:5 zhou:1 avoid:1 gatech:1 release:1 focus:1 improvement:9 rank:9 likelihood:5 believing:1 greatly:1 contrast:4 adversarial:19 underneath:1 huijuan:1 sense:1 baseline:4 helpful:1 am:1 rupesh:1 dependent:1 kottur:2 entire:1 transferring:3 typically:3 brockett:2 mth:2 hidden:3 perona:1 pixel:1 overall:1 augment:1 denoted:1 lucas:1 development:1 art:7 spatial:2 softmax:11 brox:1 saurabh:1 piotr:2 sampling:2 park:1 look:3 denton:1 nearly:1 unsupervised:2 broad:1 yu:2 siqi:1 future:1 yoshua:6 np:19 dosovitskiy:1 richard:3 primarily:1 employ:3 mirza:1 oriented:1 few:3 simultaneously:1 recognize:1 individual:1 murphy:1 argmax:1 microsoft:1 ab:8 mlp:1 highly:3 mnih:1 alexei:1 evaluation:7 alignment:1 farley:1 held:1 antol:1 amenable:1 visdial:10 encourage:1 worker:3 arthur:1 experience:1 closer:1 vely:2 old:1 instance:1 column:1 modeling:2 asking:1 phrase:1 artistic:1 usefulness:2 wonder:1 successful:1 galley:2 johnson:1 sumit:1 too:2 encoders:1 answer:62 teacher:1 learnt:1 cho:1 st:1 fritz:2 international:1 tianlin:1 lstm:11 person:1 fundamental:1 lee:1 ritter:1 jos:3 michael:4 bethge:1 quickly:1 ilya:1 gans:6 na:2 concrete:4 aaai:2 central:1 containing:1 huang:1 worse:1 zhao:1 style:2 return:1 michel:2 li:5 dialogue:9 nevatia:1 wojciech:1 diversity:1 de:2 chandar:1 kate:2 explicitly:1 ranking:1 tsung:1 piece:1 performed:1 try:1 lowe:3 break:1 jason:2 mario:2 weinan:1 h1:1 start:1 sort:4 questioner:3 option:12 metz:1 portion:1 maintains:1 contribution:4 minimize:1 convolutional:4 characteristic:1 judgment:1 serge:1 sitting:1 yew:1 tejani:1 yes:10 identify:1 produced:1 iid:1 ren:1 venugopalan:1 lu:3 straight:3 ary:1 history:23 moura:2 tended:1 sharing:1 trevor:1 facebook:2 attentive:6 energy:3 turk:1 james:1 chintala:2 couple:1 sampled:3 ledig:1 gain:2 dataset:13 jeremie:1 wh:2 mitchell:3 recall:6 vlad:2 conversation:3 color:1 improves:3 ask:3 subtle:2 knowledge:22 guesswhat:1 back:3 feed:1 alexandre:1 higher:4 supervised:1 day:1 totz:1 response:33 modal:3 improved:1 wei:4 zisserman:1 evaluated:1 done:1 charlin:3 just:1 smola:1 alykhan:1 hand:2 receives:5 lstms:1 mehdi:1 canned:1 chat:1 pineau:4 logistic:2 gray:1 mostafazadeh:1 perhaps:1 lei:1 mary:1 xiaodong:2 phillip:1 ye:1 requiring:1 building:1 matt:1 counterpart:3 concept:1 fidler:1 kyunghyun:1 assigned:1 semantic:2 illustrated:1 deal:1 white:2 round:7 latch:1 game:1 self:7 encourages:2 during:1 backpropagating:1 shixiang:1 whye:1 bansal:1 chin:1 performs:2 l1:1 temperature:2 zhiheng:1 image:61 novel:6 recently:5 parikh:6 predominantly:1 common:2 mt:1 ji:1 teen:1 hugo:1 anitha:1 overview:1 he:3 dahua:1 significant:3 ai:3 automatic:4 rd:4 language:5 had:1 maximizing:2 sanja:1 satwik:2 similarity:12 longer:1 v0:1 gt:3 base:1 sergio:2 dominant:1 recent:1 mrr:6 produce:11 aishwarya:1 driven:1 coco:6 discard:1 forcing:1 hay:1 outperforming:1 continue:3 success:1 joelle:4 yi:2 scoring:4 captured:1 seen:1 additional:6 preserving:1 florian:2 isola:1 deng:2 dai:1 novelty:1 maximize:1 paradigm:4 signal:2 multiple:2 conducive:2 alan:1 technical:1 taesung:1 cross:3 chia:1 retrieval:3 long:1 lin:3 zweig:1 mle:30 a1:1 laplacian:1 involving:1 variant:5 essentially:1 metric:18 vision:9 arxiv:44 grounded:5 pyramid:1 answerer:2 addition:3 preserved:1 separately:1 receive:1 jiwei:1 source:1 crucial:1 appropriately:1 conversing:1 unlike:2 sure:2 subject:1 tend:5 strub:2 pooling:1 flow:1 leveraging:1 grammatically:1 seem:1 supersedes:1 yang:2 chopra:1 leverage:5 split:1 enough:1 bengio:7 rendering:1 embeddings:1 variety:3 restaurant:3 followup:1 architecture:5 andriy:1 knowing:1 vgg:1 multiclass:1 whether:1 accelerating:1 render:1 returned:1 karen:1 speaking:1 repeatedly:1 deep:9 jie:1 useful:4 fake:5 malinowski:1 clear:1 johannes:1 dbatra:1 vqa:2 hrea:2 repeating:1 karpathy:1 fool:1 involve:3 mid:1 sohn:1 differentiability:4 wreck:1 unpaired:1 generate:1 http:1 outperform:3 piot:1 overly:1 correctly:1 diverse:4 discrete:11 key:2 serban:4 mango:2 ht:6 backward:1 ram:1 relaxation:5 pietro:1 merely:1 year:1 convert:1 package:1 jose:1 you:5 communicate:1 raquel:1 powerful:2 yann:1 forrest:1 home:1 draw:1 huszar:1 capturing:1 layer:4 courville:8 replaces:1 g:10 occur:1 yangfeng:1 fei:4 your:2 vanderwende:1 alex:1 scene:1 yong:1 toshev:1 aspect:1 generates:3 formulating:1 kumar:1 leon:1 conversational:1 truncate:1 combination:2 across:5 subhashini:1 slightly:1 kusner:1 rob:1 modification:1 happens:4 quoc:1 intuitively:2 iccv:1 jiasen:4 ln:1 remains:1 previously:1 hern:1 bing:1 mechanism:9 describing:3 turn:1 needed:1 know:3 end:20 photo:1 available:4 operation:1 doll:2 hierarchical:5 away:1 generic:5 caiming:1 wat:1 nicholas:1 xiong:1 alternative:2 jang:1 shetty:1 thomas:1 include:1 gan:4 cryptic:1 concatenated:1 especially:1 build:1 jake:1 chieh:1 objective:4 question:37 disengage:1 primary:2 rt:1 traditional:1 antoine:1 gradient:13 separate:1 reinforce:1 concatenation:1 decoder:4 entity:1 parametrized:1 chris:3 maddison:1 collected:1 discriminant:1 urtasun:1 reason:6 stanislaw:1 unstable:1 marcus:3 consensus:1 bleu:2 besides:1 latching:1 useless:1 ozair:1 z3:4 code:1 minimizing:3 providing:4 zichao:1 equivalently:1 lg:4 liang:1 unfortunately:1 setup:4 robert:1 hao:1 negative:3 ba:1 design:1 reliably:1 policy:3 twenty:1 perform:2 teh:1 discretize:1 convolution:1 hqi:3 datasets:1 neuron:2 observation:1 beat:1 hinton:1 communication:1 head:1 auli:2 paraphrase:1 david:1 introduced:3 pair:10 trainable:2 mechanical:1 connection:1 discriminator:31 z1:4 coherent:1 learned:8 textual:1 established:1 nip:2 justin:1 recurring:2 poole:1 able:2 hendricks:2 pattern:3 mateusz:1 regime:1 max:1 memory:5 including:2 reliable:3 green:1 hot:1 natural:3 localizes:1 mn:5 zhu:3 shop:2 improve:1 github:1 technology:1 mathieu:1 categorical:5 jun:2 coupled:5 utterance:2 text:6 prior:3 understanding:1 discovery:1 epoch:2 val:2 theis:1 l2:2 relative:1 georgios:1 seqgan:1 dolan:2 expect:1 loss:29 highlight:1 mixed:1 generation:11 geoffrey:2 ramakrishna:1 ingredient:1 generator:15 localized:1 downloaded:1 agent:9 sufficient:1 consistent:3 metal:1 informativeness:1 bank:1 systematically:1 roukos:1 bordes:1 translation:2 pi:1 penalized:1 token:2 summary:1 last:2 keeping:1 infeasible:1 dis:11 side:1 lisa:2 understand:2 institute:1 wide:1 face:2 gan1:3 differentiating:1 emerge:1 benefit:2 overcome:1 feedback:6 vocabulary:1 world:3 valid:2 forward:2 reinforcement:2 adaptive:1 far:1 erhan:1 kishore:1 correlate:1 approximate:1 implicitly:2 multilingual:1 hongyuan:1 harm:2 belongie:1 vedantam:1 discriminative:31 sentinel:1 abhishek:2 fergus:2 don:2 continuous:6 latent:2 why:1 table:9 promising:1 nature:1 jiasenlu:2 transfer:18 learn:2 improving:1 necessarily:1 european:3 domain:2 zitnick:4 marc:1 da:4 pk:2 protocol:2 ebgan:1 main:2 aurelio:1 wenzhe:1 big:1 scored:1 motivation:2 noise:1 junhua:1 body:1 xu:5 fig:6 augmented:3 georgia:1 deployed:2 aid:1 position:1 mao:1 guiding:1 candidate:14 perceptual:8 answering:12 late:2 learns:6 donahue:1 ian:2 rk:2 dumitru:1 specific:1 symbol:2 list:15 abadie:1 gupta:2 fusion:1 workshop:1 socher:1 albeit:1 effectively:2 corr:8 importance:2 agt:8 zhenhai:1 perceptually:3 conditioned:3 vries:2 margin:2 gumbel:13 chen:3 monroe:1 mf:1 suited:1 yang1:1 intersection:1 entropy:3 backpropagate:1 lucy:1 likely:1 gao:6 rohrbach:3 visual:42 devi:6 vinyals:3 iandola:1 bo:1 talking:1 radford:1 springer:3 kan:1 truth:13 abc:1 weston:2 conditional:3 sorted:1 goal:5 viewed:1 jianwei:2 towards:1 jeff:2 man:1 content:1 tiger:2 hard:1 butter:1 specifically:12 uniformly:1 sampler:7 hre:1 total:1 batra:4 pas:3 called:2 discriminate:1 meaningful:3 saenko:2 aaron:8 formally:3 wq:2 kihyuk:1 lu1:1 alexander:2 dissimilar:1 oriol:3 evaluate:3 audio:1 bench:1 scratch:1 srivastava:1 |
6,230 | 6,636 | MaskRNN: Instance Level Video Object
Segmentation
Yuan-Ting Hu
UIUC
[email protected]
Jia-Bin Huang
Virginia Tech
[email protected]
Alexander G. Schwing
UIUC
[email protected]
Abstract
Instance level video object segmentation is an important technique for video editing
and compression. To capture the temporal coherence, in this paper, we develop
MaskRNN, a recurrent neural net approach which fuses in each frame the output
of two deep nets for each object instance ? a binary segmentation net providing
a mask and a localization net providing a bounding box. Due to the recurrent
component and the localization component, our method is able to take advantage
of long-term temporal structures of the video data as well as rejecting outliers.
We validate the proposed algorithm on three challenging benchmark datasets,
the DAVIS-2016 dataset, the DAVIS-2017 dataset, and the Segtrack v2 dataset,
achieving state-of-the-art performance on all of them.
1
Introduction
Instance level video object segmentation of complex scenes is a challenging problem with applications
in areas such as object identification, video editing, and video compression. With the recent release
of the DAVIS dataset [39], the task of segmenting multiple object instances from videos has gained
considerable attention. However, just like for classical foreground-background segmentation, deforming shapes, fast movements, and multiple objects occluding each other pose significant challenges to
instance level video object segmentation.
Classical techniques [5, 10, 11, 17, 21, 41, 20, 44, 49] for video object segmentation often rely on
geometry and assume rigid scenes. Since these assumptions are often violated in practice, visually
apparent artifacts are commonly observed. To temporally and spatially smooth object mask estimates,
graphical model based techniques [22, 2, 14, 45, 47, 46] have been proposed in the past. While
graphical models enable an effective label propagation across the entire video sequences, they often
tend to be sensitive to parameters.
Recently, deep learning based approaches [7, 26, 23, 6, 25] have been applied to video object
segmentation. Early work in this direction predicts the segmentation mask frame by frame [7]. Later,
prediction of the current frame incorpoerates additional cues from the preceding frame using optical
flow [23, 26, 25], semantic segmentations [6], or mask propagation [26, 25]. Importantly, all these
methods only address the foreground-background segmentation of a single object and are not directly
applicable to instance level segmentation of multiple objects in videos.
In contrast to the aforementioned methods, in this paper, we develop MaskRNN, a framework that
deals with instance level segmentation of multiple objects in videos. We use a bottom-up approach
where we first track and segment individual objects before merging the results. To capture the
temporal structure, our approach employs a recurrent neural net while the segmentation of individual
objects is based on predictions of binary segmentation masks confined to a predicted bounding box.
We evaluate our approach on the DAVIS-2016 dataset [37], the DAVIS-2017 dataset [39], and the
Segtrack v2 dataset [30]. On all three we observe state-of-the-art performance.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
Table 1: Comparisons with the state-of-the-art deep learning based video object segmentation
algorithms.
Method
Using flow
Temporal information
Location prior
Semantic prior
Post-processing
Finetuning on the 1st frame
2
OSVOS
[7]
MaskProp
[26]
FusionSeg
[23]
LucidTracker
[25]
SemanticProp
[6]
Ours
No
No
No
No
Boundary snapping
Yes
Yes
Short-term
Previous mask
No
denseCRF
Yes
Yes
Short-term
No
No
No
No
Yes
Short-term
Previous mask
No
denseCRF
Yes
No
No
No
Yes
No
Yes
Yes
Long-term (RNN)
Previous mask+Bounding box
No
No
Yes
Related Work
Video object segmentation has been studied extensively in recent years [45, 30, 34, 40, 29, 28, 36, 48,
16, 46, 37, 23, 6, 25]. In the following, we group the literature into two categories: (1) graph-based
approaches and (2) deep learning methods.
Video object segmentation via spatio-temporal graphs: Methods in this category construct a
three-dimensional spatio-temporal graph [45, 30, 16, 28] to model the inter- and the intra-frame
relationship of pixels or superpixels in a video. Evidence about a pixels assignment to the foreground
or background is then propagated along this spatio-temporal graph, to determine which pixels are
to be labeled as foreground and which pixel corresponds to the background of the observed scene.
Graph-based approaches are able to accept different degrees of human supervision. For example,
interactive video object segmentation approaches allow users to annotate the foreground segments
in several key frames to generate accurate results by propagating the user-specified masks to the
entire video [40, 13, 34, 31, 22]. Semi-supervised video object segmentation techniques [4, 16, 45,
22, 46, 33] require only one mask for the first frame of the video. Also, there are unsupervised
methods [9, 28, 50, 36, 35, 12, 48] that do not require manual annotation. Since constructing and
exploring the 3D spatio-temporal graphs is computationally expensive, the graph-based methods are
typically slow, and the running time of the graph-based video object segmentation is often far from
real time.
Video object segmentation via deep learning: With the success of deep nets on semantic segmentation [32, 42], deep learning based approaches for video object segmentation [7, 26, 23, 6, 25]
have been intensively studied recently and often yield state-of-the-art performance, outperforming
graph-based methods. Generally, the employed deep nets are pre-trained on object segmentation
datasets. In the semi-supervised setting where the ground truth mask of the first frame of a video
is given, the network parameters are then finetuned on the given ground truth of the first frame of
a particular video, to improve the results and the specificity of the network. Additionally, contour
cues [7] and semantic segmentation information [7] can be incorporated into the framework. Besides
those cues, optical flow between adjacent frames is another important key information for video
data. Several methods [26, 23, 25] utilize the magnitude of the optical flow between adjacent frames.
However, these methods do not explicitly model the location prior, which is important for object
tracking. In addition, these methods focus on separating foreground from background and do not
consider instance level segmentation of multiple objects in a video sequence.
In Tab. 1, we provide a feature-by-feature comparison of our video object segmentation technique
with representative state-of-the-art approaches. We note that the developed method is the only one
that takes long-term temporal information into account via back-propagation through time using a
recurrent neural net. In addition, the discussed method is the only one that estimates the bounding
boxes in addition to the segmentation masks, allowing us to incorporate a location prior of the tracked
object.
3
Instance Level Video Object Segmentation
Next, we present MaskRNN, a joint multi-object video segmentation technique, which performs
instance level object segmentation by combining binary segmentation with effective object tracking
via bounding boxes. To benefit from temporal dependencies, we employ a recurrent neural net
component to connect prediction over time in a unifying framework. In the following, we first provide
a general outline of the developed approach illustrated in Fig. 1 and detail the individual components
subsequently.
2
Figure 1: An illustration for the proposed algorithm. We show an example video with 2 objects (left).
Our method predicts the binary segmentation for each object using 2 deep nets (Section 3.3), one for
each object, which perform binary segmentation and object localization. The output instance-level
segmentation mask is obtained by combining the binary segmentation masks (Section 3.2).
3.1
Overview
We consider a video sequence I = {I1 , I2 , ..., IT } which consists of T frames It , t ? {1, . . . , T }.
Throughout, we assume the ground truth segmentation mask of the N object instances of interest
to be given for the first frame I1 . We refer to the ground truth segmentation mask of the first frame
via y?1 ? {0, 1, ..., N}H?W , where N is the number of object instances, and H and W are the height
and width of the video frames. In multi-instance video object segmentation, the goal is to predict
y2 , . . . , yT ? {0, . . . , N}H?W , which are the segmentation masks corresponding to frames I2 to IT .
The proposed method is outlined in Fig. 1. Motivated by the time-dependence of the frames in the
video sequence we formulate the task of instance level semantic video segmentation as a recurrent
neural net, where the prediction of the previous frame influences prediction of the current frame.
Beyond the prediction yt?1 for the previous frame t ? 1, our approach also takes into account both the
previous and the current frames, i.e., It?1 and It . We compute the optical flow from the two images.
We then use the predicted optical flow (i) as input feature to the neural nets and (ii) to warp the
previous prediction to roughly align with the current frame.
The warped prediction, the optical flow itself, and the appearance of the current frame are then
used as input for N deep nets, one for each of the N objects. Each of the deep nets consists of two
parts, a binary segmentation net which predicts a segmentation mask, and an object localization net
which performs bounding box regression. The latter is used to alleviate outliers. Both, bounding box
regression and segmentation map are merged into a binary segmentation mask bti ? [0, 1]H?W denoting
the foreground-background probability maps for each of the N object instances i ? {1, . . . , N}. The
binary semantic segmentations for all N objects are subsequently merged using an arg max operation.
The prediction for the current frame, i.e., yt , is computed via thresholding. Note that we combine the
binary predictions only at test time.
In the following, we first describe our fusion operation in detail, before discussing the deep net
performing binary segmentation and object localization.
3.2
Multiple instance level segmentation
Predicting the segmentation mask yt for the t-th frame, can be viewed as a multi-class prediction
problem, i.e., assigning to every pixel in the video a label, indicating whether the pixel p represents an
object instance (ytp = {1, ..., N}) or whether the pixel is considered background (ytp = 0). Following
a recent technique for instance level image segmentation [18], we cast this multi-class prediction
problem into multiple binary segmentations, one per object instance.
Assume availability of binary segmentation masks bti ? [0, 1]H?W which provide for each object
instance i ? {1, . . . , N} the probability that a pixel should be considered foreground or background.
To combine the binary segmentations bti into one final prediction yt such that every pixel is assigned
to only one object label, is achieved by assigning the class with the largest probability for every pixel.
3
Figure 2: An illustration of the binary object segmentation network and the object localization
network as described in Section 3.3. The binary segmentation network is a two-stream network
including appearance stream and flow stream. The inputs of the appearance stream are the current
i ). The inputs of the flow stream are the flow magnitude and the warped mask,
frame It and ?t (bt?1
i
?t (bt?1 ). The object localization net refines the bounding box proposal to estimate the location prior.
To compute the binary segmentation mask bti , the output of appearance stream and the flow stream
are linearly combined and the responses outside the refined bounding box are discarded.
To be more specific, we assign class label i ? {1, . . . , N} to the pixel if the probability for class i at
the pixel (indicated by bti ) is largest among the N probability maps for the N object instances. Note
that this operation is similar to a pooling operation, and permits back-propagation.
3.3
Binary Segmentation
To obtain the binary segmentations bti ? [0, 1]H?W employed in the fusion step, N deep nets are used,
one for each of the N considered object instances. One of the N deep nets is illustrated in Fig. 2. It
consists of two components, the binary segmentation net and the object localization net, which are
discussed in greater detail in the following.
Binary Segmentation Net: The objective for each of the binary segmentation nets is to predict the
foreground-background mask bti ? [0, 1]H?W for its corresponding object instance i ? {1, . . . , N}. To
achieve this task, the binary segmentation net is split into two streams, i.e., the appearance stream
and the flow stream. The input of the appearance stream is the concatenation of the current frame It
and the warped prediction of the previous frame yt?1 , denoted as ?t?1,t (yt?1 ). The warping function
?t?1,t (.) transforms the input based on the optical flow field from frame It?1 to frame It . The input of
the flow stream is the concatenation of the magnitude of the flow field from It to It?1 and It to It+1
and, again, the warped prediction of the previous frame ?t?1,t (yt?1 ). The architecture of both streams
is identical and follows the subsequent description.
The network architecture is inspired by [7] where the bottom of the network follows the structure of
the VGG-16 network [43]. The intermediate representations of the VGG-16 network, right before the
max-pooling layers and after the ReLU layers, are extracted, upsampled by bilinear interpolation and
linearly combined to form a single channel feature representation which has the same size as the input
image. By linearly combining the two representations, one from the appearance stream and the other
one from the flow stream, and by taking the sigmoid function on the combined single channel feature
response, we obtain a probability map which indicates the probability bti ? [0, 1]H?W of a pixel in the
t-th frame being foreground, i.e., corresponding to the i-th object. The network architecture of the
appearance stream is shown in Fig. 2 (right panel). During training, we use the weighted binary cross
entropy loss as suggested in [7].
Note that all the operations in our network are differentiable. Hence, we can train the developed
network end-to-end via back-propagation through time.
4
Table 2: Contribution of different components of our algorithm evaluated on DAVIS-2016 and
DAVIS-2017 dataset. The performance is in term of IoU (%).
Component
Enable (X) / Disable
AStream
X
X
X
X
X
X
Segmentation net FStream
X
X
X
X
X
Warp mask
X
X
X
X
Train
X
X
X
Localization net
Apply
X
X
RNN
X
DAVIS-2016 IoU(%), w/o Finetuning 54.17 55.87 56.88 52.29 53.90 56.32
DAVIS-2016 IoU(%), w/ Finetuning
76.63 79.77 79.92 78.43 80.10 80.38
DAVIS-2017 IoU(%), w/o Finetuning 41.29 43.33 44.52 38.95 41.57 45.53
DAVIS-2017 IoU(%), w/ Finetuning
58.66 59.46 59.71 56.12 60.41 60.51
Object Localization Net: Usage of an object localization net is inspired by tracking approaches
which regularize the prediction by assuming that the object is less likely to move drastically between
temporally adjacent frames. The object localization network computes the location for the i-th object
in the current frame via bounding box regression. First, we find the bounding box proposal on
i ). Similarly to the bounding box regression in Fast-RCNN [15], with the
the warped mask ?t (bt?1
bounding box proposal as the region of interest, we use the conv5_3 feature in the appearance stream
of the segmentation net to perform RoI-pooling, followed by two fully connected layers. Their output
is used to regress the bounding box position. We refer the reader to [15] for more details on bounding
box regression.
Given the bounding box, a pixel is classified as foreground if it is predicted as foreground by the
segmentation net and if it is inside a bounding box which is enlarged by a factor of 1.25 compared
to the predicted of the localization net. The estimated bounding box is then used to restrict the
segmentation to avoid outliers which are far away from the object.
3.4
Training and Finetuning
Our framework outlined in the preceding sections and illustrated in Fig. 1 can be trained end-to-end
via back-propagation through time given a training sequence. Note that back-propagation through
time is used because of the recurrence relation that connects multiple frames of the video sequence.
To further improve the predictive performance, we follow the protocol [39] for the semi-supervised
setting of video object segmentation and finetune our networks using the ground truth segmentation
mask provided for the first frame. Specifically, we further optimize the binary segmentation net
and localization net based on the given ground truth. Note that it is not possible to adjust the entire
architecture since only a single ground truth frame is provided in the supervised setting.
4
Implementation Details
In the following, we describe the implementation details of our approach, as well as the training data.
We also provide details about the offline training and online training in our experimental setup.
Training data: We use the training set of the DAVIS dataset to pre-train the appearance network
for general-purpose object segmentation. The DAVIS-2016 dataset [37] contains 30 training videos
and 20 testing videos and the DAVIS-2017 dataset [39] consists of 60 training videos and 30 testing
videos. Note that the annotation of the DAVIS-2016 dataset contains only one single object per video.
For a fair evaluation on the DAVIS-2016 and DAVIS-2017 datasets, the object segmentation net and
localization nets are trained on the training set of each dataset separately. During testing, the network
is further finetuned online on the given ground-truth of the first frame since we assume the ground
truth segmentation mask of the first frame, i.e., y?1 , to be available.
Offline training: During offline training, we first optimize the networks on static images. We
?
found it useful to randomly perturb the ground-truth segmentation mask yt?1
locally, to simulate
the imperfect prediction of the last frame. The random perturbation includes dilation, deformation,
resizing, rotation and translation. After having trained both the binary segmentation net and the
object localization net on single frames, we further optimize the segmentation net by taking long-term
5
Table 3: The quantitative evaluation on the validation set of DAVIS dataset [37]. The evaluation
matrics are the IoU measurement J , boundary precision F , and time stability T . Following [37],
we also report the recall and the decay of performance over time for J and F measurements.
Semi-supervised
OSVOS MSK VPN OFL BVS FCP JMP HVS SEA TSP OURS
[7]
[26] [24] [46] [33] [38] [13] [16] [1]
[8]
J
Mean M ?
Recall O ?
Decay D ?
79.8
93.6
14.9
79.7
93.1
8.9
70.2
82.3
12.4
68.0
75.6
26.4
60.0
66.9
28.9
58.4
71.5
-2.0
57.0
62.6
39.4
54.6
61.4
23.6
50.4 31.9
53.1 30.0
36.4 38.1
80.4
96.0
4.4
F
Mean M ?
Recall O ?
Decay D ?
80.6
92.6
15.0
75.4
87.1
9.0
65.5
69.0
14.4
63.4
70.4
27.2
58.8
67.9
21.3
49.2
49.5
-1.1
53.1
54.2
38.4
52.9
61.0
22.7
48.0 29.7
46.3 23.0
34.5 35.7
82.3
93.2
8.8
T
Mean M ?
37.8
21.8
32.4
22.2
34.7
30.6
15.9
36.0
15.4 41.7
19.0
information into account, i.e., training using the recurrence relation. We consider 7 frames at a time
due to the memory limitation imposed by the GPU.
During offline training all networks are optimized for 10 epochs using the Adam solver [27] and the
learning rate is gradually decayed during training, starting from 10?5 . Note that we use the pre-trained
flowNet2.0 [19] for optical flow computation. During training, we apply data augmentation with
randomly resizing, rotating, cropping, and left-right flipping the images and masks.
Online finetuning: In the semi-supervised setting of video object segmentation, the ground-truth
segmentation mask of the first frame is available. The object segmentation net and the localization
net are further finetuned on the first frame of the testing video sequence. We set the learning rate
to 10?5 . We train the network for 200 iterations, and the learning rate is gradually decayed over
time. To enrich the variation of the training data, for online finetuning the same data augmentation
techniques are applied as in offline training, namely randomly resizing, rotating, cropping and flipping
the images. Note that the RNN is not employed during online finetuning since only a single frame of
training data is available.
5
Experimental Results
Next, we first describe the evaluation metrics before we present an ablation study of our approach,
quantitative results, and qualitative results.
5.1
Evaluation Metrics
Intersection over union: We use the common mean intersection over union (IoU) metric which
calculates the average across all frames of the dataset. The IoU metric is particularly challenging for
small sized foreground objects.
Contour accuracy [37]: Besides an accurate object overlap measured by IoU, we are also interested
in an accurate delineation of the foreground objects. To assess the delineation quality of our approach,
we measure the precision, P, and the recall R of the two sets of points on the contours of the ground
truth segment and the output segment via a bipartite graph matching. The contour accuracy is
2PR
calculated as P+R
.
Temporal stability [37]: The temporal stability estimates the degree of deformation needed to
transform the segmentation masks from one frame to the next. The temporal stability is measured by
the dissimilarity of the shape context descriptors [3] which describe the points on the contours of the
segmentation between the two adjacent frames.
5.2
Ablation study
We validate the contributions of the components in our method by presenting an ablation study
summarized in Tab. 2 on two datasets, DAVIS-2016 and DAVIS-2017. We mark the enabled
components using the ?X? symbol. We analyze the contribution of the binary segmentation net
6
Table 4: The quantitative evaluation on DAVIS-2017 dataset [39] and SegTrack v2 dataset [30].
DAVIS-2017
SegTrack v2
OSVOS [7] OFL [46] OURS OSVOS [7] MSK [26] OFL [46] OURS
IoU(%)
52.1
54.9
60.5
61.9
67.4
67.5
72.1
including the appearance stream (?AStream?), the flow stream (?FStream?) and whether to warp
the input mask, yt?1 , based on the optical flow field (?Warp mask?). In addition, we analyze the
effects of the object localization net. Specifically, we assess the occurring performance changes of
two configurations: (i) by only adding the bounding box regression loss into the objective function
(?Train?), i.e., both the segmentation net and the object localization net are trained but only the
segmentation net is deployed; (ii) by training and applying the object localization net (?Apply?). The
contribution of the recurrent training (?RNN?) is also illustrated. The performances with and without
online finetuning as described in Section 4 are shown for each dataset as well.
In Tab. 2, we generally observe that online finetuning is important as the network is adjusted to the
specific object appearance in the current video.
For the segmentation net, the combination of the appearance stream and the flow stream performs
better than using only the appearance stream. This is due to the fact that the optical flow magnitude
provided in the flow stream provides complementary information by encoding motion boundaries,
which helps to discover moving objects in the cluttered background. The performance can be
further improved by using the optical flow to warp the mask so that the input to both streams of the
segmentation net also takes the motion into account.
For the localization net, we first show that adding the bounding box regression loss decreases
the performance of the segmentation net (adding ?Train? configuration). However, by applying the
bounding box to restrict the segmentation mask improves the results beyond the performance achieved
by only applying the segmentation net.
Training the network using the recurrence relationship further improves the results as the network
produces more consistent segmentation masks over time.
5.3
Quantitative evaluation
We compare the performance of our approach to several baselines on two tasks: foregroundbackground video object segmentation and multiple instance-level video object segmentation. More
specifically, we use DAVIS-2016 [37] for evaluating foreground-background segmentation, and
DAVIS-2017 [39] and Segtrack v2 [30] datasets for evaluating multiple instance-level segmentation.
The three datasets serve as a good testbed as they contain challenging variations, such as drastic
appearance changes, fast motion, and occlusion. We compare the performance of our approach to
several state-of-the-art benchmarks. We assess performance on the validation set when using the
DAVIS datasets and we use the whole dataset for Segtrack v2 as no split into train and validation
sets is available. The results on DAVIS-2016 are summarized in Tab. 3, where we report the IoU,
the contour accuracy, and the time stability metrics following [37]. The results on DAVIS-2017 and
SegTrack v2 are summarized in Tab. 4.
Foreground-background video object segmentation: We use the DAVIS-2016 dataset to evaluate
the performance of foreground-background video object segmentation. The DAVIS-2016 dataset
contains 30 training videos and 20 validation videos. The network is first trained on the 30 training
videos and finetuned on the first frame of the 20 validation videos, respectively during testing. The
performance evaluation is reported in Tab. 3. We outperform the other state-of-the-art semi-supervised
methods by 0.6%. Note that OSVOS [7], MSK [26], VPN [24] are also deep learning approach. In
contrast to our approach, these methods don?t employ the location prior.
Instance-level video object segmentation: We use the DAVIS-2017 and the Segtrack v2 datasets
to evaluate the performance of instance-level video object segmentation. The DAVIS-2017 dataset
contains 60 training videos and 30 validation videos. The Segtrack v2 dataset contains 14 videos.
There are 2.27 objects per video on average in the DAVIS-2017 dataset and 1.74 in the Segtrack v2
dataset. Again, as for DAVIS-2016, the network is trained on the training set and then finetuned using
the groundtruth of the given first frame. Since the Segtrack v2 dataset does not provide a training
set, we use the DAVIS-2017 training set to optimize and finetune the deep nets. The performance
evaluation is reported in Tab. 4. We outperform other state-of-the-art semi-supervised methods by
5.6% and 4.6% on DAVIS-2017 and Segtrack v2, respectively.
7
Figure 3: Visual results of our approach on DAVIS-2016 (1st and 2nd row), DAVIS-2017 (3rd and
4th row) and Segtrack v2 dataset (5th and 6th row).
Figure 4: Failure cases of our approach. The 1st and the 3rd column shows the results of the beginning
frames. Our method fails to track the object instances as shown in the 2nd and 4th column.
5.4
Qualitative evaluation
We visualize some of the qualitative results of our approach in Fig. 3 and Fig. 4. In Fig. 3, we show
some successful cases of our algorithm on the DAVIS and Segtrack datasets. We observe that the
proposed method accurately keeps track of the foreground objects even with complex motion and
cluttered background. We also observe accurate instance level segmentation of multiple objects which
occlude each other. In Fig. 4, we visualize two failure cases of our approach. Reasons for failures
are the similar appearance of instances of interest as can be observed for the leftmost two figures.
Another reason for failure is large variations in scale and viewpoint as shown for the two figures on
the right of Fig. 4.
6
Conclusion
We proposed MaskRNN, a recurrent neural net based approach for instance-level video object
segmentation. Due to the recurrent component and the combination of segmentation and localization
nets, our approach takes advantage of the long-term temporal information and the location prior to
improve the results.
Acknowledgments: This material is based upon work supported in part by the National Science
Foundation under Grant No. 1718221. We thank NVIDIA for providing the GPUs used in this
research.
8
References
[1] S. Avinash Ramakanth and R. Venkatesh Babu. SeamSeg: Video object segmentation using patch seams.
In Proc. CVPR, 2014. 6
[2] V. Badrinarayanan, F. Galasso, and R. Cipolla. Label propagation in video sequences. In Proc. CVPR,
2010. 1
[3] S. Belongie, J. Malik, and J. Puzicha. Shape matching and object recognition using shape contexts. TPAMI,
2002. 6
[4] W. Brendel and S. Todorovic. Video object segmentation by tracking regions. In Proc. ICCV, 2009. 2
[5] S. Brutzer, B. Hoeferlin, and G. Heidemann. Evaluation of background subtraction techniques for video
surveillance. In Proc. CVPR, 2011. 1
[6] S. Caelles, Y. Chen, J. Pont-Tuset, and L. Van Gool. Semantically-guided video object segmentation. arXiv
preprint arXiv:1704.01926, 2017. 1, 2
[7] S. Caelles, K.-K. Maninis, J. Pont-Tuset, L. Leal-Taix?, D. Cremers, and L. Van Gool. One-shot video
object segmentation. In Proc. CVPR, 2017. 1, 2, 4, 6, 7
[8] J. Chang, D. Wei, and J. W. Fisher. A video representation using temporal superpixels. In Proc. CVPR,
2013. 6
[9] H.-T. Cheng and N. Ahuja. Exploiting nonlocal spatiotemporal structure for video segmentation. In Proc.
CVPR, 2012. 2
[10] A. Criminisi, G. Cross, A. Blake, and V. Kolmogorov. Bilayer segmentation of live video. In Proc. CVPR,
2006. 1
[11] A. Elgammal, R. Duraiswami, D. Harwood, and L. Davis. Background and foreground modeling using
nonparametric kernel density estimation for visual surveillance. In Proc. IEEE, 2002. 1
[12] A. Faktor and M. Irani. Video segmentation by non-local consensus voting. In BMVC, 2014. 2
[13] Q. Fan, F. Zhong, D. Lischinski, D. Cohen-Or, and B. Chen. JumpCut: Non-successive mask transfer and
interpolation for video cutout. ACM TOG (Proc. SIGGRAPH), 2015. 2, 6
[14] F. Galasso, N. Nagaraja, T. Cardenas, T. Brox, and B. Schiele. A unified video segmentation benchmark:
Annotation, metrics and analysis. In Proc. ICCV, 2013. 1
[15] R. Girshick. Fast R-CNN. In Proc. CVPR, 2015. 5
[16] M. Grundmann, V. Kwatra, M. Han, and I. Essa. Efficient hierarchical graph-based video segmentation. In
Proc. CVPR, 2010. 2, 6
[17] E. Haymanand and J. O. Eklundh. Statistical background subtraction for a mobile observer. In Proc. ICCV,
2003. 1
[18] K. He, G. Gkioxari, P. Doll?r, and R. Girshick. Mask r-cnn. In Proc. ICCV, 2017. 3
[19] E. Ilg, N. Mayer, T. Saikia, M. Keuper, A. Dosovitskiy, and T. Brox. Flownet 2.0: Evolution of optical
flow estimation with deep networks. In Proc. CVPR, 2017. 6
[20] M. Irani and P. Anandan. A unified approach to moving object detection in 2d and 3d scenes. PAMI, 1998.
1
[21] M. Irani, B. Rousso, and S. Peleg. Computing occluding and transparent motions. IJCV, 1994. 1
[22] S. D. Jain and K. Grauman. Supervoxel-consistent foreground propagation in video. In Proc. ECCV, 2014.
1, 2
[23] S. D. Jain, B. Xiong, and K. Grauman. FusionSeg: Learning to combine motion and appearance for fully
automatic segmention of generic objects in videos. In Proc. CVPR, 2017. 1, 2
[24] V. Jampani, R. Gadde, and P. V. Gehler. Video propagation networks. In Proc. CVPR, 2017. 6, 7
[25] A. Khoreva, R. Benenson, E. Ilg, T. Brox, and B. Schiele. Lucid data dreaming for object tracking. arXiv
preprint arXiv:1703.09554, 2017. 1, 2
[26] A. Khoreva, F. Perazzi, R. Benenson, B. Schiele, and A.Sorkine-Hornung. Learning video object segmentation from static images. In Proc. CVPR, 2017. 1, 2, 6, 7
[27] D. Kingma and J. Ba. Adam: A method for stochastic optimization. In Proc. ICLR, 2014. 6
[28] Y. J. Lee, J. Kim, and K. Grauman. Key-segments for video object segmentation. In Proc. ICCV, 2011. 2
[29] J. Lezama, K. Alahari, J. Sivic, and I. Laptev. Track to the future: Spatio-temporal video segmentation
with long-range motion cues. In Proc. CVPR, 2011. 2
[30] F. Li, T. Kim, A. Humayun, D. Tsai, and J. Rehg. Video segmentation by tracking many figure- ground
segments. In Proc. ICCV, 2013. 1, 2, 7
[31] W. Li, F. Viola, J. Starck, G. J. Brostow, and N. D. Campbell. Roto++: Accelerating professional
rotoscoping using shape manifolds. ACM TOG (Proc. SIGGRAPH), 2016. 2
[32] J. Long, E. Shelhamer, and T. Darrell. Fully convolutional networks for semantic segmentation. In Proc.
CVPR, 2015. 2
[33] N. Maerki, F. Perazzi, O. Wang, and A. Sorkine-Hornung. Bilateral space video segmentation. In Proc.
CVPR, 2016. 2, 6
[34] N. Nagaraja, F. Schmidt, and T. Brox. Video segmentation with just a few strokes. In Proc. ICCV, 2015. 2
[35] P. Ochs, J. Malik, and T. Brox. Segmentation of moving objects by long term video analysis. TPAMI, 2014.
2
9
[36] A. Papazoglou and V. Ferrari. Fast object segmentation in unconstrained video. In Proc. ICCV, 2013. 2
[37] F. Perazzi, J. Pont-Tuset, B. McWilliams, L. V. Gool, M. Gross, and A. Sorkine-Hornung. A benchmark
dataset and evaluation methodology for video object segmentation. In Proc. CVPR, 2016. 1, 2, 5, 6, 7
[38] F. Perazzi, O. Wang, M. Gross, and A. Sorkine-Hornung. Fully connected object proposals for video
segmentation. In Proc. ICCV, 2015. 6
[39] J. Pont-Tuset, F. Perazzi, S. Caelles, P. Arbel?ez, A. Sorkine-Hornung, and L. Van Gool. The 2017 davis
challenge on video object segmentation. arXiv preprint arXiv:1704.00675, 2017. 1, 5, 7
[40] B. L. Price, B. S. Morse, and S. Cohen. LIVEcut: Learning-based interactive video segmentation by
evaluation of multiple propagated cues. In Proc. ICCV, 2009. 2
[41] Y. Ren, C. S. Chua, and Y. K. Ho. Statistical background modeling for non-stationary camera. PRL, 2003.
1
[42] A. G. Schwing and R. Urtasun.
Fully Connected Deep Structured Networks.
In
https://arxiv.org/abs/1503.02351, 2015. 2
[43] K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition. In
Proc. ICLR, 2015. 4
[44] P. H. S. Torr and A. Zisserman. Concerning bayesian motion segmentation, model averaging, matching
and the trifocal tensor. In Proc. ECCV, 1998. 1
[45] D. Tsai, M. Flagg, and J. Rehg. Motion coherent tracking with multi-label mrf optimization. In Proc.
BMVC, 2010. 1, 2
[46] Y.-H. Tsai, M.-H. Yang, and M. J. Black. Video Segmentation via Object Flow. In Proc. CVPR, 2016. 1, 2,
6, 7
[47] S. Vijayanarasimhan and K. Grauman. Active frame selection for label propagation in videos. In Proc.
ECCV, 2012. 1
[48] F. Xiao and Y. J. Lee. Track and segment: An iterative unsupervised approach for video object proposals.
In Proc. CVPR, 2016. 2
[49] C. Yuan, G. Medioni, J. Kang, and I. Cohen. Detecting motion regions in the presence of a strong parallax
from a moving camera by multiview geometric constraints. PAMI, 2007. 1
[50] D. Zhang, O. Javed, and M. Shah. Video object segmentation through spatially accurate and temporally
dense extraction of primary object regions. In Proc. CVPR, 2013. 2
10
| 6636 |@word cnn:2 compression:2 nd:2 hu:1 shot:1 configuration:2 contains:5 denoting:1 ours:4 past:1 current:10 assigning:2 gpu:1 refines:1 subsequent:1 shape:5 hvs:1 occlude:1 stationary:1 cue:5 beginning:1 short:3 chua:1 provides:1 detecting:1 location:7 successive:1 org:1 zhang:1 height:1 along:1 brostow:1 yuan:2 consists:4 qualitative:3 ijcv:1 combine:3 inside:1 parallax:1 inter:1 mask:39 roughly:1 uiuc:2 multi:5 inspired:2 delineation:2 solver:1 provided:3 discover:1 panel:1 jmp:1 developed:3 unified:2 keuper:1 temporal:16 quantitative:4 every:3 voting:1 interactive:2 grauman:4 mcwilliams:1 grant:1 segmenting:1 before:4 local:1 bilinear:1 encoding:1 interpolation:2 pami:2 black:1 studied:2 challenging:4 aschwing:1 range:1 acknowledgment:1 camera:2 testing:5 practice:1 union:2 area:1 rnn:4 matching:3 pre:3 specificity:1 upsampled:1 selection:1 context:2 influence:1 applying:3 live:1 vijayanarasimhan:1 optimize:4 gkioxari:1 map:4 imposed:1 yt:10 attention:1 starting:1 cluttered:2 formulate:1 importantly:1 regularize:1 rehg:2 enabled:1 stability:5 cutout:1 ferrari:1 variation:3 user:2 jampani:1 expensive:1 finetuned:5 particularly:1 recognition:2 ochs:1 predicts:3 labeled:1 gehler:1 observed:3 bottom:2 preprint:3 wang:2 capture:2 region:4 connected:3 movement:1 decrease:1 gross:2 harwood:1 schiele:3 lezama:1 trained:8 segment:7 laptev:1 predictive:1 serve:1 localization:22 bipartite:1 upon:1 tog:2 finetuning:11 joint:1 siggraph:2 kolmogorov:1 train:7 dreaming:1 jain:2 fast:5 effective:2 describe:4 outside:1 refined:1 apparent:1 cvpr:20 hornung:5 resizing:3 cardenas:1 simonyan:1 galasso:2 transform:1 itself:1 tsp:1 final:1 online:7 advantage:2 sequence:8 differentiable:1 net:55 tpami:2 essa:1 arbel:1 combining:3 ablation:3 achieve:1 description:1 validate:2 exploiting:1 darrell:1 sea:1 cropping:2 produce:1 adam:2 object:108 help:1 develop:2 recurrent:9 pose:1 propagating:1 measured:2 strong:1 predicted:4 peleg:1 msk:3 iou:11 direction:1 guided:1 merged:2 subsequently:2 criminisi:1 stochastic:1 human:1 enable:2 material:1 bin:1 require:2 assign:1 transparent:1 alleviate:1 adjusted:1 exploring:1 considered:3 ground:13 roi:1 visually:1 bvs:1 blake:1 lischinski:1 predict:2 visualize:2 early:1 purpose:1 estimation:2 proc:39 applicable:1 label:7 sensitive:1 ilg:2 largest:2 weighted:1 avoid:1 zhong:1 surveillance:2 mobile:1 release:1 focus:1 indicates:1 superpixels:2 tech:1 contrast:2 baseline:1 kim:2 rigid:1 entire:3 typically:1 accept:1 bt:3 relation:2 pont:4 i1:2 interested:1 pixel:14 arg:1 aforementioned:1 among:1 denoted:1 enrich:1 art:8 brox:5 field:3 construct:1 having:1 beach:1 extraction:1 identical:1 represents:1 unsupervised:2 foreground:20 future:1 report:2 dosovitskiy:1 employ:3 few:1 randomly:3 national:1 individual:3 geometry:1 connects:1 occlusion:1 ab:1 detection:1 interest:3 intra:1 evaluation:13 adjust:1 accurate:5 rotating:2 deformation:2 girshick:2 lucid:1 instance:33 column:2 modeling:2 ytp:2 assignment:1 successful:1 virginia:1 reported:2 dependency:1 connect:1 spatiotemporal:1 combined:3 st:4 decayed:2 density:1 lee:2 again:2 augmentation:2 flownet:1 huang:1 warped:5 li:2 account:4 summarized:3 availability:1 includes:1 babu:1 vpn:2 cremers:1 explicitly:1 stream:24 later:1 bilateral:1 observer:1 tab:7 analyze:2 annotation:3 jia:1 nagaraja:2 contribution:4 ass:3 brendel:1 accuracy:3 convolutional:2 descriptor:1 yield:1 yes:10 identification:1 bayesian:1 rejecting:1 accurately:1 ren:1 classified:1 stroke:1 manual:1 failure:4 regress:1 static:2 propagated:2 dataset:28 intensively:1 recall:4 improves:2 sorkine:5 segmentation:126 back:5 campbell:1 finetune:2 supervised:8 follow:1 methodology:1 seam:1 response:2 improved:1 wei:1 editing:2 duraiswami:1 evaluated:1 box:21 bmvc:2 alahari:1 zisserman:2 just:2 propagation:11 artifact:1 indicated:1 quality:1 usa:1 usage:1 effect:1 contain:1 y2:1 avinash:1 evolution:1 hence:1 assigned:1 spatially:2 irani:3 semantic:7 illustrated:4 deal:1 i2:2 adjacent:4 during:8 width:1 recurrence:3 maninis:1 davis:40 leftmost:1 presenting:1 outline:1 multiview:1 starck:1 performs:3 motion:10 fusionseg:2 image:8 recently:2 sigmoid:1 rotation:1 common:1 tracked:1 overview:1 cohen:3 discussed:2 he:1 significant:1 refer:2 measurement:2 rd:2 automatic:1 outlined:2 unconstrained:1 similarly:1 illinois:2 moving:4 han:1 supervision:1 bti:8 align:1 recent:3 supervoxel:1 nvidia:1 binary:27 outperforming:1 success:1 vt:1 discussing:1 additional:1 greater:1 preceding:2 disable:1 employed:3 anandan:1 subtraction:2 determine:1 semi:7 ii:2 multiple:12 smooth:1 cross:2 long:9 concerning:1 post:1 calculates:1 prediction:17 mrf:1 regression:7 metric:6 arxiv:7 annotate:1 iteration:1 ofl:3 kernel:1 confined:1 achieved:2 proposal:5 background:18 addition:4 separately:1 heidemann:1 benenson:2 humayun:1 pooling:3 tend:1 flow:25 yang:1 prl:1 intermediate:1 split:2 presence:1 relu:1 architecture:4 restrict:2 elgammal:1 imperfect:1 faktor:1 vgg:2 whether:3 motivated:1 accelerating:1 todorovic:1 deep:19 generally:2 useful:1 transforms:1 nonparametric:1 extensively:1 locally:1 category:2 generate:1 http:1 outperform:2 estimated:1 track:5 per:3 group:1 key:3 achieving:1 utilize:1 fuse:1 graph:11 flownet2:1 year:1 throughout:1 reader:1 groundtruth:1 patch:1 coherence:1 layer:3 followed:1 cheng:1 fan:1 constraint:1 scene:4 simulate:1 performing:1 optical:12 saikia:1 gpus:1 structured:1 combination:2 across:2 outlier:3 gradually:2 pr:1 iccv:10 computationally:1 needed:1 drastic:1 end:4 available:4 operation:5 permit:1 doll:1 apply:3 observe:4 hierarchical:1 v2:13 away:1 generic:1 xiong:1 schmidt:1 professional:1 ho:1 shah:1 running:1 graphical:2 unifying:1 matric:1 ting:1 perturb:1 classical:2 warping:1 objective:2 move:1 malik:2 tensor:1 flipping:2 primary:1 dependence:1 iclr:2 thank:1 separating:1 concatenation:2 perazzi:5 manifold:1 consensus:1 urtasun:1 reason:2 assuming:1 besides:2 relationship:2 illustration:2 providing:3 setup:1 conv5_3:1 ba:1 implementation:2 perform:2 allowing:1 javed:1 datasets:9 discarded:1 benchmark:4 viola:1 incorporated:1 frame:54 perturbation:1 venkatesh:1 cast:1 namely:1 specified:1 optimized:1 mayer:1 sivic:1 coherent:1 testbed:1 kang:1 kingma:1 nip:1 address:1 able:2 beyond:2 suggested:1 challenge:2 max:2 including:2 video:96 memory:1 gool:4 medioni:1 overlap:1 rely:1 predicting:1 improve:3 badrinarayanan:1 temporally:3 prior:7 literature:1 epoch:1 geometric:1 morse:1 loss:3 fully:5 limitation:1 validation:6 rcnn:1 foundation:1 shelhamer:1 degree:2 consistent:2 xiao:1 thresholding:1 viewpoint:1 translation:1 row:3 eccv:3 supported:1 last:1 drastically:1 offline:5 allow:1 warp:5 taking:2 benefit:1 van:3 boundary:3 calculated:1 leal:1 evaluating:2 tuset:4 contour:6 computes:1 commonly:1 far:2 nonlocal:1 eklundh:1 keep:1 active:1 belongie:1 spatio:5 don:1 iterative:1 dilation:1 table:4 additionally:1 roto:1 channel:2 transfer:1 ca:1 fcp:1 complex:2 constructing:1 protocol:1 dense:1 linearly:3 bounding:21 whole:1 fair:1 complementary:1 enlarged:1 fig:10 representative:1 deployed:1 slow:1 ahuja:1 precision:2 fails:1 position:1 densecrf:2 specific:2 symbol:1 decay:3 evidence:1 fusion:2 merging:1 adding:3 gained:1 magnitude:4 dissimilarity:1 occurring:1 chen:2 entropy:1 intersection:2 appearance:17 likely:1 gadde:1 ez:1 visual:2 tracking:7 chang:1 cipolla:1 kwatra:1 corresponds:1 truth:12 extracted:1 acm:2 goal:1 viewed:1 sized:1 price:1 fisher:1 considerable:1 change:2 specifically:3 torr:1 semantically:1 averaging:1 schwing:2 experimental:2 deforming:1 occluding:2 indicating:1 puzicha:1 mark:1 latter:1 alexander:1 violated:1 tsai:3 incorporate:1 evaluate:3 |
6,231 | 6,637 | Gated Recurrent Convolution Neural Network for
OCR
Jianfeng Wang?
Beijing University of Posts and Telecommunications
Beijing 100876, China
[email protected]
Xiaolin Hu
Tsinghua National Laboratory for Information Science and Technology (TNList)
Department of Computer Science and Technology
Center for Brain-Inspired Computing Research (CBICR)
Tsinghua University, Beijing 100084, China
[email protected]
Abstract
Optical Character Recognition (OCR) aims to recognize text in natural images.
Inspired by a recently proposed model for general image classification, Recurrent
Convolution Neural Network (RCNN), we propose a new architecture named
Gated RCNN (GRCNN) for solving this problem. Its critical component, Gated
Recurrent Convolution Layer (GRCL), is constructed by adding a gate to the
Recurrent Convolution Layer (RCL), the critical component of RCNN. The gate
controls the context modulation in RCL and balances the feed-forward information
and the recurrent information. In addition, an efficient Bidirectional Long ShortTerm Memory (BLSTM) is built for sequence modeling. The GRCNN is combined
with BLSTM to recognize text in natural images. The entire GRCNN-BLSTM
model can be trained end-to-end. Experiments show that the proposed model
outperforms existing methods on several benchmark datasets including the IIIT-5K,
Street View Text (SVT) and ICDAR.
1
Introduction
Reading text in scene images can be regarded as an image sequence recognition task. It is an important
problem which has drawn much attention in the past decades. There are two types of scene text
recognition tasks: constrained and unconstrained. In constrained text recognition, there is a fixed
lexicon or dictionary with known length during inference. In unconstrained text recognition, each
word is recognized without a dictionary. Most of the previous works are about the first task.
In recent years, deep neural networks have gained great success in many computer vision tasks
[39, 19, 34, 42, 9, 11]. The fast development of deep neural networks inspires researchers to use
them to solve the problem of scene text recognition. For example, an end-to-end architecture which
combines a convolutional network with a recurrent network is proposed [30]. In this framework, a
plain seven-layer-CNN is used as a feature extractor for the input image while a recurrent neural
network (RNN) is used for image sequence modeling. For another example, to let the recurrent
network focus on the most important segments of incoming features, an end-to-end system which
integrates attention mechanism, recurrent network and recursive CNN is developed [21].
?
This work was done when Jianfeng Wang was an intern at Tsinghua University.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
Figure 1: Illustration of using RCL with T = 2 for OCR.
A Recurrent Convolution Neural Network (RCNN) is proposed for general image classification
[22], which simulates an anatomical fact that recurrent connections are ubiquitously existent in the
neocortex. It is believed that the recurrent synapses that exist in the neocortex play an important role
in context modulation during visual recognition. A feed-forward model can only capture the context
in higher layers where units have larger receptive fields, but this information cannot modulate the
units in lower layers which is responsible for recognizing smaller objects. Hence, using recurrent
connections within a convolutional layer (called Recurrent Convolution Layer or RCL) can bring
context information to all units in this layer. This implements a nonclassical receptive field [12, 18]
of a biological neuron as the effective receptive field is larger than that determined by feedforward
connections. However, with increasing iterations, the size of the effective receptive field will increase
unboundedly, which contradicts the biological fact. One needs a mechanism to constrain the growth
of the effective receptive field.
In addition, from the viewpoint of performance enhancing, one also needs to control the context
modulation of neurons in RCNN. For example, in Figure 1, it is seen that for recognizing a character,
not all of the context are useful. When the network recognizes the character "h", the recurrent
kernel which covers the other parts of the character is beneficial. However, when the recurrent
kernel is enlarged to the parts of other characters, such as "p", the context carried by the kernel is
unnecessary. Therefore, we need to weaken the signal that comes from unrelated context and combine
the feed-forward information with recurrent information in a flexible way.
To achieve the above goals, we introduce a gate to control the context modulation in each RCL layer,
which leads to the Gated Recurrent Convolution Neural Network (GRCNN).
In addition, the recurrent neural network is adopted in the recognition of words in natural images
since it is good at sequence modeling. In this work, we choose the Long Short Term Memory (LSTM)
as the top layer of the proposed model, which is trained in an end-to-end fashion.
2
Related Work
OCR is one of the most important challenges in computer vision and many methods have been
proposed for this task. The word locations and scores of detected characters are input into the
Pictorial Structure formulation to acquire an optimal configuration of a particular word [26]. A
complete OCR system that contains text detection as well as text recognition is designed, and it can
be well applied to both unconstrained and constrained text recognition task [3].
The OCR can be understood as a classification task by treating each word in the lexicon as an
object category [37]. Another classical method, the Conditional Random Fields, is proposed in text
recognition [17, 25, 31]. Besides those conventional methods, some deep network-based methods are
also proposed. The Convolutional Neural Network (CNN) is used to extract shared features, which
are fed into a character classifier [16]. To further improve the performance, more than one objective
functions are used in a CNN-based OCR system [15]. A CNN is combined with the Conditional
Random Field graphical model, which can be jointly optimized through back-propagation [14].
Moreover, some works introduced RNN, such as LSTM, to recognize constrained and unconstrained
words [30]. The attention mechanism is applied to a stacked RNN on the top of the recursive CNN
[21].
2
Figure 2: Illustration of GRCL with T = 2. The convolutional kernels in the same color use the same
weights.
Many new deep neural networks for general image classification have been proposed in these years.
The closely related model to the proposed model in this paper is RCNN [22], which is inspired by
the observation of abundant recurrent synapses in the brain. It adds recurrent connections within the
standard convolutional layers, and the recurrent connections improve the network capacity for object
recognition. When unfolded in time, it becomes a CNN with many shortcuts from the bottom layer
to upper layers. RCNN has been used to solve other problems such as scene labelling [23], action
recognition [35], and speech processing [43]. One related model to RCNN is the recursive neural
network [32], in which a recursive layer is unfolded into a stack of layers with tied weights. It can be
regarded as an RCNN without shortcuts. Another related model is the DenseNet [11], in which every
lower layer is connected to the upper layers. It can be regarded as an RCNN with more shortcuts and
without weight sharing. The idea of shortcut have also been explored in the residual network [9].
3
3.1
GRCNN-BLSTM Model
Recurrent Convolution Layer and RCNN
The RCL [22] is a module with recurrent connections in the convolutional layer of CNN. Consider a
generic RNN model with feed-forward input u(t). The internal state x(t) can be defined as:
x(t) = F(u(t), x(t ? 1), ?)
(1)
where the function F describes the nonlinearity of RNN (e.g. ReLU unit) and ? is the parameter. The
state of RCL evolves over discrete time steps:
x(t) = F((wf ? u(t) + wr ? x(t ? 1))
(2)
where "*" denotes convolution, u(t) and x(t ? 1) denote the feed-forward input and recurrent input
respectively, wf and wr denote the feed-forward weights and recurrent weights respectively.
Multiple RCLs together with other types of layers can be stacked into a deep model. A CNN that
contains RCL is called RCNN [22].
3
Figure 3: Overall pipeline of the architecture.
3.2
Gated Recurrent Convolution Layer and GRCNN
The Gated Recurrent Convolution Layer (GRCL) is the essential module in our framework. This
module is equipped with a gate to control the context modulation in RCL and it can weaken or even
cut off irrelevant context information. The gate of GRCL can be written as follows:
(
G(t) =
0
sigmoid(BN (wgf
? u(t)) +
t=0
BN (wgr
? x(t ? 1)))
(3)
t>0
Inspired by the Gated Recurrent Unit (GRU) [4], we let the controlling gate receive the signals from
the feed-forward input as well as the states at the last time step. We use two 1 ? 1 kernels, wgf
and wgr , to convolve with the feed-forward input and recurrent input separately. wgf denotes the
feed-forward weights for the gate and wgr denotes the recurrent weights for the gate. The recurrent
weights are shared over all time steps (Figure 2). Batch normalization (BN) [13] is used to improve
the performance and accelerate convergence. The GRCL can be described by:
(
x(t) =
ReLU (BN (wf ? u(t))
t=0
ReLU (BN (wf ? u(t)) + BN (BN (wr ? x(t ? 1)) G(t)))
t>0
(4)
In the equations, "" denotes element-wise multiplication. Batch normalization (BN) is applied after
each convolution and element-wise multiplication operation. The parameters and statistics in BN
are not shared over different time steps, though the recurrent weights are shared. It is assumed that
the input to GRCL is the same over time t, which is denoted by u(0). It is the output of the previous
layer. This assumption means that the feed-forward part contributes equally at each time step. It
is important to clarify that the time step in GRCL is not identical to the time associated with the
sequential data. The time steps denote the iterations in processing the input.
Figure 2 shows the diagram of the GRCL with T = 2. When t = 0, only the feed-forward
computation takes place. At t = 1, the gate?s output, which is determined by the feed-forward input
and the states at the previous time step (t = 0), acts on the recurrent component. It can modulate
the recurrent signals. When the output of the gate is 1, it becomes the standard RCL. When the
output of the gate is 0, the recurrent signal is dropped and it becomes the standard convolutional layer.
Therefore, the GRCL is a generalization of RCL and it can adjust context modulation dynamically.
The effective receptive field (RF) of each GRCL unit in the previous layer?s feature maps expands
while the iteration number increases. However, unlike the RCL, some regions that contain unrelated
information in large effective RF cannot provide strong signal to the center of RF. This mimics the
fact that human eyes only care about the context information that helps the recognition of the objects.
Multiple GRCLs together with other types of layers can be stacked into a deep model. Hereafter, a
CNN that contains GRCL is called a GRCNN.
3.3
Overall Architecture
The architecture consists of three parts: feature sequence extraction, sequence modelling, and
transcription. Figure 3 shows the overall pipeline which can be trained end-to-end.
Feature Sequence Extraction: We use the GRCNN in the first part and there are no fully-connected
layers. The input to the network is a whole image and the image is resized to fixed length and height.
4
Table 1: The GRCNN configuration
Conv MaxPool GRCL MaxPool GRCL MaxPool GRCL MaxPool Conv
3?3
2?2
3?3
2?2
3?3
2?2
3?3
2?2
2?2
num: 64
num: 64
num: 128
num: 256
num: 512
sh:1 sw:1 sh:2 sw:2 sh:1 sw:1 sh:2 sw:2 sh:1 sw:1 sh:2 sw:1 sh:1 sw:1 sh:2 sw:1 sh:1 sw:1
ph:1 pw:1ph:0 pw:0ph:1 pw:1ph:0 pw:0ph:1 pw:1ph:0 pw:1ph:1 pw:1ph:0 pw:1ph:0 pw:0
Specifically, the feature map in the last layer is sliced from left to right by column to form a feature
sequence. Therefore, the i-th feature vector is formed by concatenating the i-th columns of all of the
maps. We add some max pooling layers to the network in order to ensure the width of each column is
1. Each feature vector in the sequence represents a rectangle region of the input image, and it can be
regarded as the image descriptor for that region. Comparing the GRCL with the basic convolutional
layer, we find that each feature vector generated by the GRCL represents a larger region. This feature
vector contains more information than the feature vector generated by the basic convolutional layer,
and it is beneficial for the recognition of text.
Sequence Modeling: An LSTM [10] is used on the top of feature extraction module for sequence
modeling. The peephole LSTM is first proposed in [5], whose gates not only receive the inputs from
the previous layer, but also from the cell state. We add the peephole LSTM to the top of GRCNN and
investigate the effect of peephole connections to the whole network?s performance. The inputs to the
gates of LSTM can be written as:
i = ?(Wxi xt + Whi ht?1 + ?1 Wci ct?1 + bi ),
(5)
f = ?(Wxf xt + Whf ht?1 + ?2 Wcf ct?1 + bf ),
(6)
o = ?(Wxo xt + Who ht?1 + ?3 Wco ct + bo ),
(7)
?i ? {0, 1}.
(8)
?i is defined as an indication factor and its value is 0 or 1. When ?i is equal to 1, the gate receives the
modulation of the cell?s state.
However, LSTM only considers past events. In fact, the context information from both directions are
often complementary. Therefore, we use stacked bidirectional LSTM [29] in our architecture.
Transcription: The last part is transcription which converts per-frame predictions to real labels. The
Connectionist Temporal Classification (CTC) [8] method is used. Denote the dataset by S = {(I, z)},
where I is a training image and z is the corresponding ground truth label sequence. The objective
function to be minimized is defined as follows:X
O=?
logp(z|I).
(9)
(I,z)?S
Given an input image I, the prediction of RNN at each time step is denoted by ?t . The sequence
? may contain blanks and repeated labels (e.g. (a ? b ? ?b)=(ab ? ?bb)) and we need a concise
representation l (the two examples are both reduced to abb). Define a function ? which maps ? to l
by removing blanks and repeated labels. Then
X
p(l|I) =
logp(?|I),
(10)
?:?(?)=l
p(?|I) =
T
Y
y?t t ,
(11)
t=1
where y?t t denotes the probability of generating label ?t at time step t.
After training, for lexicon-free transcription, the predicted label sequence for a test image I is obtained
by [30]:
l? = ?(arg max p(?|I)).
(12)
?
The lexicon-based method needs a dictionary or lexicon. Each test image is associated with a fix
length lexicon D. The result is obtained by choosing the sequence in the lexicon that has highest
conditional probability [30]:
l? = arg max p(l|I).
(13)
l?D
5
Table 2: Model analysis over the IIIT5K and SVT (%). Mean and standard deviation of the results
are reported.
(a) GRCNN analysis
Model
IIIT5K
SVT
Plain CNN
77.21?0.54 77.69?0.59
RCNN(1 iter) 77.64?0.58 78.23?0.56
RCNN(2 iters) 78.17?0.56 79.11?0.63
RCNN(3 iters) 78.94?0.61 79.76?0.59
GRCNN(1 iter) 77.92?0.57 78.67?0.53
GRCNN(2 iters) 79.42?0.63 79.89?0.64
GRCNN(3 iters) 80.21?0.57 80.98?0.60
4
4.1
(b) LSTM?s variants analysis
LSTM variants
IIIT5K
SVT
LSTM{?1 =0,?2 =0,?3 =0} 77.92?0.57 78.67?0.53
LSTM-F{?1 =0,?2 =1,?3 =0} 77.26?0.61 78.23?0.53
LSTM-I{?1 =1,?2 =0,?3 =0} 76.84?0.58 76.89?0.63
LSTM-O{?1 =0,?2 =0,?3 =1} 76.91?0.64 78.65?0.56
LSTM-A{?1 =1,?2 =1,?3 =1} 76.52?0.66 77.88?0.59
Experiments
Datasets
ICDAR2003: ICDAR2003 [24] contains 251 scene images and there are 860 cropped images of the
words. We perform unconstrained text recognition and constrained text recognition on this dataset.
Each image is associated with a 50-word lexicon defined by wang et al. [36]. The full lexicon is
composed of all per-image lexicons.
IIIT5K: This dataset has 3000 cropped testing word images and 2000 cropped training images
collected from the Internet [31]. Each image has a lexicon of 50 words and a lexicon of 1000 words.
Street View Text (SVT): This dataset has 647 cropped word images from Google Street View [36].
We use the 50-word lexicon defined by Wang et al [36] in our experiment.
Synth90k: This dataset contains around 7 million training images, 800k validation images and 900k
test images [15]. All of the word images are generated by a synthetic text engine and are highly
realistic.
When evaluating the performance of our model on those benchmark dataset, we follow the evaluation
protocol in [36]. We perform recognition on the words that contain only alphanumeric characters
(A-Z and 0-9) and at least three characters. All recognition results are case-insensitive.
4.2
Implementation Details
The configuration of the network is listed in Table 1, where "sh" denotes the stride of the kernel along
the height; "sw" denotes the stride along the width; "ph" and "pw" denote the padding value of height
and width respectively; and "num" denotes the number of feature maps. The input is a gray-scale
image which is resized to 100?32. Before input to the network, the pixel values are rescaled to
the range (-1, 1). The final output of the feature extractor is a feature sequence of 26 frames. The
recurrent layer is a bidirectional LSTM with 512 units without dropout. The ADADELTA method
[41] is used for training with the parameter ?=0.9. The batch size is set to 192 and training is stopped
after 300k iterations.
All of the networks and related LSTM variants are trained on the training set of Synth90k. The
validation set of Synth90k is used for model selection. When a model is selected in this way, its
parameters are fixed and it is directly tested on other datasets (ICDAR2003, IIIT5K and SVT datasets)
without finetuning. The code and pre-trained model will be released at https://github.com/
Jianfeng1991/GRCNN-for-OCR.
4.3
Explorative Study
We empirically analyze the performance of the proposed model. The results are listed in Table 2. To
ensure robust comparison, for each configuration, after convergence during training, a different model
is saved at every 3000 iterations. We select ten models which perform the best on the Synth90k?s
validation set, and report the mean accuracy as well as the standard deviation on each tested dataset.
6
Table 3: The text recognition accuracies in natural images. "50","1k" and "Full" denote the lexicon
size used for lexicon-based recognition task. The dataset without lexicon size means the unconstrained
text recognition
Method
ABBYY [36]
wang et al. [36]
Mishra et al. [25]
Novikova et al. [27]
wang et al. [38]
Bissacco et al. [3]
Goel et al. [6]
Alsharif [2]
Almazan et al. [1]
Lee et al. [20]
Yao et al. [40]
Rodriguez et al. [28]
Jaderberg et al. [16]
Su and Lu et al. [33]
Gordo [7]
Jaderberg et al. [14]
Baoguang et al. [30]
Chen-Yu et al. [21]
ResNet-BLSTM
Ours
SVT-50
35.0%
57.0%
73.2%
72.9%
70.0%
90.4%
77.3%
74.3%
89.2%
80.0%
75.9%
70.0%
86.1%
83.0%
90.7%
93.2%
96.4%
96.3%
96.0%
96.3%
SVT IIIT5K-50 IIIT5K-1k IIIT5K IC03-50 IC03-Full IC03
24.3%
56.0%
55.0%
76.0%
62.0%
81.8%
67.8%
64.1%
57.5%
82.8%
90.0%
84.0%
78.0%
89.7%
93.1%
88.6%
91.2%
82.1%
88.0%
76.0%
80.2%
69.3%
88.5%
80.3%
76.1%
57.4%
96.2%
91.5%
92.0%
82.0%
93.3%
86.6%
71.1%
95.5%
89.6%
97.8%
97.0% 89.6%
80.8%
97.6%
94.4%
78.2% 98.7%
97.6% 89.4%
80.7%
96.8%
94.4%
78.4% 97.9%
97.0% 88.7%
80.2%
97.5%
94.9%
79.2% 98.1%
97.3% 89.9%
81.5% 98.0%
95.6% 80.8% 98.8%
97.8% 91.2%
First, a purely feed-forward CNN is constructed for comparison. To make this CNN have approximately the same number of parameters as GRCNN and RCNN, we use two convolutional layers
to replace each GRCL in Table 1, and each of them has the same number of feature maps as the
corresponding GRCL. Besides, this plain CNN has the same depth as the GRCNN with T = 1. The
results show that the plain CNN has lower accuracy than both RCNN and GRCNN.
Second, we compare GRCNN and RCNN to investigate the effect of adding gate to the RCL.
RCNN is constructed by replacing GRCL in Table 1 with RCL. Each RCL in RCNN has the same
number of feature maps as the corresponding GRCL. Batch normalization is also inserted after each
convolutional kernel in RCL. We fix T , and compare these two models. The results in Table 2(a)
show that each GRCNN model outperforms the corresponding RCNN on both IIIT5K and SVT.
Those results show the advantage of the introduced gate in the model.
Furthermore, we explore the effect of iterations in GRCL. From Table 2(a), we can conclude that
having more iterations is beneficial to GRCL. The increments of accuracy between each iteration
number are 1.50%, 0.79% on IIIT5K and 1.22%, 1.09% on SVT, respectively. This is reasonable
since GRCL with more iterations is able to receive more context information.
Finally, we compare various peephole LSTM units in the bidirectional LSTM for processing feature
sequences. Five types of LSTM variants are compared: full peephole LSTM (LSTM-A), input gate
peephole LSTM (LSTM-I), output gate peephole LSTM (LSTM-O), forget gate peephole LSTM
(LSTM-F) and none peephole LSTM (LSTM). In the feature extraction part, we use GRCNN with
T = 1 which is described in Table 1. Table 2(b) shows that the LSTM without peephole connections
(?1 = ?2 = ?3 = 0) gives the best result.
4.4
Comparison with the state-of-the-art
We use the GRCNN described in Table 1 as the feature extractor. Since having more iterations
is helpful to GRCL, we use GRCL with T = 5 in GRCNN. For sequence learning, we use the
bidirectional LSTM without peephole connections. The training details are described in Sec.4.2.
The best model on the validation set of Synth90k is selected for comparison. Note that this model
is trained on the training set of Synth90k and not finetuned with respect to any dataset. Table 3
shows the results. The proposed method outperforms most existing models for both constrained and
unconstrained text recognition.
7
Figure 4: Lexicon-free recognition results by the proposed GRCL-BLSTM framework on SVT,
ICDAR03 and IIIT5K
Moreover, we do an extra experiment by building a residual block in feature extraction part. We
use the ResNet-20 [9] in this experiment, since it has similar depth with GRCNN with T = 5. The
implementation is similar to what we have discussed in Sec.4.2. The results of ResNet-BLSTM are
listed in Table 3. This ResNet-based framework performs worse than the GRCNN-based framework.
The result indicates that GRCNN is more suitable for scene text recognition.
Some examples predicted by the proposed method under unconstrained scenario are shown in Figure
4. The correctly predicted examples are shown in the left. It is seen that our model can recognize
some long words that have missing part, such as "MEMENTO" in which "N" is not clear. Some other
bend words are also recognized perfectly, for instance, "COTTAGE" and "BADGER". However,
there are some words that cannot be distinguished precisely and some of them are showed in the right
side in Figure 4. The characters which are combined closely may lead to bad recognition results.
For the word "ARMADA", the network cannot accurately split "A" and "R", leading to a missing
character in the result. Moreover, some special symbols whose shapes are similar to the English
characters affect the results. For example, the symbol in "BLOOM" looks like the character "O", and
the word is incorrectly recognized as "OBLOOM". Finally, some words that have strange-shaped
character are also difficult to be recognized, such as "SBULT" (the ground truth is "SALVADOR").
5
Conclusion
we propose a new architecture named GRCNN which uses a gate to modulate recurrent connections
in a previous model RCNN. GRCNN is able to choose the context information dynamically and
combine the feed-forward part with recurrent part flexibly. The unrelated context information coming
from the recurrent part is inhibited by the gate. In addition, through experiments we find the LSTM
without peephole connection is suitable for scene text recognition. The experiments on scene text
recognition benchmarks demonstrate the superior performance of the proposed method.
Acknowledgements
This work was supported in part by the National Basic Research Program (973 Program) of China
under grant no. 2013CB329403, the National Natural Science Foundation of China under grant nos.
91420201, 61332007, 61621136008 and 61620106010, and in part by a grant from Sensetime.
8
References
[1] J. Almazan, A. Gordo, A. Fornes, and E. Valveny. Word spotting and recognition with embedded
attributes. IEEE Transactions on Pattern Analysis & Machine Intelligence, 36(12):2552?2566,
2014.
[2] O. Alsharif and J. Pineau. End-to-end text recognition with hybrid hmm maxout models.
Computer Science, 2013.
[3] A. Bissacco, M. Cummins, Y. Netzer, and H. Neven. Photoocr: Reading text in uncontrolled
conditions. In ICCV, pages 785?792, 2014.
[4] K. Cho, B. V. Merrienboer, C. Gulcehre, D. Bahdanau, F. Bougares, H. Schwenk, and Y. Bengio.
Learning phrase representations using rnn encoder-decoder for statistical machine translation.
Computer Science, 2014.
[5] F. A. Gers and J. Schmidhuber. Recurrent nets that time and count. In International Joint
Conference on Neural Networks, pages 189?194, 2000.
[6] V. Goel, A. Mishra, K. Alahari, and C. V. Jawahar. Whole is greater than sum of parts: Recognizing scene text words. In International Conference on Document Analysis and Recognition,
pages 398?402, 2013.
[7] A. Gordo. Supervised mid-level features for word image representation. In CVPR, pages
2956?2964, 1998.
[8] A. Graves and F. Gomez. Connectionist temporal classification: labelling unsegmented sequence
data with recurrent neural networks. In ICML, pages 369?376, 2006.
[9] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. In CVPR,
pages 770?778, 2016.
[10] S. Hochreiter and J. Schmidhuber. Long short-term memory. Neural Computation, 9(8):1735,
1997.
[11] G. Huang, Z. Liu, K. Q. Weinberger, and V. D. M. Laurens. Densely connected convolutional
networks. In CVPR, 2017.
[12] D. H. Hubel and T. N. Wiesel. Receptive fields and functional architecture in two nonstriate
visual areas (18 and 19) of the cat. Journal of Neurophysiology, 28:229?89, 1965.
[13] S. Ioffe and C. Szegedy. Batch normalization: Accelerating deep network training by reducing
internal covariate shift. In ICML, pages 448?456, 2015.
[14] M. Jaderberg, K. Simonyan, A. Vedaldi, and A. Zisserman. Deep structured output learning for
unconstrained text recognition. In ICLR, 2014.
[15] M. Jaderberg, K. Simonyan, A. Vedaldi, and A. Zisserman. Synthetic data and artificial neural
networks for natural scene text recognition. Workshop on Deep Learning, NIPS, 2014.
[16] M. Jaderberg, A. Vedaldi, and A. Zisserman. Deep features for text spotting. In ECCV, pages
512?528, 2014.
[17] C. V. Jawahar, K. Alahari, and A. Mishra. Top-down and bottom-up cues for scene text
recognition. In CVPR, pages 2687?2694, 2014.
[18] H. E. Jones, K. L. Grieve, W. Wang, and A. M. Sillito. Surround suppression in primate V1.
Journal of Neurophysiology, 86(4):2011, 2001.
[19] A. Krizhevsky, I. Sutskever, and G. E. Hinton. Imagenet classification with deep convolutional
neural networks. In NIPS, pages 1097?1105, 2012.
[20] C. Y. Lee, A. Bhardwaj, W. Di, V. Jagadeesh, and R. Piramuthu. Region-based discriminative
feature pooling for scene text recognition. In CVPR, pages 4050?4057, 2014.
9
[21] C. Y. Lee and S. Osindero. Recursive recurrent nets with attention modeling for ocr in the wild.
In CVPR, pages 2231?2239, 2016.
[22] M. Liang and X. Hu. Recurrent convolutional neural network for object recognition. In CVPR,
pages 3367?3375, 2015.
[23] M. Liang, X. Hu, and Zhang B. Convolutional neural networks with intra-layer recurrent
connections for scene labeling. In NIPS, 2015.
[24] S. M. Lucas, A. Panaretos, L. Sosa, A. Tang, S. Wong, and R. Young. Icdar 2003 robust reading
competitions. In International Conference on Document Analysis and Recognition, page 682,
2003.
[25] A. Mishra, K. Alahari, and C. V. Jawahar. Scene text recognition using higher order language
priors. In BMVC, 2013.
[26] L. Neumann and J. Matas. Real-time scene text localization and recognition. In CVPR, pages
3538?3545, 2012.
[27] T. Novikova, O. Barinova, V. Lempitsky, and V. Lempitsky. Large-lexicon attribute-consistent
text recognition in natural images. In ECCV, pages 752?765, 2012.
[28] J. A. Rodriguez-Serrano, A. Gordo, and F. Perronnin. Label embedding: A frugal baseline for
text recognition. International Journal of Computer Vision, 113(3):193?207, 2015.
[29] M. Schuster and K. K. Paliwal. Bidirectional recurrent neural networks. IEEE Press, 1997.
[30] B. Shi, X. Bai, and C. Yao. An end-to-end trainable neural network for image-based sequence
recognition and its application to scene text recognition. IEEE Transactions on Pattern Analysis
& Machine Intelligence, 39(11):2298?2304, 2017.
[31] C. Shi, C. Wang, B. Xiao, Y. Zhang, S. Gao, and Z. Zhang. Scene text recognition using
part-based tree-structured character detection. In CVPR, pages 2961?2968, 2013.
[32] R. Socher, C. D. Manning, and A. Y. Ng. Learning continuous phrase representations and
syntactic parsing with recursive neural networks. In NIPS, 2010.
[33] B. Su and S. Lu. Accurate scene text recognition based on recurrent neural network. In ACCV,
pages 35?48, 2014.
[34] C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and
A. Rabinovich. Going deeper with convolutions. In CVPR, pages 1?9, 2014.
[35] J. Wang, W. Wang, X. Chen, R. Wang, and W. Gao. Deep alternative neural network: Exploring
contexts as early as possible for action recognition. In NIPS, 2016.
[36] K. Wang, B. Babenko, and S. Belongie. End-to-end scene text recognition. In ICCV, pages
1457?1464, 2012.
[37] K. Wang and S. Belongie. Word spotting in the wild. In ECCV, pages 591?604, 2010.
[38] T. Wang, D. J. Wu, A. Coates, and A. Y. Ng. End-to-end text recognition with convolutional
neural networks. In International Conference on Pattern Recognition, pages 3304?3308, 2013.
[39] S. Xie, R. Girshick, P. Doll?r, Z. Tu, and K. He. Aggregated residual transformations for deep
neural networks. CVPR, 2017.
[40] C. Yao, X. Bai, B. Shi, and W. Liu. Strokelets: A learned multi-scale representation for scene
text recognition. In CVPR, pages 4042?4049, 2014.
[41] M. D. Zeiler. Adadelta: An adaptive learning rate method. Computer Science, 2012.
[42] M. D. Zeiler and R. Fergus. Visualizing and understanding convolutional networks. In ECCV,
2014.
[43] Y. Zhao, X. Jin, and X. Hu. Recurrent convolutional neural network for speech processing. In
IEEE International Conference on Acoustics, Speech and Signal Processing, 2017.
10
| 6637 |@word neurophysiology:2 cnn:15 pw:10 wco:1 wiesel:1 bf:1 hu:4 bn:9 concise:1 tnlist:1 bai:2 configuration:4 contains:6 score:1 hereafter:1 liu:3 ours:1 document:2 outperforms:3 existing:2 past:2 blank:2 com:2 comparing:1 mishra:4 sosa:1 babenko:1 gmail:1 written:2 parsing:1 explorative:1 realistic:1 alphanumeric:1 shape:1 designed:1 treating:1 intelligence:2 selected:2 cue:1 bissacco:2 short:2 num:6 lexicon:18 location:1 zhang:4 five:1 height:3 along:2 constructed:3 consists:1 alsharif:2 combine:3 wild:2 introduce:1 grieve:1 multi:1 brain:2 inspired:4 unfolded:2 equipped:1 increasing:1 becomes:3 conv:2 unrelated:3 moreover:3 what:1 rcls:1 developed:1 transformation:1 temporal:2 every:2 act:1 expands:1 growth:1 classifier:1 control:4 unit:8 grant:3 before:1 understood:1 svt:11 dropped:1 tsinghua:4 modulation:7 approximately:1 china:4 dynamically:2 bi:1 range:1 responsible:1 testing:1 recursive:6 block:1 implement:1 area:1 rnn:7 vedaldi:3 word:25 pre:1 sensetime:1 cannot:4 selection:1 bend:1 context:18 wong:1 conventional:1 map:7 center:2 missing:2 shi:3 attention:4 flexibly:1 regarded:4 embedding:1 increment:1 controlling:1 play:1 us:1 element:2 adadelta:2 recognition:51 finetuned:1 cut:1 xiaolin:1 bottom:2 role:1 module:4 inserted:1 wang:14 capture:1 region:5 connected:3 sun:1 jagadeesh:1 highest:1 rescaled:1 existent:1 trained:6 solving:1 segment:1 purely:1 localization:1 accelerate:1 finetuning:1 joint:1 schwenk:1 various:1 cat:1 stacked:4 fast:1 effective:5 novikova:2 detected:1 artificial:1 labeling:1 jianfeng:2 choosing:1 whose:2 whi:1 larger:3 solve:2 cvpr:12 tested:2 encoder:1 statistic:1 simonyan:2 peephole:12 syntactic:1 jointly:1 final:1 sequence:20 indication:1 advantage:1 net:2 propose:2 nonclassical:1 coming:1 serrano:1 tu:1 wgr:3 achieve:1 competition:1 sutskever:1 convergence:2 unboundedly:1 valveny:1 neumann:1 generating:1 object:5 help:1 resnet:4 recurrent:48 strong:1 predicted:3 come:1 direction:1 laurens:1 closely:2 saved:1 attribute:2 human:1 fix:2 generalization:1 merrienboer:1 biological:2 whf:1 exploring:1 clarify:1 around:1 ground:2 great:1 gordo:4 dictionary:3 early:1 released:1 integrates:1 label:7 jawahar:3 aim:1 resized:2 focus:1 modelling:1 indicates:1 suppression:1 baseline:1 wf:4 helpful:1 inference:1 perronnin:1 neven:1 entire:1 strokelets:1 going:1 pixel:1 overall:3 classification:7 flexible:1 arg:2 denoted:2 lucas:1 development:1 constrained:6 iters:4 art:1 special:1 field:9 equal:1 extraction:5 beach:1 having:2 shaped:1 identical:1 represents:2 ng:2 yu:1 look:1 icml:2 jones:1 mimic:1 minimized:1 connectionist:2 report:1 inhibited:1 composed:1 national:3 recognize:4 densely:1 pictorial:1 ab:1 detection:2 investigate:2 highly:1 intra:1 evaluation:1 adjust:1 sh:10 accurate:1 baoguang:1 netzer:1 tree:1 abundant:1 girshick:1 weaken:2 stopped:1 instance:1 column:3 modeling:6 cover:1 logp:2 rabinovich:1 phrase:2 deviation:2 recognizing:3 krizhevsky:1 inspires:1 osindero:1 reported:1 synthetic:2 combined:3 cho:1 st:1 bhardwaj:1 lstm:33 international:6 lee:3 off:1 maxpool:4 together:2 yao:3 choose:2 huang:1 worse:1 cb329403:1 zhao:1 leading:1 szegedy:2 stride:2 rcl:16 sec:2 view:3 analyze:1 jia:1 formed:1 accuracy:4 convolutional:18 descriptor:1 who:1 accurately:1 lu:2 none:1 ren:1 researcher:1 synapsis:2 sharing:1 associated:3 wxo:1 wgf:3 di:1 dataset:9 color:1 back:1 feed:14 bidirectional:6 higher:2 supervised:1 follow:1 xie:1 zisserman:3 bmvc:1 formulation:1 done:1 though:1 salvador:1 furthermore:1 alahari:3 receives:1 replacing:1 su:2 unsegmented:1 propagation:1 google:1 rodriguez:2 pineau:1 gray:1 usa:1 effect:3 building:1 contain:3 hence:1 laboratory:1 visualizing:1 during:3 width:3 abb:1 complete:1 demonstrate:1 performs:1 bring:1 image:36 wise:2 recently:1 sigmoid:1 superior:1 functional:1 ctc:1 empirically:1 insensitive:1 million:1 discussed:1 he:2 bougares:1 anguelov:1 surround:1 unconstrained:9 nonlinearity:1 language:1 add:3 recent:1 showed:1 irrelevant:1 scenario:1 schmidhuber:2 paliwal:1 success:1 ubiquitously:1 iiit:1 seen:2 greater:1 care:1 goel:2 recognized:4 aggregated:1 signal:6 cummins:1 multiple:2 full:4 blstm:7 believed:1 long:5 post:1 equally:1 prediction:2 variant:4 basic:3 vision:3 enhancing:1 iteration:10 kernel:7 normalization:4 hochreiter:1 cell:2 receive:3 addition:4 cropped:4 separately:1 diagram:1 extra:1 unlike:1 pooling:2 simulates:1 bahdanau:1 feedforward:1 split:1 bengio:1 almazan:2 affect:1 relu:3 architecture:8 perfectly:1 idea:1 cn:1 shift:1 photoocr:1 accelerating:1 padding:1 speech:3 action:2 deep:14 useful:1 clear:1 listed:3 neocortex:2 ten:1 ph:10 mid:1 category:1 reduced:1 http:1 exist:1 coates:1 wci:1 wr:3 per:2 correctly:1 anatomical:1 discrete:1 iter:2 drawn:1 densenet:1 ht:3 bloom:1 rectangle:1 v1:1 year:2 beijing:3 convert:1 sum:1 telecommunication:1 named:2 place:1 reasonable:1 strange:1 wu:1 dropout:1 layer:35 ct:3 internet:1 uncontrolled:1 gomez:1 precisely:1 constrain:1 scene:20 optical:1 department:1 structured:2 manning:1 wxi:1 smaller:1 beneficial:3 describes:1 character:15 contradicts:1 piramuthu:1 evolves:1 primate:1 iccv:2 pipeline:2 equation:1 icdar:2 mechanism:3 count:1 fed:1 end:18 adopted:1 gulcehre:1 operation:1 doll:1 ocr:9 generic:1 icdar2003:3 distinguished:1 batch:5 weinberger:1 alternative:1 gate:21 top:5 denotes:8 convolve:1 ensure:2 recognizes:1 graphical:1 zeiler:2 sw:10 classical:1 objective:2 matas:1 receptive:7 iclr:1 capacity:1 street:3 hmm:1 decoder:1 seven:1 considers:1 collected:1 length:3 besides:2 code:1 reed:1 illustration:2 balance:1 acquire:1 sermanet:1 liang:2 difficult:1 implementation:2 gated:7 perform:3 upper:2 convolution:13 neuron:2 datasets:4 observation:1 benchmark:3 accv:1 jin:1 incorrectly:1 hinton:1 frame:2 stack:1 frugal:1 introduced:2 gru:1 wxf:1 connection:12 optimized:1 imagenet:1 engine:1 acoustic:1 learned:1 nip:6 able:2 spotting:3 pattern:3 reading:3 challenge:1 program:2 built:1 including:1 memory:3 rf:3 max:3 critical:2 event:1 natural:7 suitable:2 hybrid:1 residual:4 improve:3 github:1 technology:2 eye:1 carried:1 shortterm:1 extract:1 text:41 prior:1 understanding:1 acknowledgement:1 multiplication:2 graf:1 embedded:1 fully:1 validation:4 rcnn:22 foundation:1 vanhoucke:1 consistent:1 xiao:1 viewpoint:1 translation:1 eccv:4 supported:1 last:3 free:2 english:1 side:1 deeper:1 plain:4 wcf:1 evaluating:1 depth:2 forward:14 adaptive:1 erhan:1 transaction:2 bb:1 jaderberg:5 transcription:4 hubel:1 incoming:1 ioffe:1 unnecessary:1 assumed:1 conclude:1 discriminative:1 belongie:2 fergus:1 continuous:1 decade:1 sillito:1 table:14 robust:2 ca:1 contributes:1 protocol:1 whole:3 repeated:2 sliced:1 complementary:1 enlarged:1 fashion:1 gers:1 concatenating:1 tied:1 extractor:3 young:1 tang:1 removing:1 down:1 bad:1 xt:3 covariate:1 barinova:1 symbol:2 explored:1 essential:1 workshop:1 socher:1 adding:2 sequential:1 gained:1 labelling:2 chen:2 forget:1 explore:1 intern:1 gao:2 visual:2 bo:1 truth:2 conditional:3 modulate:3 goal:1 lempitsky:2 maxout:1 shared:4 replace:1 shortcut:4 determined:2 specifically:1 reducing:1 called:3 select:1 internal:2 trainable:1 schuster:1 |
6,232 | 6,638 | Towards Accurate Binary Convolutional Neural
Network
Xiaofan Lin
Cong Zhao
Wei Pan*
DJI Innovations Inc, Shenzhen, China
{xiaofan.lin, cong.zhao, wei.pan}@dji.com
Abstract
We introduce a novel scheme to train binary convolutional neural networks (CNNs)
? CNNs with weights and activations constrained to {-1,+1} at run-time. It has been
known that using binary weights and activations drastically reduce memory size
and accesses, and can replace arithmetic operations with more efficient bitwise operations, leading to much faster test-time inference and lower power consumption.
However, previous works on binarizing CNNs usually result in severe prediction
accuracy degradation. In this paper, we address this issue with two major innovations: (1) approximating full-precision weights with the linear combination of
multiple binary weight bases; (2) employing multiple binary activations to alleviate information loss. The implementation of the resulting binary CNN, denoted
as ABC-Net, is shown to achieve much closer performance to its full-precision
counterpart, and even reach the comparable prediction accuracy on ImageNet and
forest trail datasets, given adequate binary weight bases and activations.
1
Introduction
Convolutional neural networks (CNNs) have achieved state-of-the-art results on real-world applications such as image classification [He et al., 2016] and object detection [Ren et al., 2015], with the
best results obtained with large models and sufficient computation resources. Concurrent to these
progresses, the deployment of CNNs on mobile devices for consumer applications is gaining more
and more attention, due to the widespread commercial value and the exciting prospect.
On mobile applications, it is typically assumed that training is performed on the server and test
or inference is executed on the mobile devices [Courbariaux et al., 2016, Esser et al., 2016]. In
the training phase, GPUs enabled substantial breakthroughs because of their greater computational
speed. In the test phase, however, GPUs are usually too expensive to deploy. Thus improving the
test-time performance and reducing hardware costs are likely to be crucial for further progress,
as mobile applications usually require real-time, low power consumption and fully embeddable.
As a result, there is much interest in research and development of dedicated hardware for deep
neural networks (DNNs). Binary neural networks (BNNs) [Courbariaux et al., 2016, Rastegari et al.,
2016], i.e., neural networks with weights and perhaps activations constrained to only two possible
values (e.g., -1 or +1), would bring great benefits to specialized DNN hardware for three major
reasons: (1) the binary weights/activations reduce memory usage and model size 32 times compared
to single-precision version; (2) if weights are binary, then most multiply-accumulate operations can
be replaced by simple accumulations, which is beneficial because multipliers are the most space
and power-hungry components of the digital implementation of neural networks; (3) furthermore, if
both activations and weights are binary, the multiply-accumulations can be replaced by the bitwise
operations: xnor and bitcount Courbariaux et al. [2016]. This could have a big impact on dedicated
deep learning hardware. For instance, a 32-bit floating point multiplier costs about 200 Xilinx FPGA
slices [Govindu et al., 2004], whereas a 1-bit xnor gate only costs a single slice. Semiconductor
? indicates corresponding author.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
manufacturers like IBM [Esser et al., 2016] and Intel [Venkatesh et al., 2016] have been involved in
the research and development of related chips.
However, binarization usually cause severe prediction accuracy degradation, especially on complex
tasks such as classification on ImageNet dataset. To take a closer look, Rastegari et al. [2016] shows
that binarizing weights causes the accuracy of Resnet-18 drops from 69.3% to 60.8% on ImageNet
dataset. If further binarize activations, the accuracy drops to 51.2%. Similar phenomenon can also be
found in literatures such as [Hubara et al., 2016]. Clearly there is a considerable gap between the
accuracy of a full-precision model and a binary model.
This paper proposes a novel scheme for binarizing CNNs, which aims to alleviate, or even eliminate
the accuracy degradation, while still significantly reducing inference time, resource requirement and
power consumption. The paper makes the following major contributions.
? We approximate full-precision weights with the linear combination of multiple binary weight
bases. The weights values of CNNs are constrained to { 1, +1}, which means convolutions
can be implemented by only addition and subtraction (without multiplication), or bitwise
operation when activations are binary as well. We demonstrate that 3?5 binary weight bases
are adequate to well approximate the full-precision weights.
? We introduce multiple binary activations. Previous works have shown that the quantization
of activations, especially binarization, is more difficult than that of weights [Cai et al., 2017,
Courbariaux et al., 2016]. By employing five binary activations, we have been able to reduce
the Top-1 and Top-5 accuracy degradation caused by binarization to around 5% on ImageNet
compared to the full precision counterpart.
It is worth noting that the multiple binary weight bases/activations scheme is preferable to the fixedpoint quantization in previous works. In those fixed-point quantized networks one still needs to
employ arithmetic operations, such as multiplication and addition, on fixed-point values. Even though
faster than floating point, they still require relatively complex logic and can consume a lot of power.
Detailed discussions can be found in Section 5.2.
Ideally, combining more binary weight bases and activations always leads to better accuracy and will
eventually get very close to that of full-precision networks. We verify this on ImageNet using Resnet
network topology. This is the first time a binary neural network achieves prediction accuracy
comparable to its full-precision counterpart on ImageNet.
2
Related work
Quantized Neural Networks: High precision parameters are not very necessary to reach high
performance in deep neural networks. Recent research efforts (e.g., [Hubara et al., 2016]) have
considerably reduced a large amounts of memory requirement and computation complexity by using
low bitwidth weights and activations. Zhou et al. [2016] further generalized these schemes and
proposed to train CNNs with low bitwidth gradients. By performing the quantization after network
training or using the ?straight-through estimator (STE)" [Bengio et al., 2013], these works avoided the
issues of non-differentiable optimization. While some of these methods have produced good results
on datasets such as CIFAR-10 and SVHN, none has produced low precision networks competitive
with full-precision models on large-scale classification tasks, such as ImageNet. In fact, [Zhou et al.,
2016] and [Hubara et al., 2016] experiment with different combinations of bitwidth for weights and
activations, and show that the performance of their highly quantized networks deteriorates rapidly
when the weights and activations are quantized to less than 4-bit numbers. Cai et al. [2017] enhance
the performance of a low bitwidth model by addressing the gradient mismatch problem, nevertheless
there is still much room for improvement.
Binarized Neural Networks: The binary representation for deep models is not a new topic. At the
emergence of artificial neural networks, inspired biologically, the unit step function has been used as
the activation function [Toms, 1990]. It is known that binary activation can use spiking response for
event-based computation and communication (consuming energy only when necessary) and therefore
is energy-efficient [Esser et al., 2016]. Recently, Courbariaux et al. [2016] introduce BinarizedNeural-Networks (BNNs), neural networks with binary weights and activations at run-time. Different
from their work, Rastegari et al. [2016] introduce simple, efficient, and accurate approximations
to CNNs by binarizing the weights and even the intermediate representations in CNNs. All these
works drastically reduce memory consumption, and replace most arithmetic operations with bitwise
operations, which potentially lead to a substantial increase in power efficiency.
2
In all above mentioned works, binarization significantly reduces accuracy. Our experimental results
on ImageNet show that we are close to filling the gap between the accuracy of a binary model and its
full-precision counterpart. We relied on the idea of finding the best approximation of full-precision
convolution using multiple binary operations, and employing multiple binary activations to allow
more information passing through.
3
Binarization methods
In this section, we detail our binarization method, which is termed ABC-Net (Accurate-BinaryConvolutional) for convenience. Bear in mind that during training, the real-valued weights are
reserved and updated at every epoch, while in test-time only binary weights are used in convolution.
3.1
Weight approximation
Consider a L-layer CNN architecture. Without loss of generality, we assume the weights of each
convolutional layer are tensors of dimension (w, h, cin , cout ), which represents filter width, filter
height, input-channel and output-channel respectively. We propose two variations of binarization
method for weights at each layer: 1) approximate weights as a whole and 2) approximate weights
channel-wise.
3.1.1
Approximate weights as a whole
At each layer, in order to constrain a CNN to have binary weights, we estimate the real-value weight
filter W 2 Rw?h?cin ?cout using the linear combination of M binary filters B1 , B2 , ? ? ? , BM 2
{ 1, +1}w?h?cin ?cout such that W ? ?1 B1 +?2 B2 +? ? ?+?M BM . To find an optimal estimation,
a straightforward way is to solve the following optimization problem:
minJ(?, B) = ||w
?,B
B?||2 ,
s.t. Bij 2 { 1, +1},
(1)
where B = [vec(B1 ), vec(B2 ), ? ? ? , vec(BM )], w = vec(W ) and ? = [?1 , ?2 , ? ? ? , ?M ]T . Here
the notation vec(?) refers to vectorization.
Although a local minimum solution to (1) can be obtained by numerical methods, one could not
backpropagate through it to update the real-value weight filter W . To address this issue, assuming the
mean and standard deviation of W are mean(W ) and std(W ) respectively, we fix Bi ?s as follows:
? + ui std(W )), i = 1, 2, ? ? ? , M,
Bi = Fui (W ) := sign(W
(2)
? = W mean(W ), and ui is a shift parameter. For example, one can choose ui ?s to
where W
be ui = 1 + (i 1) M2 1 , i = 1, 2, ? ? ? , M, to shift evenly over the range [ std(W ), std(W )],
or leave it to be trained by the network. This is based on the observation that the full-precision
weights tend to have a symmetric, non-sparse distribution, which is close to Gaussian. To gain more
intuition and illustrate the approximation effectiveness, an example is visualized in Section S2 of the
supplementary material.
With Bi ?s chosen, (1) becomes a linear regression problem
min J(?) = ||w
?
B?||2 ,
(3)
in which Bi ?s serve as the bases in the design/dictionary matrix. We can then back-propagate
through Bi ?s using the ?straight-through estimator? (STE) [Bengio et al., 2013]. Assume c as the
cost function, A and O as the input and output tensor of a convolution respectively, the forward and
backward approach of an approximated convolution during training can be computed as follows:
Forward: B1 , B2 , ? ? ? , BM = Fu1 (W ), Fu2 (W ), ? ? ? , FuM (W ),
(4)
Solve (3) for ?,
(5)
O=
M
X
(6)
?m Conv(Bm , A).
m=1
@c
@c
Backward:
=
@W
@O
M
X
@O @Bm
?m
@B
m @W
m=1
!
3
@c
=
@O
STE
M
X
@O
?m
@B
m
m=1
!
=
M
X
m=1
?m
@c
.
@Bm
(7)
In test-time, only (6) is required. The block structure of this approximated convolution layer is shown
on the left side in Figure 1. With suitable hardwares and appropriate implementations, the convolution
can be efficiently computed. For example, since the weight values are binary, we can implement the
convolution with additions and subtractions (thus without multiplications). Furthermore, if the input
A is binary as well, we can implement the convolution with bitwise operations: xnor and bitcount
[Rastegari et al., 2016]. Note that the convolution with each binary filter can be computed in parallel.
Figure 1: An example of the block structure of the convolution in ABC-Net. M = N = 3. On the
left is the structure of the approximated convolution (ApproxConv). ApproxConv is expected to approximate the conventional full-precision convolution with linear combination of binary convolutions
(BinConv), i.e., convolution with binary and weights. On the right is the overall block structure of the
convolution in ABC-Net. The input is binarized using different functions Hv1 , Hv2 , Hv3 , passed into
the corresponding ApproxConv?s and then summed up after multiplying their corresponding n ?s.
With the input binarized, the BinConv?s can be implemented with highly efficient bitwise operations.
There are 9 BinConv?s in this example and they can work in parallel.
3.1.2
Approximate weights channel-wise
Alternatively, we can estimate the real-value weight filter Wi 2 Rw?h?cin of each output channel i 2 {1, 2, ? ? ? , cout } using the linear combination of M binary filters Bi1 , Bi2 , ? ? ? , BiM 2
{ 1, +1}w?h?cin such that Wi ? ?i1 Bi1 + ?i2 Bi2 + ? ? ? + ?iM BiM . Again, to find an optimal
estimation, we solve a linear regression problem analogy to (3) for each output channel. After
convolution, the results are concatenated together along the output-channel dimension. If M = 1,
this approach reduces to the Binary-Weights-Networks (BWN) proposed in [Rastegari et al., 2016].
Compared to weights approximation as a whole, the channel-wise approach approximates weights
more elaborately, however no extra cost is needed during inference. Since this approach requires
more computational resources during training, we leave it as a future work and focus on the former
approximation approach in this paper.
3.2
Multiple binary activations and bitwise convolution
As mentioned above, a convolution can be implemented without multiplications when weights are
binarized. However, to utilize the bitwise operation, the activations must be binarized as well, as they
are the inputs of convolutions.
Similar to the activation binarization procedure in [Zhou et al., 2016], we binarize activations after
passing it through a bounded activation function h, which ensures h(x) 2 [0, 1]. We choose the
bounded rectifier as h. Formally, it can be defined as:
hv (x) = clip(x + v, 0, 1),
(8)
where v is a shift parameter. If v = 0, then hv is the clip activation function in [Zhou et al., 2016].
We constrain the binary activations to either 1 or -1. In order to transform the real-valued activation
R into binary activation, we use the following binarization function:
Hv (R) := 2Ihv (R)
4
0.5
1,
(9)
where I is the indicator function. The conventional forward and backward approach of the activation
can be given as follows:
Forward: A = Hv (R).
@c
@c
Backward:
=
I0?R
@R
@A
v?1 .
(using STE)
(10)
Here denotes the Hadamard product. As can be expected, binaizing activations as above is kind of
crude and leads to non-trivial losses in accuracy, as shown in Rastegari et al. [2016], Hubara et al.
[2016]. While it is also possible to approximate activations with linear regression, as that of weights,
another critical challenge arises ? unlike weights, the activations always vary in test-time inference.
Luckily, this difficulty can be avoided by exploiting the statistical structure of the activations of deep
networks.
Our scheme can be described as follows. First of all, to keep the distribution of activations relatively
stable, we resort to batch normalization [Ioffe and Szegedy, 2015]. This is a widely used normalization
technique, which forces the responses of each network layer to have zero mean and unit variance. We
apply this normalization before activation. Secondly, we estimate the real-value activation R using
the linear combination of N binary activations A1 , A2 , ? ? ? , AN such that R ? 1 A1 + 2 A2 +
? ? ? + N AN , where
A1 , A2 , ? ? ? , AN = Hv1 (R), Hv2 (R), ? ? ? , HvN (R).
(11)
Different from that of weights, the parameters n ?s and vn ?s (n = 1, ? ? ? , N ) here are both trainable,
just like the scale and shift parameters in batch normalization. Without the explicit linear regression
approach, n ?s and vn ?s are tuned by the network itself during training and fixed in test-time. They
are expected to learn and utilize the statistical features of full-precision activations.
The resulting network architecture outputs multiple binary activations A1 , A2 , ? ? ? , AN and their
corresponding coefficients 1 , 2 , ? ? ? , N , which allows more information passing through compared
to the former one. Combining with the weight approximation, the whole convolution scheme is given
by:
!
M
N
M X
N
X
X
X
Conv(W , R) ? Conv
? m Bm ,
=
?m n Conv (Bm , An ) , (12)
n An
m=1
n=1
m=1 n=1
which suggests that it can be implemented by computing M ? N bitwise convolutions in parallel.
An example of the whole convolution scheme is shown in Figure 1.
3.3
Training algorithm
A typical block in CNN contains several different layers, which are usually in the following order:
(1) Convolution, (2) Batch Normalization, (3) Activation and (4) Pooling. The batch normalization
layer [Ioffe and Szegedy, 2015] normalizes the input batch by its mean and variance. The activation
is an element-wise non-linear function (e.g., Sigmoid, ReLU). The pooling layer applies any type of
pooling (e.g., max,min or average) on the input batch. In our experiment, we observe that applying
max-pooling on binary input returns a tensor that most of its elements are equal to +1, resulting in a
noticeable drop in accuracy. Similar phenomenon has been reported in Rastegari et al. [2016] as well.
Therefore, we put the max-pooling layer before the batch normalization and activation.
Since our binarization scheme approximates full-precision weights, using the full-precision pre-train
model serves as a perfect initialization. However, fine-tuning is always required for the weights
to adapt to the new network structure. The training procedure, i.e., ABC-Net, is summarized in
Section S1 of the supplementary material.
It is worth noting that as M increases, the shift parameters get closer and the bases of the linear
combination are more correlated, which sometimes lead to rank deficiency when solving (3). This
can be tackled with the `2 regularization.
4
Experiment results
In this section, the proposed ABC-Net was evaluated on the ILSVRC12 ImageNet classification
dataset [Deng et al., 2009], and visual perception of forest trails datasets for mobile robots [Giusti
et al., 2016] in Section S6 of supplementary material.
5
4.1
Experiment results on ImageNet dataset
The ImageNet dataset contains about 1.2 million high-resolution natural images for training that
spans 1000 categories of objects. The validation set contains 50k images. We use Resnet ([He et al.,
2016]) as network topology. The images are resized to 224x224 before fed into the network. We
report our classification performance using Top-1 and Top-5 accuracies.
4.1.1
Effect of weight approximation
We first evaluate the weight approximation technique by examining the accuracy improvement for a
binary model. To eliminate variables, we leave the activations being full-precision in this experiment.
Table 1 shows the prediction accuracy of ABC-Net on ImageNet with different choices of M . For
comparison, we add the results of Binary-Weights-Network (denoted ?BWN?) reported in Rastegari
et al. [2016] and the full-precision network (denoted ?FP?). The BWN binarizes weights and leaves
the activations being full-precision as we do. All results in this experiment use Resnet-18 as network
topology. It can be observed that as M increases, the accuracy of ABC-Net converges to its fullprecision counterpart. The Top-1 gap between them reduces to only 0.9 percentage point when
M = 5, which suggests that this approach nearly eliminates the accuracy degradation caused by
binarizing weights.
Table 1: Top-1 (left) and Top-5 (right) accuracy of ABC-Net on ImageNet, using full-precision
activation and different choices of the number of binary weight bases M .
Top-1
Top-5
BWN
60.8%
83.0%
M =1
62.8%
84.4%
M =2
63.7%
85.2%
M =3
66.2%
86.7%
M =5
68.3%
87.9%
FP
69.3%
89.2%
For interested readers, Figure S4 in section S5 of the supplementary material shows that the relationship between accuracy and M appears to be linear. Also, in Section S2 of the supplementary material,
a visualization of the approximated weights is provided.
4.1.2
Configuration space exploration
We explore the configuration space of combinations of number of weight bases and activations. Table
2 presents the results of ABC-Net with different configurations. The parameter settings for these
experiments are provided in Section S4 of the supplementary material.
Table 2: Prediction accuracy (Top-1/Top-5) for ImageNet with different choices of M and N in a
ABC-Net (approximate weights as a whole). ?res18?, ?res34? and ?res50? are short for Resnet-18,
Resnet-34 and Resnet-50 network topology respectively. M and N refer to the number of weight
bases and activations respectively.
Network
res18
res18
res18
res18
res18
res18
res18
res18
res34
res34
res34
res34
res50
res50
M -weight base N -activation base
1
1
3
1
3
3
3
5
5
1
5
3
5
5
Full Precision
1
1
3
3
5
5
Full Precision
5
5
Full Precision
Top-1
42.7%
49.1%
61.0%
63.1%
54.1%
62.5%
65.0%
69.3%
52.4%
66.7%
68.4%
73.3%
70.1%
76.1%
Top-5
67.6%
73.8%
83.2%
84.8%
78.1%
84.2%
85.9%
89.2%
76.5%
87.4%
88.2%
91.3%
89.7%
92.8%
Top-1 gap
26.6%
20.2%
8.3%
6.2%
15.2%
6.8%
4.3%
20.9%
6.6%
4.9%
6.0%
-
Top-5 gap
21.6%
15.4%
6.0%
4.4%
11.1%
5.0%
3.3%
14.8%
3.9%
3.1%
3.1%
-
As balancing between multiple factors like training time and inference time, model size and accuracy
is more a problem of practical trade-off, there will be no definite conclusion as which combination of
6
(M, N ) one should choose. In general, Table 2 shows that (1) the prediction accuracy of ABC-Net
improves greatly as the number of binary activations increases, which is analogous to the weight
approximation approach; (2) larger M or N gives better accuracy; (3) when M = N = 5, the Top-1
gap between the accuracy of a full-precision model and a binary one reduces to around 5%. To gain a
visual understanding and show the possibility of extensions to other tasks such object detection, we
print the a sample of feature maps in Section S3 of supplementary material.
4.1.3
Comparison with the state-of-the-art
Table 3: Classification test accuracy of CNNs trained on ImageNet with Resnet-18 network topology.
?W? and ?A? refer to the weight and activation bitwidth respectively.
Model
Full-Precision Resnet-18 [full-precision weights and activation]
BWN [full-precision activation] Rastegari et al. [2016]
DoReFa-Net [1-bit weight and 4-bit activation] Zhou et al. [2016]
XNOR-Net [binary weight and activation] Rastegari et al. [2016]
BNN [binary weight and activation] Courbariaux et al. [2016]
ABC-Net [5 binary weight bases, 5 binary activations]
ABC-Net [5 binary weight bases, full-precision activations]
W
32
1
1
1
1
1
1
A
32
32
4
1
1
1
32
Top-1
69.3%
60.8%
59.2%
51.2%
42.2%
65.0%
68.3%
Top-5
89.2%
83.0%
81.5%
73.2%
67.1%
85.9%
87.9%
Table 3 presents a comparison between ABC-Net and several other state-of-the-art models, i.e.,
full-precision Resnet-18, BWN and XNOR-Net in [Rastegari et al., 2016], DoReFa-Net in [Zhou
et al., 2016] and BNN in [Courbariaux et al., 2016] respectively. All comparative models use Resnet18 as network topology. The full-precision Resnet-18 achieves 69.3% Top-1 accuracy. Although
Rastegari et al. [2016]?s BWN model and DeReFa-Net perform well, it should be noted that they use
full-precision and 4-bit activation respectively. Models (XNOR-Net and BNN) that used both binary
weights and activations achieve much less satisfactory accuracy, and is significantly outperformed
by ABC-Net with multiple binary weight bases and activations. It can be seen that ABC-Net has
achieved state-of-the-art performance as a binary model.
One might argue that 5-bit width quantization scheme could reach similar accuracy as that of ABCNet with 5 weight bases and 5 binary activations. However, the former one is less efficient and
requires distinctly more hardware resource. More detailed discussions can be found in Section 5.2.
5
5.1
Discussion
Why adding a shift parameter works?
Intuitively, the multiple binarized weight bases/activations scheme works because it allows more
information passing through. Consider the case that a real value, say 1.5, is passed to a binarized
function f (x) = sign(x), where sign maps a positive x to 1 and otherwise -1. In that case, the
outputs of f (1.5) is 1, which suggests that the input value is positive. Now imagine that we have two
binarization function f1 (x) = sign(x) and f2 (x) = sign(x 2). In that case f1 outputs 1 and f2
outputs -1, which suggests that the input value is not only positive, but also must be smaller than 2.
Clearly we see that each function contributes differently to represent the input and more information
is gained from f2 compared to the former case.
From another point of view, both coefficients ( ?s) and shift parameters are expected to learn and
utilize the statistical features of full-precision tensors, just like the scale and shift parameters in batch
normalization. If we have more binarized weight bases/activations, the network has the capacity to
approximate the full-precision one more precisely. Therefore, it can be deduced that when M or N
is large enough, the network learns to tune itself so that the combination of M weight bases or N
binarized activations can act like the full-precision one.
5.2
Advantage over the fixed-point quantization scheme
It should be noted that there are key differences between the multiple binarization scheme (M
binarized weight bases or N binarized activations) proposed in this paper and the fixed-point quantization scheme in the previous works such as [Zhou et al., 2016, Hubara et al., 2016], though at first
Courbariaux et al. [2016] did not report their result on ImageNet. We implemented and presented the result.
7
thought K-bit width quantization seems to share the same memory requirement with K binarizations.
Specifically, our K binarized weight bases/activations is preferable to the fixed K-bit width scheme
for the following reasons:
(1) The K binarization scheme preserves binarization for bitwise operations. One or several bitwise
operations is known to be more efficient than a fixed-point multiplication, which is a major reason
that BNN/XNOR-Net was proposed.
(2) A K-bit width multiplier consumes more resources than K 1-bit multipliers in a digital chip: it
requires more than K bits to store and compute, otherwise it could easily overflow/underflow. For
example, if a real number is quantized to a 2-bit number, a possible choice is in range {0,1,2,4}.
In this 2-bit multiplication, when both numbers are 4, it outputs 4 ? 4 = 16, which is not within
the range. In [Zhou et al., 2016], the range of activations is constrained within [0,1], which seems
to avoid this situation. However, fractional numbers do not solve this problem, severe precision
deterioration will appear during the multiplication if there are no extra resources. The fact that the
complexity of a multiplier is proportional to THE SQUARE of bit-widths can be found in literatures
(e.g., sec 3.1.1. in [Grabbe et al., 2003]). In contrast, our K binarization scheme does not have this
issue ? it always outputs within the range {-1,1}. The saved hardware resources can be further used
for parallel computing.
(3) A binary activation can use spiking response for event-based computation and communication
(consuming energy only when necessary) and therefore is energy-efficient [Esser et al., 2016]. This
can be employed in our scheme, but not in the fixed K-bit width scheme. Also, we have mentioned the
fact that K-bit width multiplier consumes more resources than K 1-bit multipliers. It is noteworthy
that these resources include power.
To sum up, K-bit multipliers are the most space and power-hungry components of the digital
implementation of DNNs. Our scheme could bring great benefits to specialized DNN hardware.
5.3
Further computation reduction in run-time
On specialized hardware, the following operations in our scheme can be integrated with other
operations in run-time and further reduce the computation requirement.
(1) Shift operations. The existence of shift parameters seem to require extra additions/subtractions
(see (2) and (8)). However, the binarization operation with a shift parameter can be implemented
as a comparator where the shift parameter is the number for comparison, e.g., Hv (R) =
?
1, R 0.5 v;
1, R < 0.5 v. (0.5 v is a constant), so no extra additions/subtractions are involved.
(2) Batch normalization. In run-time, a batch normalization is simply an affine function, say,
BN(R) = aR + b, whose scale and shift parameters a, b are fixed and can be integrated with
vn ?s. More specifically, ?
a batch normalization can be integrated
into a binarization operation as
?
1, aR + b 0.5 v;
1, R (0.5 v b)/a;
follow: Hv (BN(R)) =
Therefore,
1, aR + b < 0.5 v. =
1, R < (0.5 v b)/a.
there will be no extra cost for the batch normalization.
6
Conclusion and future work
We have introduced a novel binarization scheme for weights and activations during forward and
backward propagations called ABC-Net. We have shown that it is possible to train a binary CNN with
ABC-Net on ImageNet and achieve accuracy close to its full-precision counterpart. The binarization
scheme proposed in this work is parallelizable and hardware friendly, and the impact of such a method
on specialized hardware implementations of CNNs could be major, by replacing most multiplications
in convolution with bitwise operations. The potential to speed-up the test-time inference might be
very useful for real-time embedding systems. Future work includes the extension of those results
to other tasks such as object detection and other models such as RNN. Also, it would be interesting
to investigate using FPGA/ASIC or other customized deep learning processor [Liu et al., 2016] to
implement ABC-Net at run-time.
7
Acknowledgement
We acknowledge Mr Jingyang Xu for helpful discussions.
8
References
Y. Bengio, N. L?onard, and A. Courville. Estimating or propagating gradients through stochastic neurons for
conditional computation. arXiv preprint arXiv:1308.3432, 2013.
Z. Cai, X. He, J. Sun, and N. Vasconcelos. Deep learning with low precision by half-wave gaussian quantization.
arXiv preprint arXiv:1702.00953, 2017.
M. Courbariaux, I. Hubara, D. Soudry, R. El-Yaniv, and Y. Bengio. Binarized neural networks: Training deep
neural networks with weights and activations constrained to+ 1 or-1. arXiv preprint arXiv:1602.02830, 2016.
J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei. Imagenet: A large-scale hierarchical image database.
In Computer Vision and Pattern Recognition, 2009. CVPR 2009. IEEE Conference on, pages 248?255. IEEE,
2009.
S. K. Esser, P. A. Merolla, J. V. Arthur, A. S. Cassidy, R. Appuswamy, A. Andreopoulos, D. J. Berg, J. L.
McKinstry, T. Melano, D. R. Barch, et al. Convolutional networks for fast, energy-efficient neuromorphic
computing. Proceedings of the National Academy of Sciences, page 201604850, 2016.
A. Giusti, J. Guzzi, D. Ciresan, F.-L. He, J. P. Rodriguez, F. Fontana, M. Faessler, C. Forster, J. Schmidhuber,
G. Di Caro, D. Scaramuzza, and L. Gambardella. A machine learning approach to visual perception of forest
trails for mobile robots. IEEE Robotics and Automation Letters, 2016.
G. Govindu, L. Zhuo, S. Choi, and V. Prasanna. Analysis of high-performance floating-point arithmetic on fpgas.
In Parallel and Distributed Processing Symposium, 2004. Proceedings. 18th International, page 149. IEEE,
2004.
C. Grabbe, M. Bednara, J. Teich, J. von zur Gathen, and J. Shokrollahi. Fpga designs of parallel high performance
gf (2233) multipliers. In ISCAS (2), pages 268?271. Citeseer, 2003.
K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. In Proceedings of the IEEE
Conference on Computer Vision and Pattern Recognition, pages 770?778, 2016.
I. Hubara, M. Courbariaux, D. Soudry, R. El-Yaniv, and Y. Bengio. Quantized neural networks: Training neural
networks with low precision weights and activations. arXiv preprint arXiv:1609.07061, 2016.
S. Ioffe and C. Szegedy. Batch normalization: Accelerating deep network training by reducing internal covariate
shift. arXiv preprint arXiv:1502.03167, 2015.
D. Kingma and J. Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
S. Liu, Z. Du, J. Tao, D. Han, T. Luo, Y. Xie, Y. Chen, and T. Chen. Cambricon: An instruction set architecture
for neural networks. In Proceedings of the 43rd International Symposium on Computer Architecture, pages
393?405. IEEE Press, 2016.
N. Qian. On the momentum term in gradient descent learning algorithms. Neural networks, 12(1):145?151,
1999.
M. Rastegari, V. Ordonez, J. Redmon, and A. Farhadi. Xnor-net: Imagenet classification using binary convolutional neural networks. In European Conference on Computer Vision, pages 525?542. Springer, 2016.
S. Ren, K. He, R. Girshick, and J. Sun. Faster r-cnn: Towards real-time object detection with region proposal
networks. In Advances in neural information processing systems, pages 91?99, 2015.
D. Toms. Training binary node feedforward neural networks by back propagation of error. Electronics letters,
26(21):1745?1746, 1990.
G. Venkatesh, E. Nurvitadhi, and D. Marr. Accelerating deep convolutional networks using low-precision and
sparsity. arXiv preprint arXiv:1610.00324, 2016.
S. Zhou, Y. Wu, Z. Ni, X. Zhou, H. Wen, and Y. Zou. Dorefa-net: Training low bitwidth convolutional neural
networks with low bitwidth gradients. arXiv preprint arXiv:1606.06160, 2016.
9
| 6638 |@word cnn:6 version:1 seems:2 instruction:1 propagate:1 bn:2 teich:1 citeseer:1 reduction:1 electronics:1 configuration:3 contains:3 liu:2 tuned:1 bitwise:12 com:1 luo:1 activation:73 must:2 numerical:1 drop:3 update:1 half:1 leaf:1 device:2 short:1 quantized:6 node:1 zhang:1 five:1 height:1 along:1 symposium:2 introduce:4 expected:4 shokrollahi:1 inspired:1 farhadi:1 becomes:1 conv:4 provided:2 notation:1 bounded:2 estimating:1 kind:1 finding:1 every:1 binarized:13 act:1 friendly:1 preferable:2 unit:2 appear:1 before:3 positive:3 local:1 semiconductor:1 soudry:2 noteworthy:1 might:2 initialization:1 china:1 suggests:4 deployment:1 bim:2 bi:5 range:5 practical:1 block:4 implement:3 definite:1 procedure:2 rnn:1 significantly:3 thought:1 onard:1 pre:1 refers:1 get:2 convenience:1 close:4 put:1 applying:1 accumulation:2 conventional:2 map:2 straightforward:1 attention:1 resolution:1 qian:1 m2:1 estimator:2 marr:1 enabled:1 s6:1 embedding:1 variation:1 analogous:1 updated:1 imagine:1 commercial:1 deploy:1 trail:3 element:2 expensive:1 approximated:4 recognition:3 std:4 database:1 observed:1 preprint:8 embeddable:1 hv:6 cong:2 region:1 ensures:1 sun:3 trade:1 prospect:1 cin:5 consumes:2 substantial:2 mentioned:3 intuition:1 complexity:2 ui:4 ideally:1 trained:2 solving:1 serve:1 binarizing:5 efficiency:1 f2:3 easily:1 chip:2 bitwidth:7 differently:1 train:4 fast:1 artificial:1 whose:1 supplementary:7 valued:2 solve:4 consume:1 widely:1 larger:1 say:2 otherwise:2 cvpr:1 emergence:1 transform:1 itself:2 fontana:1 advantage:1 differentiable:1 net:29 cai:3 propose:1 product:1 ste:4 combining:2 hadamard:1 rapidly:1 achieve:3 bitcount:2 academy:1 exploiting:1 yaniv:2 requirement:4 comparative:1 adam:1 perfect:1 leave:3 converges:1 object:5 resnet:11 illustrate:1 cassidy:1 propagating:1 noticeable:1 progress:2 implemented:6 fu2:1 saved:1 cnns:12 filter:8 luckily:1 exploration:1 stochastic:2 material:7 require:3 dnns:2 fix:1 f1:2 alleviate:2 bi1:2 secondly:1 im:1 extension:2 around:2 great:2 bnns:2 major:5 achieves:2 dictionary:1 vary:1 a2:4 estimation:2 outperformed:1 hubara:7 concurrent:1 clearly:2 always:4 gaussian:2 aim:1 zhou:10 avoid:1 resized:1 mobile:6 focus:1 improvement:2 rank:1 indicates:1 greatly:1 contrast:1 helpful:1 inference:7 el:2 i0:1 typically:1 eliminate:2 integrated:3 resnet18:1 dnn:2 i1:1 x224:1 interested:1 tao:1 issue:4 classification:7 overall:1 denoted:3 development:2 proposes:1 constrained:5 art:4 breakthrough:1 summed:1 equal:1 vasconcelos:1 beach:1 represents:1 look:1 filling:1 nearly:1 govindu:2 hv1:2 future:3 report:2 employ:1 wen:1 preserve:1 national:1 floating:3 replaced:2 phase:2 iscas:1 mckinstry:1 detection:4 interest:1 highly:2 possibility:1 multiply:2 investigate:1 severe:3 accurate:3 closer:3 necessary:3 arthur:1 girshick:1 instance:1 ar:3 neuromorphic:1 cost:6 addressing:1 deviation:1 fpga:3 examining:1 too:1 reported:2 bwn:7 considerably:1 st:1 deduced:1 international:2 off:1 dong:1 fui:1 enhance:1 together:1 again:1 von:1 choose:3 resort:1 zhao:2 leading:1 return:1 li:2 szegedy:3 potential:1 b2:4 summarized:1 automation:1 sec:1 coefficient:2 inc:1 includes:1 caused:2 performed:1 view:1 lot:1 competitive:1 relied:1 wave:1 parallel:6 contribution:1 square:1 ni:1 accuracy:31 convolutional:8 variance:2 reserved:1 efficiently:1 shenzhen:1 produced:2 ren:3 none:1 multiplying:1 worth:2 straight:2 processor:1 minj:1 reach:3 parallelizable:1 energy:5 involved:2 di:1 gain:2 dataset:5 fractional:1 improves:1 back:2 appears:1 xie:1 follow:1 tom:2 response:3 wei:2 evaluated:1 though:2 generality:1 furthermore:2 just:2 merolla:1 replacing:1 propagation:2 rodriguez:1 widespread:1 ordonez:1 perhaps:1 usa:1 usage:1 effect:1 verify:1 multiplier:9 counterpart:6 former:4 regularization:1 symmetric:1 satisfactory:1 i2:1 xnor:8 bnn:4 during:7 width:8 noted:2 xilinx:1 generalized:1 cout:4 demonstrate:1 dedicated:2 bring:2 hungry:2 svhn:1 image:6 wise:4 novel:3 recently:1 sigmoid:1 specialized:4 dji:2 spiking:2 million:1 caro:1 he:6 approximates:2 nurvitadhi:1 accumulate:1 s5:1 refer:2 vec:5 tuning:1 rd:1 elaborately:1 esser:5 access:1 stable:1 robot:2 han:1 base:22 add:1 recent:1 termed:1 store:1 asic:1 server:1 schmidhuber:1 binary:60 seen:1 minimum:1 greater:1 mr:1 deng:2 employed:1 subtraction:4 gambardella:1 arithmetic:4 full:35 multiple:13 reduces:4 faster:3 adapt:1 long:1 lin:2 cifar:1 a1:4 impact:2 prediction:7 regression:4 vision:3 arxiv:16 normalization:13 sometimes:1 represent:1 deterioration:1 achieved:2 robotics:1 zur:1 proposal:1 whereas:1 addition:5 fine:1 appuswamy:1 crucial:1 extra:5 eliminates:1 unlike:1 pooling:5 tend:1 effectiveness:1 seem:1 noting:2 intermediate:1 bengio:5 enough:1 feedforward:1 relu:1 architecture:4 topology:6 ciresan:1 reduce:5 idea:1 shift:14 dorefa:3 passed:2 giusti:2 effort:1 accelerating:2 passing:4 cause:2 adequate:2 deep:11 useful:1 detailed:2 fu1:1 tune:1 amount:1 s4:2 hardware:11 visualized:1 clip:2 rw:2 reduced:1 category:1 percentage:1 s3:1 sign:5 deteriorates:1 key:1 nevertheless:1 utilize:3 backward:5 sum:1 run:6 ilsvrc12:1 letter:2 reader:1 wu:1 vn:3 comparable:2 bit:19 layer:10 tackled:1 courville:1 precisely:1 deficiency:1 constrain:2 fei:2 speed:2 min:2 span:1 performing:1 relatively:2 gpus:2 combination:11 beneficial:1 smaller:1 pan:2 wi:2 biologically:1 s1:1 intuitively:1 resource:9 visualization:1 eventually:1 needed:1 mind:1 fed:1 serf:1 operation:20 manufacturer:1 apply:1 observe:1 hierarchical:1 appropriate:1 batch:13 gate:1 existence:1 fpgas:1 top:19 denotes:1 include:1 binarizes:1 concatenated:1 especially:2 overflow:1 approximating:1 tensor:4 print:1 forster:1 gradient:5 capacity:1 consumption:4 evenly:1 topic:1 argue:1 binarize:2 reason:3 trivial:1 consumer:1 assuming:1 relationship:1 innovation:2 difficult:1 executed:1 potentially:1 ba:1 implementation:5 design:2 perform:1 convolution:25 observation:1 datasets:3 neuron:1 acknowledge:1 descent:1 situation:1 communication:2 introduced:1 venkatesh:2 required:2 imagenet:19 kingma:1 nip:1 address:2 able:1 zhuo:1 usually:5 perception:2 mismatch:1 pattern:2 andreopoulos:1 fp:2 sparsity:1 challenge:1 gaining:1 memory:5 max:3 ihv:1 power:8 event:2 suitable:1 bi2:2 critical:1 difficulty:1 force:1 indicator:1 natural:1 customized:1 residual:1 scheme:22 gf:1 binarization:19 epoch:1 literature:2 understanding:1 acknowledgement:1 multiplication:8 loss:3 fully:1 bear:1 interesting:1 proportional:1 analogy:1 digital:3 validation:1 affine:1 sufficient:1 exciting:1 courbariaux:10 prasanna:1 share:1 balancing:1 ibm:1 normalizes:1 drastically:2 side:1 allow:1 sparse:1 distinctly:1 benefit:2 slice:2 distributed:1 dimension:2 world:1 author:1 forward:5 fixedpoint:1 avoided:2 bm:9 employing:3 approximate:10 logic:1 keep:1 ioffe:3 b1:4 assumed:1 consuming:2 alternatively:1 vectorization:1 why:1 table:7 channel:8 learn:2 ca:1 rastegari:13 contributes:1 forest:3 improving:1 du:1 complex:2 european:1 zou:1 did:1 big:1 whole:6 s2:2 xu:1 intel:1 precision:42 momentum:1 explicit:1 crude:1 learns:1 bij:1 choi:1 rectifier:1 covariate:1 socher:1 quantization:8 adding:1 gained:1 barch:1 gap:6 chen:2 backpropagate:1 simply:1 likely:1 hvn:1 explore:1 visual:3 applies:1 springer:1 underflow:1 abc:20 comparator:1 conditional:1 towards:2 room:1 replace:2 considerable:1 typical:1 specifically:2 reducing:3 redmon:1 degradation:5 called:1 experimental:1 hv2:2 formally:1 berg:1 internal:1 arises:1 evaluate:1 trainable:1 phenomenon:2 correlated:1 |
6,233 | 6,639 | Semi-Supervised Learning for Optical Flow
with Generative Adversarial Networks
Wei-Sheng Lai1
Jia-Bin Huang2
Ming-Hsuan Yang1,3
2
3
University of California, Merced
Virginia Tech
Nvidia Research
1
2
{wlai24|mhyang}@ucmerced.edu
[email protected]
1
Abstract
Convolutional neural networks (CNNs) have recently been applied to the optical
flow estimation problem. As training the CNNs requires sufficiently large amounts
of labeled data, existing approaches resort to synthetic, unrealistic datasets. On
the other hand, unsupervised methods are capable of leveraging real-world videos
for training where the ground truth flow fields are not available. These methods,
however, rely on the fundamental assumptions of brightness constancy and spatial
smoothness priors that do not hold near motion boundaries. In this paper, we
propose to exploit unlabeled videos for semi-supervised learning of optical flow
with a Generative Adversarial Network. Our key insight is that the adversarial
loss can capture the structural patterns of flow warp errors without making explicit
assumptions. Extensive experiments on benchmark datasets demonstrate that the
proposed semi-supervised algorithm performs favorably against purely supervised
and baseline semi-supervised learning schemes.
1
Introduction
Optical flow estimation is one of the fundamental problems in computer vision. The classical formulation builds upon the assumptions of brightness constancy and spatial smoothness [15, 25]. Recent
advancements in this field include using sparse descriptor matching as guidance [4], leveraging dense
correspondences from hierarchical features [2, 39], or adopting edge-preserving interpolation techniques [32]. Existing classical approaches, however, involve optimizing computationally expensive
non-convex objective functions.
With the rapid growth of deep convolutional neural networks (CNNs), several approaches have
been proposed to solve optical flow estimation in an end-to-end manner. Due to the lack of the
large-scale ground truth flow datasets of real-world scenes, existing approaches [8, 16, 30] rely on
training on synthetic datasets. These synthetic datasets, however, do not reflect the complexity of
realistic photometric effects, motion blur, illumination, occlusion, and natural image noise. Several
recent methods [1, 40] propose to leverage real-world videos for training CNNs in an unsupervised
setting (i.e., without using ground truth flow). The main idea is to use loss functions measuring
brightness constancy and spatial smoothness of flow fields as a proxy for losses using ground truth
flow. However, the assumptions of brightness constancy and spatial smoothness often do not hold
near motion boundaries. Despite the acceleration in computational speed, the performance of these
approaches still does not match up to the classical flow estimation algorithms.
With the limited quantity and unrealistic of ground truth flow and the large amounts of real-world
unlabeled data, it is thus of great interest to explore the semi-supervised learning framework. A
straightforward approach is to minimize the End Point Error (EPE) loss for data with ground truth
flow and the loss functions that measure classical brightness constancy and smoothness assumptions
for unlabeled training images (Figure 1 (a)). However, we show that such an approach is sensitive
to the choice of parameters and may sometimes decrease the accuracy of flow estimation. Prior
work [1, 40] minimizes a robust loss function (e.g., Charbonnier function) on the flow warp error
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
EPE loss
CNN
EPE loss
CNN
Ground truth flow
Predicted flow
Predicted flow
Labeled data
Flow
warp error
Unlabeled data
Smoothness loss
Predicted flow
Labeled data
Unlabeled data
Brightness
constancy loss
CNN
Ground truth flow
Adversarial
loss
CNN
Predicted flow
Flow
warp error
(a) Baseline semi-supervised learning
Flow
warp error
(b) The proposed semi-supervised learning
Figure 1: Semi-supervised learning for optical flow estimation. (a) A baseline semi-supervised
algorithm utilizes the assumptions of brightness constancy and spatial smoothness to train CNN from
unlabeled data (e.g., [1, 40]). (b) We train a generative adversarial network to capture the structure
patterns in flow warp error images without making any prior assumptions.
(i.e., the difference between the first input image and the warped second image) by modeling the
brightness constancy with a Laplacian distribution. As shown in Figure 2, although robust loss
functions can fit the likelihood of the per-pixel flow warp error well, the spatial structure in the warp
error images cannot be modeled by simple distributions. Such structural patterns often arise from
occlusion and dis-occlusion caused by large object motion, where the brightness constancy assumption
does not hold. A few approaches have been developed to cope with such brightness inconsistency
problem using the Fields-of-Experts (FoE) [37] or a Gaussian Mixture Model (GMM) [33]. However,
the inference of optical flow entails solving time-consuming optimization problems.
In this work, our goal is to leverage both the labeled and the unlabeled data without making explicit
assumptions on the brightness constancy and flow smoothness. Specifically, we propose to impose an
adversarial loss [12] on the flow warp error image to replace the commonly used brightness constancy
loss. We formulate the optical flow estimation as a conditional Generative Adversarial Network
(GAN) [12]. Our generator takes the input image pair and predicts the flow. We then compute the flow
warp error image using a bilinear sampling layer. We learn a discriminator to distinguish between the
flow warp error from predicted flow and ground truth optical flow fields. The adversarial training
scheme encourages the generator to produce the flow warp error images that are indistinguishable
from the ground truth. The adversarial loss serves as a regularizer for both labeled and unlabeled data
(Figure 1 (b)). With the adversarial training, our network learns to model the structural patterns of
flow warp error to refine the motion boundary. During the test phase, the generator can efficiently
predict optical flow in one feed-forward pass.
We make the following three contributions:
? We propose a generative adversarial training framework to learn to predict optical flow by leveraging both labeled and unlabeled data in a semi-supervised learning framework.
? We develop a network to capture the spatial structure of the flow warp error without making
primitive assumptions on brightness constancy or spatial smoothness.
? We demonstrate that the proposed semi-supervised flow estimation method outperforms the purely
supervised and baseline semi-supervised learning when using the same amount of ground truth
flow and network parameters.
2
Related Work
In the following, we discuss the learning-based optical flow algorithms, CNN-based semi-supervised
learning approaches, and generative adversarial networks within the context of this work.
Optical flow. Classical optical flow estimation approaches typically rely on the assumptions of
brightness constancy and spatial smoothness [15, 25]. Sun et al. [36] provide a unified review of
classical algorithms. Here we focus our discussion on recent learning-based methods in this field.
Learning-based methods aim to learn priors from natural image sequences without using hand-crafted
assumptions. Sun et al. [37] assume that the flow warp error at each pixel is independent and use a
set of linear filters to learn the brightness inconsistency. Rosenbaum and Weiss [33] use a GMM to
learn the flow warp error at the patch level. The work of Rosenbaum et al. [34] learns patch priors
2
Input image 2
Negative log likelihood
Input image 1
-0.15
Ground truth optical flow
Ground truth flow warp error
Flow warp error
Gaussian
Laplacian
Lorentzian
-0.1
-0.05
0
0.05
Flow warp error
0.1
0.15
Negative log likelihood
Figure 2: Modeling the distribution of flow warp error. The robust loss functions, e.g., Lorentzian
or Charbonnier functions, can model the distribution of per-pixel flow warp error well. However, the
spatial pattern resulting from large motion and occlusion cannot be captured by simple distributions.
to model the local flow statistics. These approaches incorporate the learned priors into the classical
formulation and thus require solving time-consuming alternative optimization to infer the optical flow.
Furthermore, the limited amount of training data (e.g., Middlebury [3] or Sintel [5]) may not fully
demonstrate the capability of learning-based optical flow algorithms. In contrast, we train a deep
CNN with large datasets (FlyingChairs [8] and KITTI [10]) in an end-to-end manner. Our model can
predict flow efficiently in a single feed-forward pass.
The FlowNet [8] presents a deep CNN approach for learning optical flow. Even though the network
is trained on a large dataset with ground truth flow, strong data augmentation and the variational
refinement are required. Ilg et al. [16] extend the FlowNet by stacking multiple networks and using
more training data with different motion types including complex 3D motion and small displacements.
To handle large motion, the SPyNet approach [30] estimates flow in a classical spatial pyramid
framework by warping one of the input images and predicting the residual flow at each pyramid level.
A few attempts have recently been made to learn optical flow from unlabeled videos in an unsupervised
manner. The USCNN method [1] approximates the brightness constancy with a Taylor series
expansion and trains a deep network using the UCF101 dataset [35]. Yu et al. [40] enables the backpropagation of the warping function using the bilinear sampling layer from the spatial transformer
network [18] and explicitly optimizes the brightness constancy and spatial smoothness assumptions.
While Yu et al. [40] demonstrate comparable performance with the FlowNet on the KITTI dataset,
the method requires significantly more sophisticated data augmentation techniques and different
parameter settings for each dataset. Our approach differs from these methods in that we use both
labeled and unlabeled data to learn optical flow in a semi-supervised framework.
Semi-supervised learning. Several methods combine the classification objective with unsupervised
reconstruction losses for image recognition [31, 41]. In low-level vision tasks, Kuznietsov et al. [21]
train a deep CNN using sparse ground truth data for single-image depth estimation. This method
optimizes a supervised loss for pixels with ground truth depth value as well as an unsupervised
image alignment cost and a regularization cost. The image alignment cost resembles the brightness
constancy, and the regularization cost enforces the spatial smoothness on the predicted depth maps.
We show that adopting a similar idea to combine the EPE loss with image reconstruction and
smoothness losses may not improve flow accuracy. Instead, we use the adversarial training scheme
for learning to model the structural flow warp error without making assumptions on images or flow.
Generative adversarial networks. The GAN framework [12] has been successfully applied to
numerous problems, including image generation [7, 38], image inpainting [28], face completion [23],
image super-resolution [22], semantic segmentation [24], and image-to-image translation [17, 42].
Within the scope of domain adaptation [9, 14], the discriminator learns to differentiate the features
from the two different domains, e.g., synthetic, and real images. Kozi?nski et al. [20] adopt the
adversarial training framework for semi-supervised learning on the image segmentation task where the
discriminator is trained to distinguish between the predictions produced from labeled and unlabeled
data. Different from Kozi?nski et al. [20], our discriminator learns to distinguish the flow warp errors
between using the ground truth flow and using the estimated flow. The generator thus learns to model
the spatial structure of flow warp error images and can improve flow estimation accuracy around
motion boundaries.
3
3
Semi-Supervised Optical Flow Estimation
In this section, we describe the semi-supervised learning approach for optical flow estimation, the
design methodology of the proposed generative adversarial network for learning the flow warp error,
and the use of the adversarial loss to leverage labeled and unlabeled data.
3.1
Semi-supervised learning
We address the problem of learning optical flow by using both labeled data (i.e., with the ground truth
dense optical flow) and unlabeled data (i.e., raw videos). Given a pair of input images {I1 , I2 }, we
train a deep network to generate the dense optical flow field f = [u, v]. For labeled data with the
ground truth optical flow (denoted by f? = [?
u, v?]), we optimize the EPE loss between the predicted
and ground truth flow:
q
2
2
LEPE (f, f?) = (u ? u
?) + (v ? v?) .
(1)
For unlabeled data, existing work [40] makes use of the classical brightness constancy and spatial
smoothness to define the image warping loss and flow smoothness loss:
Lwarp (I1 , I2 , f ) = ? (I1 ? W (I2 , f )) ,
Lsmooth (f ) = ?(?x u) + ?(?y u) + ?(?x v) + ?(?y v),
(2)
(3)
where ?x and ?y are horizontal and vertical gradient operators and ?(?) is the robust penalty function.
The warping function W (I2 , f ) uses the bilinear sampling [18] to warp I2 according to the flow
field f . The difference I1 ? W (I2 , f ) is the flow warp error as shown in Figure 2. Minimizing
Lwarp (I1 , I2 , f ) enforces the flow warp error to be close to zero at every pixel.
A baseline semi-supervised learning approach is to minimize LEPE for labeled data and minimize
Lwarp and Lsmooth for unlabeled data:
X
X
(j) (j)
LEPE f (i) , f?(i) +
?w Lwarp I1 , I2 , f (j) + ?s Lsmooth f (j) ,
(4)
i?Dl
j?Du
where Dl and Du represent labeled and unlabeled datasets, respectively. However, the commonly
used robust loss functions (e.g., Lorentzian and Charbonnier) assume that the error is independent
at each pixel and thus cannot model the structural patterns of flow warp error caused by occlusion.
Minimizing the combination of the supervised loss in (1) and unsupervised losses in (2) and (3)
may degrade the flow accuracy, especially when large motion present in the input image pair. As a
result, instead of using the unsupervised losses based on classical assumptions, we propose to impose
an adversarial loss on the flow warp images within a generative adversarial network. We use the
adversarial loss to regularize the flow estimation for both labeled and unlabeled data.
3.2
Adversarial training
Training a GAN involves optimizing the two networks: a generator G and a discriminator D. The
generator G takes a pair of input images to generate optical flow. The discriminator D performs
binary classification to distinguish whether a flow warp error image is produced by the estimated
flow from the generator G or by the ground truth flow. We denote the flow warp error image from the
ground truth flow and generated flow by y? = I1 ? W(I2 , f?) and y = I1 ? W(I2 , f ), respectively.
The objective function to train the GAN can be expressed as:
Ladv (y, y?) = Ey?[log D(?
y )] + Ey [log (1 ? D(y))].
(5)
We incorporate the adversarial loss with the supervised EPE loss and solve the following minmax
problem for optimizing G and D:
min max LEPE (G) + ?adv Ladv (G, D),
G
D
(6)
where ?adv controls the relative importance of the adversarial loss for optical flow estimation.
Following the standard procedure for GAN training, we alternate between the following two steps
to solve (6): (1) update the discriminator D while holding the generator G fixed and (2) update
generator G while holding the discriminator D fixed.
4
Frozen
Updated
Labeled data
Predicted flow
Flow warp error
Ground truth
flow
Ground truth
flow warp error
%
? "#$
(Eq. 7)
Generator G
Discriminator D
(a) Update discriminator D using labeled data
? "#" (Eq. 1)
Updated
Frozen
Predicted flow
Labeled data
Ground truth flow
Flow warp error
'
? $%&
(Eq. 9)
Generator G
Discriminator D
Predicted flow
Unlabeled data
Flow warp error
(b) Update generator G using both labeled and unlabeled data
Figure 3: Adversarial training procedure. Training a generative adversarial network involves the
alternative optimization of the discriminator D and generator G.
Updating discriminator D. We train the discriminator D to classify between the ground truth
flow warp error (real samples, labeled as 1) and the flow warp error from the predicted flow (fake
samples, labeled as 0). The maximization of (5) is equivalent to minimizing the binary cross-entropy
loss LBCE (p, t) = ?t log(p) ? (1 ? t) log(1 ? p) where p is the output from the discriminator and t
is the target label. The adversarial loss for updating D is defined as:
LD
?) = LBCE (D(?
y ), 1) + LBCE (D(y), 0)
adv (y, y
= ? log D(?
y ) ? log(1 ? D(y)).
(7)
As the ground truth flow is required to train the discriminator, only the labeled data Dl is involved in
this step. By fixing G in (6), we minimize the following loss function for updating D:
X
(i) (i)
LD
? ).
(8)
adv (y , y
i?Dl
Updating generator G. The goal of the generator is to ?fool? the discriminator by producing flow
to generate realistic flow warp error images. Optimizing (6) with respect to G becomes minimizing
log(1 ? D(y)). As suggested by Goodfellow et al. [12], one can instead minimize ? log(D(y)) to
speed up the convergence. The adversarial loss for updating G is then equivalent to the binary cross
entropy loss that assigns label 1 to the generated flow warp error y:
LG
(9)
adv (y) = LBCE (D(y), 1) = ? log(D(y)).
By combining the adversarial loss with the supervised EPE loss, we minimize the following function
for updating G:
X
X
LEPE f (i) , f?(i) + ?adv LG (y (i) ) +
?adv LG (y (j) ).
(10)
adv
i?Dl
adv
j?Du
We note that the adversarial loss is computed for both labeled and unlabeled data, and thus guides
the flow estimation for image pairs without the ground truth flow. Figure 3 illustrates the two main
steps to update the generator D and the discriminator G in the proposed semi-supervised learning
framework.
5
3.3
Network architecture and implementation details
Generator. We construct a 5-level SPyNet [30] as our generator. Instead of using simple stacks
of convolutional layers as sub-networks [30], we choose the encoder-decoder architecture with skip
connections to effectively increase the receptive fields. Each convolutional layer has a 3 ? 3 spatial
support and is followed by a ReLU activation. We present the details of our SPyNet architecture in
the supplementary material.
Discriminator. As we aim to learn the local structure of flow warp error at motion boundaries, it
is more effective to penalize the structure at the scale of local patches instead of the whole image.
Therefore, we use the PatchGAN [17] architecture as our discriminator. The PatchGAN is a fully
convolutional classifier that classifies whether each N ? N overlapping patch is real or fake. The
PatchGAN has a receptive field of 47 ? 47 pixels.
Implementation details. We implement the proposed method using the Torch framework [6]. We
use the Adam solver [19] to optimize both the generator and discriminator with ?1 = 0.9, ?2 = 0.999
and the weight decay of 1e ? 4. We set the initial learning rate as 1e ? 4 and then multiply by 0.5
every 100k iterations after the first 200k iterations. We train the network for a total of 600k iterations.
We use the FlyingChairs dataset [8] as the labeled dataset and the KITTI raw videos [10] as the
unlabeled dataset. In each mini-batch, we randomly sample 4 image pairs from each dataset. We
randomly augment the training data in the following ways: (1) Scaling between [1, 2], (2) Rotating
within [?17? , 17? ], (3) Adding Gaussian noise with a sigma of 0.1, (4) Using color jitter with respect
to brightness, contrast and saturation uniformly sampled from [0, 0.04]. We then crop images to
384 ? 384 patches and normalize by the mean and standard deviation computed from the ImageNet
dataset [13]. The source code is publicly available on http://vllab.ucmerced.edu/wlai24/
semiFlowGAN.
4
Experimental Results
We evaluate the performance of optical flow estimation on five benchmark datasets. We conduct
ablation studies to analyze the contributions of individual components and present comparisons with
the state-of-the-art algorithms including classical variational algorithms and CNN-based approaches.
4.1
Evaluated datasets and metrics
We evaluate the proposed optical flow estimation method on the benchmark datasets: MPI-Sintel [5],
KITTI 2012 [11], KITTI 2015 [27], Middlebury [3] and the test set of FlyingChairs [8]. The MPISintel and FlyingChairs are synthetic datasets with dense ground truth flow. The Sintel dataset
provides two rendered sets, Clean and Final, that contain both small displacements and large motion.
The training and test sets contain 1041 and 552 image pairs, respectively. The FlyingChairs test set
is composed of 640 image pairs with similar motion statistics to the training set. The Middlebury
dataset has only eight image pairs with small motion. The images from the KITTI 2012 and KITTI
2015 datasets are collected from driving real-world scenes with large forward motion. The ground
truth optical flow is obtained from a 3D laser scanner and thus only covers about 50% of image pixels.
There are 194 image pairs in the KITTI 2012 dataset, and 200 image pairs in the KITTI 2015 dataset.
We compute the average EPE (1) on pixels with the ground truth flow available for each dataset. On
the KITTI-2015 dataset, we also compute the Fl score [27], which is the ratio of pixels that have EPE
greater than 3 pixels and 5% of the ground truth value.
4.2
Ablation study
We conduct ablation studies to analyze the contributions of the adversarial loss and the proposed
semi-supervised learning with different training schemes.
Adversarial loss. We adjust the weight of the adversarial loss ?adv in (10) to validate the effect
of the adversarial training. When ?adv = 0, our method falls back to the fully supervised learning
setting. We show the quantitative evaluation in Table 1. Using larger values of ?adv may decrease the
performance and cause visual artifacts as shown in Figure 4. We therefore choose ?adv = 0.01.
6
Table 1: Analysis on adversarial loss. We train the proposed model using different weights for the
adversarial loss in (10).
?adv
Sintel-Clean
EPE
Sintel-Final
EPE
KITTI 2012
EPE
0
0.01
0.1
1
3.51
3.30
3.57
3.93
4.70
4.68
4.73
5.18
7.69
7.16
8.25
13.89
KITTI 2015
EPE
Fl-all
17.19
16.02
16.82
21.07
40.82%
38.77%
42.78%
63.43%
FlyingChairs
EPE
2.15
1.95
2.11
2.21
Table 2: Analysis on receptive field of discriminator. We vary the number of strided convolutional
layers in the discriminator to achieve different size of receptive fields.
# Strided
convolutions
Receptive field
Sintel-Clean
EPE
Sintel-Final
EPE
KITTI 2012
EPE
d=2
d=3
d=4
23 ? 23
47 ? 47
95 ? 95
3.66
3.30
3.70
4.90
4.68
5.00
7.38
7.16
7.54
KITTI 2015
EPE
Fl-all
16.28
16.02
16.38
40.19%
38.77%
41.52%
FlyingChairs
EPE
2.15
1.95
2.16
Receptive fields of discriminator. The receptive field of the discriminator is equivalent to the size
of patches used for classification. The size of the receptive field is determined by the number of
strided convolutional layers, denoted by d. We test three different values, d = 2, 3, 4, which are
corresponding to the receptive field of 23?23, 47?47, and 95?95, respectively. As shown in Table 2,
the network with d = 3 performs favorably against other choices on all benchmark datasets. Using
too large or too small patch sizes might not be able to capture the structure of flow warp error well.
Therefore, we design our discriminator to have a receptive field of 47 ? 47 pixels.
Training schemes. We train the same network (i.e., our generator G) with the following training
schemes: (a) Supervised: minimizing the EPE loss (1) on the FlyingChairs dataset. (b) Unsupervised:
minimizing the classical brightness constancy (2) and spatial smoothness (3) using the Charbonnier
loss function on the KITTI raw dataset. (c) Baseline semi-supervised: minimizing the combination
of supervised and unsupervised losses (4) on the FlyingChairs and KITTI raw datasets. For the
semi-supervised setting, we evaluate different combinations of ?w and ?s in Table 3. We note that it
is not easy to run grid search to find the best parameter combination for all evaluated datasets. We
choose ?w = 1 and ?s = 0.01 for the baseline semi-supervised and unsupervised settings.
We provide the quantitative evaluation of the above training schemes in Table 4 and visual comparisons
in Figure 5 and 6. As images in KITTI 2015 have large forward motion, there are large occluded/disoccluded regions, particularly on the image and moving object boundaries. The brightness constancy
does not hold in these regions. Consequently, minimizing the image warping loss (2) results in
inaccurate flow estimation. Compared to the fully supervised learning, our method further refines the
motion boundaries by modeling the flow warp error. By incorporating both labeled and unlabeled
data in training, our method effectively reduces EPEs on the KITTI 2012 and 2015 datasets.
Training on partially labeled data. We further analyze the effect of the proposed semi-supervised
method by reducing the amount of labeled training data. Specifically, we use 75%, 50% and 25%
Input images
?adv = 0
?adv = 0.01
Ground truth flow
?adv = 0.1
?adv = 1
Figure 4: Comparisons of adversarial loss ?adv . Using larger value of ?adv does not necessarily
improve the performance and may cause unwanted visual artifacts.
7
Table 3: Evaluation for baseline semi-supervised setting. We test different combinations of ?w
and ?s in (4). We note that it is difficult to find the best parameters for all evaluated datasets.
?w
?s
Sintel-Clean
EPE
Sintel-Final
EPE
KITTI 2012
EPE
1
1
1
0.1
0.01
0
0.1
0.01
0.01
0.01
3.77
3.75
3.69
3.64
3.57
5.02
5.05
4.86
4.81
4.82
10.90
11.82
10.38
10.15
8.63
KITTI 2015
EPE
Fl-all
18.52
19.98
18.07
18.94
18.87
FlyingChairs
EPE
39.94%
43.18%
39.33%
40.85 %
42.63 %
2.25
2.19
2.11
2.17
2.22
Table 4: Analysis on different training schemes. ?Chairs? represents the FlyingChairs dataset and
?KITTI? denotes the KITTI raw dataset. The baseline semi-supervised settings cannot improve the
flow accuracy as the brightness constancy assumption does not hold on occluded regions. In contrast,
our approach effectively utilizes the unlabeled data to improve the performance.
Method
Training Datasets
Sintel-Clean
EPE
Sintel-Final
EPE
KITTI 2012
EPE
Supervised
Unsupervised
Baseline semi-supervised
Proposed semi-supervised
Chairs
KITTI
Chairs + KITTI
Chairs + KITTI
3.51
8.01
3.69
3.30
4.70
8.97
4.86
4.68
7.69
16.54
10.38
7.16
KITTI 2015
EPE
Fl
17.19
25.53
18.07
16.02
40.82%
54.40%
39.33%
38.77%
FlyingChairs
EPE
2.15
6.66
2.11
1.95
of labeled data with ground truth flow from the FlyingChairs dataset and treat the remaining part as
unlabeled data to train the proposed semi-supervised method. We also train the purely supervised
method with the same amount of labeled data for comparisons. Table 5 shows that the proposed semisupervised method consistently outperforms the purely supervised method on the Sintel, KITTI2012
and KITTI2015 datasets. The performance gap becomes larger when using less labeled data, which
demonstrates the capability of the proposed method on utilizing the unlabeled data.
4.3 Comparisons with the state-of-the-arts
In Table 6, we compare the proposed algorithm with four variational methods: EpicFlow [32],
DeepFlow [39], LDOF [4] and FlowField [2], and four CNN-based algorithms: FlowNetS [8],
FlowNetC [8], SPyNet [30] and FlowNet 2.0 [16]. We further fine-tune our model on the Sintel
training set (denoted by ?+ft?) and compare with the fine-tuned results of FlowNetS, FlowNetC,
SPyNet, and FlowNet2. We note that the SPyNet+ft is also fine-tuned on the Driving dataset [26]
for evaluating on the KITTI2012 and KITTI2015 datasets, while other methods are fine-tuned on
the Sintel training data. The FlowNet 2.0 has significantly more network parameters and uses more
training datasets (e.g., FlyingThings3D [26]) to achieve the state-of-the-art performance. We show
that our model achieves competitive performance with the FlowNet and SPyNet when using the same
amount of ground truth flow (i.e., FlyingChairs and Sintel datasets). We present more qualitative
comparisons with the state-of-the-art methods in the supplementary material.
4.4 Limitations
As the images in the KITTI raw dataset are captured in driving scenes and have a strong prior of
forward camera motion, the gain of our semi-supervised learning over the supervised setting is mainly
on the KITTI 2012 and 2015 datasets. In contrast, the Sintel dataset typically has moving objects
with various types of motion. Exploring different types of video datasets, e.g., UCF101 [35] or
DAVIS [29], as the source of unlabeled data in our semi-supervised learning framework is a promising
future direction to improve the accuracy on general scenes.
Table 5: Training on partial labeled data. We use 75%, 50% and 25% of data with ground truth
flow from the FlyingChair dataset as labeled data and treat the remaining part as unlabeled data. The
proposed semi-supervised method consistently outperforms the purely supervised method.
Method
Amount of
labeled data
Sintel-Clean
EPE
Sintel-Final
EPE
KITTI 2012
EPE
Supervised
Proposed semi-supervised
75%
4.35
3.58
5.40
4.81
8.22
7.30
17.43
16.46
41.62%
41.00%
1.96
2.20
Supervised
Proposed semi-supervised
50%
4.48
3.67
5.46
4.92
9.34
7.39
18.71
16.64
42.14%
40.48%
2.04
2.28
Supervised
Proposed semi-supervised
25%
4.91
3.95
5.78
5.00
10.60
7.40
19.90
16.61
43.79%
40.68%
2.09
2.33
8
KITTI 2015
EPE
Fl-all
FlyingChairs
EPE
Input images
Unsupervised
Supervised
Ground truth flow
Baseline semi-supervised
Proposed semi-supervised
Figure 5: Comparisons of training schemes. The proposed method learns the flow warp error
using the adversarial training and improve the flow accuracy on motion boundary.
Baseline semi-supervised
Ground truth
Proposed semi-supervised
Figure 6: Comparisons of flow warp error. The baseline semi-supervised approach penalizes the
flow warp error on occluded regions and thus produce inaccurate flow.
Table 6: Comparisons with state-of-the-arts. We report the average EPE on six benchmark
datasets and the Fl score on the KITTI 2015 dataset.
Middlebury
Train
EPE
Method
5
Sintel-Clean
Train
Test
EPE
EPE
Sintel-Final
Train
Test
EPE
EPE
KITTI 2012
Train
Test
EPE
EPE
Train
EPE
KITTI 2015
Train
Test
Fl-all
Fl-all
Chairs
Test
EPE
EpicFlow
DeepFlow
LDOF
FlowField
[32]
[39]
[4]
[2]
0.31
0.25
0.44
0.27
2.27
2.66
4.64
1.86
4.12
5.38
7.56
3.75
3.57
4.40
5.96
3.06
6.29
7.21
9.12
5.81
3.47
4.58
10.94
3.33
3.8
5.8
12.4
3.5
9.27
10.63
18.19
8.33
27.18%
26.52%
38.11%
24.43%
27.10%
29.18%
-
2.94
3.53
3.47
-
FlowNetS
FlowNetC
SpyNet
FlowNet2
[8]
[8]
[30]
[16]
1.09
1.15
0.33
0.35
4.50
4.31
4.12
2.02
7.42
7.28
6.69
3.96
5.45
5.87
5.57
3.14
8.43
8.81
8.43
6.02
8.26
9.35
9.12
4.09
-
15.44
12.52
20.56
10.06
52.86%
47.93%
44.78%
30.37%
-
2.71
2.19
2.63
1.68
FlowNetS + ft [8]
FlowNetC + ft [8]
SpyNet + ft [30]
FlowNet2 + ft [16]
0.98
0.93
0.33
0.35
(3.66)
(3.78)
(3.17)
(1.45)
6.96
6.85
6.64
4.16
(4.44)
(5.28)
(4.32)
(2.01)
7.76
8.51
8.36
5.74
7.52
8.79
4.13
3.61
9.10
4.7
-
9.84
28.20%
-
3.04
2.27
3.07
-
Ours
Ours + ft
0.37
0.32
3.30
(2.41)
6.28
6.27
4.68
(3.16)
7.61
7.31
7.16
5.23
7.5
6.8
16.02
14.69
38.77%
30.30%
39.71%
31.01 %
1.95
2.41
Conclusions
In this work, we propose a generative adversarial network for learning optical flow in a semisupervised manner. We use a discriminative network and an adversarial loss to learn the structural
patterns of the flow warp error without making assumptions on brightness constancy and spatial
smoothness. The adversarial loss serves as guidance for estimating optical flow from both labeled and
unlabeled datasets. Extensive evaluations on benchmark datasets validate the effect of the adversarial
loss and demonstrate that the proposed method performs favorably against the purely supervised and
the straightforward semi-supervised learning approaches for learning optical flow.
Acknowledgement
This work is supported in part by the NSF CAREER Grant #1149783, gifts from Adobe and NVIDIA.
9
References
[1] A. Ahmadi and I. Patras. Unsupervised convolutional neural networks for motion estimation.
In ICIP, 2016.
[2] C. Bailer, B. Taetz, and D. Stricker. Flow fields: Dense correspondence fields for highly accurate
large displacement optical flow estimation. In ICCV, 2015.
[3] S. Baker, D. Scharstein, J. Lewis, S. Roth, M. J. Black, and R. Szeliski. A database and
evaluation methodology for optical flow. IJCV, 92(1):1?31, 2011.
[4] T. Brox and J. Malik. Large displacement optical flow: descriptor matching in variational
motion estimation. TPAMI, 33(3):500?513, 2011.
[5] D. J. Butler, J. Wulff, G. B. Stanley, and M. J. Black. A naturalistic open source movie for
optical flow evaluation. In ECCV, 2012.
[6] R. Collobert, K. Kavukcuoglu, and C. Farabet. Torch7: A Matlab-like environment for machine
learning. In BigLearn, NIPS Workshop, 2011.
[7] E. L. Denton, S. Chintala, and R. Fergus. Deep generative image models using a laplacian
pyramid of adversarial networks. In NIPS, 2015.
[8] P. Fischer, A. Dosovitskiy, E. Ilg, P. H?usser, C. Haz?rba?s, V. Golkov, P. van der Smagt,
D. Cremers, and T. Brox. FlowNet: Learning optical flow with convolutional networks. In
ICCV, 2015.
[9] Y. Ganin, E. Ustinova, H. Ajakan, P. Germain, H. Larochelle, F. Laviolette, M. Marchand, and
V. Lempitsky. Domain-adversarial training of neural networks. Journal of Machine Learning
Research, 17(59):1?35, 2016.
[10] A. Geiger, P. Lenz, C. Stiller, and R. Urtasun. Vision meets robotics: The KITTI dataset. The
International Journal of Robotics Research, 32(11):1231?1237, 2013.
[11] A. Geiger, P. Lenz, and R. Urtasun. Are we ready for autonomous driving? The KITTI vision
benchmark suite. In CVPR, 2012.
[12] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and
Y. Bengio. Generative adversarial nets. In NIPS, 2014.
[13] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. In CVPR,
2016.
[14] J. Hoffman, D. Wang, F. Yu, and T. Darrell. FCNs in the wild: Pixel-level adversarial and
constraint-based adaptation. arXiv, 2016.
[15] B. K. Horn and B. G. Schunck. Determining optical flow. Artificial intelligence, 17(1-3):185?
203, 1981.
[16] E. Ilg, N. Mayer, T. Saikia, M. Keuper, A. Dosovitskiy, and T. Brox. FlowNet 2.0: Evolution of
optical flow estimation with deep networks. In CVPR, 2017.
[17] P. Isola, J.-Y. Zhu, T. Zhou, and A. A. Efros. Image-to-image translation with conditional
adversarial networks. In CVPR, 2017.
[18] M. Jaderberg, K. Simonyan, and A. Zisserman. Spatial transformer networks. In NIPS, 2015.
[19] D. Kingma and J. Ba. ADAM: A method for stochastic optimization. In ICLR, 2015.
[20] M. Kozi?nski, L. Simon, and F. Jurie. An adversarial regularisation for semi-supervised training
of structured output neural networks. arXiv, 2017.
[21] Y. Kuznietsov, J. St?ckler, and B. Leibe. Semi-supervised deep learning for monocular depth
map prediction. In CVPR, 2017.
[22] C. Ledig, L. Theis, F. Husz?r, J. Caballero, A. Cunningham, A. Acosta, A. Aitken, A. Tejani,
J. Totz, Z. Wang, and W. Shi. Photo-realistic single image super-resolution using a generative
adversarial network. In CVPR, 2017.
[23] Y. Li, S. Liu, J. Yang, and M.-H. Yang. Generative face completion. In CVPR, 2017.
[24] P. Luc, C. Couprie, S. Chintala, and J. Verbeek. Semantic segmentation using adversarial
networks. In NIPS Workshop on Adversarial Training, 2016.
[25] B. D. Lucas and T. Kanade. An iterative image registration technique with an application to
stereo vision. In International Joint Conference on Artificial Intelligence, 1981.
[26] N. Mayer, E. Ilg, P. H?usser, P. Fischer, D. Cremers, A. Dosovitskiy, and T. Brox. A large
dataset to train convolutional networks for disparity, optical flow, and scene flow estimation. In
CVPR, 2016.
[27] M. Menze and A. Geiger. Object scene flow for autonomous vehicles. In CVPR, 2015.
10
[28] D. Pathak, P. Krahenbuhl, J. Donahue, T. Darrell, and A. A. Efros. Context encoders: Feature
learning by inpainting. In CVPR, 2016.
[29] F. Perazzi, J. Pont-Tuset, B. McWilliams, L. Van Gool, M. Gross, and A. Sorkine-Hornung. A
benchmark dataset and evaluation methodology for video object segmentation. In CVPR, 2016.
[30] A. Ranjan and M. J. Black. Optical flow estimation using a spatial pyramid network. In CVPR,
2017.
[31] A. Rasmus, M. Berglund, M. Honkala, H. Valpola, and T. Raiko. Semi-supervised learning with
ladder networks. In NIPS, 2015.
[32] J. Revaud, P. Weinzaepfel, Z. Harchaoui, and C. Schmid. EpicFlow: Edge-preserving interpolation of correspondences for optical flow. In CVPR, 2015.
[33] D. Rosenbaum and Y. Weiss. Beyond brightness constancy: Learning noise models for optical
flow. arXiv, 2016.
[34] D. Rosenbaum, D. Zoran, and Y. Weiss. Learning the local statistics of optical flow. In NIPS,
2013.
[35] K. Soomro, A. R. Zamir, and M. Shah. UCF101: A dataset of 101 human actions classes from
videos in the wild. CRCV-TR-12-01, 2012.
[36] D. Sun, S. Roth, and M. J. Black. A quantitative analysis of current practices in optical flow
estimation and the principles behind them. IJCV, 106(2):115?137, 2014.
[37] D. Sun, S. Roth, J. Lewis, and M. Black. Learning optical flow. In ECCV, 2008.
[38] X. Wang and A. Gupta. Generative image modeling using style and structure adversarial
networks. In ECCV, 2016.
[39] P. Weinzaepfel, J. Revaud, Z. Harchaoui, and C. Schmid. DeepFlow: Large displacement optical
flow with deep matching. In ICCV, 2013.
[40] J. J. Yu, A. W. Harley, and K. G. Derpanis. Back to basics: Unsupervised learning of optical
flow via brightness constancy and motion smoothness. In ECCV Workshops, 2016.
[41] Y. Zhang, K. Lee, and H. Lee. Augmenting supervised neural networks with unsupervised
objectives for large-scale image classification. In ICML, 2016.
[42] J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros. Unpaired image-to-image translation using
cycle-consistent adversarial networks. In ICCV, 2017.
11
| 6639 |@word cnn:11 fcns:1 open:1 brightness:26 inpainting:2 tr:1 ld:2 initial:1 minmax:1 series:1 score:2 liu:1 disparity:1 tuned:3 ours:2 outperforms:3 existing:4 current:1 activation:1 refines:1 realistic:3 blur:1 enables:1 update:5 generative:16 intelligence:2 advancement:1 provides:1 zhang:2 five:1 qualitative:1 ijcv:2 combine:2 wild:2 manner:4 aitken:1 rapid:1 ming:1 solver:1 becomes:2 gift:1 classifies:1 estimating:1 baker:1 minimizes:1 developed:1 unified:1 keuper:1 suite:1 quantitative:3 every:2 growth:1 unwanted:1 classifier:1 demonstrates:1 ustinova:1 control:1 haz:1 grant:1 mcwilliams:1 producing:1 local:4 treat:2 despite:1 bilinear:3 middlebury:4 meet:1 interpolation:2 ajakan:1 might:1 black:5 resembles:1 ucmerced:2 limited:2 jurie:1 horn:1 camera:1 enforces:2 practice:1 implement:1 differs:1 backpropagation:1 procedure:2 displacement:5 lorentzian:3 significantly:2 matching:3 ucf101:3 naturalistic:1 cannot:4 unlabeled:29 close:1 operator:1 context:2 transformer:2 optimize:2 equivalent:3 map:2 roth:3 shi:1 ranjan:1 straightforward:2 primitive:1 convex:1 patchgan:3 formulate:1 resolution:2 hsuan:1 assigns:1 pouget:1 insight:1 utilizing:1 regularize:1 handle:1 autonomous:2 updated:2 target:1 us:2 goodfellow:2 expensive:1 recognition:2 updating:6 particularly:1 merced:1 predicts:1 labeled:33 database:1 constancy:23 ft:7 wang:3 capture:4 zamir:1 region:4 revaud:2 adv:20 sun:5 cycle:1 decrease:2 gross:1 environment:1 complexity:1 occluded:3 warde:1 trained:2 zoran:1 solving:2 purely:6 upon:1 joint:1 various:1 epe:45 regularizer:1 train:21 laser:1 describe:1 effective:1 artificial:2 supplementary:2 solve:3 larger:3 cvpr:13 hornung:1 encoder:1 statistic:3 fischer:2 simonyan:1 final:7 differentiate:1 sequence:1 tpami:1 frozen:2 net:1 propose:6 reconstruction:2 huang2:1 adaptation:2 combining:1 ablation:3 achieve:2 validate:2 normalize:1 convergence:1 darrell:2 produce:2 stiller:1 adam:2 object:5 kitti:36 develop:1 completion:2 augmenting:1 fixing:1 ganin:1 eq:3 strong:2 predicted:11 rosenbaum:4 involves:2 skip:1 larochelle:1 direction:1 cnns:4 filter:1 stochastic:1 human:1 material:2 bin:1 require:1 exploring:1 hold:5 scanner:1 sufficiently:1 around:1 ground:38 caballero:1 great:1 scope:1 predict:3 driving:4 efros:3 vary:1 adopt:1 achieves:1 estimation:26 lenz:2 label:2 honkala:1 sensitive:1 ilg:4 successfully:1 hoffman:1 biglearn:1 gaussian:3 rba:1 aim:2 super:2 husz:1 zhou:1 focus:1 consistently:2 likelihood:3 mainly:1 tech:1 contrast:4 adversarial:52 baseline:13 inference:1 inaccurate:2 typically:2 lbce:4 torch:1 cunningham:1 smagt:1 pont:1 i1:8 pixel:13 classification:4 denoted:3 augment:1 lucas:1 spatial:21 art:5 weinzaepfel:2 brox:4 field:20 construct:1 beach:1 sampling:3 represents:1 park:1 yu:4 unsupervised:15 denton:1 icml:1 photometric:1 future:1 report:1 mirza:1 dosovitskiy:3 few:2 strided:3 randomly:2 composed:1 individual:1 phase:1 occlusion:5 attempt:1 harley:1 interest:1 highly:1 multiply:1 evaluation:7 adjust:1 alignment:2 mixture:1 farley:1 behind:1 accurate:1 deepflow:3 edge:2 capable:1 partial:1 conduct:2 taylor:1 rotating:1 penalizes:1 guidance:2 classify:1 modeling:4 cover:1 measuring:1 maximization:1 stacking:1 cost:4 deviation:1 virginia:1 too:2 encoders:1 synthetic:5 nski:3 st:2 fundamental:2 international:2 lee:2 augmentation:2 reflect:1 flownet:8 choose:3 berglund:1 resort:1 warped:1 expert:1 style:1 li:1 cremers:2 caused:2 explicitly:1 collobert:1 vehicle:1 spynet:9 analyze:3 competitive:1 flowfield:2 capability:2 simon:1 jia:1 contribution:3 minimize:6 publicly:1 accuracy:7 convolutional:10 descriptor:2 efficiently:2 tejani:1 raw:6 kavukcuoglu:1 produced:2 ren:1 foe:1 farabet:1 against:3 involved:1 chintala:2 sampled:1 gain:1 dataset:29 ledig:1 color:1 usser:2 stanley:1 segmentation:4 sorkine:1 sophisticated:1 back:2 feed:2 supervised:67 methodology:3 zisserman:1 wei:4 totz:1 formulation:2 evaluated:3 though:1 furthermore:1 sheng:1 hand:2 horizontal:1 overlapping:1 lack:1 artifact:2 semisupervised:2 usa:1 effect:4 contain:2 evolution:1 regularization:2 semantic:2 i2:10 indistinguishable:1 during:1 encourages:1 davis:1 mpi:1 demonstrate:5 ladv:2 performs:4 motion:24 image:63 variational:4 recently:2 extend:1 he:1 approximates:1 smoothness:18 grid:1 moving:2 entail:1 recent:3 optimizing:4 optimizes:2 menze:1 nvidia:2 binary:3 vt:1 inconsistency:2 der:1 preserving:2 captured:2 greater:1 impose:2 isola:2 ey:2 semi:46 multiple:1 harchaoui:2 infer:1 reduces:1 match:1 cross:2 long:1 laplacian:3 adobe:1 prediction:2 verbeek:1 crop:1 basic:1 vision:5 metric:1 arxiv:3 iteration:3 sometimes:1 adopting:2 represent:1 pyramid:4 robotics:2 penalize:1 epes:1 fine:4 source:3 flow:158 leveraging:3 structural:6 near:2 leverage:3 golkov:1 yang:2 bengio:1 easy:1 fit:1 relu:1 architecture:4 idea:2 whether:2 six:1 torch7:1 penalty:1 stereo:1 soomro:1 cause:2 action:1 matlab:1 deep:11 fake:2 fool:1 involve:1 tune:1 amount:8 unpaired:1 generate:3 http:1 epicflow:3 nsf:1 estimated:2 per:2 key:1 four:2 gmm:2 clean:7 registration:1 flownet2:3 run:1 jitter:1 utilizes:2 patch:7 geiger:3 sintel:20 scaling:1 krahenbuhl:1 comparable:1 layer:6 fl:9 followed:1 distinguish:4 courville:1 correspondence:3 marchand:1 refine:1 constraint:1 scene:6 speed:2 min:1 chair:5 optical:50 rendered:1 saikia:1 structured:1 according:1 alternate:1 combination:5 acosta:1 making:6 iccv:4 computationally:1 monocular:1 discus:1 end:5 serf:2 photo:1 available:3 eight:1 leibe:1 hierarchical:1 alternative:2 batch:1 ahmadi:1 shah:1 denotes:1 remaining:2 include:1 gan:5 laviolette:1 exploit:1 build:1 especially:1 classical:12 warping:5 objective:4 malik:1 quantity:1 receptive:10 gradient:1 iclr:1 valpola:1 decoder:1 degrade:1 perazzi:1 collected:1 urtasun:2 ozair:1 code:1 modeled:1 mini:1 ratio:1 minimizing:8 rasmus:1 lg:3 difficult:1 holding:2 favorably:3 sigma:1 negative:2 ba:1 design:2 implementation:2 vertical:1 convolution:1 datasets:27 benchmark:8 stack:1 pair:11 required:2 germain:1 extensive:2 connection:1 discriminator:26 imagenet:1 icip:1 mayer:2 california:1 learned:1 kingma:1 nip:8 address:1 able:1 suggested:1 beyond:1 pattern:7 saturation:1 including:3 max:1 video:9 gool:1 unrealistic:2 pathak:1 natural:2 rely:3 predicting:1 residual:2 zhu:2 scheme:9 improve:7 movie:1 ladder:1 numerous:1 raiko:1 ready:1 schmid:2 prior:7 review:1 acknowledgement:1 theis:1 determining:1 relative:1 regularisation:1 loss:54 fully:4 generation:1 limitation:1 generator:20 proxy:1 consistent:1 principle:1 translation:3 eccv:4 supported:1 dis:1 guide:1 warp:47 szeliski:1 fall:1 face:2 sparse:2 van:2 boundary:8 depth:4 world:5 evaluating:1 tuset:1 forward:5 commonly:2 refinement:1 made:1 mhyang:1 cope:1 scharstein:1 jaderberg:1 consuming:2 discriminative:1 butler:1 fergus:1 search:1 iterative:1 table:12 promising:1 learn:9 kanade:1 robust:5 ca:1 career:1 expansion:1 du:3 complex:1 necessarily:1 domain:3 charbonnier:4 dense:5 main:2 whole:1 noise:3 arise:1 derpanis:1 wulff:1 xu:1 crafted:1 sub:1 explicit:2 learns:6 stricker:1 donahue:1 decay:1 abadie:1 gupta:1 dl:5 incorporating:1 workshop:3 adding:1 effectively:3 importance:1 illumination:1 illustrates:1 gap:1 patras:1 yang1:1 entropy:2 explore:1 visual:3 schunck:1 expressed:1 partially:1 truth:38 lewis:2 conditional:2 lempitsky:1 goal:2 acceleration:1 consequently:1 couprie:1 replace:1 luc:1 specifically:2 determined:1 uniformly:1 reducing:1 crcv:1 total:1 pas:2 experimental:1 support:1 incorporate:2 evaluate:3 |
6,234 | 664 | Forecasting Demand for Electric Power
Jen-Lun Yuan and Terrence L. Fine
School of Electrical Engineering
Cornell University
Ithaca, NY 14853
Abstract
We are developing a forecaster for daily extremes of demand for
electric power encountered in the service area of a large midwestern utility and using this application as a testbed for approaches
to input dimension reduction and decomposition of network training. Projection pursuit regression representations and the ability
of algorithms like SIR to quickly find reasonable weighting vectors
enable us to confront the vexing architecture selection problem by
reducing high-dimensional gradient searchs to fitting single-input
single-output (SISO) subnets. We introduce dimension reduction
algorithms, to select features or relevant subsets of a set of many
variables, based on minimizing an index of level-set dispersions
(closely related to a projection index and to SIR), and combine
them with backfitting to implement a neural network version of
projection pursuit. The performance achieved by our approach,
when trained on 1989, 1990 data and tested on 1991 data, is comparable to that achieved in our earlier study of backpropagation
trained networks.
1
Introduction
Our work has the intertwined goals of:
(i) contributing to the improvement of the short-term electrical load (demand)
forecasts used by electric utilities to buy and sell power and ensure that they can
meet demand;
739
740
Yuan and Fine
(ii) reducing the computational burden entailed in gradient-based training of neural
networks and thereby enabling the exploration of architectures;
(iii) improving prospects for good statistical generalization by use of rational methods for reducing complexity through the identification of good small subsets of
variables drawn from a large set of candidate predictor variables (feature selection);
(iv) benchmarking backpropagation and neural networks as an approach to the
applied problem of load forecasting.
Our efforts proceed in the context of a problem suggested by the operational needs
of a particular electric utility to make daily forecasts of short-term load or demand.
Forecasts are made at midday (1 p.m.) on a weekday t ( Monday - Thursday), for
the next evening peak e(t) (occuring usually about 8 p.m. in the winter), the daily
minimum d(t + 1) (occuring about 4 a.m. the next morning) and the morning
peak m( t + 1) (about noon ). In addition, on Friday we are to forecast these
three variables for the weekend through the Monday morning peak. These daily
extremes of demand are illustrated in an excerpt from our hourly load data plotted
in Figure 1.
4600
4400
f~
4200
i
J
4000
3600
5am
3400
0
5
lam
I~am
10
15
Rpm
20
2S
6am
' lam
Ilam
30
35
9jlm
40
45
50
number of houri
Figure 1: Hourly demand for two consecutive days showing the intended forecasting
variables.
In this paper, we focus on forecasting these extremal demands up to three days
ahead (e.g. forecasting on Fridays). Neural network-based forecasters are developed which parallel the recently proposed method of slicing inverse regression (SIR)
(Li [1991]) and then use backfitting (Hastie and Tibshirani [1990]) to implement a
training algorithm for a projection pursuit model (Friedman [1987]' Huber [1985])
that can be implemented with a single hidden layer network. Our data consists of
hourly integrated system demand (MWH) and hourly temperatures measured at
three cities in the service area of a large midwestern utility during 1989-91. We use
1989 and 1990 for a training set and test over the whole of 1991, with the exception
of holidays that occur so infrequently that we have no training base.
Forecasting Demand for Electric Power
2
2.1
Baseline Performance
Previous Work on Load Forecasting
Since demand is a process which does not have a known physical or mathematical
model, we do not know the best achievable forecasting performance, and we are led
to making comparisons with methods and results reported elsewhere. There is a
substantial literature on short-term load forecasting, with Gross et al. [1987] and
Willis et al. [1984] providing good reviews of approaches based upon such statistical methods as linear least squares regression -1Ild Box-Jenkins and ARMAX time
series models. Many utilities rely upon the seemingly seat-of-the-pants estimates
produced by individuals who have been long employed at this task and who extrapolate from a large historical data base. In the past few years there have been several
efforts to employ neural networks trained through backpropagation. In two such
recent studies conducted at the Univ. of Washington an average peak error of 2.04%
was reported by Damborg et al. [1990] and an hourly load error of about 2.2% was
given by Connor et al. [1991]. However, the accuracies reported in the literature are
difficult to compare with since utilities are exposed to different operating conditions
(e.g., weather, residential/industrial balance). To provide a benchmark for the error
performance achieved by our method, we evaluated three basic forecasting models
on our data. These methods are based on a pair of features made plausible by the
scatter plots shown in Figure 2.
.
5000~~--~---r--~
.
......
.. ,It ..
:
.
o
??
' ....
~
aD
'.
4b-:~.
10
.Ib.
:
~
~
~
~
4200
??
,
,
~
.\
4600 _..... .......... .. :""..: .......; . ..
.
~
~
.',"
:
:
.....,... .... .,..,-:JO" " " ....... ........ .
4400
.
'0
4800
...... " ":" ... ", . .
? ?? ? ? ? ?
,
:
~
~
? ~ 0?
:
:c
0
~
4800
4600
50oo~--......----'r----.....,
,
0;
4400
. '.f.. ; :?'
.... ....... .. ,...~ ....1. .. ~ ......~ ..
~. ~,
.~.t"".
??
4200
4000
3800
.. ..
"
D.
3800
.
...so "
,
.
3~~'-OO--4000
..........~4-:'-5oo-:---:-5000"":---S-:'5oo
m(t)MWH
3~'----~--~~--~
-SO
o
100
temperalUre(t) oP
Figure 2: Evening peaks ('IUe.-Fri.,1989-90) vs. morning peaks and temperatures.
2.2
Feature Selection and Homogeneous Data Types
Demand depends on predictable calendar factors such as the season, day-of-theweek and time-of-day considerations. We grouped separately Mondays, 'IUesdays
through Fridays, Saturdays, and Sundays, as well as holidays. In contrast to all
of the earlier work on this problem, we ignored seasonal considerations and let the
network and training algorithm adjust as needed. The advantage of this was the
ability to form larger training data sets. We thus constructed twelve networks, one
741
742
Yuan and Fine
type
m(t+1)
e(t)
d(t+1)
Monday
Tue.-Fri.
Saturday
Sunday
m(t-3)
m(t-1)
m(t-1)
m(t-2}
m(t)
m(t)
m(t-1)
m(t-2)
d(t-3)
d(t-1)
d(t-1)
d(t- 2)
Table 1: Most recent peaks of a two- feature set
type
Monday
The.-Fri.
Saturday
Sunday
LLS
m(t+1}
LOESS
BP
LLS
e(t)
LOESS
BP
LLS
d(t+1)
LOESS
BP
3.78
3.01
3.37
4.83
2.45
2.44
2.60
3.28
2.42
1.98
2.36
3.79
1.73
1.89
4.54
4.89
2.43
3.04
3.76
2.74
1.59
1.65
3.10
3.81
4.40
3.29
3.48
4.26
3.30
3.81
3.25
2.44
2.69
2.49
2.06
3.03
Table 2: Forecasting accuracies (percentage absolute error) for three basic methods
for each pair consisting of one of these four types of days and one of the three
daily extremes to be forecast. Demand also depends heavily upon weather which is
the primary random factor affecting forecasts. This dependency can be seen in the
scatter plots of current demand vs. previous demand and temperature in Figure 2,
particularly in the projection onto the 'current demand-temperature' plane which
shows a pronounced "U"-shaped nonlinearity. A two-feature set consisting of the
most recent peaks and average temperatures over the three cities and the preceding
six hours is employed for testing all three models (Table 1).
2.3
Benchmark Results
The three basic forecasting models using the two-featured set are:
1) linear regression model fitted to the data in Figure 2;
2) demand vs. temperature models which roughly model the "U-shaped" nonlinear
relationship, (LOESS with .5 span was employed for scatter plot smoothing);
3) backpropagation trained neural networks using 5 logistic nodes in a single hidden
layer.
The test set errors are given in Table 2. Note that among these three models,
BP-trained neural networks gives superior test set performance on all but Sundays.
These models all give results comparable to those obtained in our earlier work on
forecasting demands for Thesday-Friday using autoregressive neural networks (Yuan
and Fine [1992]).
Forecasting Demand for Electric Power
3
Projection Pursuit Training
Satisfactory forecasting performance of the neural networks described above relies
on the appropriate choice of feature sets and network architectures. Unfortunately,
BP can only address the problem of appropriate architecture and relevant feature
sets through repeated time-consuming experiments. Modeling of high-dimensional
input features using gradient search requires extensive computation. We were thus
prompted to look at other network structures and at training algorithms that could
make it easier to explore architecture and training problems. Our initial attempt
combined the dimension reduction algorithm cf SIR (Li [1991]), currently replaced
by an algorithm of our devising sketched in Section 4, and backfitting (Hastie
et.al [1990]) to implement a neural network version of projection pursuit regression (PPR).
3.1
The Algorithm
A general nonlinear regression model for a forecast variable y in terms of a vector
x of input variables and model noise ?, independent of x, is gi ven by
y = f({3ix,{3~x, .. ,{3~X,?)
(*).
A least mean square predictor is the conditional expectation E(ylx). The projection
pursuit model/approximation of this conditional expectation is given in terms of a
family of SISO functions 2 1 ,22, ", 2k by
k
E(ylx)
= I: 2 i ({3:x) + {3o.
i=l
A single hidden layer neural network can approximate this representation by introducing subnets whose summed outputs approximate the individual 2 j .
We train such a 'projection pursuit network' with nodes partitioned into subnets,
representing the 2 i , by training the subnets individually in rotation. In this we
follow the statistical regression notion of backfitting. The subnet 2i is trained to
predict the residuals resulting from the difference between the weighted outputs of
the other k - 1 subnets and the true value of the demand variable. After a number
of training cycles one then proceeds to the next subnet and repeats the process. The
inputs to each subnet 2i are the low-dimensional projections {3:x of the regression
model. One termination criteria for determining the number of subnets k is to stop
adding subnets when the projection appears to be normally distributed; results
of Diaconis and Freedman point out that 'most' projections will be so distributed
and thus are 'uninteresting'. The directions {3i can be found by minimizing some
projection index which defines interesting projections that deviates from Gaussian
distributions (e.g., Friedman [1987]). Each {3i determines the weights connecting
the input to sub net 2 i .The whole projection pursuit regression process is simplified
by decoupling the direction {3 search from training the SISO subnets. Albeit, its
success depends upon an ability to rapidly discern the significant directions {3i.
3.2
Implementations
There are several variants in the implementation of projection pursuit training algorithms. General PPR procedure can be implemented in one stage by computa-
743
744
Yuan and Fine
type
m(t+l)
e(t)
d(t+1)
Monday
Tue.-Fri.
Saturday
Sunday
2.35/3.45
2.37/2.83
2.67/3.16
3.15/5.38
1.25/1.60
1.65/1.66
2.78/3.96
2.63/3.67
2.76/3.49
2.15/2.66
2.57/3.04
2.29/3.61
Table 3: Forecasting performance (training/testing percentage error) of projection
pursuit trained networks
tionally intensive numerical methods, or in a two-stage heuristic (finding f3i, then
Bi ) as proposed here. It can be implemented with or without back fitting after the
PPR phase is done. Intrator [1992] has recently suggested incorporating the projection index into the objective function and then running an overall BPA. Other
variants in training each Bi net include using nonparametric smoothing techniques
such as LOESS or kernel methods. BP training can then be applied only in the
last stage to fit the smoothed curves so obtained. The complexity of each subnet
is then largely determined by the smoothing parameters, like window sizes, inherent in most nonparametric smoothing techniques. Another practical advantage of
this process is that one can incorporate easily fixed functions of a single variable
(e.g. linear nodes or quadratic nodes) when one's prior knowledge of the data
source suggests that such components may be present. Our current implementation
employs the two-stage algorithm with simple (either one or two nodes) logistic Bi
subnets. Each SISO Bi net runs a BP algorithm to fit the data. The directions f3i
are calculated based on minimizing a projection index (dispersion of level-sets, described in Section 4) which can be executed in a direct fashion. One can encourage
the convergence of backfitting by using a relaxation parameter (like a momentum
parameter in BPA ) to control the amount updated in the current direction. Training (fitting) of each (SISO) 3 i net can be carried out more efficiently than running
BP based on high-dimensional inputs, for example, it is less expensive to evaluate
the Hessian matrices in a Bi net than in a full BPA networks.
3.3
Forecasting Results
Experimental results were obtained using the two-component feature data sets
which gave the earlier baseline performance. To calibrate the performance we employed in all twelve projection pursuit trained networks an uniform architecture of
three subnets ( a (1,2, 2)-logistic network), matching the 5 nodes of the BP network of Section 2. The number of backfitting cycles was set to 20 with a relaxation
parameter w = 0.1. BPA was employed for fitting each Binet. The training/testing
percentage absolute errors are given in Table 3. The limited data sets in the cases
of individual days (Monday, Saturday, Sunday) led to failure in generalization that
could have been prevented by using one or two, rather than three, subnets.
Forecasting Demand for Electric Power
4
4.1
Dimension Reduction
Index of Level-Set Dispersion
A key step in the projection pursuit training algorithm is to find for each 3 i net
the projection direction f3i' an instance of the important problem of economically
choosing input features/variables in constructing a forecasting model. In general,
the fewer the number of input features, the more likely are the results to generalize
from training set performance to test set performance- reduction in variance at the
possible expense of increase in bias. Our controlled size subnet projection pursuit
training algorithm deals with part of the complexity problem, provided that the
input features are fixed. We turn now to our approach to finding input features
or search directions based on minimizing an index of dispersion of level-sets. Li
[1991] proposed taking an inverse ('slicing the y's') point of view to estimate the
directions f3i. The justification provided for this so-called slicing inverse regression
(SIR) method, however, requires that the input or feature vector x be elliptically
symmetrically distributed, and this is not likely to be the case in our electric load
forecasting problem. The basic idea behind minimizing dispersion of level-sets
is that from Eq. (*) we see that a fixed value of y, and small noise ?, implies
a highly constrained set of values for f3ix, ... ,f3~x, while leaving unconstrained the
components of x that lie in the subspace B~ orthogonal to that space B spanned by
the f3i.. Hence, if one has a good number of i.i.d. observations sharing a similar value
of the response y, then there should be more dispersion of input vectors projected
into Bl.. than along the projections into B. We implement this by quantizing the
observed y values into, say, H slices, with Lh denoting the h_th level-set containing
those inputs with y-value in the h_th slice, and X-h is their sample mean. The f3 are
then picked as the the eigenvector associated with the smallest eigenvalue of the
centered covariance matrix:
H
L L
(Xi - x"h)(Xi. - Xh)'.
h=l xiELh
4.2
Implementations
In practical implementations, one may discard both extremes of the family of H
level sets (trimming) to avoid large response values when it is believed that they
may correspond to large magnitudes of input components. One should also standardize initially the input data to a unit sample covariance matrix. Otherwise, our
results will reflect the distribution of x rather than the functional relationship of
Eq. (*). We have applied this projection index both in finding the f3i. during projection pursuit training and in reducing a high-dimensional feature set to a lowdimensional feature set. We have implemented such a feature selection scheme for
forecasting the Monday - Friday evening peaks. The initial feature set consists of
thirteen hourly loads from lam to 1pm, thirteen hourly temperatures from lam to
1pm and the temperature around the peak times. Three eigenvectors of the centered covariance matrix were chosen, thereby reducing a 27-dimensional feature set
to a 3-dimensional one. We then ran a standard BPA on this reduced featured set
and tested on the 1991 data. We obtained a percentage absolute error of 1.6% (rms
error about 100 MWH), which is as good as all of our previous efforts.
745
746
Yuan and Fine
Acknow ledgements
Partial support for this research was provided by NSF Grant No. ECS-9017493.
We wish to thank Prof. J. Hwang, Mathematics Department, Cornell, for initial
discussions of SIR and are grateful to Dr. P.D. Yeshakul, American Electric Service
Corp., for providing the data set and instructing us patiently in the lore of shortterm load forecasting.
References
Connor, J., L. Atlas, R.D. Martin [1991], Recurrent networks and NARMA modeling, NIPS 91.
Damborg, M., M.EI-Sharkawi, R. Marks II [1990], Potential of artificial neural networks in power system operation, Proc. 1990 IEEE Inter.Symp. on Circuits and
Systems, 4, 2933-2937.
Friedman, J. [1987], Exploratory projection pursuit, J. Amer. Stat. Assn., 82,
249-266.
Gross,G., F. Galiana [1987], Short-term load forecasting, Proc. IEEE, 75, 15581573.
Hastie, T., R. Tibshirani [1990], Generalized Additive Models,Chapman and Hall.
Huber, P. [1985]' Projection pursuit, The Annals of Statistics, 13, 435-475.
Intrator, N. [1992] Combinining exploratory projection pursuit and projection pursuit regression with applicatons to neural networks, To appear in Neural Computation.
Li, K.-C. [1991] Slicing inverse regression for dimension reduction, Journal of American Statistical Assoc., 86.
Willis, H.1., J.F.D. Northcote-Green [1984], Comparison tests of fourteen load
forecasting methods, IEEE Trans. on Power Apparatus and Systems, PAS-103,
1190-1197.
Yuan, J-1., T.L.Fine [1992]' Forecasting demand for electric power using autoregressive neural networks, Proc. Con! on Info. Sci. and Systems, Princeton, NJ.
| 664 |@word economically:1 version:2 achievable:1 termination:1 forecaster:2 decomposition:1 covariance:3 weekday:1 thereby:2 reduction:6 initial:3 series:1 denoting:1 past:1 current:4 scatter:3 additive:1 numerical:1 plot:3 atlas:1 v:3 fewer:1 devising:1 plane:1 short:4 node:6 monday:8 mathematical:1 along:1 constructed:1 direct:1 saturday:5 yuan:7 consists:2 backfitting:6 fitting:4 combine:1 symp:1 introduce:1 inter:1 huber:2 roughly:1 window:1 provided:3 circuit:1 eigenvector:1 developed:1 finding:3 nj:1 jlm:1 computa:1 assoc:1 control:1 normally:1 unit:1 grant:1 appear:1 hourly:7 service:3 engineering:1 apparatus:1 meet:1 suggests:1 limited:1 bi:5 practical:2 testing:3 implement:4 backpropagation:4 procedure:1 area:2 featured:2 weather:2 projection:30 matching:1 onto:1 selection:4 context:1 slicing:4 seat:1 spanned:1 exploratory:2 notion:1 holiday:2 justification:1 updated:1 annals:1 heavily:1 homogeneous:1 pa:1 infrequently:1 expensive:1 particularly:1 standardize:1 observed:1 midday:1 electrical:2 cycle:2 prospect:1 ran:1 gross:2 substantial:1 predictable:1 complexity:3 trained:8 grateful:1 exposed:1 upon:4 easily:1 weekend:1 train:1 univ:1 artificial:1 choosing:1 whose:1 heuristic:1 larger:1 plausible:1 say:1 otherwise:1 calendar:1 ability:3 statistic:1 gi:1 seemingly:1 advantage:2 eigenvalue:1 quantizing:1 net:6 lam:4 lowdimensional:1 relevant:2 rapidly:1 pronounced:1 convergence:1 oo:4 recurrent:1 stat:1 subnets:11 measured:1 op:1 school:1 eq:2 implemented:4 implies:1 direction:8 closely:1 exploration:1 centered:2 enable:1 subnet:5 generalization:2 around:1 hall:1 predict:1 consecutive:1 smallest:1 siso:5 proc:3 currently:1 extremal:1 individually:1 grouped:1 city:2 weighted:1 gaussian:1 sunday:6 rather:2 avoid:1 season:1 cornell:2 focus:1 seasonal:1 improvement:1 narma:1 industrial:1 contrast:1 baseline:2 am:3 integrated:1 initially:1 hidden:3 sketched:1 overall:1 among:1 loess:5 smoothing:4 summed:1 constrained:1 f3:2 shaped:2 washington:1 chapman:1 sell:1 ven:1 look:1 lls:3 few:1 employ:2 inherent:1 winter:1 diaconis:1 individual:3 replaced:1 intended:1 consisting:2 phase:1 friedman:3 attempt:1 trimming:1 highly:1 adjust:1 entailed:1 extreme:4 behind:1 encourage:1 partial:1 daily:5 lh:1 orthogonal:1 iv:1 plotted:1 fitted:1 instance:1 earlier:4 modeling:2 calibrate:1 introducing:1 subset:2 predictor:2 uninteresting:1 uniform:1 conducted:1 reported:3 dependency:1 combined:1 peak:10 twelve:2 terrence:1 connecting:1 quickly:1 jo:1 reflect:1 containing:1 dr:1 american:2 friday:5 li:4 potential:1 ad:1 depends:3 view:1 mwh:3 picked:1 parallel:1 square:2 lore:1 accuracy:2 variance:1 who:2 largely:1 efficiently:1 correspond:1 generalize:1 identification:1 produced:1 sharing:1 failure:1 associated:1 con:1 rational:1 stop:1 knowledge:1 back:1 appears:1 day:6 follow:1 response:2 amer:1 evaluated:1 box:1 done:1 stage:4 ei:1 ild:1 nonlinear:2 morning:4 defines:1 logistic:3 hwang:1 true:1 binet:1 hence:1 satisfactory:1 illustrated:1 deal:1 during:2 criterion:1 generalized:1 occuring:2 willis:2 temperature:8 consideration:2 recently:2 superior:1 rotation:1 functional:1 physical:1 fourteen:1 significant:1 connor:2 unconstrained:1 pm:2 mathematics:1 nonlinearity:1 tionally:1 operating:1 base:2 recent:3 discard:1 corp:1 success:1 seen:1 minimum:1 preceding:1 employed:5 fri:4 ii:2 full:1 believed:1 long:1 prevented:1 controlled:1 variant:2 regression:12 basic:4 confront:1 expectation:2 kernel:1 achieved:3 addition:1 affecting:1 fine:7 separately:1 source:1 leaving:1 ithaca:1 symmetrically:1 iii:1 fit:2 gave:1 architecture:6 hastie:3 f3i:6 idea:1 intensive:1 six:1 rms:1 utility:6 forecasting:25 effort:3 proceed:1 hessian:1 armax:1 elliptically:1 ignored:1 eigenvectors:1 ylx:2 amount:1 nonparametric:2 reduced:1 assn:1 percentage:4 nsf:1 tibshirani:2 ledgements:1 intertwined:1 key:1 four:1 drawn:1 relaxation:2 year:1 residential:1 run:1 inverse:4 discern:1 family:2 reasonable:1 excerpt:1 rpm:1 comparable:2 layer:3 bpa:5 quadratic:1 encountered:1 ahead:1 occur:1 bp:9 span:1 martin:1 department:1 developing:1 partitioned:1 making:1 damborg:2 turn:1 needed:1 know:1 pursuit:18 jenkins:1 operation:1 intrator:2 appropriate:2 running:2 ensure:1 cf:1 include:1 prof:1 bl:1 objective:1 primary:1 gradient:3 subspace:1 thank:1 sci:1 tue:2 index:8 relationship:2 prompted:1 providing:2 minimizing:5 balance:1 difficult:1 unfortunately:1 executed:1 thirteen:2 expense:1 info:1 acknow:1 implementation:5 observation:1 dispersion:6 benchmark:2 enabling:1 smoothed:1 pair:2 extensive:1 instructing:1 testbed:1 hour:1 nip:1 trans:1 address:1 suggested:2 proceeds:1 usually:1 green:1 power:9 rely:1 residual:1 representing:1 scheme:1 carried:1 shortterm:1 ppr:3 deviate:1 review:1 literature:2 prior:1 contributing:1 determining:1 sir:6 interesting:1 pant:1 elsewhere:1 repeat:1 last:1 bias:1 taking:1 absolute:3 distributed:3 slice:2 curve:1 dimension:5 calculated:1 autoregressive:2 made:2 projected:1 simplified:1 historical:1 ec:1 approximate:2 buy:1 consuming:1 xi:2 search:4 evening:3 table:6 noon:1 decoupling:1 operational:1 improving:1 electric:10 constructing:1 whole:2 noise:2 freedman:1 repeated:1 benchmarking:1 fashion:1 ny:1 lun:1 sub:1 momentum:1 wish:1 xh:1 candidate:1 lie:1 ib:1 weighting:1 ix:1 load:12 jen:1 showing:1 burden:1 incorporating:1 albeit:1 adding:1 galiana:1 magnitude:1 sharkawi:1 demand:22 forecast:7 easier:1 led:2 explore:1 likely:2 determines:1 relies:1 conditional:2 goal:1 determined:1 reducing:5 called:1 experimental:1 exception:1 select:1 thursday:1 support:1 mark:1 incorporate:1 evaluate:1 princeton:1 tested:2 extrapolate:1 |
6,235 | 6,640 | Learning a Multi-View Stereo Machine
Abhishek Kar
UC Berkeley
[email protected]
Christian H?ne
UC Berkeley
[email protected]
Jitendra Malik
UC Berkeley
[email protected]
Abstract
We present a learnt system for multi-view stereopsis. In contrast to recent learning
based methods for 3D reconstruction, we leverage the underlying 3D geometry of
the problem through feature projection and unprojection along viewing rays. By
formulating these operations in a differentiable manner, we are able to learn the
system end-to-end for the task of metric 3D reconstruction. End-to-end learning
allows us to jointly reason about shape priors while conforming to geometric
constraints, enabling reconstruction from much fewer images (even a single image)
than required by classical approaches as well as completion of unseen surfaces. We
thoroughly evaluate our approach on the ShapeNet dataset and demonstrate the
benefits over classical approaches and recent learning based methods.
1
Introduction
Multi-view stereopsis (MVS) is classically posed as the following problem - given a set of images
with known camera poses, it produces a geometric representation of the underlying 3D world. This
representation can be a set of disparity maps, a 3D volume in the form of voxel occupancies, signed
distance fields etc. An early example of such a system is the stereo machine from Kanade et al. [26]
that computes disparity maps from images streams from six video cameras. Modern approaches
focus on acquiring the full 3D geometry in the form of volumetric representations or polygonal
meshes [48]. The underlying principle behind MVS is simple - a 3D point looks locally similar when
projected to different viewpoints [29]. Thus, classical methods use the basic principle of finding
dense correspondences in images and triangulate to obtain a 3D reconstruction.
The question we try to address in this work is can we learn a multi-view stereo system? For the
binocular case, Becker and Hinton [1] demonstrated that a neural network can learn to predict a depth
map from random dot stereograms. A recent work [28] shows convincing results for binocular stereo
by using an end-to-end learning approach with binocular geometry constraints.
In this work, we present Learnt Stereo Machines (LSM) - a system which is able to reconstruct object
geometry as voxel occupancy grids or per-view depth maps from a small number of views, including
just a single image. We design our system inspired by classical approaches while learning each
component from data embedded in an end to end system. LSMs have built in projective geometry,
enabling reasoning in metric 3D space and effectively exploiting the geometric structure of the MVS
problem. Compared to classical approaches, which are designed to exploit a specific cue such as
silhouettes or photo-consistency, our system learns to exploit the cues that are relevant to the particular
instance while also using priors about shape to predict geometry for unseen regions.
Recent work from Choy et al. [5] (3D-R2N2) trains convolutional neural networks (CNNs) to
predict object geometry given only images. While this work relied primarily on semantic cues for
reconstruction, our formulation enables us to exploit strong geometric cues. In our experiments, we
demonstrate that a straightforward way of incorporating camera poses for volumetric occupancy
prediction does not lead to expected gains, while our geometrically grounded method is able to
effectively utilize the additional information.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
Fused Feature
Grid
Recurrent Fusion
Unprojection
Image
Encoder
2D Feature
Maps
Final Grid
Voxel Occupancy Grid
3D Grid Reasoning
Projection
3D Feature
Grids
Depth Maps
Figure 1: Overview of a Learnt Stereo Machine (LSM). It takes as input one or more views and camera poses.
The images are processed through a feature encoder which are then unprojected into the 3D world frame using a
differentiable unprojection operation. These grids {Gfi }n
i=1 are then matched in a recurrent manner to produce a
fused grid G p which is then transformed by a 3D CNN into G o . LSMs can produce two kinds of outputs - voxel
occupancy grids (Voxel LSM) decoded from G o or per-view depth maps (Depth LSM) decoded after a projection
operation.
Classical multi-view stereopsis is traditionally able to handle both objects and scenes - we only
showcase our system for the case of objects with scenes left for future work. We thoroughly evaluate
our system on the synthetic ShapeNet [3] dataset. We compare to classical plane sweeping stereo,
visual hulls and several challenging learning-based baselines. Our experiments show that we are able
to reconstruct objects with fewer images than classical approaches. Compared to recent learning
based reconstruction approaches, our system is able to better use camera pose information leading to
significantly large improvements while adding more views. Finally, we show successful generalization
to unseen object categories demonstrating that our network goes beyond semantic cues and strongly
uses geometric information for unified single and multi-view 3D reconstruction.
2
Related Work
Extracting 3D information from images is one of the classical problems in computer vision. Early
works focused on the problem of extracting a disparity map from a binocular image pair [36]. We
refer the reader to [47] for an overview of classical binocular stereo matching algorithms. In the multiview setting, early work focused on using silhouette information via visual hulls [32], incorporating
photo-consistency to deal with concavities (photo hull) [29], and shape refinement using optimization
[55, 50, 7, 15]. [39, 35, 54] directly reason about viewing rays in a voxel grid, while [34] recovers
a quasi dense point cloud. In our work, we aim to learn a multi-view stereo machine grounded in
geometry, that learns to use these classical constraints while also being able to reason about semantic
shape cues from the data. Another approach to MVS involves representing the reconstruction as a
collection of depth maps [6, 57, 41, 13, 40]. This allows recovery of fine details for which a consistent
global estimate may be hard to obtain. These depth maps can then be fused using a variety of different
techniques [38, 8, 33, 59, 30]. Our learnt system is able to produce a set of per-view depth maps
along with a globally consistent volumetric representation which allows us to preserve fine details
while conforming to global structure.
Learning has been used for multi-view reconstruction in the form of shape priors for objects [2, 9, 58,
20, 27, 52], or semantic class specific surface priors for scenes [22, 17, 45]. These works use learnt
shape models and either directly fit them to input images or utilize them in a joint representation
that fuses semantic and geometric information. Most recently, CNN based learning methods have
been proposed for 3D reconstruction by learning image patch similarity functions [60, 18, 23] and
end-to-end disparity regression from stereo pairs [37, 28]. Approaches which predict shape from
a single image have been proposed in form of direct depth map regression [46, 31, 10], generating
multiple depth maps from novel viewpoints [51], producing voxel occupancies [5, 16], geometry
images [49] and point clouds [11]. [12] study a related problem of view interpolation, where a rough
depth estimate is obtained within the system.
A line of recent works, complementary to ours, has proposed to incorporate ideas from multi-view
geometry in a learning framework to train single view prediction systems [14, 56, 53, 42, 61] using
multiple views as supervisory signal. These works use the classical cues of photo-consistency and
2
De
2
pth
pla
2-D World Grid
ne
s
1
z=
z=
3
2-D Feature Grid
z=
z=3
z=2
z=1
1-D Projections
1-D
Feature Map
Sampling locations
1-D
Canvas
Camera
Camera
(a) Projection
(b) Unprojection
Figure 2: Illustrations of projection and unprojection operations between 1D maps and 2D grids. (a) The
projection operation samples values along the ray at equally spaced z-values into a 1D canvas/image. The
sampled features (shown by colors here) at the z planes are stacked into channels to form the projected feature
map. (b) The unprojection operation takes features from a feature map (here in 1-D) and places them along rays
at grid blocks where the respective rays intersect. Best viewed in color.
silhouette consistency only during training - their goal during inference is to only perform single
image shape prediction. In contrast, we also use geometric constraints during inference to produce
high quality outputs.
Closest to our work is the work of Kendall et al. [28] which demonstrates incorporating binocular
stereo geometry into deep networks by formulating a cost volume in terms of disparities and regressing
depth values using a differentiable arg-min operation. We generalize to multiple views by tracing
rays through a discretized grid and handle variable number of views via incremental matching using
recurrent units. We also propose a differentiable projection operation which aggregates features along
viewing rays and learns a nonlinear combination function instead of using the differentiable arg-min
which is susceptible to multiple modes. Moreover, we can also infer 3D from a single image during
inference.
3
Learnt Stereo Machines
Our goal in this paper is to design an end-to-end learnable system that produces a 3D reconstruction
given one or more input images and their corresponding camera poses. To this end, we draw inspiration from classical geometric approaches where the underlying guiding principle is the following - the
reconstructed 3D surface has to be photo-consistent with all the input images that depict this particular
surface. Such approaches typically operate by first computing dense features for correspondence
matching in image space. These features are then assembled into a large cost volume of geometrically
feasible matches based on the camera pose. Finally, the optimum of this matching volume (along
with certain priors) results in an estimate of the 3D volume/surface/disparity maps of the underlying
shape from which the images were produced.
Our proposed system, shown in Figure 1, largely follows the principles mentioned above. It uses
a discrete grid as internal representation of the 3D world and operates in metric 3D space. The
input images {Ii }ni=1 are first processed through a shared image encoder which produces dense
feature maps {Fi }ni=1 , one for each image. The features are then unprojected into 3D feature grids
{Gif }ni=1 by rasterizing the viewing rays with the known camera poses {Pi }ni=1 . This unprojection
operation aligns the features along epipolar lines, enabling efficient local matching. This matching
is modelled using a recurrent neural network which processes the unprojected grids sequentially to
produce a grid of local matching costs G p . This cost volume is typically noisy and is smoothed in
an energy optimization framework with a data term and smoothness term. We model this step by a
feed forward 3D convolution-deconvolution CNN that transforms G p into a 3D grid G o of smoothed
costs taking context into account. Based on the desired output, we propose to either let the final
grid be a volumetric occupancy map or a grid of features which is projected back into 2D feature
3
maps {Oi }ni=1 using the given camera poses. These 2D maps are then mapped to a view specific
representation of the shape such as a per view depth/disparity map. The key components of our
system are the differentiable projection and unprojection operations which allow us to learn the
system end-to-end while injecting the underlying 3D geometry in a metrically accurate manner. We
refer to our system as a Learnt Stereo Machine (LSM). We present two variants - one that produces
per voxel occupancy maps (Voxel LSM) and another that outputs a depth map per input image (Depth
LSM) and provide details about the components and the rationale behind them below.
2D Image Encoder. The first step in a stereo algorithm is to compute a good set of features to
match across images. Traditional stereo algorithms typically use raw patches as features. We model
this as a feed forward CNN with a convolution-deconvolution architecture with skip connections
(UNet) [44] to enable the features to have a large enough receptive field while at the same time having
access to lower level features (using skip connections) whenever needed. Given images {Ii }ni=1 , the
feature encoder produces dense feature maps {Fi }ni=1 in 2D image space, which are passed to the
unprojection module along with the camera parameters to be lifted into metric 3D space.
Differentiable Unprojection. The goal of the unprojection operation is to lift information from
2D image frame to the 3D world frame. Given a 2D point p, its feature representation F(p) and our
global 3D grid representation, we replicate F(p) along the viewing ray for p into locations along
the viewing ray in the metric 3D grid (a 2D illustration is presented in Figure 2). In the case of
perspective projection specified by an intrinsic camera matrix K and an extrinsic camera matrix [R|t],
the unprojection operation uses this camera pose to trace viewing rays in the world and copy the
image features into voxels in this 3D world grid. Instead of analytically tracing rays, given the centers
th
k
V
of blocks in our 3D grid {Xwk }N
block by projecting {Xw
} using
k=1 , we compute the feature for k
0
k
0
the camera projection equations pk = K[R|t]Xw into the image space. pk is a continuous quantity
whereas F is defined on at discrete 2D locations. Thus, we use the differentiable bilinear sampling
operation to sample from the discrete grid [25] to obtain the feature at Xwk .
Such an operation has the highly desirable property that features from pixels in multiple images
that may correspond to the same 3D world point unproject to the same location in the 3D grid trivially enforcing epipolar constraints. As a result, any further processing on these unprojected grids
has easy access to corresponding features to make matching decisions foregoing the need for long
range image connections for feature matching in image space. Also, by projecting discrete 3D points
into 2D and bilinearly sampling from the feature map rather than analytically tracing rays in 3D,
we implicitly handle the issue where the probability of a grid voxel being hit by a ray decreases
with distance from the camera due to their projective nature. In our formulation, every voxel gets
a ?soft" feature assigned based on where it projects back in the image, making the feature grids
G f smooth and providing stable gradients. This geometric procedure of lifting features from 2D
maps into 3D space is in contrast with recent learning based approaches [5, 51] which either reshape
flattened feature maps into 3D grids for subsequent processing or inject pose into the system using
fully connected layers. This procedure effectively saves the network from having to implicitly learn
projective geometry and directly bakes this given fact into the system. In LSMs, we use this operation
to unproject the feature maps {Fi }ni=1 in image space produced by the feature encoder into feature
grids {Gif }ni=1 that lie in metric 3D space.
For single image prediction, LSMs cannot match features from multiple images to reason about where
to place surfaces. Therefore, we append geometric features along the rays during the projection and
unprojection operation to facilitate single view prediction. Specifically, we add the depth value and
the ray direction at each sampling point.
Recurrent Grid Fusion. The 3D feature grids {Gif }ni=1 encode information about individual input
images and need to be fused to produce a single grid so that further stages may reason jointly over all
the images. For example, a simple strategy to fuse them would be to just use a point-wise function e.g. max or average. This approach poses an issue where the combination is too spatially local and
early fuses all the information from the individual grids. Another extreme is concatenating all the
feature grids before further processing. The complexity of this approach scales linearly with the
number of inputs and poses issues while processing a variable number of images. Instead, we choose
to processed the grids in a sequential manner using a recurrent neural network. Specifically, we use a
3D convolutional variant of the Gated Recurrent Unit (GRU) [24, 4, 5] which combines the grids
4
{Gif }ni=1 using 3D convolutions (and non-linearities) into a single grid G p . Using convolutions helps
us effectively exploit neighborhood information in 3D space for incrementally combining the grids
while keeping the number of parameters low. Intuitively, this step can be thought of as mimicking
incremental matching in MVS where the hidden state of the GRU stores a running belief about the
matching scores by matching features in the observations it has seen. One issue that arises is that we
now have to define an ordering on the input images, whereas the output should be independent of the
image ordering. We tackle this issue by randomly permuting the image sequences during training
while constraining the output to be the same. During inference, we empirically observe that the final
output has very little variance with respect to ordering of the input image sequence.
3D Grid Reasoning. Once the fused grid G p is constructed, a classical multi-view stereo approach
would directly evaluate the photo-consistency at the grid locations by comparing the appearance of
the individual views and extract the surface at voxels where the images agree. We model this step with
a 3D UNet that transforms the fused grid G p into G o . The purpose of this network is to use shape cues
present in G p such as feature matches and silhouettes as well as build in shape priors like smoothness
and symmetries and knowledge about object classes enabling it to produce complete shapes even
when only partial information is visible. The UNet architecture yet again allows the system to use
large enough receptive fields for doing multi-scale matching while also using lower level information
directly when needed to produce its final estimate G o . In the case of full 3D supervision (Voxel LSM),
this grid can be made to represent a per voxel occupancy map. G o can also be seen as a feature grid
containing the final representation of the 3D world our system produces from which views can be
rendered using the projection operation described below.
Differentiable Projection. Given a 3D feature grid G and a camera P, the projection operation
produces a 2D feature map O by gathering information along viewing rays. The direct method would
be to trace rays for every pixel and accumulate information from all the voxels on the ray?s path.
Such an implementation would require handling the fact that different rays can pass through different
number of voxels on their way. For example, one can define a reduction function along the rays to
aggregate information (e.g. max, mean) but this would fail to capture spatial relationships between
the ray features. Instead, we choose to adopt a plane sweeping approach where we sample from
z
locations on depth planes at equally spaced z-values {zk }N
k=1 along the ray.
Consider a 3D point Xw that lies along the ray corresponding to a 2D point p in the projected
feature grid at depth zw - i.e. p = K[R|t]Xw and z(Xw ) = zw . The corresponding feature O(p) is
computed by sampling from the grid G at the (continuous) location Xw . This sampling can be done
differentiably in 3D using trilinear interpolation. In practice, we use nearest neighbor interpolation
in 3D for computational efficiency. Samples along each ray are concatenated in ascending z-order
to produce the 2D map O where the features are stacked along the channel dimension. Rays in this
feature grid can be trivially traversed by just following columns along the channel dimension allowing
us to learn the function to pool along these rays by using 1x1 convolutions on these feature maps and
progressively reducing the number of feature channels.
Architecture Details. As mentioned above, we present two versions of LSMs - Voxel LSM (VLSM) and Depth LSM (D-LSM). Given one or more images and cameras, Voxel LSM (V-LSM)
produces a voxel occupancy grid whereas D-LSM produces a depth map per input view. Both
systems share the same set of CNN architectures (UNet) for the image encoder, grid reasoning and
the recurrent pooling steps. We use instance normalization for all our convolution operations and
layer normalization for the 3D convolutional GRU. In V-LSM, the final grid G o is transformed into
a probabilistic voxel occupancy map V ? Rvh ?vw ?vd by a 3D convolution followed by softmax
operation. We use simple binary cross entropy loss between ground truth occupancy maps and V. In
D-LSM, G o is first projected into 2D feature maps {Oi }ni=1 which are then transformed into metric
depth maps {di }ni=1 by 1x1 convolutions to learn the reduction function along rays followed by
deconvolution layers to upsample the feature map back to the size of the input image. We use absolute
L1 error in depth to train D-LSM. We also add skip connections between early layers of the image
encoder and the last deconvolution layers producing depth maps giving it access to high frequency
information in the images.
5
Figure 3: Voxel grids produced by V-LSM for example image sequences alongside a learning based baseline
which uses pose information in a fully connected manner. V-LSM produces geometrically meaningful reconstructions (e.g. the curved arm rests instead of perpendicular ones (in R2N2) in the chair on the top left and
the siren lights on top of the police car) instead of relying on purely semantic cues. More visualizations in
supplementary material.
4
Experiments
In this section, we demonstrate the ability of LSMs to learn 3D shape reconstruction in a geometrically
accurate manner. First, we present quantitative results for V-LSMs on the ShapeNet dataset [3] and
compare it to various baselines, both classical and learning based. We then show that LSMs generalize
to unseen object categories validating our hypothesis that LSMs go beyond object/class specific priors
and use photo-consistency cues to perform category-agnostic reconstruction. Finally, we present
qualitative and quantitative results from D-LSM and compare it to traditional multi-view stereo
approaches.
Dataset and Metrics. We use the synthetic ShapeNet dataset [3] to generate posed image-sets,
ground truth 3D occupancy maps and depth maps for all our experiments. More specifically, we
use a subset of 13 major categories (same as [5]) containing around 44k 3D models resized to lie
within the unit cube centered at the origin with a train/val/test split of [0.7, 0.1, 0.2]. We generated a
large set of realistic renderings for the models sampled from a viewing sphere with ?az ? [0, 360)
and ?el ? [?20, 30] degrees and random lighting variations. We also rendered the depth images
corresponding to each rendered image. For the volumetric ground truth, we voxelize each of the
models at a resolution of 32 ? 32 ? 32. In order to evaluate the outputs of V-LSM, we binarize the
probabilities at a fixed threshold (0.4 for all methods except visual hull (0.75)) and use the voxel
intersection over union (IoU) as the similarity measure. To aggregate the per model IoU, we compute
a per class average and take the mean as a per dataset measure. All our models are trained in a class
agnostic manner.
Implementation. We use 224 ? 224 images to train LSMs with a shape batch size of 4 and
4 views per shape. Our world grid is at a resolution of 323 . We implemented our networks in
Tensorflow and trained both the variants of LSMs for 100k iterations using Adam. The projection and
unprojection operations are trivially implemented on the GPU with batched matrix multiplications
and bilinear/nearest sampling enabling inference at around 30 models/sec on a GTX 1080Ti. We
unroll the GRU for upto 4 time steps while training and apply the trained models for arbitrary number
of views at test time.
Multi-view Reconstruction on ShapeNet. We evaluate V-LSMs on the ShapeNet test set and
compare it to the following baselines - a visual hull baseline which uses silhouettes to carve out
volumes, 3D-R2N2 [5], a previously proposed system which doesn?t use camera pose and performs
multi-view reconstruction, 3D-R2N2 w/pose which is an extension of 3D-R2N2 where camera pose
is injected using fully connected layers. For the experiments, we implemented the 3D-R2N2 system
6
2
3
4
3D-R2N2 [5]
55.6
59.6
61.3
62.0
Visual Hull
3D-R2N2 w/pose
V-LSM
18.0
55.1
61.5
36.9
59.4
72.1
47.0
61.2
76.2
52.4
62.1
78.2
V-LSM w/bg
60.5
69.8
73.7
75.6
3D-R2N2 w/pose
V-LSM
25
Gap in Performance
1
# Views
Table 1: Mean Voxel IoU on the ShapeNet test set. Note
that the original 3D-R2N2 system does not use camera
pose whereas the 3D-R2N2 w/pose system is trained
with pose information. V-LSM w/bg refers to voxel LSM
trained and tested with random images as backgrounds
instead of white backgrounds only.
20
15
10
5
1
2
3
4
5
6
Number of Views
7
8
Figure 4: Generalization performance for V-LSM
and 3D-R2N2 w/pose measured by gap in voxel
IoU when tested on unseen object categories.
Figure 5: Qualitative results for per-view depth map prediction on ShapeNet. We show the depth maps predicted
by Depth-LSM (visualized with shading from a shifted viewpoint) and the point cloud obtained by unprojecting
them into world coordinates.
(and the 3D-R2N2 w/pose) and trained it on our generated data (images and voxel grids). Due to the
difference in training data/splits and the implementation, the numbers are not directly comparable
to the ones reported in [5] but we observe similar performance trends. For the 3D-R2N2 w/pose
system, we use the camera pose quaternion as the pose representation and process it through 2 fully
connected layers before concatenating it with the feature passed into the LSTM. Table 1 reports the
mean voxel IoU (across 13 categories) for sequences of {1, 2, 3, 4} views. The accuracy increases
with number of views for all methods but it can be seen that the jump is much less for the R2N2
methods indicating that it already produces a good enough estimate at the beginning but fails to
effectively use multiple views to improve its reconstruction significantly. The R2N2 system with
naively integrated pose fails to improve over the base version, completely ignoring it in favor of
just image-based information. On the other hand, our system, designed specifically to exploit these
geometric multi-view cues improves significantly with more views. Figure 3 shows some example
reconstructions for V-LSM and 3D-R2N2 w/pose. Our system progressively improves based on
the viewpoint it receives while the R2N2 w/pose system makes very confident predictions early on
(sometimes ?retrieving" a completely different instance) and then stops improving as much. As we
use a geometric approach, we end up memorizing less and reconstruct when possible. More detailed
results can be found in the supplementary material.
Generalization. In order to test how well LSMs learn to generalize to unseen data, we split our
data into 2 parts with disjoint sets of classes - split 1 has data from 6 classes while split 2 has data
from the other 7. We train three V-LSMs - trained on split 1 (V-LSM-S1), on split 2 (V-LSM-S2) and
both splits combined (V-LSM-All). The quantity we are interested in is the change in performance
when we test the system on a category it hasn?t seen during training. We use the difference in test IoU
of a category C between V-LSM-All and V-LSM-S1 if C is not in split 1 and vice versa. Figure 4
shows the mean of this quantity across all classes as the number of views change. It can be seen that
for a single view, the difference in performance is fairly high and as we see more views, the difference
7
in performance decreases - indicating that our system has learned to exploit category agnostic shape
cues. On the other hand, the 3D-R2N2 w/pose system fails to generalize with more views. Note that
the V-LSMs have been trained with a time horizon of 4 but are evaluated till upto 8 steps here.
Mean Voxel IoU
Sensitivity to noisy camera pose and masks.
80
We conducted experiments to quantify the efAccurate Pose
Noisy Pose
Noise:0
Noise:0
fects of noisy camera pose and segmentations
Noise:10
Noise:10
75
Noise:20
Noise:20
on performance for V-LSMs. We evaluated
70
models trained with perfect poses on data with
perturbed camera extrinsics and observed that
65
performance degrades (as expected) yet still re60
mains better than the baseline (at 10? noise).
55
We also trained new models with synthetically
perturbed extrinsics and achieve significantly
50
higher robustness to noisy poses while maintain45
Views:1
Views:2
Views:3
Views:4
ing competitive performance (Figure 6). This is
illustrated in Figure 6. The perturbation is intro- Figure 6: Sensitivity to noise in camera pose estimates
duced by generating a random rotation matrix for V-LSM for systems trained with and without pose
which rotates the viewing axis by a max angular perturbation.
magnitude ? while still pointing at the object of
interest.
We also trained LSMs on images with random images backgrounds (V-LSM w/bg in Table 1) rather
than only white backgrounds and saw a very small drop in performance. This shows that our method
learns to match features rather than relying heavily on perfect segmentations.
Multi-view Depth Map Prediction. We show qualitative results from Depth LSM in Figure 5. We
manage to obtain thin structures in challenging examples (chairs/tables) while predicting consistent
geometry for all the views. We note that the skip connections from the image to last layers for D-LSM
do help in directly using low level image features while producing depth maps. The depth maps
are viewed with shading in order to point out that we produce metrically accurate geometry. The
unprojected point clouds also align well with each other showing the merits of jointly predicting the
depth maps from a global volume rather than processing them independently.
Comparision to Plane Sweeping. We qualitatively compare D-LSM to the popular plane sweeping
(PS) approach [6, 57] for stereo matching. Figure 7 shows the unprojected point clouds from per
view depths maps produced using PS and D-LSM using 5 and 10 images. We omit an evaluation with
less images as plane sweeping completely fails with fewer images. We use the publicly available
implementation for the PS algorithm [19] and use 5x5 zero mean normalized cross correlation as
matching windows with 300 depth planes. We can see that our approach is able to produce much
cleaner point clouds with less input images. It is robust to texture-less areas where traditional stereo
algorithms fail (e.g. the car windows) by using shape priors to reason about them. We also conducted
a quantitative comparison using PS and D-LSM with 10 views (D-LSM
was trained using only four
?
images). The evaluation region is limited to a depth range of ? 3/2 (maximally possible depth
range) around the origin as the original models lie in a unit cube centered at the origin. Furthermore,
pixels where PS is not able to provide a depth estimate are not taken into account. Note that all these
choices disadvantage our method. We compute the per depth map error as the median absolute depth
difference for the valid pixels, aggregate to a per category mean error and report the average of the
per category means for PS as 0.051 and D-LSM as 0.024. Please refer to the supplementary material
for detailed results.
5
Discussion
We have presented Learnt Stereo Machines (LSM) - an end-to-end learnt system that performs
multi-view stereopsis. The key insight of our system is to use ideas from projective geometry
to differentiably transfer features between 2D images and the 3D world and vice-versa. In our
experiments we showed the benefits of our formulation over direct methods - we are able to generalize
to new object categories and produce compelling reconstructions with fewer images than classical
8
(a) PS 5 Images
(b) LSM 5 Images
(c) PS 10 Images
(d) LSM 10 Images
(e) PS 20 Images
Figure 7: Comparison between Depth-LSM and plane sweeping stereo (PS) with varying numbers of images.
systems. However, our system also has some limitations. We discuss some below and describe how
they lead to future work.
A limiting factor in our current system is the coarse resolution (323 ) of the world grid. Classical
algorithms typically work on much higher resolutions frequently employing special data structures
such as octrees. We can borrow ideas from recent works [43, 21] which show that CNNs can predict
such high resolution volumes. We also plan to apply LSMs to more general geometry than objects,
eventually leading to a system which can reconstruct single/multiple objects and entire scenes. The
main challenge in this setup is to find the right global grid representation. In scenes for example, a
grid in terms of a per-view camera frustum might be more appropriate than a global aligned euclidean
grid.
In our experiments we evaluated classical multi-view 3D reconstruction where the goal is to produce
3D geometry from images with known poses. However, our system is more general and the projection
modules can be used wherever one needs to move between 2D image and 3D world frames. Instead
of predicting just depth maps from our final world representation, one can also predict other view
specific representations such as silhouettes or pixel wise part segmentation labels etc. We can also
project the final world representation into views that we haven?t observed as inputs (we would omit
the skip connections from the image encoder to make the projection unconditional). This can be used
to perform view synthesis grounded in 3D.
Acknowledgments
This work was supported in part by NSF Award IIS- 1212798 and ONR MURI-N00014-10-1-0933.
Christian H?ne is supported by an ?Early Postdoc.Mobility? fellowship No. 165245 from the Swiss
National Science Foundation. The authors would like to thank David Fouhey, Saurabh Gupta and
Shubham Tulsiani for valuable discussions and Fyusion Inc. for providing GPU hours for the work.
References
[1] S. Becker and G. E. Hinton. Self-organizing neural network that discovers surfaces in random-dot
stereograms. Nature, 1992.
[2] V. Blanz and T. Vetter. A morphable model for the synthesis of 3d faces. In Conference on Computer
Graphics and Interactive Techniques, 1999.
[3] A. X. Chang, T. Funkhouser, L. Guibas, P. Hanrahan, Q. Huang, Z. Li, S. Savarese, M. Savva, S. Song,
H. Su, J. Xiao, L. Yi, and F. Yu. Shapenet: An information-rich 3d model repository. arXiv preprint
arXiv:1512.03012, 2015.
[4] K. Cho, B. van Merrienboer, C. Gulcehre, F. Bougares, H. Schwenk, and Y. Bengio. Learning phrase
representations using rnn encoder-decoder for statistical machine translation. In Conference on Empirical
Methods in Natural Language Processing (EMNLP 2014), 2014.
[5] C. B. Choy, D. Xu, J. Gwak, K. Chen, and S. Savarese. 3d-r2n2: A unified approach for single and
multi-view 3d object reconstruction. In European Conference on Computer Vision (ECCV), 2016.
[6] R. T. Collins. A space-sweep approach to true multi-image matching. In Conference on Computer Vision
and Pattern Recognition (CVPR), 1996.
9
[7] D. Cremers and K. Kolev. Multiview stereo and silhouette consistency via convex functionals over convex
domains. Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 33(6):1161?1174, 2011.
[8] B. Curless and M. Levoy. A volumetric method for building complex models from range images. In
Conference on Computer Graphics and Interactive Techniques, 1996.
[9] A. Dame, V. A. Prisacariu, C. Y. Ren, and I. Reid. Dense reconstruction using 3d object shape priors. In
Conference on Computer Vision and Pattern Recognition (CVPR), 2013.
[10] D. Eigen, C. Puhrsch, and R. Fergus. Depth map prediction from a single image using a multi-scale deep
network. In Neural Information Processing Systems (NIPS), 2014.
[11] H. Fan, H. Su, and L. Guibas. A point set generation network for 3d object reconstruction from a single
image. In Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
[12] J. Flynn, I. Neulander, J. Philbin, and N. Snavely. Deepstereo: Learning to predict new views from the
world?s imagery. In Conference on Computer Vision and Pattern Recognition (CVPR), 2016.
[13] Y. Furukawa and J. Ponce. Accurate, dense, and robust multi-view stereopsis. Transactions on Pattern
Analysis and Machine Intelligence (TPAMI), 2010.
[14] R. Garg, G. Carneiro, and I. Reid. Unsupervised cnn for single view depth estimation: Geometry to the
rescue. In European Conference on Computer Vision (ECCV), 2016.
[15] P. Gargallo, E. Prados, and P. Sturm. Minimizing the reprojection error in surface reconstruction from
images. In International Conference on Computer Vision (ICCV), pages 1?8, 2007.
[16] R. Girdhar, D. F. Fouhey, M. Rodriguez, and A. Gupta. Learning a predictable and generative vector
representation for objects. In European Conference on Computer Vision (ECCV), 2016.
[17] C. Haene, C. Zach, A. Cohen, and M. Pollefeys. Dense semantic 3d reconstruction. Transactions on
Pattern Analysis and Machine Intelligence (TPAMI), 2016.
[18] X. Han, T. Leung, Y. Jia, R. Sukthankar, and A. C. Berg. Matchnet: Unifying feature and metric learning
for patch-based matching. In Conference on Computer Vision and Pattern Recognition (CVPR), 2015.
[19] C. H?ne, L. Heng, G. H. Lee, A. Sizov, and M. Pollefeys. Real-time direct dense matching on fisheye
images using plane-sweeping stereo. In International Conference on 3D Vision (3DV), 2014.
[20] C. H?ne, N. Savinov, and M. Pollefeys. Class specific 3d object shape priors using surface normals. In
Conference on Computer Vision and Pattern Recognition (CVPR), 2014.
[21] C. H?ne, S. Tulsiani, and J. Malik. Hierarchical surface prediction for 3d object reconstruction. In
International Conference on 3D Vision (3DV), 2017.
[22] C. H?ne, C. Zach, A. Cohen, R. Angst, and M. Pollefeys. Joint 3d scene reconstruction and class
segmentation. In Conference on Computer Vision and Pattern Recognition (CVPR), 2013.
[23] W. Hartmann, S. Galliani, M. Havlena, K. Schindler, and L. V. Gool. Learned multi-patch similarity. In
International Conference on Computer Vision, (ICCV), 2017.
[24] S. Hochreiter and J. Schmidhuber. Long short-term memory. Neural computation, 1997.
[25] M. Jaderberg, K. Simonyan, A. Zisserman, and K. Kavukcuoglu. Spatial transformer networks. In Neural
Information Processing Systems (NIPS), 2015.
[26] T. Kanade, H. Kano, S. Kimura, A. Yoshida, and K. Oda. Development of a video-rate stereo machine. In
International Conference on Intelligent Robots and Systems (IROS), 1995.
[27] A. Kar, S. Tulsiani, J. Carreira, and J. Malik. Category-specific object reconstruction from a single image.
In Conference on Computer Vision and Pattern Recognition (CVPR), 2015.
[28] A. Kendall, H. Martirosyan, S. Dasgupta, P. Henry, R. Kennedy, A. Bachrach, and A. Bry. End-to-end
learning of geometry and context for deep stereo regression. In International Conference on Computer
Vision (ICCV), 2017.
[29] K. N. Kutulakos and S. M. Seitz. A theory of shape by space carving. International Journal of Computer
Vision (IJCV), 38(3):199?218, 2000.
10
[30] P. Labatut, J.-P. Pons, and R. Keriven. Efficient multi-view reconstruction of large-scale scenes using
interest points, delaunay triangulation and graph cuts. In International Conference on Computer Vision,
(ICCV), 2007.
[31] L. Ladicky, J. Shi, and M. Pollefeys. Pulling things out of perspective. In Conference on Computer Vision
and Pattern Recognition (CVPR), 2014.
[32] A. Laurentini. The visual hull concept for silhouette-based image understanding. Transactions on Pattern
Analysis and Machine Intelligence (TPAMI), 16(2):150?162, 1994.
[33] V. Lempitsky and Y. Boykov. Global optimization for shape fitting. In Conference on Computer Vision
and Pattern Recognition, (CVPR), 2007.
[34] M. Lhuillier and L. Quan. A quasi-dense approach to surface reconstruction from uncalibrated images.
Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2005.
[35] S. Liu and D. B. Cooper. Statistical inverse ray tracing for image-based 3d modeling. Transactions on
Pattern Analysis and Machine Intelligence (TPAMI), 36(10):2074?2088, 2014.
[36] D. Marr and T. Poggio. Cooperative computation of stereo disparity. In From the Retina to the Neocortex,
pages 239?243. 1976.
[37] N. Mayer, E. Ilg, P. Hausser, P. Fischer, D. Cremers, A. Dosovitskiy, and T. Brox. A large dataset to train
convolutional networks for disparity, optical flow, and scene flow estimation. In Conference on Computer
Vision and Pattern Recognition (CVPR), 2016.
[38] P. Merrell, A. Akbarzadeh, L. Wang, P. Mordohai, J.-M. Frahm, R. Yang, D. Nist?r, and M. Pollefeys.
Real-time visibility-based fusion of depth maps. In International Conference on Computer Vision (ICCV),
2007.
[39] T. Pollard and J. L. Mundy. Change detection in a 3-d world. In Conference on Computer Vision and
Pattern Recognition (CVPR), 2007.
[40] M. Pollefeys, D. Nist?r, J.-M. Frahm, A. Akbarzadeh, P. Mordohai, B. Clipp, C. Engels, D. Gallup, S.-J.
Kim, P. Merrell, C. Salmi, S. Sinha, B. Talton, L. Wang, Q. Yang, H. Stewenius, R. Yang, G. Welch, and
H. Towles. Detailed real-time urban 3d reconstruction from video. International Journal of Computer
Vision (IJCV), 78(2):143?167, 2008.
[41] M. Pollefeys, L. Van Gool, M. Vergauwen, F. Verbiest, K. Cornelis, J. Tops, and R. Koch. Visual modeling
with a hand-held camera. International Journal of Computer Vision (IJCV), 59(3):207?232, 2004.
[42] D. J. Rezende, S. A. Eslami, S. Mohamed, P. Battaglia, M. Jaderberg, and N. Heess. Unsupervised learning
of 3d structure from images. In Neural Information Processing Systems (NIPS), 2016.
[43] G. Riegler, A. O. Ulusoy, H. Bischof, and A. Geiger. Octnetfusion: Learning depth fusion from data. In
International Conference on 3D Vision (3DV), 2017.
[44] O. Ronneberger, P.Fischer, and T. Brox. U-net: Convolutional networks for biomedical image segmentation.
In Medical Image Computing and Computer-Assisted Intervention (MICCAI), 2015.
[45] N. Savinov, C. H?ne, L. Ladicky, and M. Pollefeys. Semantic 3d reconstruction with continuous regularization and ray potentials using a visibility consistency constraint. In Conference on Computer Vision and
Pattern Recognition (CVPR), 2016.
[46] A. Saxena, J. Schulte, and A. Y. Ng. Depth estimation using monocular and stereo cues. In Neural
Information Processing Systems (NIPS), 2005.
[47] D. Scharstein and R. Szeliski. A taxonomy and evaluation of dense two-frame stereo correspondence
algorithms. International Journal of Computer Vision (IJCV), 47(1-3):7?42, 2002.
[48] S. M. Seitz, B. Curless, J. Diebel, D. Scharstein, and R. Szeliski. A comparison and evaluation of multiview stereo reconstruction algorithms. In Conference on Computer Vision and Pattern Recognition (CVPR),
2006.
[49] A. Sinha, J. Bai, and K. Ramani. Deep learning 3d shape surfaces using geometry images. In European
Conference on Computer Vision (ECCV), 2016.
[50] S. N. Sinha, P. Mordohai, and M. Pollefeys. Multi-view stereo via graph cuts on the dual of an adaptive
tetrahedral mesh. In International Conference on Computer Vision (ICCV), 2007.
11
[51] M. Tatarchenko, A. Dosovitskiy, and T. Brox. Multi-view 3d models from single images with a convolutional network. In European Conference on Computer Vision (ECCV), 2016.
[52] S. Tulsiani, A. Kar, J. Carreira, and J. Malik. Learning category-specific deformable 3d models for object
reconstruction. Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2016.
[53] S. Tulsiani, T. Zhou, A. A. Efros, and J. Malik. Multi-view supervision for single-view reconstruction via
differentiable ray consistency. In Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
[54] A. O. Ulusoy, A. Geiger, and M. J. Black. Towards probabilistic volumetric reconstruction using ray
potentials. In International Conference on 3D Vision (3DV), 2015.
[55] G. Vogiatzis, P. H. Torr, and R. Cipolla. Multi-view stereo via volumetric graph-cuts. In Conference on
Computer Vision and Pattern Recognition (CVPR), 2005.
[56] X. Yan, J. Yang, E. Yumer, Y. Guo, and H. Lee. Perspective transformer nets: Learning single-view 3d
object reconstruction without 3d supervision. In Neural Information Processing Systems (NIPS), 2016.
[57] R. Yang, G. Welch, and G. Bishop. Real-time consensus-based scene reconstruction using commodity
graphics hardware. In Computer Graphics Forum, 2003.
[58] S. Yingze Bao, M. Chandraker, Y. Lin, and S. Savarese. Dense object reconstruction with semantic priors.
In Conference on Computer Vision and Pattern Recognition (CVPR), 2013.
[59] C. Zach, T. Pock, and H. Bischof. A globally optimal algorithm for robust tv-l 1 range image integration.
In International Conference on Computer Vision, (ICCV), 2007.
[60] J. Zbontar and Y. LeCun. Stereo matching by training a convolutional neural network to compare image
patches. Journal of Machine Learning Research (JMLR), 17(1-32):2, 2016.
[61] T. Zhou, M. Brown, N. Snavely, and D. G. Lowe. Unsupervised learning of depth and ego-motion from
video. In Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
12
| 6640 |@word cnn:6 version:2 repository:1 replicate:1 choy:2 seitz:2 shading:2 reduction:2 bai:1 liu:1 disparity:9 score:1 ours:1 current:1 comparing:1 yet:2 conforming:2 gpu:2 mesh:2 subsequent:1 realistic:1 visible:1 shape:23 christian:2 enables:1 visibility:2 designed:2 drop:1 depict:1 progressively:2 cue:13 fewer:4 intelligence:7 generative:1 plane:10 beginning:1 short:1 coarse:1 lsm:47 location:7 shubham:1 along:20 constructed:1 direct:4 retrieving:1 qualitative:3 ijcv:4 combine:1 oda:1 ray:32 fitting:1 manner:7 mask:1 expected:2 frequently:1 multi:28 discretized:1 inspired:1 globally:2 relying:2 little:1 window:2 project:2 underlying:6 matched:1 moreover:1 linearity:1 agnostic:3 kind:1 gif:4 unified:2 finding:1 flynn:1 kimura:1 berkeley:6 every:2 quantitative:3 ti:1 saxena:1 tackle:1 interactive:2 commodity:1 demonstrates:1 hit:1 unit:4 medical:1 omit:2 intervention:1 producing:3 reid:2 before:2 local:3 pock:1 bilinear:2 eslami:1 riegler:1 path:1 interpolation:3 cornelis:1 signed:1 might:1 garg:1 black:1 challenging:2 limited:1 projective:4 bilinearly:1 range:5 perpendicular:1 deepstereo:1 carving:1 acknowledgment:1 camera:29 pla:1 lecun:1 practice:1 block:3 union:1 swiss:1 procedure:2 intersect:1 area:1 rnn:1 empirical:1 yan:1 significantly:4 thought:1 projection:18 matching:19 vetter:1 refers:1 ronneberger:1 get:1 cannot:1 context:2 transformer:2 sukthankar:1 map:54 demonstrated:1 center:1 shi:1 straightforward:1 go:2 yoshida:1 independently:1 convex:2 focused:2 resolution:5 bachrach:1 welch:2 recovery:1 insight:1 borrow:1 marr:1 handle:3 traditionally:1 variation:1 coordinate:1 limiting:1 heavily:1 us:5 hypothesis:1 origin:3 trend:1 bry:1 recognition:18 ego:1 showcase:1 cut:3 muri:1 cooperative:1 observed:2 cloud:6 module:2 preprint:1 wang:2 capture:1 region:2 connected:4 ordering:3 decrease:2 valuable:1 uncalibrated:1 mentioned:2 predictable:1 stereograms:2 complexity:1 trained:13 purely:1 efficiency:1 completely:3 joint:2 schwenk:1 various:1 carneiro:1 train:7 stacked:2 describe:1 aggregate:4 lift:1 neighborhood:1 posed:2 supplementary:3 foregoing:1 cvpr:18 reconstruct:4 encoder:10 ability:1 favor:1 blanz:1 unseen:6 simonyan:1 fischer:2 jointly:3 noisy:5 final:8 sequence:4 differentiable:10 tpami:7 net:2 reconstruction:39 propose:2 relevant:1 combining:1 aligned:1 organizing:1 till:1 achieve:1 deformable:1 bao:1 az:1 exploiting:1 reprojection:1 optimum:1 p:10 produce:24 generating:2 incremental:2 adam:1 perfect:2 object:25 help:2 recurrent:8 completion:1 pose:39 measured:1 nearest:2 tulsiani:5 strong:1 implemented:3 predicted:1 involves:1 skip:5 quantify:1 iou:7 direction:1 cnns:2 hull:7 centered:2 viewing:10 enable:1 material:3 require:1 generalization:3 merrienboer:1 traversed:1 extension:1 assisted:1 around:3 koch:1 ground:3 guibas:2 normal:1 predict:7 pointing:1 major:1 efros:1 early:7 adopt:1 purpose:1 battaglia:1 estimation:3 injecting:1 label:1 saw:1 ilg:1 vice:2 rough:1 aim:1 rather:4 zhou:2 resized:1 lifted:1 varying:1 encode:1 rezende:1 focus:1 ponce:1 improvement:1 contrast:3 shapenet:9 baseline:6 kim:1 inference:5 el:1 leung:1 typically:4 integrated:1 entire:1 hidden:1 transformed:3 quasi:2 interested:1 pixel:5 arg:2 issue:5 mimicking:1 hartmann:1 dual:1 development:1 plan:1 spatial:2 special:1 softmax:1 brox:3 fairly:1 uc:3 cube:2 once:1 having:2 beach:1 sampling:7 saurabh:1 schulte:1 field:3 ng:1 look:1 yu:1 unsupervised:3 thin:1 triangulate:1 future:2 report:2 intelligent:1 haven:1 primarily:1 fouhey:2 modern:1 randomly:1 retina:1 dosovitskiy:2 preserve:1 national:1 individual:3 geometry:21 detection:1 interest:2 highly:1 regressing:1 evaluation:4 lsms:18 extreme:1 light:1 behind:2 permuting:1 unconditional:1 held:1 accurate:4 partial:1 poggio:1 respective:1 mobility:1 euclidean:1 savarese:3 desired:1 sinha:3 instance:3 column:1 soft:1 compelling:1 modeling:2 disadvantage:1 pons:1 phrase:1 cost:5 subset:1 successful:1 keriven:1 conducted:2 too:1 xwk:2 graphic:4 reported:1 perturbed:2 learnt:9 synthetic:2 combined:1 thoroughly:2 st:1 confident:1 lstm:1 sensitivity:2 cho:1 international:16 probabilistic:2 lee:2 pool:1 synthesis:2 fused:6 again:1 imagery:1 manage:1 containing:2 choose:2 huang:1 emnlp:1 classically:1 zbontar:1 inject:1 leading:2 li:1 account:2 potential:2 de:1 sec:1 inc:1 jitendra:1 hasn:1 cremers:2 mv:5 bg:3 stream:1 stewenius:1 view:69 try:1 philbin:1 kendall:2 doing:1 lowe:1 competitive:1 relied:1 jia:1 oi:2 ni:13 publicly:1 convolutional:7 variance:1 largely:1 accuracy:1 spaced:2 correspond:1 trilinear:1 generalize:5 modelled:1 raw:1 curless:2 neulander:1 produced:4 kavukcuoglu:1 ren:1 lighting:1 kennedy:1 whenever:1 aligns:1 volumetric:8 energy:1 frequency:1 mohamed:1 di:1 recovers:1 gain:1 sampled:2 dataset:7 stop:1 popular:1 color:2 knowledge:1 car:2 improves:2 segmentation:5 savinov:2 ramani:1 yumer:1 back:3 feed:2 higher:2 zisserman:1 maximally:1 formulation:3 done:1 evaluated:3 strongly:1 furthermore:1 just:5 binocular:6 miccai:1 stage:1 angular:1 correlation:1 canvas:2 hand:3 receives:1 sturm:1 biomedical:1 su:2 nonlinear:1 incrementally:1 rodriguez:1 mode:1 quality:1 pulling:1 supervisory:1 building:1 usa:1 facilitate:1 normalized:1 true:1 gtx:1 brown:1 unroll:1 analytically:2 inspiration:1 assigned:1 spatially:1 regularization:1 semantic:9 illustrated:1 deal:1 white:2 funkhouser:1 x5:1 during:8 self:1 please:1 kutulakos:1 doesn:1 multiview:3 complete:1 demonstrate:3 performs:2 l1:1 motion:1 reasoning:4 image:99 wise:2 novel:1 recently:1 fi:3 discovers:1 boykov:1 rotation:1 empirically:1 overview:2 cohen:2 volume:9 accumulate:1 refer:3 bougares:1 versa:2 smoothness:2 grid:62 consistency:9 trivially:3 language:1 gwak:1 dot:2 henry:1 robot:1 access:3 stable:1 similarity:3 surface:13 etc:2 add:2 supervision:3 base:1 align:1 closest:1 morphable:1 recent:8 showed:1 perspective:3 delaunay:1 triangulation:1 schmidhuber:1 store:1 certain:1 n00014:1 integration:1 kar:3 binary:1 onr:1 hanrahan:1 yi:1 furukawa:1 seen:5 additional:1 signal:1 ii:3 full:2 multiple:8 desirable:1 infer:1 smooth:1 ing:1 match:5 cross:2 long:3 sphere:1 lin:1 equally:2 award:1 prediction:10 variant:3 basic:1 regression:3 vision:39 metric:9 arxiv:2 iteration:1 grounded:3 represent:1 normalization:2 sometimes:1 hochreiter:1 whereas:4 background:4 fine:2 fellowship:1 girdhar:1 median:1 zw:2 operate:1 rest:1 pooling:1 validating:1 thing:1 quan:1 flow:2 unprojected:6 extracting:2 vw:1 yang:5 leverage:1 synthetically:1 constraining:1 r2n2:20 enough:3 easy:1 variety:1 split:9 fit:1 rendering:1 bengio:1 architecture:4 qualitatively:1 idea:3 six:1 passed:2 becker:2 stereo:32 song:1 pollard:1 deep:4 heess:1 detailed:3 cleaner:1 gfi:1 transforms:2 neocortex:1 locally:1 hardware:1 processed:3 category:14 visualized:1 generate:1 nsf:1 shifted:1 rescue:1 extrinsic:1 per:18 disjoint:1 discrete:4 pollefeys:10 dasgupta:1 key:2 four:1 demonstrating:1 threshold:1 urban:1 schindler:1 iros:1 utilize:2 fuse:3 graph:3 geometrically:4 inverse:1 injected:1 place:2 reader:1 patch:5 geiger:2 draw:1 decision:1 comparable:1 layer:8 dame:1 followed:2 correspondence:3 fan:1 comparision:1 constraint:6 ladicky:2 scene:9 carve:1 min:2 formulating:2 chair:2 tatarchenko:1 optical:1 rendered:3 tv:1 combination:2 kano:1 across:3 mordohai:3 making:1 s1:2 wherever:1 haene:1 memorizing:1 projecting:2 intuitively:1 iccv:7 gathering:1 dv:4 taken:1 monocular:1 equation:1 agree:1 visualization:1 previously:1 discus:1 fail:2 eventually:1 needed:2 merit:1 ascending:1 end:20 photo:7 gulcehre:1 available:1 operation:21 apply:2 observe:2 kolev:1 hierarchical:1 upto:2 reshape:1 appropriate:1 save:1 batch:1 robustness:1 eigen:1 extrinsics:2 original:2 top:3 running:1 unifying:1 xw:6 exploit:6 giving:1 concatenated:1 build:1 classical:18 forum:1 sweep:1 malik:6 move:1 question:1 quantity:3 already:1 intro:1 receptive:2 strategy:1 degrades:1 snavely:2 traditional:3 matchnet:1 gradient:1 distance:2 thank:1 mapped:1 rotates:1 vd:1 decoder:1 consensus:1 binarize:1 reason:6 enforcing:1 relationship:1 illustration:2 providing:2 convincing:1 minimizing:1 setup:1 susceptible:1 taxonomy:1 trace:2 append:1 design:2 implementation:4 perform:3 gated:1 allowing:1 convolution:8 observation:1 mundy:1 enabling:5 nist:2 curved:1 hinton:2 frame:5 perturbation:2 smoothed:2 sweeping:7 police:1 arbitrary:1 tetrahedral:1 duced:1 diebel:1 david:1 pair:2 required:1 specified:1 gru:4 connection:6 mayer:1 bischof:2 puhrsch:1 hausser:1 learned:2 tensorflow:1 hour:1 nip:6 assembled:1 address:1 able:11 beyond:2 alongside:1 below:3 pattern:25 differentiably:2 challenge:1 built:1 including:1 max:3 video:4 epipolar:2 belief:1 gool:2 memory:1 natural:1 predicting:3 arm:1 representing:1 occupancy:13 siren:1 improve:2 ne:8 axis:1 extract:1 prior:11 geometric:12 voxels:4 understanding:1 val:1 multiplication:1 embedded:1 fully:4 loss:1 rationale:1 generation:1 limitation:1 foundation:1 degree:1 consistent:4 xiao:1 principle:4 viewpoint:4 heng:1 pi:1 share:1 translation:1 eccv:5 concept:1 supported:2 last:2 copy:1 keeping:1 allow:1 szeliski:2 neighbor:1 taking:1 face:1 absolute:2 tracing:4 benefit:2 van:2 depth:48 dimension:2 world:18 valid:1 rich:1 computes:1 concavity:1 author:1 forward:2 adaptive:1 levoy:1 collection:1 refinement:1 projected:5 made:1 pth:1 voxel:25 jump:1 employing:1 functionals:1 reconstructed:1 transaction:7 scharstein:2 implicitly:2 jaderberg:2 silhouette:8 global:7 sequentially:1 chandraker:1 abhishek:1 fergus:1 stereopsis:5 continuous:3 table:4 kanade:2 learn:10 channel:4 robust:3 nature:2 ca:1 zk:1 symmetry:1 ignoring:1 improving:1 transfer:1 european:5 postdoc:1 complex:1 domain:1 pk:2 dense:12 main:2 linearly:1 s2:1 noise:8 complementary:1 x1:2 xu:1 frahm:2 batched:1 cooper:1 fails:4 decoded:2 guiding:1 zach:3 concatenating:2 lie:4 jmlr:1 learns:4 specific:8 bishop:1 showing:1 learnable:1 gupta:2 fusion:4 naively:1 incorporating:3 deconvolution:4 intrinsic:1 polygonal:1 adding:1 effectively:5 flattened:1 sequential:1 metrically:2 lifting:1 magnitude:1 texture:1 horizon:1 gap:2 chen:1 entropy:1 intersection:1 savva:1 appearance:1 visual:7 upsample:1 chang:1 cipolla:1 acquiring:1 truth:3 lempitsky:1 viewed:2 goal:4 towards:1 shared:1 feasible:1 hard:1 change:3 carreira:2 specifically:4 except:1 operates:1 reducing:1 torr:1 han:1 pas:1 ulusoy:2 meaningful:1 indicating:2 berg:1 internal:1 guo:1 arises:1 quaternion:1 collins:1 incorporate:1 evaluate:5 tested:2 handling:1 |
6,236 | 6,641 | Phase Transitions in the Pooled Data Problem
Jonathan Scarlett and Volkan Cevher
Laboratory for Information and Inference Systems (LIONS)
?cole Polytechnique F?d?rale de Lausanne (EPFL)
{jonathan.scarlett,volkan.cevher}@epfl.ch
Abstract
In this paper, we study the pooled data problem of identifying the labels associated with a large collection of items, based on a sequence of pooled tests revealing
the counts of each label within the pool. In the noiseless setting, we identify an
exact asymptotic threshold on the required number of tests with optimal decoding, and prove a phase transition between complete success and complete failure.
In addition, we present a novel noisy variation of the problem, and provide an
information-theoretic framework for characterizing the required number of tests
for general random noise models. Our results reveal that noise can make the problem considerably more difficult, with strict increases in the scaling laws even at
low noise levels. Finally, we demonstrate similar behavior in an approximate recovery setting, where a given number of errors is allowed in the decoded labels.
1
Introduction
Consider the following setting: There exists a large population of items, each of which has an
associated label. The labels are initially unknown, and are to be estimated based on pooled tests.
Each pool consists of some subset of the population, and the test outcome reveals the total number
of items corresponding to each label that are present in the pool (but not the individual labels). This
problem, which we refer to as the pooled data problem, was recently introduced in [1,2], and further
studied in [3, 4]. It is of interest in applications such as medical testing, genetics, and learning with
privacy constraints, and has connections to the group testing problem [5] and its linear variants [6,7].
The best known bounds on the required number of tests under optimal decoding were given in [3];
however, the upper and lower bounds therein do not match, and can exhibit a large gap. In this
paper, we completely close these gaps by providing a new lower bound that exactly matches the
upper bound of [3]. These results collectively reveal a phase transition between success and failure,
with the probability of error vanishing when the number of tests exceeds a given threshold, but
tending to one below that threshold. In addition, we explore the novel aspect of random noise in the
measurements, and show that this can significantly increase the required number of tests. Before
summarizing these contributions in more detail, we formally introduce the problem.
1.1
Problem setup
We consider a large population of items [p] = {1, . . . , p}, each of which has an associated label
in [d] = {1, . . . , d}. We let ? = (?1 , . . . , ?d ) denote a vector containing the proportions of items
having each label, and we assume that the vector of labels itself, = ( 1 , . . . , p ), is uniformly
distributed over the sequences consistent with these proportions:
? Uniform(B(?)),
where B(?) is the set of length-p sequences whose empirical distribution is ?.
(1)
The goal is to recover based on a sequence of pooled tests. The i-th test is represented by a
(i)
(possibly random) vector X (i) 2 {0, 1}p , whose j-th entry Xj indicates whether the j-th item is
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
Necessary for Pe 6! 1
(this paper)
p
p
1
p
?
max
f (r)
? f (1)
?
max
f (r)
log p r2{1,...,d 1}
log p 2
log p r2{1,...,d 1}
Table 1: Necessary and sufficient conditions on the number of tests n in the noiseless setting. The
function f (r) is defined in (5). Asymptotic multiplicative 1 + o(1) terms are omitted.
Sufficient for Pe ! 0 [3]
Necessary for Pe 6! 1 [3]
Noisy testing
Noisy testing
Noisy testing
Noiseless testing
(SNR = ?(1))
(SNR = (log p)?(1) )
(SNR = p?(1) )
? p ?
? p ?
?
?
p
?
?
?
? p log p
log p
log p
log log p
Table 2: Necessary and sufficient conditions on the number of tests n in the noisy setting. SNR
denotes the signal-to-noise ratio, and the noise model is given in Section 2.2.
included in the i-th test. We define a measurement matrix X 2 {0, 1}n?p whose i-th row is given
by X (i) for i = 1, . . . , n, where n denotes the total number of tests. We focus on the non-adaptive
testing scenario, where the entire matrix X must be specified prior to performing any tests.
(i)
(i)
In the noiseless setting, the i-th test outcome is a vector Y (i) = (Y1 , . . . , Yd ), with t-th entry
(i)
= Nt ( , X (i) ),
(2)
P
where for t = 1, . . . , d, we let Nt ( , X) = j2[p] 1{ j = t \ Xj = 1} denote the number of
items with label t that are included in the test described by X 2 {0, 1}p . More generally, in the
possible presence of noise, the i-th observation is randomly generated according to
Yt
Y (i) | X (i) ,
? PY |N1 ( ,X (i) )...Nd ( ,X (i) )
(3)
for some conditional probability mass function PY |N1 ,...,Nd (or density function in the case of continuous observations). We assume that the observations Y (i) (i = 1, . . . , n) are conditionally independent given X, but otherwise make no assumptions on PY |N1 ,...,Nd . Clearly, the noiseless model
(2) falls under this more general setup.
Similarly to X, we let Y denote an n ? d matrix of observations, with the i-th row being Y (i) . Given
X and Y, a decoder outputs an estimate ? of , and the error probability is given by
Pe = P[ ? 6= ],
(4)
where the probability is with respect to , X, and Y. We seek to find conditions on the number of
tests n under which Pe attains a certain target value in the limit as p ! 1, and our main results
provide necessary conditions (i.e., lower bounds on n) for this to occur. As in [3], we focus on the
case that d and ? are fixed and do not depend on p.1
1.2
Contributions and comparisons to existing bounds
Our focus in this paper is on information-theoretic bounds on the required number of tests that hold
regardless of practical considerations such as computation and storage. Among the existing works in
the literature, the one most relevant to this paper is [3], whose bounds strictly improve on the initial
bounds in [1]. The same authors also proved a phase transition for a practical algorithm based on
approximate message passing [4], but the required number of tests is in fact significantly larger than
the information-theoretic threshold (specifically, linear in p instead of sub-linear).
Table 1 gives a summary of the bounds from [3] and our contributions in the noiseless setting. To
define the function f (r) therein, we introduce the additional notation that for r = {1, . . . , d 1},
(r)
(r)
? (r) = (?1 , . . . , ?r ) is a vector whose first entry sums the largest d r + 1 entries of ?, and
whose remaining entries coincide with the remaining r 1 entries of ?. We have
2(H(?) H(? (r) ))
,
d r
r2{1,...,d 1}
meaning that the entries in Table 1 corresponding to the results of [3] are given as follows:
f (r) =
max
1
(5)
More precisely, ? should be rounded to the nearest empirical distribution (e.g., in `1 -norm) for sequences
2 [d]p of length p; we leave such rounding implicit throughout the paper.
2
1.4
Random :
Uniform :
Highly non-uniform :
1.2
f(r)
1
0.8
0.6
0.4
0.2
0
1
2
3
4
5
6
7
8
9
r
Figure 1: The function f (r) in (5), for several choices of ?, with d = 10.
The random ? are
drawn uniformly on the probability simplex, and the highly non-uniform choice of ? is given by ? =
(0.49, 0.49, 0.0025, . . . , 0.0025). When the maximum is achieved at r = 1, the bounds of [3] coincide up
to a factor of two, whereas if the maximum is achieved for r > 1 then the gap is larger.
? (Achievability) When the entries of X are i.i.d. on Bernoulli(q) for some q 2 (0, 1) (not
depending on p), there exists a decoder such that Pe ! 0 as p ! 1 with
?
?
p
2(H(?) H(? (r) ))
n?
max
(1 + ?)
(6)
log p r2{1,...,d 1}
d r
for arbitrarily small ? > 0.
? (Converse) In order to achieve Pe 6! 1 as p ! 1, it is necessary that
?
?
p
H(?)
n
(1 ?)
log p d 1
(7)
for arbitrarily small ? > 0.
Unfortunately, these bounds do not coincide. If the maximum in (6) is achieved by r = 1 (which
occurs, for example, when ? is uniform [3]), then the gap only amounts to a factor of two. However,
as we show in Figure 1, if we compute the bounds for some ?random? choices of ? then the gap is
typically larger (i.e., r = 1 does not achieve the maximum), and we can construct choices where the
gap is significantly larger. Closing these gaps was posed as a key open problem in [3].
We can now summarize our contributions as follows:
1. We give a lower bound that exactly matches (6), thus completely closing the above-mentioned
gaps in the existing bounds and solving the open problem raised in [3]. More specifically,
(r)
))
we show that Pe ! 1 whenever n ? logp p maxr2{1,...,d 1} 2(H(?)d H(?
(1 ?) for
r
some ? > 0, thus identifying an exact phase transition ? a threshold above which the error
probability vanishes, but below which the error probability tends to one.
2. We develop a framework for understanding variations of the problem consisting of random
noise, and give an example of a noise model where the scaling laws are strictly higher compared to the noiseless case. A summary is given in Table 2; the case SNR = (log p)?(1) reveals
a strict increase in the scaling laws even when the signal-to-noise ratio grows unbounded, and
the case SNR = ?(1) reveals that the required number of tests increases from sub-linear to
super-linear in the dimension when the signal-to-noise ratio is constant.
3. In the supplementary material, we discuss how our lower bounds extend readily to the approximate recovery criterion, where we only require to be identified up to a certain Hamming
distance. However, for clarity, we focus on exact recovery throughout the paper.
In a recent independent work [8], an adversarial noise setting was introduced. This turns out to
be fundamentally different to our noisy setting. In particular, the results of [8] state that exact recovery is impossible, and even with approximate recovery, a huge number of tests (i.e., higher than
1/2+o(1)
polynomial) is needed unless = O qmax
, where qmax is the maximum allowed reconstruction error measured by the Hamming distance, and is maximum adversarial noise amplitude. Of
course, both random and adversarial noise are of significant interest, depending on the application.
3
Notation. For a positive integer d, we write [d] = {1, . . . , d}. We use standard information-theoretic
notations for the (conditional) entropy and mutual information, e.g., H(X), H(Y |X), I(X; Y |Z)
[9]. All logarithms have base e, and accordingly, all of the preceding information measures are in
units of nats. The Gaussian distribution with mean ? and variance 2 is denoted by N(?, 2 ). We
use the standard asymptotic notations O(?), o(?), ?(?), !(?) and ?(?).
2
Main results
In this section, we present our main results for the noiseless and noisy settings. The proofs are given
in Section 3, as well as the supplementary material.
2.1
Phase transition in the noiseless setting
The following theorem proves that the upper bound given in (6) is tight. Recall that for r =
(r)
(r)
{1, . . . , d 1}, ? (r) = (?1 , . . . , ?r ) is a vector whose first entry sums the largest d r + 1
entries of ?, and whose remaining entries coincide with the remaining r 1 entries of ?.
Theorem 1. (Noiseless setting) Consider the pooled data problem described in Section 1.1 with a
given number of labels d and label proportion vector ? (not depending on the dimension p). For any
decoder, in order to achieve Pe 6! 1 as p ! 1, it is necessary that
?
?
p
2(H(?) H(? (r) ))
n
max
(1 ?)
(8)
log p r2{1,...,d 1}
d r
for arbitrarily small ? > 0.
Combined with (6), this result reveals an exact phase transition on the required number of measurer
))
ments: Denoting n? = logp p maxr2{1,...,d 1} 2(H(?)d H(?
, the error probability vanishes for
r
?
?
n n (1 + ?), tends to one for n ? n (1 ?), regardless of how small ? is chosen to be.
Remark 1. Our model assumes that is uniformly distributed over the sequences with empirical
distribution ?, whereas [3] assumes that is i.i.d. on ?. However, Theorem 1 readily extends to
the latter setting: Under the i.i.d. model, once we condition on a given empirical distribution, the
conditional distribution of is uniform. As a result, the converse bound for the i.i.d. model follows
directly from Theorem 1 by basic concentration and the continuity of the entropy function.
2.2
Information-theoretic framework for the noisy setting
We now turn to general noise models of the form (3), and provide necessary conditions for the noisy
pooled data problem in terms of the mutual information. General characterizations of this form were
provided previously for group testing [10, 11] and other sparse recovery problems [12, 13].
Our general result is stated in terms of a maximization over a vector parameter ` = (`1 , . . . , `d ) with
`t 2 {0, . . . , ?t p} for all t. We will see in the proof that `t represents the number of items of type t
that are unknown to the decoder after p?t `t are revealed by a genie. We define the following:
? Given ` and , we let S` be a random set of indices in [p] such that for each t 2 [d], the set
contains `t indices corresponding to entries where equals t. Specifically, we define S` to be
uniformly distributed over all such sets. Moreover, we define S`c = [p] \ S` .
? Given the above definitions, we define
?
j 2 S`c
j
c
(9)
S` =
? otherwise,
where ? can be thought of as representing an unknown value. Hence, knowing S`c amounts to
knowing the labels of all items in the set S`c .
? We define |B` (?)| to be the number of sequences 2 [d]p that coincide with a given S`c on
the entries not equaling ?, while also having empirical distribution ? overall. This number
does not depend on the specific choice of S`c . As an example, when `t = p?t for all t, we have
S` = [p], S`c = (?, . . . , ?), and |B` (?)| = |B(?)|, defined following (1)
? We let k`k0 denote the number of values in (`1 , . . . , `d ) that are positive.
4
With these definitions, we have the following result for general random noise models.
Theorem 2. (Noisy setting) Consider the pooled data problem described in Section 1.1 under a
general observation model of the form (3), with a given number of labels d and label proportion
vector ?. For any decoder, in order to achieve Pe ? for a given 2 (0, 1), it is necessary that
n
max
` : k`k0
log |B` (?)| (1
) log 2
Pn
.
(i)
| S`c , X (i) )
i=1 I( ; Y
(10)
1
2
n
In order to obtain more explicit bounds on n from (10), one needs to characterize the mutual information terms, ideally forming an upper bound that does not depend on the distribution of the
measurement matrix X. We do this for some specific models below; however, in general it can be a
difficult task. The following corollary reveals that if the entries of X are i.i.d. on Bernoulli(q) for
some q 2 (0, 1) (as was assumed in [3]), then we can simplify the bound.
Corollary 1. (Noisy setting with Bernoulli testing) Suppose that the entries of X are i.i.d. on
Bernoulli(q) for some q 2 (0, 1). Under the setup of Theorem 2, it is necessary that
n
max
` : k`k0 2
log |B` (?)| (1
) log 2
,
I(X0,` ; Y |X1,` )
(11)
where (X0,` , X1,` , Y ) are distributed as follows: (i) X0,` (respectively, X1,` ) is a concatenation
of the vectors X0,` (1), . . . , X0,` (d) (respectively, X1,` (1), . . . , X1,` (d)), the t-th of which contains
`t (respectively, ?t p `t ) entries independently drawn from Bernoulli(q); (ii) Letting each Nt
(t = 1, . . . , d) be the total number of ones in X0,` (t) and X1,` (t) combined, the random variable Y
is drawn from PY |N1 ,...,Nd according to (3).
As well as being simpler to evaluate, this corollary may be of interest in scenarios where one does
not have complete freedom in designing X, and one instead insists on using Bernoulli testing. For
instance, one may not know how to optimize X, and accordingly resort to generating it at random.
Example 1: Application to the noiseless setting. In the supplementary material, we show that in
the noiseless setting, Theorem 2 recovers a weakened version of Theorem 1 with 1 ? replaced by
1
o(1) in (8). Hence, while Theorem 2 does not establish a phase transition, it does recover
the exact threshold on the number of measurements required to obtain Pe ! 0.
An overview of the proof of this claim is as follows. We restrict the maximum in (10) to choices
of ` where each `t equals either its minimum value 0 or its maximum value p?t . Since we
are in the noiseless setting, each mutual information term reduces to the conditional entropy of
(i)
(i)
(i)
Y (i) = (Y1 , . . . , Yd ) given S`c and X (i) . For the values of t such that `t = 0, the value Yt is
(i)
deterministic (i.e., it has zero entropy), whereas for the values of t such that `t = p?t , the value Yt
follows a hypergeometric distribution, whose entropy behaves as 12 log p (1 + o(1)).
In the case that X is i.i.d. on Bernoulli(q), we can use Corollary 1 to obtain the following necessary
condition for Pe ? as as p ! 1, proved in the supplementary material:
?
?
p
2(H(?) H(? r ))
n
max
(1
o(1))
(12)
log(pq(1 q)) r2{1,...,d 1}
d r
for any q = q(p) such that both q and 1 q behave as ! p1 . Hence, while q = ?(1) recovers the
threshold in (8), the required number of tests strictly increases when q = o(1), albeit with a mild
logarithmic dependence.
Example 2: Group testing. To highlight the versatility of Theorem 2 and Corollary 1, we show
that the latter recovers the lower bounds given in the group testing framework of [11].
Set d = 2, and let label 1 represent ?defective? items, and label 2 represent ?non-defective? items.
Let PY |N1 N2 be of the form PY |N1 with Y 2 {0, 1}, meaning the observations are binary and
depend only on the number of defective items in the test. For brevity, let k = p?1 denote the total
number of defective items, so that p?2 = p k is the number of non-defective items.
Letting `2 = p
k in (11), and letting `1 remain arbitrary, we obtain the necessary condition
n
max
`1 2{1,...,k}
log
p k+`1
`1
(1
)
I(X0,`1 ; Y |X1,`1 )
5
log 2
,
(13)
where X0,`1 is a shorthand for X0,` with ` = (`1 , p k), and similarly for X1,`1 . This matches
the lower bound given in [11] for Bernoulli testing with general noise models, for which several
corollaries for specific models were also given.
Example 3: Gaussian noise. To give a concrete example of a noisy setting, consider the case that
we observe the values in (2), but with each such value corrupted by independent Gaussian noise:
(i)
Yt
(i)
= Nt ( , X (i) ) + Zt ,
(14)
(i)
Zt
where
? N(0, p 2 ) for some 2 > 0. Note that given X (i) , the values Nt themselves have
variance at most proportional to p (e.g., see Appendix C), so 2 = ?(1) can be thought of as the
constant signal-to-noise ratio (SNR) regime.
In the supplementary material, we prove the following bounds for this model:
? By letting each `t in (10) equal its minimum or maximum value analogously to the noiseless case
above, we obtain the following necessary condition for Pe ? as p ! 1:
?
?
pG H(?G )
P
n
max
(1
o(1)),
(15)
?t
1
G?[d] : |G| 2
t2G 2 log 1 + 4 2 )
P
where pG := t2G ?t p, and ?G has entries P 0 ?t ? 0 for t 2 G. Hence, we have the following:
t 2G
t
? In the case that 2 = p c for some c 2 (0, 1), each summand in the denominator simplifies
to 2c log p (1 + o(1)), and we deduce that compared to the noiseless case (cf., (8)), the
asymptotic number of tests increases by at least a constant factor of 1c .
? In the case that 2 = (log p) c for some c > 0, each summand in the denominator simplifies
to 2c log log p (1 + o(1)), and we deduce that compared to the noiseless case, the asymptotic
log p
number of tests increases by at least a factor of c log
log p . Hence, we observe a strict increase
in the scaling laws despite the fact that the SNR grows unbounded.
? While (15) also provides an ?(p) lower bound for the case 2 = ?(1), we can in fact do
better via a different choice of ` (see below).
? By letting `1 = p?1 , `2 = 1, and `t = 0 for t = 3, . . . , d, we obtain the necessary condition
n
4p 2 log p (1
o(1))
(16)
for Pe ? as p ! 1. Hence, if 2 = ?(1), we require n = ?(p log p); this is super-linear in
the dimension, in contrast with the sub-linear ? logp p behavior observed in the noiseless case.
Note that this choice of ` essentially captures the difficulty in identifying a single item, namely,
the one corresponding to `2 = 1.
These findings are summarized in Table 2; see also the supplementary material for extensions to the
approximate recovery setting.
Remark 2. While it may seem unusual to add continuous noise to discrete observations, this still
captures the essence of the noisy pooled data problem, and simplifies the evaluation of the mutual
information terms in (10). Moreover, this converse bound immediately implies the same bound for
the discrete model in which the noise consists of adding a Gaussian term, rounding, and clipping to
{0, . . . , p}, since the decoder could always choose to perform these operations as pre-processing.
3
Proofs
Here we provide the proof of Theorem 1, along with an overview of the proof of Theorem 2. The
remaining proofs are given in the supplementary material.
3.1
Proof of Theorem 1
Step 1: Counting typical outcomes. We claim that it suffices to consider the case that X is deterministic and ? is a deterministic function of Y; to see this, we note that when either of these are
random we have Pe = EX, ? [P [error]], and the average is lower bounded by the minimum.
(i)
The following lemma, proved in the supplementary material, shows
p that for any X , each entry of
(i)
the corresponding outcome Y lies in an interval of length O p log p with high probability.
6
Lemma 1. For any deterministic test vector X 2 {0, 1}p , and for uniformly distributed on B(?),
we have for each t 2 [d] that
h
i
p
2
P Nt ( , X) E[Nt ( , X)] > p log p ? 2 .
(17)
p
(i)
By Lemma 1 and the union bound, we have with probability at least 1 2nd
p2 that Nt ( , X )
p
(i)
E[Nt ( , X )] ? p log p for all i 2 [n] and t 2 [d]. Letting this event be denoted by A, we have
Pe
P[A \ no error]
P[A]
1
2nd
p2
P[A \ no error].
(18)
Next, letting Y( ) 2 [p]n?d denote Y explicitly as a function of and similarly for ?(Y) 2 [d]p ,
and letting YA denote the set of matrices Y under which the event A occurs, we have
X
1
P[A \ no error] =
1{Y(b) 2 YA \ ?(Y(b)) = b}
(19)
|B(?)|
b2B(?)
|YA |
?
,
|B(?)|
(20)
where (20) follows since each each Y 2 YA can only be counted once in the summation of (19),
due to the condition ?(Y(b)) = b.
Step 2: Bounding the set cardinalities. By a standard combinatorial argument (e.g., [14, Ch. 2])
and the fact that ? is fixed as p ! 1, we have
|B(?)| = ep(H(?)+o(1)) .
(21)
To bound |YA |, first note that the entries of each Y (i) 2 [p]d sum to a deterministic value, namely, the
number of ones in X (i) . Hence, each Y 2 YA is uniquely described by a sub-matrix of Y 2 [p]n?d
of size n ? (d 1). Moreover, since YA onlypincludes matrices under which A occurs, each value
in this sub-matrix only takes one of at most 2 p log p + 1 values. As a result, we have
p
n(d 1)
|YA | ? 2 p log p + 1
,
(22)
and combining (18)?(22) gives
p
n(d
2 p log p + 1
ep(H(?)+o(1))
Pe
1)
2nd
.
p2
(23)
pH(?)
p
Since d is constant, it immediately follows that Pe ! 1 whenever n ? (d 1) log(2
(1 ?)
p log p+1)
p
1
for some ? > 0. Applying log(2 p log p + 1) = 2 log p (1 + o(1)), we obtain the following
necessary condition for Pe 6! 1:
n
(d
2pH(?)
(1
1) log p
(24)
?).
This yields the term in (8) corresponding to r = 1.
Step 3: Genie argument. Let G be a subset of [d] of cardinality at least two, and define Gc = [d]\G.
Moreover, define Gc to be a length-p vector with
?
c
j
j 2G
( Gc ) j =
(25)
?
j 2 G,
where the symbol ? can be thought of as representing an unknown value. We consider a modified
setting in which a genie reveals Gc to the decoder, i.e., the decoder knows the labels of all items
for which the label lies in Gc , and is only left to estimate those in G. This additional knowledge
can only make the pooled data problem easier, and hence, any lower bound in this modified setting
remains valid in the original setting.
(i)
(i)
In the genie-aided setting, instead of receiving the full observation vector Y (i) = (Y1 , . . . , Yd ),
(i)
it is equivalent to only be given {Yj : j 2 G}, since the values in Gc are uniquely determined
7
from Gc and X (i) . This means that the genie-aided
Psetting can be cast in the original setting with
modified parameters: (i) p is replaced by pG =
t2G ?t p, the number of items with unknown
labels; (ii) d is replaced by |G|, the number of distinct remaining labels; (iii) ? is replaced by ?G ,
defined to be a |G|-dimensional probability vector with entries equaling P 0 ?t ? 0 (t 2 G).
t 2G
Due to this equivalence, the condition (24) yields the necessary condition n
and maximizing over all G with |G| 2 gives
2pG H(?G )
n
max
1 ? .
1) log p
G?[d] : |G| 2 (|G|
t
2pG H(?G )
(|G| 1) log p (1
?),
(26)
Step 4: Simplification. Define r = d |G| + 1. We restrict the maximum in (26) to sets G indexing
the highest |G| = d r + 1 values of ?, and consider the following process for sampling from ?:
? Draw a sample v from ? (r) (defined above Theorem 1);
? If v corresponds to the first entry of ? (r) , then draw a random sample from ?G and output it as
a label (i.e., the labels have conditional probability proportional to the top |G| entries of ?);
? Otherwise, if v corresponds to one of the other entries of ? (r) , then output v as a label.
By Shannon?s property ofP
entropy for sequentially-generated random variables
[15, p. 10], we find
P
that H(?) = H(? (r) ) +
?
H(?
).
Moreover,
since
p
=
p
?
?
,
this can be written
t
G
G
j
t2G
t2G
as pG H(?G ) = p H(?) H(? (r) ) . Substituting into (26), noting that |G| 1 = d r by the
definition of r, and maximizing over r = 1, . . . , d 1, we obtain the desired result (8).
3.2
Overview of proof of Theorem 2
We can interpret the pooled data problem as a communication problem in which a ?message?
(i)
(i)
is sent over a ?channel? PY |N1 ,...,Nd via ?codewords? of the form {(N1 , . . . , Nd )}ni=1 that are
constructed by summing various columns of X. As a result, it is natural to use Fano?s inequality [9,
Ch. 7] to lower bound the error probability in terms of information content (entropy) of and the
amount of information that Y reveals about (mutual information).
However, a naive application of Fano?s inequality only recovers the bound in (10) with ` = p?.
To handle the other possible choices of `, we again consider a genie-aided setting in which, for
each t 2 [d], the decoder is informed of p?t `t of the items whose label equals t. Hence, it only
remains to identify the remaining `t items of each type. This genie argument is a generalization of
that used in the proof of Theorem 1, in which each `t was either equal to its minimum value zero
or its maximum value p?t . In Example 3 of Section 2, we saw that this generalization can lead to a
strictly better lower bound in certain noisy scenarios.
The complete proof of Theorem 2 is given in the supplementary material.
4
Conclusion
We have provided novel information-theoretic lower bounds for the pooled data problem. In the
noiseless setting, we provided a matching lower bound to the upper bound of [3], establishing an
exact threshold indicating a phase transition between success and failure. In the noisy setting, we
provided a characterization of general noise models in terms of the mutual information. In the special
case of Gaussian noise, we proved an inherent added difficulty compared to the noiseless setting,
with strict increases in the scaling laws even when the signal-to-noise ratio grows unbounded.
An interesting direction for future research is to provide upper bounds for the noisy setting, potentially establishing the tightness of Theorem 2 for general noise models. This appears to be challenging using existing techniques; for instance, the pooled data problem bears similarity to group testing
with linear sparsity, whereas existing mutual information based upper bounds for group testing are
limited to the sub-linear regime [10, 11, 16]. In particular, the proofs of such bounds are based on
concentration inequalities which, when applied to the linear regime, lead to additional requirements
on the number of tests that prevent tight performance characterizations.
Acknowledgment: This work was supported in part by the European Commission under Grant
ERC Future Proof, SNF Sinergia project CRSII2-147633, SNF 200021-146750, and EPFL Fellows
Horizon2020 grant 665667.
8
References
[1] I.-H. Wang, S. L. Huang, K. Y. Lee, and K. C. Chen, ?Data extraction via histogram and
arithmetic mean queries: Fundamental limits and algorithms,? in IEEE Int. Symp. Inf. Theory,
July 2016, pp. 1386?1390.
[2] I.-H. Wang, S. L. Huang, and K. Y. Lee, ?Extracting sparse data via histogram queries,? in
Allerton Conf. Comm., Control, and Comp., 2016.
[3] A. E. Alaoui, A. Ramdas, F. Krzakala, L. Zdeborova, and M. I. Jordan, ?Decoding from pooled
data: Sharp information-theoretic bounds,? 2016, http://arxiv.org/abs/1611.09981.
[4] ??, ?Decoding from pooled data: Phase transitions of message passing,? 2017,
http://arxiv.org/abs/1702.02279.
[5] D.-Z. Du and F. K. Hwang, Combinatorial group testing and its applications, ser. Series on
Applied Mathematics. World Scientific, 1993.
[6] A. Seb?o, ?On two random search problems,? J. Stat. Plan. Inf., vol. 11, no. 1, pp. 23?31, 1985.
[7] M. Malyutov and H. Sadaka, ?Maximization of ESI. Jaynes principle for testing significant
inputs of linear model,? Rand. Opt. Stoch. Eq., vol. 6, no. 4, pp. 339?358, 1998.
[8] W.-N. Chen and I.-H. Wang, ?Partial data extraction via noisy histogram queries: Information
theoretic bounds,? in IEEE Int. Symp. Inf. Theory (ISIT), 2017.
[9] T. M. Cover and J. A. Thomas, Elements of Information Theory. John Wiley & Sons, Inc.,
2006.
[10] M. Malyutov, ?The separating property of random matrices,? Math. Notes Acad. Sci. USSR,
vol. 23, no. 1, pp. 84?91, 1978.
[11] G. Atia and V. Saligrama, ?Boolean compressed sensing and noisy group testing,? IEEE Trans.
Inf. Theory, vol. 58, no. 3, pp. 1880?1901, March 2012.
[12] C. Aksoylar, G. K. Atia, and V. Saligrama, ?Sparse signal processing with linear and nonlinear
observations: A unified Shannon-theoretic approach,? IEEE Trans. Inf. Theory, vol. 63, no. 2,
pp. 749?776, Feb. 2017.
[13] J. Scarlett and V. Cevher, ?Limits on support recovery with probabilistic models: An
information-theoretic framework,? IEEE Trans. Inf. Theory, vol. 63, no. 1, pp. 593?620, 2017.
[14] I. Csisz?r and J. K?rner, Information Theory: Coding Theorems for Discrete Memoryless Systems, 2nd ed. Cambridge University Press, 2011.
[15] C. E. Shannon, ?A mathematical theory of communication,? Bell Syst. Tech. Journal, vol. 27,
pp. 379?423, July and Oct. 1948.
[16] J. Scarlett and V. Cevher, ?Phase transitions in group testing,? in Proc. ACM-SIAM Symp. Disc.
Alg. (SODA), 2016.
[17] W. Hoeffding, ?Probability inequalities for sums of bounded random variables,? J. Amer. Stat.
Assoc., vol. 58, no. 301, pp. 13?30, 1963.
[18] J. Massey, ?On the entropy of integer-valued random variables,? in Int. Workshop on Inf. Theory, 1988.
[19] G. Reeves and M. Gastpar, ?The sampling rate-distortion tradeoff for sparsity pattern recovery
in compressed sensing,? IEEE Trans. Inf. Theory, vol. 58, no. 5, pp. 3065?3092, May 2012.
[20] ??, ?Approximate sparsity pattern recovery: Information-theoretic lower bounds,? IEEE
Trans. Inf. Theory, vol. 59, no. 6, pp. 3451?3465, June 2013.
[21] J. Scarlett and V. Cevher, ?How little does non-exact recovery help in group tesitng?? in IEEE
Int. Conf. Acoust. Sp. Sig. Proc. (ICASSP), New Orleans, 2017.
[22] ??, ?On the difficulty of selecting Ising models with approximate recovery,? IEEE Trans.
Sig. Inf. Proc. over Networks, vol. 2, no. 4, pp. 625?638, 2016.
[23] J. C. Duchi and M. J. Wainwright, ?Distance-based and continuum Fano inequalities with
applications to statistical estimation,? 2013, http://arxiv.org/abs/1311.2669.
9
| 6641 |@word mild:1 version:1 polynomial:1 proportion:4 norm:1 nd:10 open:2 seek:1 pg:6 initial:1 contains:2 series:1 selecting:1 denoting:1 existing:5 nt:9 jaynes:1 must:1 readily:2 written:1 john:1 item:19 accordingly:2 vanishing:1 volkan:2 characterization:3 provides:1 math:1 allerton:1 org:3 simpler:1 unbounded:3 mathematical:1 along:1 constructed:1 prove:2 consists:2 shorthand:1 psetting:1 symp:3 introduce:2 krzakala:1 privacy:1 x0:9 behavior:2 p1:1 themselves:1 little:1 cardinality:2 provided:4 project:1 notation:4 moreover:5 bounded:2 mass:1 informed:1 unified:1 finding:1 acoust:1 horizon2020:1 fellow:1 zdeborova:1 exactly:2 assoc:1 ser:1 control:1 unit:1 medical:1 converse:3 grant:2 before:1 positive:2 tends:2 limit:3 acad:1 despite:1 establishing:2 yd:3 therein:2 studied:1 weakened:1 equivalence:1 challenging:1 lausanne:1 limited:1 practical:2 acknowledgment:1 testing:19 yj:1 union:1 orleans:1 snf:2 empirical:5 t2g:5 bell:1 significantly:3 revealing:1 thought:3 matching:1 pre:1 close:1 seb:1 storage:1 impossible:1 applying:1 py:7 optimize:1 equivalent:1 deterministic:5 yt:4 maximizing:2 regardless:2 independently:1 identifying:3 recovery:12 immediately:2 population:3 handle:1 variation:2 target:1 suppose:1 exact:8 designing:1 sig:2 element:1 ising:1 observed:1 ep:2 wang:3 capture:2 equaling:2 highest:1 mentioned:1 vanishes:2 comm:1 nats:1 ideally:1 esi:1 depend:4 solving:1 tight:2 completely:2 icassp:1 k0:3 represented:1 various:1 distinct:1 query:3 outcome:4 whose:10 larger:4 posed:1 supplementary:9 valued:1 tightness:1 otherwise:3 compressed:2 distortion:1 noisy:18 itself:1 sequence:7 reconstruction:1 saligrama:2 j2:1 relevant:1 combining:1 achieve:4 csisz:1 requirement:1 generating:1 leave:1 help:1 depending:3 develop:1 stat:2 measured:1 nearest:1 eq:1 p2:3 implies:1 direction:1 material:9 require:2 suffices:1 generalization:2 opt:1 isit:1 summation:1 strictly:4 extension:1 hold:1 claim:2 substituting:1 continuum:1 omitted:1 estimation:1 proc:3 label:26 combinatorial:2 cole:1 saw:1 largest:2 clearly:1 gaussian:5 always:1 super:2 modified:3 pn:1 corollary:6 focus:4 stoch:1 june:1 bernoulli:8 indicates:1 tech:1 contrast:1 adversarial:3 attains:1 summarizing:1 inference:1 epfl:3 entire:1 typically:1 initially:1 overall:1 among:1 denoted:2 ussr:1 plan:1 raised:1 special:1 mutual:8 equal:5 construct:1 once:2 having:2 beach:1 sampling:2 extraction:2 represents:1 future:2 simplex:1 fundamentally:1 simplify:1 summand:2 inherent:1 randomly:1 individual:1 replaced:4 phase:11 consisting:1 versatility:1 n1:8 ab:3 freedom:1 interest:3 message:3 huge:1 highly:2 evaluation:1 partial:1 necessary:16 unless:1 logarithm:1 desired:1 cevher:5 instance:2 column:1 boolean:1 cover:1 logp:3 maximization:2 clipping:1 subset:2 entry:24 snr:8 uniform:6 rounding:2 characterize:1 commission:1 corrupted:1 considerably:1 combined:2 st:1 density:1 fundamental:1 siam:1 lee:2 probabilistic:1 receiving:1 decoding:4 pool:3 rounded:1 analogously:1 concrete:1 again:1 containing:1 choose:1 possibly:1 huang:2 hoeffding:1 conf:2 resort:1 syst:1 de:1 pooled:16 summarized:1 coding:1 int:4 inc:1 explicitly:1 multiplicative:1 recover:2 contribution:4 ni:1 variance:2 yield:2 identify:2 disc:1 comp:1 whenever:2 ed:1 definition:3 failure:3 pp:12 associated:3 proof:13 recovers:4 hamming:2 proved:4 recall:1 knowledge:1 genie:7 amplitude:1 appears:1 higher:2 insists:1 rand:1 amer:1 implicit:1 nonlinear:1 continuity:1 reveal:2 hwang:1 grows:3 scientific:1 usa:1 hence:9 memoryless:1 laboratory:1 b2b:1 conditionally:1 uniquely:2 essence:1 criterion:1 complete:4 polytechnique:1 theoretic:11 demonstrate:1 duchi:1 meaning:2 consideration:1 novel:3 recently:1 tending:1 behaves:1 overview:3 extend:1 interpret:1 refer:1 measurement:4 significant:2 cambridge:1 reef:1 mathematics:1 similarly:3 fano:3 closing:2 erc:1 pq:1 similarity:1 deduce:2 base:1 add:1 feb:1 recent:1 inf:10 scenario:3 certain:3 inequality:5 binary:1 success:3 arbitrarily:3 minimum:4 additional:3 preceding:1 signal:6 ii:2 arithmetic:1 full:1 july:2 reduces:1 exceeds:1 match:4 long:1 ofp:1 variant:1 basic:1 denominator:2 noiseless:19 essentially:1 arxiv:3 histogram:3 represent:2 achieved:3 addition:2 whereas:4 interval:1 strict:4 sent:1 alaoui:1 seem:1 jordan:1 integer:2 extracting:1 presence:1 counting:1 revealed:1 iii:1 noting:1 xj:2 identified:1 restrict:2 simplifies:3 knowing:2 tradeoff:1 whether:1 passing:2 remark:2 generally:1 amount:3 ph:2 http:3 estimated:1 scarlett:5 write:1 discrete:3 vol:11 group:10 key:1 threshold:8 drawn:3 clarity:1 prevent:1 massey:1 sum:4 qmax:2 soda:1 extends:1 throughout:2 draw:2 appendix:1 scaling:5 bound:42 simplification:1 occur:1 constraint:1 precisely:1 aspect:1 argument:3 performing:1 according:2 march:1 remain:1 son:1 indexing:1 previously:1 remains:2 discus:1 count:1 turn:2 needed:1 know:2 letting:8 unusual:1 operation:1 observe:2 original:2 thomas:1 denotes:2 remaining:7 assumes:2 cf:1 top:1 prof:1 establish:1 added:1 occurs:3 codewords:1 concentration:2 dependence:1 exhibit:1 distance:3 separating:1 concatenation:1 decoder:9 sci:1 length:4 index:2 providing:1 ratio:5 difficult:2 setup:3 unfortunately:1 potentially:1 stated:1 zt:2 unknown:5 perform:1 upper:7 observation:9 behave:1 communication:2 y1:3 gc:7 arbitrary:1 sharp:1 introduced:2 namely:2 required:10 specified:1 cast:1 connection:1 hypergeometric:1 nip:1 trans:6 lion:1 below:4 pattern:2 rale:1 regime:3 sparsity:3 summarize:1 max:11 wainwright:1 event:2 difficulty:3 natural:1 representing:2 improve:1 naive:1 rner:1 prior:1 literature:1 understanding:1 asymptotic:5 law:5 highlight:1 bear:1 interesting:1 proportional:2 sufficient:3 consistent:1 principle:1 row:2 genetics:1 summary:2 achievability:1 course:1 supported:1 fall:1 characterizing:1 sparse:3 distributed:5 dimension:3 transition:11 valid:1 world:1 crsii2:1 author:1 collection:1 adaptive:1 coincide:5 counted:1 approximate:7 sequentially:1 reveals:7 summing:1 assumed:1 continuous:2 search:1 table:6 channel:1 ca:1 alg:1 du:1 european:1 sp:1 main:3 bounding:1 noise:26 n2:1 ramdas:1 allowed:2 defective:5 x1:8 wiley:1 sub:6 decoded:1 explicit:1 lie:2 pe:19 theorem:19 specific:3 symbol:1 r2:6 sensing:2 ments:1 exists:2 workshop:1 albeit:1 adding:1 gap:8 easier:1 chen:2 entropy:8 logarithmic:1 explore:1 forming:1 collectively:1 ch:3 corresponds:2 acm:1 oct:1 conditional:5 goal:1 content:1 aided:3 included:2 specifically:3 typical:1 uniformly:5 determined:1 lemma:3 total:4 ya:8 shannon:3 indicating:1 formally:1 support:1 latter:2 jonathan:2 brevity:1 evaluate:1 ex:1 |
6,237 | 6,642 | Universal Style Transfer via Feature Transforms
Yijun Li
UC Merced
[email protected]
Zhaowen Wang
Adobe Research
[email protected]
Chen Fang
Adobe Research
[email protected]
Xin Lu
Adobe Research
[email protected]
Jimei Yang
Adobe Research
[email protected]
Ming-Hsuan Yang
UC Merced, NVIDIA Research
[email protected]
Abstract
Universal style transfer aims to transfer arbitrary visual styles to content images.
Existing feed-forward based methods, while enjoying the inference efficiency, are
mainly limited by inability of generalizing to unseen styles or compromised visual
quality. In this paper, we present a simple yet effective method that tackles these
limitations without training on any pre-defined styles. The key ingredient of our
method is a pair of feature transforms, whitening and coloring, that are embedded
to an image reconstruction network. The whitening and coloring transforms reflect
a direct matching of feature covariance of the content image to a given style
image, which shares similar spirits with the optimization of Gram matrix based
cost in neural style transfer. We demonstrate the effectiveness of our algorithm by
generating high-quality stylized images with comparisons to a number of recent
methods. We also analyze our method by visualizing the whitened features and
synthesizing textures via simple feature coloring.
1
Introduction
Style transfer is an important image editing task which enables the creation of new artistic works.
Given a pair of examples, i.e., the content and style image, it aims to synthesize an image that
preserves some notion of the content but carries characteristics of the style. The key challenge is how
to extract effective representations of the style and then match it in the content image. The seminal
work by Gatys et al. [8, 9] show that the correlation between features, i.e., Gram matrix or covariance
matrix (shown to be as effective as Gram matrix in [20]), extracted by a trained deep neural network
has remarkable ability of capturing visual styles. Since then, significant efforts have been made to
synthesize stylized images by minimizing Gram/covariance matrices based loss functions, through
either iterative optimization [9] or trained feed-forward networks [27, 16, 20, 2, 6]. Despite the recent
rapid progress, these existing works often trade off between generalization, quality and efficiency,
which means that optimization-based methods can handle arbitrary styles with pleasing visual quality
but at the expense of high computational costs, while feed-forward approaches can be executed
efficiently but are limited to a fixed number of styles or compromised visual quality.
By far, the problem of universal style transfer remains a daunting task as it is challenging to develop
neural networks that achieve generalization, quality and efficiency at the same time. The main issue
is how to properly and effectively apply the extracted style characteristics (feature correlations) to
content images in a style-agnostic manner.
In this work, we propose a simple yet effective method for universal style transfer, which enjoys
the style-agnostic generalization ability with marginally compromised visual quality and execution
efficiency. The transfer task is formulated as image reconstruction processes, with the content features
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
C
S
C
VGG
Relu_5_1
S
VGG
Relu_X_1
VGG
Relu_X_1
Whitening & Coloring
Transform (WCT)
Recons
DecoderX
VGG
Relu_4_1
S
VGG
Relu_3_1
S
Recons
DecoderX
VGG
Relu_2_1
S
Output
S
(a) Reconstruction (b) Single-level stylization
VGG
Relu_1_1
Recons
Decoder5
WCT
WCT
WCT
WCT
WCT
I5
Recons
Decoder4
I4
Recons
Decoder3
Recons
Decoder2
Recons
Decoder1
I3
I2
Output I1
(c) Multi-level stylization
Figure 1: Universal style transfer pipeline. (a) We first pre-train five decoder networks DecoderX
(X=1,2,...,5) through image reconstruction to invert different levels of VGG features. (b) With both
VGG and DecoderX fixed, and given the content image C and style image S, our method performs
the style transfer through whitening and coloring transforms. (c) We extend single-level to multi-level
stylization in order to match the statistics of the style at all levels. The result obtained by matching
higher level statistics of the style is treated as the new content to continue to match lower-level
information of the style.
being transformed at intermediate layers with regard to the statistics of the style features, in the
midst of feed-forward passes. In each intermediate layer, our main goal is to transform the extracted
content features such that they exhibit the same statistical characteristics as the style features of the
same layer and we found that the classic signal whitening and coloring transforms (WCTs) on those
features are able to achieve this goal in an almost effortless manner.
In this work, we first employ the VGG-19 network [26] as the feature extractor (encoder), and train a
symmetric decoder to invert the VGG-19 features to the original image, which is essentially the image
reconstruction task (Figure 1(a)). Once trained, both the encoder and the decoder are fixed through all
the experiments. To perform style transfer, we apply WCT to one layer of content features such that
its covariance matrix matches that of style features, as shown in Figure 1(b). The transformed features
are then fed forward into the downstream decoder layers to obtain the stylized image. In addition
to this single-level stylization, we further develop a multi-level stylization pipeline, as depicted in
Figure 1(c), where we apply WCT sequentially to multiple feature layers. The multi-level algorithm
generates stylized images with greater visual quality, which are comparable or even better with much
less computational costs. We also introduce a control parameter that defines the degree of style
transfer so that the users can choose the balance between stylization and content preservation. The
entire procedure of our algorithm only requires learning the image reconstruction decoder with no
style images involved. So when given a new style, we simply need to extract its feature covariance
matrices and apply them to the content features via WCT. Note that this learning-free scheme is
fundamentally different from existing feed-forward networks that require learning with pre-defined
styles and fine-tuning for new styles. Therefore, our approach is able to achieve style transfer
universally.
The main contributions of this work are summarized as follows:
? We propose to use feature transforms, i.e., whitening and coloring, to directly match content
feature statistics to those of a style image in the deep feature space.
? We couple the feature transforms with a pre-trained general encoder-decoder network, such
that the transferring process can be implemented by simple feed-forward operations.
? We demonstrate the effectiveness of our method for universal style transfer with high-quality
visual results, and also show its application to universal texture synthesis.
2
Related Work
Existing style transfer methods are mostly example-based [13, 25, 24, 7, 21]. The image analogy
method [13] aims to determine the relationship between a pair of images and then apply it to stylize
2
other images. As it is based on finding dense correspondence, analogy-based approaches [25, 24, 7,
21] often require that a pair of image depicts the same type of scene. Therefore these methods do not
scale to the setting of arbitrary style images well.
Recently, Gatys et al. [8, 9] proposed an algorithm for arbitrary stylization based on matching the
correlations (Gram matrix) between deep features extracted by a trained network classifier within an
iterative optimization framework. Numerous methods have since been developed to address different
aspects including speed [27, 19, 16], quality [28, 18, 32, 31], user control [10], diversity [29, 20],
semantics understanding [7, 1] and photorealism [23]. It is worth mentioning that one of the
major drawbacks of [8, 9] is the inefficiency due to the optimization process. The improvement of
efficiency in [27, 19, 16] is realized by formulating the stylization as learning a feed-forward image
transformation network. However, these methods are limited by the requirement of training one
network per style due to the lack of generalization in network design.
Most recently, a number of methods have been proposed to empower a single network to transfer
multiple styles, including a model that conditioned on binary selection units [20], a network that
learns a set of new filters for every new style [2], and a novel conditional normalization layer that
learns normalization parameters for each style [6]. To achieve arbitrary style transfer, Chen et al. [3]
first propose to swap the content feature with the closest style feature locally. Meanwhile, inspired
by [6], two following work [30, 11] turn to learn a general mapping from the style image to style
parameters. One closest related work [15] directly adjusts the content feature to match the mean and
variance of the style feature. However, the generalization ability of the learned models on unseen
styles is still limited.
Different from the existing methods, our approach performs style transfer efficiently in a feed-forward
manner while achieving generalization and visual quality on arbitrary styles. Our approach is closely
related to [15], where content feature in a particular (higher) layer is adaptively instance normalized
by the mean and variance of style feature. This step can be viewed as a sub-optimal approximation
of the WCT operation, thereby leading to less effective results on both training and unseen styles.
Moreover, our encoder-decoder network is trained solely based on image reconstruction, while [15]
requires learning such a module particularly for stylization task. We evaluate the proposed algorithm
with existing approaches extensively on both style transfer and texture synthesis tasks and present
in-depth analysis.
3
Proposed Algorithm
We formulate style transfer as an image reconstruction process coupled with feature transformation,
i.e., whitening and coloring. The reconstruction part is responsible for inverting features back to the
RGB space and the feature transformation matches the statistics of a content image to a style image.
3.1
Reconstruction decoder
We construct an auto-encoder network for general image reconstruction. We employ the VGG-19 [26]
as the encoder, fix it and train a decoder network simply for inverting VGG features to the original
image, as shown in Figure 1(a). The decoder is designed as being symmetrical to that of VGG-19
network (up to Relu_X_1 layer), with the nearest neighbor upsampling layer used for enlarging
feature maps. To evaluate with features extracted at different layers, we select feature maps at five
layers of the VGG-19, i.e., Relu_X_1 (X=1,2,3,4,5), and train five decoders accordingly. The pixel
reconstruction loss [5] and feature loss [16, 5] are employed for reconstructing an input image,
L = kIo ? Ii k22 + ?k?(Io ) ? ?(Ii )k22 ,
(1)
where Ii , Io are the input image and reconstruction output, and ? is the VGG encoder that extracts
the Relu_X_1 features. In addition, ? is the weight to balance the two losses. After training, the
decoder is fixed (i.e., will not be fine-tuned) and used as a feature inverter.
3.2
Whitening and coloring transforms
Given a pair of content image Ic and style image Is , we first extract their vectorized VGG feature
maps fc ? <C?Hc Wc and fs ? <C?Hs Ws at a certain layer (e.g., Relu_4_1), where Hc , Wc (Hs ,
3
Figure 2: Inverting whitened features. We invert the whitened VGG Relu_4_1 feature as an example.
Left: original images, Right: inverted results (pixel intensities are rescaled for better visualization).
The whitened features still maintain global content structures.
(a) Style (b) Content
(c) HM
(d) WCT
(e) Style
(f) Content
(g) HM
(h) WCT
Figure 3: Comparisons between different feature transform strategies. Results are obtained by our
multi-level stylization framework in order to match all levels of information of the style.
Ws ) are the height and width of the content (style) feature, and C is the number of channels. The
decoder will reconstruct the original image Ic if fc is directly fed into it. We next propose to use a
whitening and coloring transform to adjust fc with respect to the statistics of fs . The goal of WCT
is to directly transform the fc to match the covariance matrix of fs . It consists of two steps, i.e.,
whitening and coloring transform.
Whitening transform. Before whitening, we first center fc by subtracting its mean vector mc .
Then we transform fc linearly as in (2) so that we obtain f?c such that the feature maps are uncorrelated
>
(f?c f?c = I),
1
?
f?c = Ec Dc 2 Ec> fc ,
(2)
where Dc is a diagonal matrix with the eigenvalues of the covariance matrix fc fc> ? <C?C , and Ec
is the corresponding orthogonal matrix of eigenvectors, satisfying fc fc> = Ec Dc Ec> .
To validate what is encoded in the whitened feature f?c , we invert it to the RGB space with our
previous decoder trained for reconstruction only. Figure 2 shows two visualization examples, which
indicate that the whitened features still maintain global structures of the image contents, but greatly
help remove other information related to styles. We note especially that, for the Starry_night example
on right, the detailed stroke patterns across the original image are gone. In other words, the whitening
step helps peel off the style from an input image while preserving the global content structure. The
outcome of this operation is ready to be transformed with the target style.
Coloring transform. We first center fs by subtracting its mean vector ms , and then carry out
the coloring transform [14], which is essentially the inverse of the whitening step to transform f?c
linearly as in (3) such that we obtain f?cs which has the desired correlations between its feature maps
>
(f?cs f?cs = fs f > ),
s
1
f?cs = Es Ds2 Es> f?c ,
(3)
fs fs>
C?C
where Ds is a diagonal matrix with the eigenvalues of the covariance matrix
?<
, and Es
is the corresponding orthogonal matrix of eigenvectors. Finally we re-center the f?cs with the mean
vector ms of the style, i.e., f?cs = f?cs + ms .
To demonstrate the effectiveness of WCT, we compare it with a commonly used feature adjustment
technique, i.e., histogram matching (HM), in Figure 3. The channel-wise histogram matching [12]
method determines a mapping function such that the mapped fc has the same cumulative histogram as
fs . In Figure 3, it is clear that the HM method helps transfer the global color of the style image well
4
(a) Style
(b) Relu_1_1
(c) Relu_2_1
(d) Relu_3_1
(e) Relu_4_1
(f) Relu_5_1
Figure 4: Single-level stylization using different VGG features. The content image is from Figure 2.
(a) I5
(b) I4
(c) I1
(d) Fine-to-coarse
Figure 5: (a)-(c) Intermediate results of our coarse-to-fine multi-level stylization framework in
Figure 1(c). The style and content images are from Figure 4. I1 is the final output of our multi-level
pipeline. (d) Reversed fine-to-coarse multi-level pipeline.
but fails to capture salient visual patterns, e.g., patterns are broken into pieces and local structures are
misrepresented. In contrast, our WCT captures patterns that reflect the style image better. This can
be explained by that the HM method does not consider the correlations between features channels,
which are exactly what the covariance matrix is designed for.
After the WCT, we may blend f?cs with the content feature fc as in (4) before feeding it to the decoder
in order to provide user controls on the strength of stylization effects:
f?cs = ? f?cs + (1 ? ?) fc ,
(4)
where ? serves as the style weight for users to control the transfer effect.
3.3
Multi-level coarse-to-fine stylization
Based on the single-level stylization framework shown in Figure 1(b), we use different layers of VGG
features Relu_X_1 (X=1,2,...,5) and show the corresponding stylized results in Figure 4. It clearly
shows that the higher layer features capture more complicated local structures, while lower layer
features carry more low-level information (e.g., colors). This can be explained by the increasing size
of receptive field and feature complexity in the network hierarchy. Therefore, it is advantageous to
use features at all five layers to fully capture the characteristics of a style from low to high levels.
Figure 1(c) shows our multi-level stylization pipeline. We start by applying the WCT on Relu_5_1
features to obtain a coarse stylized result and regard it as the new content image to further adjust
features in lower layers. An example of intermediate results are shown in Figure 5. We show the
intermediate results I5 , I4 , I1 with obvious differences, which indicates that the higher layer features
first capture salient patterns of the style and lower layer features further improve details. If we reverse
feature processing order (i.e., fine-to-coarse layers) by starting with Relu_1_1, low-level information
cannot be preserved after manipulating higher level features, as shown in Figure 5(d).
4
4.1
Experimental Results
Decoder training
For the multi-level stylization approach, we separately train five reconstruction decoders for features
at the VGG-19 Relu_X_1 (X=1,2,...,5) layer. It is trained on the Microsoft COCO dataset [22] and
the weight ? to balance the two losses in (1) is set as 1.
5
(a) Style
(b) [3]
(c) [15]
(d) [27]
(e) [9]
(f) Ours
Figure 6: Results from different style transfer methods. The content images are from Figure 2-3. We
evaluate various styles including paintings, abstract styles, and styles with obvious texton elements.
We adjust the style weight of each method to obtain the best stylized effect. For our results, we set
the style weight ? = 0.6.
Table 1: Differences between our approach and other methods.
Arbitrary
Efficient
Learning-free
4.2
Chen et al. [3]
?
?
Huang et al. [15]
?
?
TNet [27]
?
?
?
?
?
DeepArt [9]
?
?
?
Ours
?
?
?
Style transfer
To demonstrate the effectiveness of the proposed algorithm, we list the differences with existing
methods in Table 1 and present stylized results in Figure 6. We adjust the style weight of other
methods to obtain the best stylized effect. The optimization-based work of [9] handles arbitrary
styles but is likely to encounter unexpected local minima issues (e.g., 5th and 6th row of Figure 6(e)).
Although the method [27] greatly improves the stylization speed, it trades off quality and generality
for efficiency, which generates repetitive patterns that overlay with the image contents (Figure 6(d)).
6
Table 2: Quantitative comparisons between different stylization methods in terms of the covariance
matrix difference (Ls ), user preference and run-time, tested on images of size 256 ? 256 and a 12GB
TITAN X.
log(Ls )
Preference/%
Time/sec
Style
Chen et al. [3]
Huang et al. [15]
TNet [27]
Gatys et al. [9]
Ours
7.4
15.7
2.1
7.0
24.9
0.20
6.8
12.7
0.18
6.7
16.4
21.2
6.3
30.3
0.83
scale = 256
scale = 768
? = 0.4
? = 0.6
? = 1.0
Figure 7: Controlling the stylization on the scale and weight.
Closest to our work on generalization are the recent methods [3, 15], but the quality of the stylized
results are less appealing. The work of [3] replaces the content feature with the most similar style
feature based on patch similarity and hence has limited capability, i.e., the content is strictly preserved
while style is not well reflected with only low-level information (e.g., colors) transferred, as shown
in Figure 6(b). In [15], the content feature is simply adjusted to have the same mean and variance
with the style feature, which is not effective in capturing high-level representations of the style. Even
learned with a set of training styles, it does not generalize well on unseen styles. Results in Figure 6(c)
indicate that the method in [15] is not effective at capturing and synthesizing salient style patterns,
especially for complicated styles where there are rich local structures and non-smooth regions.
Figure 6(f) shows the stylized results of our approach. Without learning any style, our method is able
to capture visually salient patterns in style images (e.g., the brick wall on the 6th row). Moreover,
key components in the content images (e.g., bridge, eye, mouth) are also well stylized in our results,
while other methods only transfer patterns to relatively smooth regions (e.g., sky, face). The models
and code are available at https://github.com/Yijunmaverick/UniversalStyleTransfer.
In addition, we quantitatively evaluate different methods by computing the covariance matrix difference (Ls ) on all five levels of VGG features between stylized results and the given style image. We
randomly select 10 content images from [22] and 40 style images from [17], compute the averaged
difference over all styles, and show the results in Table 2 (1st row). Quantitative results show that we
generate stylized results with lower Ls , i.e., closer to the statistics of the style.
User study. Evaluating artistic style transfer has been an open question in the community. Since
the qualitative assessment is highly subjective, we conduct a user study to evaluate 5 methods shown
in Figure 6. We use 5 content images and 30 style images, and generate 150 results based on each
content/style pair for each method. We randomly select 15 style images for each subject to evaluate.
We display stylized images by 5 compared methods side-by-side on a webpage in random order. Each
subject is asked to vote his/her ONE favorite result for each style. We finally collect the feedback
from 80 subjects of totally 1,200 votes and show the percentage of the votes each method received in
Table 2 (2nd row). The study shows that our method receives the most votes for better stylized results.
It can be an interesting direction to develop evaluation metrics based on human visual perception for
general image synthesis problems.
Efficiency. In Table 2 (3rd row), we also compare our approach with other methods in terms of
efficiency. The method by Gatys et al. [9] is slow due to loops of optimization and usually requires at
least 500 iterations to generate good results. The methods [27] and [15] are efficient as the scheme is
based on one feed-forward pass with a trained network. The approach [3] is feed-forward based but
relatively slower as the feature swapping operation needs to be carried out for thousands of patches.
Our approach is also efficient but a little bit slower than [27, 15] because we have a eigenvalue
decomposition step in WCT. But note that the computational cost on this step will not increase along
with the image size because the the dimension of covariance matrix only depends on filter numbers (or
7
(a) Content
(b) Different masks and styles
(c) Our result
Figure 8: Spatial control in transferring, which enables users to edit the content with different styles.
Figure 9: Texture synthesis. In each panel, Left: original textures, Right: our synthesized results.
Texture images are mostly from the Describable Textures Dataset (DTD) [4].
channels), which is at most 512 (Relu_5_1). Currently the decomposition step is implemented based
on CPU. Our future work includes more efficient GPU implementations of the proposed algorithm.
User Controls. Given a content/style pair, our approach is not only as simple as a one-click
transferring, but also flexible enough to accommodate different requirements from users by providing
different controls on the stylization, including the scale, weight and spatial control. The style input
on different scales will lead to different extracted statistics due to the fixed receptive field of the
network. Therefore the scale control is easily achieved by adjusting the style image size. In the
middle of Figure 7, we show two examples where the brick can be transferred in either small or
large scale. The weight control refers to controlling the balance between stylization and content
preservation. As shown on right of Figure 7, our method enjoys this flexibility in simple feed-forward
passes by simply adjusting the style weight ? in (4). However in [9] and [27], to obtain visual results
of different weight settings, a new round of time-consuming optimization or model training is needed.
Moreover, our blending directly works on deep feature space before inversion/reconstruction, which
is fundamentally different from [9, 27] where the blending is formulated as the weighted sum of the
content and style losses that may not always lead to a good balance point.
The spatial control is also highly desired when users want to edit an image with different styles
transferred on different parts of the image. Figure 8 shows an example of spatially controlling the
stylization. A set of masks M (Figure 8(b)) is additionally required as input to indicate the spatial
correspondence between content regions and styles. By replacing the content feature fc in (3) with
M fc where is a simple mask-out operation, we are able to stylize the specified region only.
4.3
Texture synthesis
By setting the content image as a random noise image (e.g., Gaussian noise), our stylization framework
can be easily applied to texture synthesis. An alternative is to directly initialize the f?c in (3) to be white
noise. Both approaches achieve similar results. Figure 9 shows a few examples of the synthesized
textures. We empirically find that it is better to run the multi-level pipeline for a few times (e.g., 3) to
get more visually pleasing results.
Our method is also able to synthesize the interpolated result of two textures. Given two texture
?
examples s1 and s2 , we first perform the WCT on the input noise and get transformed features fcs
1
?
?
?
?
and fcs2 respectively. Then we blend these two features fcs = ? fcs1 + (1 ? ?)fcs2 and feed the
8
Texture s1
Texture s2
? = 0.75
? = 0.5
? = 0.25
? = 0.5
Figure 10: Interpolation between two texture examples. Left: original textures, Middle: our interpolation results, Right: interpolated results of [9]. ? controls the weight of interpolation.
Texture
TNet [27]
Ours
Figure 11: Comparisons of diverse synthesized results between TNet [27] and our model.
combined feature into the decoder to generate mixed effects. Note that our interpolation directly
works on deep feature space. By contrast, the method in [9] generates the interpolation by matching
the weighted sum of Gram matrices of two textures at the loss end. Figure 10 shows that the result
by [9] is simply overlaid by two textures while our method generates new textural effects, e.g., bricks
in the stripe shape.
One important aspect in texture synthesis is diversity. By sampling different noise images, our
method can generate diverse synthesized results for each texture. While [27] can generate different
results driven by the input noise, the learned networks are very likely to be trapped in local optima. In
other words, the noise is marginalized out and thus fails to drive the network to generate large visual
variations. In contrast, our approach explains each input noise better because the network is unlikely
to absorb the variations in input noise since it is never trained for learning textures. We compare the
diverse outputs of our model with [27] in Figure 11. Note that the common diagonal layout is shared
across different results of [27], which causes unsatisfying visual experiences. The comparison shows
that our method achieves diversity in a more natural and flexible manner.
5
Concluding Remarks
In this work, we propose a universal style transfer algorithm that does not require learning for each
individual style. By unfolding the image generation process via training an auto-encoder for image
reconstruction, we integrate the whitening and coloring transforms in the feed-forward passes to
match the statistical distributions and correlations between the intermediate features of content and
style. We also present a multi-level stylization pipeline, which takes all level of information of a
style into account, for improved results. In addition, the proposed approach is shown to be equally
effective for texture synthesis. Experimental results demonstrate that the proposed algorithm achieves
favorable performance against the state-of-the-art methods in generalizing to arbitrary styles.
Acknowledgments
This work is supported in part by the NSF CAREER Grant #1149783, gifts from Adobe and NVIDIA.
9
References
[1] A. J. Champandard. Semantic style transfer and turning two-bit doodles into fine artworks. arXiv preprint
arXiv:1603.01768, 2016.
[2] D. Chen, L. Yuan, J. Liao, N. Yu, and G. Hua. Stylebank: An explicit representation for neural image style
transfer. In CVPR, 2017.
[3] T. Q. Chen and M. Schmidt.
arXiv:1612.04337, 2016.
Fast patch-based style transfer of arbitrary style.
arXiv preprint
[4] M. Cimpoi, S. Maji, I. Kokkinos, S. Mohamed, , and A. Vedaldi. Describing textures in the wild. In CVPR,
2014.
[5] A. Dosovitskiy and T. Brox. Generating images with perceptual similarity metrics based on deep networks.
In NIPS, 2016.
[6] V. Dumoulin, J. Shlens, and M. Kudlur. A learned representation for artistic style. In ICLR, 2017.
[7] O. Frigo, N. Sabater, J. Delon, and P. Hellier. Split and match: Example-based adaptive patch sampling for
unsupervised style transfer. In CVPR, 2016.
[8] L. A. Gatys, A. S. Ecker, and M. Bethge. Texture synthesis using convolutional neural networks. In NIPS,
2015.
[9] L. A. Gatys, A. S. Ecker, and M. Bethge. Image style transfer using convolutional neural networks. In
CVPR, 2016.
[10] L. A. Gatys, A. S. Ecker, M. Bethge, A. Hertzmann, and E. Shechtman. Controlling perceptual factors in
neural style transfer. In CVPR, 2017.
[11] G. Ghiasi, H. Lee, M. Kudlur, V. Dumoulin, and J. Shlens. Exploring the structure of a real-time, arbitrary
neural artistic stylization network. In BMVC, 2017.
[12] R. C. Gonzalez and R. E. Woods. Digital image processing (3rd edition). Prentice Hall, 2008.
[13] A. Hertzmann, C. E. Jacobs, N. Oliver, B. Curless, and D. H. Salesin. Image analogies. In SIGGRAPH,
2001.
[14] M. Hossain. Whitening and coloring transforms for multivariate gaussian random variables. Project Rhea,
2016.
[15] X. Huang and S. Belongie. Arbitrary style transfer in real-time with adaptive instance normalization. In
ICCV, 2017.
[16] J. Johnson, A. Alahi, and L. Fei-Fei. Perceptual losses for real-time style transfer and super-resolution. In
ECCV, 2016.
[17] S. Karayev, M. Trentacoste, H. Han, A. Agarwala, T. Darrell, A. Hertzmann, and H. Winnemoeller.
Recognizing image style. In BMVC, 2014.
[18] C. Li and M. Wand. Combining markov random fields and convolutional neural networks for image
synthesis. In CVPR, 2016.
[19] C. Li and M. Wand. Precomputed real-time texture synthesis with markovian generative adversarial
networks. In ECCV, 2016.
[20] Y. Li, C. Fang, J. Yang, Z. Wang, X. Lu, and M.-H. Yang. Diversified texture synthesis with feed-forward
networks. In CVPR, 2017.
[21] J. Liao, Y. Yao, L. Yuan, G. Hua, and S. B. Kang. Visual attribute transfer through deep image analogy.
arXiv preprint arXiv:1705.01088, 2017.
[22] T.-Y. Lin, M. Maire, S. Belongie, J. Hays, P. Perona, D. Ramanan, P. Doll?r, and C. L. Zitnick. Microsoft
COCO: Common objects in context. In ECCV, 2014.
[23] F. Luan, S. Paris, E. Shechtman, and K. Bala. Deep photo style transfer. In CVPR, 2017.
[24] Y. Shih, S. Paris, C. Barnes, W. T. Freeman, and F. Durand. Style transfer for headshot portraits. In
SIGGRAPH, 2014.
10
[25] Y. Shih, S. Paris, F. Durand, and W. T. Freeman. Data-driven hallucination of different times of day from a
single outdoor photo. In SIGGRAPH, 2013.
[26] K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition.
arXiv preprint arXiv:1409.1556, 2014.
[27] D. Ulyanov, V. Lebedev, A. Vedaldi, and V. Lempitsky. Texture networks: Feed-forward synthesis of
textures and stylized images. In ICML, 2016.
[28] D. Ulyanov, A. Vedaldi, and V. Lempitsky. Instance normalization: The missing ingredient for fast
stylization. arXiv preprint arXiv:1607.08022, 2016.
[29] D. Ulyanov, A. Vedaldi, and V. Lempitsky. Improved texture networks: Maximizing quality and diversity
in feed-forward stylization and texture synthesis. In CVPR, 2017.
[30] H. Wang, X. Liang, H. Zhang, D.-Y. Yeung, and E. P. Xing. Zm-net: Real-time zero-shot image manipulation network. arXiv preprint arXiv:1703.07255, 2017.
[31] X. Wang, G. Oxholm, D. Zhang, and Y.-F. Wang. Multimodal transfer: A hierarchical deep convolutional
neural network for fast artistic style transfer. In CVPR, 2017.
[32] P. Wilmot, E. Risser, and C. Barnes. Stable and controllable neural texture synthesis and style transfer
using histogram losses. arXiv preprint arXiv:1701.08893, 2017.
11
| 6642 |@word h:2 middle:2 inversion:1 advantageous:1 kokkinos:1 nd:1 open:1 rgb:2 covariance:12 decomposition:2 jacob:1 thereby:1 shot:1 accommodate:1 carry:3 shechtman:2 inefficiency:1 tuned:1 ours:4 subjective:1 existing:7 com:5 yet:2 gpu:1 shape:1 enables:2 remove:1 designed:2 generative:1 accordingly:1 coarse:6 preference:2 zhang:2 five:6 height:1 along:1 direct:1 qualitative:1 consists:1 yuan:2 wild:1 introduce:1 manner:4 mask:3 rapid:1 gatys:7 multi:13 inspired:1 ming:1 freeman:2 little:1 cpu:1 increasing:1 totally:1 gift:1 project:1 moreover:3 panel:1 agnostic:2 what:2 developed:1 finding:1 transformation:3 quantitative:2 every:1 sky:1 jimei:1 alahi:1 tackle:1 exactly:1 classifier:1 control:12 unit:1 grant:1 ramanan:1 before:3 local:5 textural:1 ulyanov:3 io:2 despite:1 solely:1 interpolation:5 ucmerced:2 challenging:1 collect:1 mentioning:1 limited:5 gone:1 averaged:1 acknowledgment:1 responsible:1 procedure:1 maire:1 universal:8 vedaldi:4 matching:6 pre:4 word:2 refers:1 get:2 cannot:1 selection:1 prentice:1 effortless:1 applying:1 seminal:1 context:1 map:5 ecker:3 center:3 missing:1 maximizing:1 layout:1 starting:1 l:4 formulate:1 resolution:1 hsuan:1 adjusts:1 shlens:2 fang:2 his:1 classic:1 handle:2 notion:1 variation:2 target:1 hierarchy:1 controlling:4 user:11 synthesize:3 element:1 satisfying:1 particularly:1 recognition:1 stripe:1 merced:2 module:1 preprint:7 wang:5 capture:6 thousand:1 region:4 trade:2 rescaled:1 yijun:1 broken:1 complexity:1 hertzmann:3 asked:1 trained:10 creation:1 efficiency:8 swap:1 easily:2 stylized:17 siggraph:3 multimodal:1 various:1 maji:1 train:5 fast:3 effective:8 outcome:1 encoded:1 cvpr:10 reconstruct:1 encoder:8 ability:3 statistic:8 simonyan:1 unseen:4 transform:11 final:1 karayev:1 eigenvalue:3 net:1 reconstruction:17 propose:5 subtracting:2 zm:1 loop:1 combining:1 flexibility:1 achieve:5 validate:1 stylization:28 webpage:1 requirement:2 optimum:1 darrell:1 generating:2 object:1 help:3 develop:3 nearest:1 received:1 progress:1 implemented:2 c:10 indicate:3 direction:1 drawback:1 closely:1 attribute:1 filter:2 human:1 explains:1 require:3 feeding:1 fix:1 generalization:7 wall:1 adjusted:1 strictly:1 blending:2 exploring:1 hall:1 ic:2 visually:2 overlaid:1 mapping:2 major:1 achieves:2 inverter:1 favorable:1 currently:1 bridge:1 edit:2 weighted:2 unfolding:1 clearly:1 always:1 gaussian:2 aim:3 i3:1 super:1 properly:1 improvement:1 indicates:1 mainly:1 greatly:2 contrast:3 adversarial:1 inference:1 entire:1 transferring:3 unlikely:1 w:2 her:1 manipulating:1 perona:1 transformed:4 i1:4 semantics:1 pixel:2 issue:2 agarwala:1 flexible:2 spatial:4 art:1 initialize:1 uc:2 brox:1 field:3 once:1 construct:1 never:1 beach:1 sampling:2 yu:1 unsupervised:1 icml:1 future:1 fundamentally:2 quantitatively:1 employ:2 few:2 dosovitskiy:1 randomly:2 preserve:1 individual:1 maintain:2 microsoft:2 pleasing:2 peel:1 highly:2 evaluation:1 adjust:4 hallucination:1 swapping:1 oliver:1 closer:1 experience:1 orthogonal:2 conduct:1 enjoying:1 desired:2 re:1 instance:3 brick:3 portrait:1 markovian:1 artistic:5 cost:4 recognizing:1 johnson:1 kudlur:2 combined:1 adaptively:1 st:2 lee:1 off:3 synthesis:15 bethge:3 lebedev:1 yao:1 reflect:2 choose:1 huang:3 style:131 leading:1 li:4 account:1 diversity:4 summarized:1 sec:1 includes:1 titan:1 unsatisfying:1 depends:1 piece:1 dumoulin:2 analyze:1 start:1 xing:1 complicated:2 capability:1 contribution:1 convolutional:5 variance:3 characteristic:4 efficiently:2 painting:1 salesin:1 generalize:1 curless:1 lu:2 marginally:1 mc:1 worth:1 drive:1 stroke:1 against:1 involved:1 mohamed:1 obvious:2 couple:1 dataset:2 adjusting:2 color:3 improves:1 back:1 coloring:15 feed:16 higher:5 day:1 reflected:1 zisserman:1 improved:2 daunting:1 editing:1 bmvc:2 generality:1 correlation:6 d:1 receives:1 replacing:1 assessment:1 lack:1 defines:1 quality:14 usa:1 effect:6 k22:2 normalized:1 hence:1 spatially:1 symmetric:1 i2:1 semantic:1 white:1 visualizing:1 round:1 width:1 m:3 demonstrate:5 performs:2 dtd:1 image:83 wise:1 novel:1 recently:2 common:2 empirically:1 extend:1 synthesized:4 significant:1 tuning:1 rd:2 stable:1 han:1 similarity:2 whitening:16 closest:3 multivariate:1 recent:3 driven:2 reverse:1 coco:2 manipulation:1 nvidia:2 certain:1 hay:1 luan:1 binary:1 continue:1 durand:2 inverted:1 preserving:1 minimum:1 greater:1 employed:1 determine:1 signal:1 preservation:2 ii:3 multiple:2 smooth:2 match:11 long:1 lin:1 equally:1 adobe:9 whitened:6 essentially:2 metric:2 liao:2 arxiv:14 histogram:4 normalization:4 repetitive:1 iteration:1 texton:1 invert:4 achieved:1 yeung:1 preserved:2 addition:4 want:1 fine:8 separately:1 wct:19 pass:3 subject:3 spirit:1 effectiveness:4 yang:4 empower:1 intermediate:6 split:1 enough:1 click:1 vgg:22 gb:1 effort:1 f:8 cause:1 remark:1 deep:10 detailed:1 eigenvectors:2 clear:1 transforms:10 locally:1 extensively:1 http:1 generate:7 overlay:1 percentage:1 nsf:1 trapped:1 per:1 diverse:3 key:3 salient:4 shih:2 achieving:1 downstream:1 sum:2 wood:1 wand:2 run:2 inverse:1 i5:3 almost:1 patch:4 gonzalez:1 comparable:1 bit:2 capturing:3 layer:22 display:1 correspondence:2 bala:1 replaces:1 barnes:2 i4:3 strength:1 fei:2 scene:1 generates:4 aspect:2 speed:2 wc:2 interpolated:2 formulating:1 concluding:1 relatively:2 transferred:3 zhaowen:1 across:2 reconstructing:1 appealing:1 describable:1 s1:2 explained:2 iccv:1 artwork:1 pipeline:7 visualization:2 remains:1 turn:1 describing:1 precomputed:1 needed:1 fed:2 serf:1 end:1 photo:2 available:1 operation:5 doll:1 apply:5 hierarchical:1 alternative:1 encounter:1 schmidt:1 slower:2 original:7 marginalized:1 especially:2 question:1 realized:1 blend:2 strategy:1 receptive:2 diagonal:3 exhibit:1 iclr:1 reversed:1 mapped:1 upsampling:1 decoder:18 code:1 relationship:1 providing:1 minimizing:1 balance:5 liang:1 executed:1 mostly:2 expense:1 ds2:1 synthesizing:2 design:1 implementation:1 perform:2 markov:1 dc:3 arbitrary:12 community:1 intensity:1 inverting:3 pair:7 required:1 specified:1 paris:3 learned:4 kang:1 nip:3 address:1 able:5 usually:1 pattern:9 perception:1 challenge:1 including:4 mouth:1 treated:1 natural:1 turning:1 scheme:2 improve:1 github:1 recons:7 eye:1 numerous:1 ready:1 carried:1 hm:5 extract:4 coupled:1 auto:2 understanding:1 embedded:1 loss:9 fully:1 mixed:1 interesting:1 limitation:1 generation:1 analogy:4 ingredient:2 remarkable:1 digital:1 integrate:1 degree:1 vectorized:1 uncorrelated:1 share:1 row:5 eccv:3 supported:1 free:2 enjoys:2 side:2 neighbor:1 face:1 regard:2 feedback:1 depth:1 dimension:1 gram:6 cumulative:1 rich:1 evaluating:1 forward:16 made:1 commonly:1 universally:1 adaptive:2 mhyang:1 far:1 ec:5 absorb:1 global:4 sequentially:1 cimpoi:1 symmetrical:1 belongie:2 oxholm:1 consuming:1 compromised:3 iterative:2 table:6 additionally:1 favorite:1 learn:1 transfer:41 channel:4 ca:1 career:1 controllable:1 hc:2 meanwhile:1 zitnick:1 main:3 dense:1 linearly:2 midst:1 s2:2 noise:9 edition:1 depicts:1 slow:1 sub:1 fails:2 explicit:1 outdoor:1 perceptual:3 extractor:1 learns:2 enlarging:1 list:1 effectively:1 texture:32 execution:1 conditioned:1 chen:6 generalizing:2 depicted:1 fc:16 simply:5 likely:2 fcs:2 visual:15 unexpected:1 adjustment:1 diversified:1 hua:2 determines:1 extracted:6 conditional:1 lempitsky:3 goal:3 formulated:2 viewed:1 shared:1 content:47 pas:1 e:3 xin:1 experimental:2 vote:4 select:3 inability:1 evaluate:6 tested:1 |
6,238 | 6,643 | On the Model Shrinkage Effect of
Gamma Process Edge Partition Models
Iku Ohama??
Issei Sato?
Takuya Kida?
Hiroki Arimura?
?
?
?
Panasonic Corp., Japan The Univ. of Tokyo, Japan Hokkaido Univ., Japan
[email protected] [email protected] {kida,arim}@ist.hokudai.ac.jp
Abstract
The edge partition model (EPM) is a fundamental Bayesian nonparametric model for extracting an overlapping structure from binary matrix. The
EPM adopts a gamma process (?P) prior to automatically shrink the number of active atoms. However, we empirically found that the model shrinkage of the EPM does not typically work appropriately and leads to an
overfitted solution. An analysis of the expectation of the EPM?s intensity
function suggested that the gamma priors for the EPM hyperparameters
disturb the model shrinkage effect of the internal ?P. In order to ensure that
the model shrinkage effect of the EPM works in an appropriate manner, we
proposed two novel generative constructions of the EPM: CEPM incorporating constrained gamma priors, and DEPM incorporating Dirichlet priors
instead of the gamma priors. Furthermore, all DEPM?s model parameters
including the infinite atoms of the ?P prior could be marginalized out, and
thus it was possible to derive a truly infinite DEPM (IDEPM) that can
be efficiently inferred using a collapsed Gibbs sampler. We experimentally
confirmed that the model shrinkage of the proposed models works well and
that the IDEPM indicated state-of-the-art performance in generalization
ability, link prediction accuracy, mixing efficiency, and convergence speed.
1
Introduction
Discovering low-dimensional structure from a binary matrix is an important problem in
relational data analysis. Bayesian nonparametric priors, such as Dirichlet process (DP) [1]
and hierarchical Dirichlet process (HDP) [2], have been widely applied to construct statistical
models with an automatic model shrinkage effect [3, 4]. Recently, more advanced stochastic
processes such as the Indian buffet process (IBP) [5] enabled the construction of statistical
models for discovering overlapping structures [6, 7], wherein each individual in a data matrix
can belong to multiple latent classes.
Among these models, the edge partition model (EPM) [8] is a fundamental Bayesian nonparametric model for extracting overlapping latent structure underlying a given binary matrix.
The EPM considers latent positive random counts for only non-zero entries in a given binary
matrix and factorizes the count matrix into two non-negative matrices and a non-negative
diagonal matrix. A link probability of the EPM for an entry is defined by transforming
the multiplication of the non-negative matrices into a probability, and thus the EPM can
capture overlapping structures with a noisy-OR manner [6]. By incorporating a gamma
process (?P) as a prior for the diagonal matrix, the number of active atoms of the EPM
shrinks automatically according to the given data. Furthermore, by truncating the infinite
atoms of the ?P with a finite number, all parameters and hyperparameters of the EPM
can be inferred using closed-form Gibbs sampler. Although, the EPM is well designed to
capture an overlapping structure and has an attractive affinity with a closed-form posterior
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
The proposed
IDEPM
successfully
found expected
5 overlapped
latent classes.
The EPM
extracted many
unexpected
latent classes.
(98 active
classes)
(a) Synthetic data
(b) EPM solution
(c) Proposed IDEPM solution
Figure 1: (Best viewed in color.) A synthetic example: (a) synthetic 90 ? 90 data (white
corresponds to one, and black to zero); (b) EPM solution; and (c) the proposed IDEPM solution. In (b) and (c), non-zero entries are colored to indicate their most probable assignment
to the latent classes.
inference, the EPM involves a critical drawback in its model shrinkage mechanism. As we
experimentally show in Sec. 5, we found that the model shrinkage effect of the EPM does
not typically work in an appropriate manner. Figure 1 shows a synthetic example. As shown
in Fig. 1a, there are five overlapping latent classes (white blocks). However, as shown in
Fig. 1b, the EPM overestimates the number of active atoms (classes) and overfits the data.
In this paper, we analyze the undesired property of the EPM?s model shrinkage mechanism
and propose novel generative constructions for the EPM to overcome the aforementioned
disadvantage. As shown in Fig. 1c, the IDEPM proposed in this paper successfully shrinks
unnecessary atoms. More specifically, we have three major contributions in this paper.
(1) We analyse the generative construction of the EPM and find a property that disturbs its
model shrinkage effect (Sec. 3). We derive the expectation of the EPM?s intensity function
(Theorem 1), which is the total sum of the infinite atoms for an entry. From the derived
expectation, we obtain a new finding that gamma priors for the EPM?s hyperparameters
disturb the model shrinkage effect of the internal ?P (Theorem 2). That is, the derived
expectation is expressed by a multiplication of the terms related to ?P and other gamma
priors. Thus, there is no guarantee that the expected number of active atoms is finite.
(2) Based on the analysis of the EPM?s intensity function, we propose two novel constructions of the EPM: the CEPM incorporating constrained gamma priors (Sec. 4.1) and the
DEPM incorporating Dirichlet priors instead of the gamma priors (Sec. 4.2). The model
shrinkage effect of the CEPM and DEPM works appropriately because the expectation of
their intensity functions depends only on the ?P prior (Sec. 4.1 and Theorem 3 in Sec. 4.2).
(3) Furthermore, for the DEPM, all model parameters, including the infinite atoms of the
?P prior, can be marginalized out (Theorem 4). Therefore, we can derive a truly infinite
DEPM (IDEPM), which has a closed-form marginal likelihood without truncating infinite
atoms, and can be efficiently inferred using collapsed Gibbs sampler [9] (Sec. 4.3).
2
The Edge Partition Model (EPM)
In this section, we review the EPM [8] as a baseline model. Let x be an I ? J binary matrix,
where an entry between i-th row and j-th column is represented by xi,j ? {0, 1}. In order to
extract an overlapping structure underlying x, the EPM [8] considers a non-negative matrix
factorization problem on latent Poisson counts as follows:
!
K
X
(1)
xi,j = I(mi,j,? ? 1), mi,j,? | U , V , ? ? Poisson
Ui,k Vj,k ?k ,
k=1
where U and V are I ? K and J ? K non-negative matrices, respectively, and ? is a K ? K
non-negative diagonal matrix. Note that I(?) is 1 if the predicate holds and is zero otherwise.
The latent counts m take positive values only for edges (non-zero entries) within a given
binary matrix and the generative model for each positive
count is equivalently expressed as
P
a sum of K Poisson random variables as mi,j,? = k mi,j,k , mi,j,k ? Poisson(Ui,k Vj,k ?k ).
This is the reason why the above model is called edge partition model. Marginalizing
m out from Eq. (1), the generative model of the EPM can be equivalently rewritten as
2
Q
xi,j | U , V , ? ? Bernoulli(1 ? k e?Ui,k Vj,k ?k ). As e?Ui,k Vj,k ?k ? [0, 1] denotes the probability that a Poisson random variable with mean Ui,k Vj,k ?k corresponds to zero, the EPM can
capture an overlapping structure with a noisy-OR manner [6].
In order to complete the Bayesian hierarchical model of the EPM, gamma priors are adopted
as Ui,k ? Gamma(a1 , b1 ) and Vj,k ? Gamma(a2 , b2 ), where a1 , a2 are shape parameters and
b1 , b2 are rate parameters for the gamma distribution, respectively. Furthermore, a gamma
process (?P) is incorporated as a Bayesian nonparametric prior for ? to make the EPM
automatically shrink its number of atoms K. Let Gamma(?0 /T, c0 ) denote a truncated ?P
with a concentration parameter ?0 and a rate parameter c0 , where T denotes a truncation
level that should be set large enough to ensure a good approximation to the true ?P. Then,
the diagonal elements of ? are drawn as ?k ? Gamma(?0 /T, c0 ) for k ? {1, . . . , T }.
The posterior inference for all parameters and hyperparameters of the EPM can be performed using Gibbs sampler (detailed in Appendix
conjugacy between
P A). Thanks to the P
gamma and Poisson distributions, given mi,?,k = j mi,j,k and m?,j,k = i mi,j,k , posterior
sampling for Ui,k and Vj,k is straightforward. As the ?P prior is approximated by a gamma
distribution, posterior sampling for ?k also can be performed straightforwardly. Given U ,
V , and ?, posterior sample for mi,j,? can be simulated using zero-truncated Poisson (ZTP)
distribution [10]. Finally, we can obtain sufficient statistics mi,j,k by partitioning mi,j,? into
T atoms using a multinomial distribution. Furthermore, all hyperparameters of the EPM
(i.e., ?0 , c0 , a1 , a2 , b1 , and b2 ) can also be sampled by assuming a gamma hyper prior
Gamma(e0 , f0 ). Thanks to the conjugacy between gamma distributions, posterior sampling
for c0 , b1 , and b2 is straightforward. For the remaining hyperparameters, we can construct
closed-form Gibbs samplers using data augmentation techniques [11, 12, 2].
3
Analysis for Model Shrinkage Mechanism
The EPM is well designed to capture an overlapping structure with a simple Gibbs inference.
However, the EPM involves a critical drawback in its model shrinkage mechanism.
For the EPM, a ?P prior is incorporated as a prior for the non-negative diagonal matrix as
?k ? Gamma(?0 /T, c0 ). From the form of the truncated ?P, thanks to the additive property
of independent gamma random P
variables, the total sum of ?k over countably infinite atoms
?
follows a gamma distribution as k=1P?k ? Gamma(?0 , c0 ), wherein the intensity function of
?
the ?P has a finite expectation as E[ k=1 ?k ] = ?c00 . Therefore, the ?P has a regularization
mechanism that automatically shrinks the number of atoms according to given observations.
However, as experimentally shown in Sec. 5, the model shrinkage mechanism of the EPM
does not work appropriately. More specifically, the EPM often overestimates the number of
active atoms and overfits the data. Thus, we analyse the intensity function of the EPM to
reveal the reason why the model shrinkage mechanism does not work appropriately.
P?
Theorem 1. The expectation of the EPM?s intensity function k=1 Ui,k Vj,k ?k for an entry
(i, j) is finite and can be expressed as follows:
#
"?
X
a2
?0
a1
?
? .
(2)
E
Ui,k Vj,k ?k =
b1
b2
c0
k=1
Proof. As U , V , and ? are independent of each other, the expected value operator is multiplicative for the EPM?s intensity function.PUsing the multiplicativity
and the low of total
P?
?
expectation, P
the proof is completed as E [ k=1 Ui,k Vj,k ?k ] = k=1 E[Ui,k ]E[Vj,k ]E[?k ] =
?
a1
a2
k=1 ?k ].
b1 ? b2 ? E[
As Eq. (2) in Theorem 1 shows, the expectation of the EPM?s intensity function is expressed
by multiplying individual expectations of a ?P and two gamma distributions. This causes
an undesirable property to the model shrinkage effect of the EPM. From Theorem 1, another
important theorem about the EPM?s model shrinkage effect is obtained as follows:
3
Theorem 2. Given an arbitrary non-negative constant
P? C, even if the expectation of the
EPM?s intensity function in Eq. (2) is fixed as E [ k=1 Ui,k Vj,k ?k ] = C, there exist cases
in which the model shrinkage effect of the ?P prior disappears.
P?
Proof. Substituting E [ k=1 Ui,k Vj,k ?k ] = C for Eq. (2), we obtain C = ab11 ? ab22 ? ?c00 . Since
a1 , a2 , b1 , and b2 are gamma random variables, even if the expectation of the EPM?s intensity
function, C, is fixed, ?c00 can take an arbitrary value so that equation C = ab11 ? ab22 ? ?c00
holds. Hence, ?0 can take an arbitrary large value such that ?0 = T ? ?
b0 . This implies that
the ?P prior for the EPM degrades to a gamma distribution without model shrinkage effect
as ?k ? Gamma(?0 /T, c0 ) = Gamma(b
?0 , c0 ).
Theorem 2 indicates that the EPM might overestimate the number of active atoms, and
lead to overfitted solutions.
4
Proposed Generative Constructions
We describe our novel generative constructions for the EPM with an appropriate model
shrinkage effect. According to the analysis described in Sec. 3, the model shrinkage mechanism of the EPM does not work because the expectation of the EPM?s intensity function
has an undesirable redundancy. This finding motivates the proposal of new generative constructions, in which the expectation of the intensity function depends only on the ?P prior.
First, we propose a naive extension of the original EPM using constrained gamma priors
(termed as CEPM). Next, we propose an another generative construction for the EPM by
incorporating Dirichlet priors instead of gamma priors (termed as DEPM). Furthermore,
for the DEPM, we derive truly infinite DEPM (termed as IDEPM) by marginalizing out all
model parameters including the infinite atoms of the ?P prior.
4.1
CEPM
In order to ensure that the EPM?s intensity function depends solely on the ?P prior, a naive
way is to introduce constraints for the hyperparameters of the gamma prior. In the CEPM,
the rate parameters of the gamma priors are constrained as b1 = C1 ? a1 and b2 = C2 ? a2 ,
respectively, where C1 > 0 and C2 > 0 are arbitrary constants. Based on the aforementioned
constraints and Theorem 1,Pthe expectation of the intensity function for the CEPM depends
?
only on the ?P prior as E[ k=1 Ui,k Vj,k ?k ] = C1?C02 c0 .
The posterior inference for the CEPM can be performed using Gibbs sampler in a manner
similar to that for the EPM. However, we can not derive closed-form samplers only for a1
and a2 because of the constraints. Thus, in this paper, posterior sampling for a1 and a2 are
performed using grid Gibbs sampling [13] (see Appendix B for details).
4.2
DEPM
We have another strategy to construct the EPM with efficient model shrinkage effect by
re-parametrizing the factorization problem. Let us denote transpose of a matrix A by A? .
According to the generative model of the EPM in Eq. (1), the original generative process
for counts m can be viewed as a matrix factorization as m ? U ?V ? . It is clear that the
optimal solution of the factorization problem is not unique. Let ?1 and ?2 be arbitrary
K ? K non-negative diagonal matrices. If a solution m ? U ?V ? is globally optimal, then
?1 ?
another solution m ? (U ?1 )(??1
is also optimal. In order to ensure that
1 ??2 )(V ?2 )
the EPM has only one optimal solution, we re-parametrize the original factorization problem
to an equivalent constrained factorization problem as follows:
m ? ??? ? ,
(3)
P
where ? denotes an I ? K non-negative matrix with l1 -constraints as Pi ?i,k = 1, ?k.
Similarly, ? denotes an J ? K non-negative matrix with l1 -constraints as j ?j,k = 1, ?k.
This parameterization ensures the uniqueness of the optimal solution for a given m because
each column of ? and ? is constrained such that it is defined on a simplex.
4
According to the factorization in Eq. (3), by incorporating Dirichlet priors instead of gamma
priors, the generative construction for m of the DEPM is as follows:
!
I
T
X
z }| {
mi,j,? | ?, ?, ? ? Poisson
?i,k ?j,k ?k , {?i,k }Ii=1 | ?1 ? Dirichlet(?1 , . . . , ?1 ),
k=1
J
z }| {
(4)
| ?2 ? Dirichlet(?2 , . . . , ?2 ), ?k | ?0 , c0 ? Gamma(?0 /T, c0 ).
P?
Theorem 3. The expectation of DEPM?s intensity
function
?
?
?
depends
sorely
k=1 i,k j,k k
P?
?0
.
on the ?P prior and can be expressed as E[ k=1 ?i,k ?j,k ?k ] = IJc
0
{?j,k }Jj=1
Proof. The expectations of Dirichlet random variables ?i,k and ?j,k are I1 and J1 , respectively. Similar to the proof for Theorem 1, using the multiplicativity of independent
random
P?
variables
and
the
low
of
total
expectation,
the
proof
is
completed
as
E
[
?
?
?k ] =
i,k
j,k
k=1
P?
P?
1
1
E[?
]E[?
]E[?
]
=
?
?
E[
?
].
i,k
j,k
k
k=1
k=1 k
I
J
Note that, if we set constants C1 = I and C2 = J for the CEPM in Sec. 4.1, then the
expectation of the intensity function for the CEPM is equivalent to that for the DEPM in
Theorem 3. Thus, in order to ensure the fairness of comparisons, we set C1 = I and C2 = J
for the CEPM in the experiments.
As the Gibbs sampler for ? and ? can be derived straightforwardly, the posterior inference
for all parameters and hyperparameters of the DEPM also can be performed via closedform Gibbs sampler (detailed in Appendix C). Differ from the CEPM, l1 -constraints in the
DEPM ensure the uniqueness of its optimal solution. Thus, the inference for the DEPM is
considered as more efficient than that for the CEPM.
4.3
Truly Infinite DEPM (IDEPM)
One remarkable property of the DEPM is that we can derive a fully marginalized likelihood
function. Similar to the beta-negative binomial topic model [13], we consider a joint distribumi,j,?
? {1, ? ? ? , T }mi,j,?
tion for mi,j,? Poisson customers and their assignments zi,j =Q
{zi,j,s }s=1
mi,j,?
to T tables as P (mi,j,? , zi,j | ?, ?, ?) = P (mi,j,? | ?, ?, ?) s=1 P (zi,j,s | mi,j,? , ?, ?, ?).
Thanks to the l1 -constraints we introduced in Eq. (3), the joint distribution P (m, z | ?, ?, ?)
has a fully factorized form (see Lemma 1 in Appendix D). Therefore, marginalizing ?, ?,
and ? out according to the prior construction in Eq. (4), we obtain an analytical marginal
likelihood P (m, z) for the truncated DEPM (see Appendix D for a detailed derivation).
Furthermore, by taking T ? ?, we can derive a closed-form marginal likelihood for the
truly infinite version of the DEPM (termed as IDEPM). In a manner similar to that in [14],
we consider the likelihood function for partition
P [z]
P instead of the assignments z. Assume
we have K+ of T atoms for which m?,?,k =
i
j mi,j,k > 0, and a partition of M (=
P P
m
)
customers
into
K
subsets.
Then,
joint
marginal likelihood of the IDEPM
i,j,?
+
i
j
for [z] and m is given by the following theorem, with the proof provided in Appendix D:
Theorem 4. The marginal likelihood function of the IDEPM is defined as P (m, [z])? =
T!
limT ?? P (m, [z]) = limT ?? (T ?K
P (m, z), and can be derived as follows:
+ )!
P (m, [z])? =
J
I Y
Y
i=1 j=1
1
mi,j,? !
?
K+
Y
k=1
I
Y
?(I?1 )
?(?1 + mi,?,k )
?(I?1 + m?,?,k ) i=1
?(?1 )
?0 Y
K+
c0
?(m?,?,k )
?(?2 + m?,j,k )
?(J?2 )
K+
? ?0
, (5)
?
?(J?2 + m?,?,k ) j=1
?(?2 )
c0 + 1
(c0 + 1)m?,?,k
k=1
k=1
P
P
P P
where mi,?,k =
j mi,j,k , m?,j,k =
i mi,j,k , and m?,?,k =
i
j mi,j,k . Note that ?(?)
denotes gamma function.
K+
Y
J
Y
From Eq. (5) in Theorem 4, we can derive collapsed Gibbs sampler [9] to perform posterior
inference for the IDEPM. Since ?, ?, and ? have been marginalized out, the only latent
variables we have to update are m and z.
5
Sampling z: Given m, similar to the Chinese restaurant process (CRP) [15], the posterior
probability that zi,j,s is assigned to k ? is given as follows:
?
\(ijs)
\(ijs)
? m\(ijs) ? ?1 +mi,?,k? ? ?2 +m?,j,k? if m\(ijs) > 0,
\(ijs)
\(ijs)
?,?,k?
k?
?
I?1 +m?,?,k?
I?2 +m?,?,k?
(6)
P (zi,j,s = k | z\(ijs) , m) ?
?
\(ijs)
1
1
?0 ? I ? J
if m?,?,k? = 0,
where the superscript \(ijs) denotes that the corresponding statistics are computed excluding the s-th customer of entry (i, j).
Sampling m: Given z, posteriors for the ? and ? are simulated as {?i,k }Ii=1 | ? ?
Dirichlet({?1 + mi,?,k }Ii=1 ) and {?j,k }Jj=1 | ? ? Dirichlet({?2 + m?,j,k }Jj=1 ) for k ?
{1, . . . , K+ }. Furthermore, the posterior sampling of the ?k for K+ active atoms can be
performed as ?k | ? ? Gamma(m?,?,k , c0 + 1). Therefore, similar to the sampler for the
EPM [8], we can update m as follows:
?(0)
if xi,j = 0,
PK +
mi,j,? | ?, ?, ? ?
(7)
ZTP( k=1 ?i,k ?j,k ?k ) if xi,j = 1,
?
(
)K+ ?
?
?
?
K+
i,k
j,k
k
?,
{mi,j,k }k=1 | mi,j,? , ?, ?, ? ? Multinomial ?mi,j,? ; PK+
(8)
? ?j,k ? ?k ?
?
?
i,k
k =1
k=1
where ?(0) denotes point mass at zero.
Sampling hyperparameters: We can construct closed-form Gibbs sampler for all hyperparameters of the IDEPM assuming a gamma prior (Gamma(e0 , f0 )). Using the additive
property
of the ?P, posterior sample for the sum of ?k over unused atoms is obtained as ??0 =
P?
k? =K+ +1 ?k? | ? ? Gamma(?0 , c0 + 1). Consequently, we obtain a closed-form posterior
PK+
?k ).
sampler for the rate parameter c0 of the ?P as c0 | ? ? Gamma(e0 + ?0 , f0 + ??0 + k=1
For all remaining hyperparameters (i.e., ?1 , ?2 , and ?0 ), we can derive posterior samplers
from Eq. (5) using data augmentation techniques [12, 8, 2, 11] (detailed in Appendix E).
5
Experimental Results
In previous sections, we theoretically analysed the reason why the model shrinkage of the
EPM does not work appropriately (Sec. 3) and proposed several novel constructions (i.e.,
CEPM, DEPM, and IDEPM) of the EPM with an efficient model shrinkage effect (Sec. 4).
The purpose of the experiments involves ascertaining the following hypotheses:
(H1) The original EPM overestimates the number of active atoms and overfits the data.
In contrast, the model shrinkage mechanisms of the CEPM and DEPM work appropriately. Consequently, the CEPM and DEPM outperform the EPM in generalization ability and link prediction accuracy.
(H2) Compared with the CEPM, the DEPM indicates better generalization ability and
link prediction accuracy because of the uniqueness of the DEPM?s optimal solution.
(H3) The IDEPM with collapsed Gibbs sampler is superior to the DEPM in generalization
ability, link prediction accuracy, mixing efficiency, and convergence speed.
Datasets: The first dataset was the Enron [16] dataset, which comprises e-mails sent
between 149 Enron employees. We extracted e-mail transactions from September 2001 and
constructed Enron09 dataset. For this dataset, xi,j = 1(0) was used to indicate whether
an e-mail was, or was not, sent by the i-th employee to the j-th employee. For larger
dataset, we used the MovieLens [17] dataset, which comprises five-point scale ratings of
movies submitted by users. For this dataset, we set xi,j = 1 when the rating was higher
than three and xi,j = 0 otherwise. We prepared two different sized MovieLens dataset:
MovieLens100K (943 users and 1,682 movies) and MovieLens1M (6,040 users and 3,706
movies). The densities of the Enron09, MovieLens100K and MovieLens1M datasets were
0.016, 0.035, and 0.026, respectively.
6
(a) Enron09
Estimated # of K
128
64
(b) MovieLens100K
128
IDEPM
DEPM-T
CEPM-T
EPM-T
64
64
32
32
16
16
16
8
8
8
4
4
4
2
2
32
2
4
8
16
32
Truncation level T
64
128
2
2
4
TDLL
TDAUC-PR
4
8
16
32
64
128
IDEPM
DEPM-T
CEPM-T
EPM-T
Truncation level T
(g) Enron09
0.350
0.300
0.250
0.200
0.150
0.100
0.050
0.000
8
16
32
Truncation level T
64
128
-0.084
-0.086 2
-0.088
-0.090
-0.092
-0.094
-0.096
-0.098
-0.100
8
16
32
Truncation level T
4
8
16
32
64
128
8
16 32 64
Truncation level T
2
4
64
128
8
16 32 64
Truncation level T
128
8
16
32
-0.072
-0.074
-0.076
-0.078
Truncation level T
Truncation level T
(i) MovieLens1M
0.440
0.420
0.400
0.380
0.360
0.340
0.320
0.300
0.390
0.370
128
128
-0.070
0.410
IDEPM
DEPM-T
CEPM-T
EPM-T
-0.068
(h) MovieLens100K
0.470
64
(f) MovieLens1M
0.430
4
4
-0.066
0.450
2
2
(e) MovieLens100K
(d) Enron09
-0.040
-0.050 2
-0.060
-0.070
-0.080
-0.090
-0.100
-0.110
-0.120
(c) MovieLens1M
128
2
4
8
16 32 64
Truncation level T
128
2
4
Figure 2: Calculated measurements as functions of the truncation level T for each dataset.
The horizontal line in each figure denotes the result obtained using the IDEPM.
Evaluating Measures: We adopted three measurements to evaluate the performance of
the models. The first is the estimated number of active atoms K for evaluating the model
shrinkage effect of each model. The second is the averaged Test Data Log Likelihood (TDLL)
for evaluating the generalization ability of each model. We calculated the averaged likelihood
that a test entry takes the actual value. For the third measurement, as many real-world
binary matrices are often sparse, we adopted the Test Data Area Under the Curve of the
Precision-Recall curve (TDAUC-PR) [18] to evaluate the link prediction ability. In order
to calculate the TDLL and TDAUC-PR, we set all the selected test entries as zero during
the inference period, because binary observations for unobserved entries are not observed
as missing values but are observed as zeros in many real-world situations.
Experimental Settings: Posterior inference for the truncated models (i.e., EPM, CEPM,
and DEPM) were performed using standard (non-collapsed) Gibbs sampler. Posterior inference for the IDEPM was performed using the collapsed Gibbs sampler derived in Sec. 4.3.
For all models, we also sampled all hyperparameters assuming the same gamma prior
(Gamma(e0 , f0 )). For the purpose of fair comparison, we set hyper-hyperparameters as
e0 = f0 = 0.01 throughout the experiments. We ran 600 Gibbs iterations for each model on
each dataset and used the final 100 iterations to calculate the measurements. Furthermore,
all reported measurements were averaged values obtained by 10-fold cross validation.
Results: Hereafter, the truncated models are denoted as EPM-T , CEPM-T , and DEPM-T
to specify the truncation level T . Figure 2 shows the calculated measurements.
(H1) As shown in Figs. 2a?c, the EPM overestimated the number of active atoms K for all
datasets especially for a large truncation level T . In contrast, the number of active atoms
K for the CEPM-T and DEPM-T monotonically converges to a specific value. This result
supports the analysis with respect to the relationship between the model shrinkage effect
and the expectation of the EPM?s intensity function, as discussed in Sec. 3. Consequently,
7
(a) Enron09
-0.050
TDLL
-0.055
0
-0.086
100 200 300 400 500 600
-0.088
0
(b) MovieLens100K
100 200 300 400 500 600 -0.066 0
-0.068
-0.060
-0.090
-0.070
-0.065
-0.092
-0.072
-0.070
-0.094
IDEPM
DEPM-128
-0.075
-0.080
-0.074
-0.076
-0.096
-0.078
-0.098
Gibbs sampling iteration
(c) MovieLens1M
100 200 300 400 500 600
-0.080
Gibbs sampling iteration
Gibbs sampling iteration
Figure 3: (Best viewed in color.) The TDLL as a function of the Gibbs iterations.
(a) Enron09
-0.050
-0.055
0
10
20
30
(b) MovieLens100K
-0.080
40
50
-0.085
0
200
400
600
800
(c) MovieLens1M
-0.065
1000
0
2000
4000
6000
8000 10000
TDLL
-0.070
-0.060
-0.090
-0.065
-0.095
-0.070
-0.075
-0.080
IDEPM
DEPM-128
DEPM-64
DEPM-32
DEPM-16
DEPM-8
DEPM-4
DEPM-2
Elapsed time (sec)
-0.075
-0.100
-0.080
-0.105
-0.110
Elapsed time (sec)
-0.085
Elapsed time (sec)
Figure 4: (Best viewed in color.) The TDLL as a function of the elapsed time (in seconds).
as shown by the TDLL (Figs. 2d?f) and TDAUC-PR (Figs. 2g?i), the CEPM and DEPM
outperformed the original EPM in both generalization ability and link prediction accuracy.
(H2) As shown in Figs. 2a?c, the model shrinkage effect of the DEPM is stronger than
that of the CEPM. As a result, the DEPM significantly outperformed the CEPM in both
generalization ability and link prediction accuracy (Figs. 2d?i). Although the CEPM slightly
outperformed the EPM, the CEPM with a larger T tends to overfit the data. In contrast,
the DEPM indicated its best performance with the largest truncation level (T = 128).
Therefore, we confirmed that the uniqueness of the optimal solution in the DEPM was
considerably important in achieving good generalization ability and link prediction accuracy.
(H3) As shown by the horizontal lines in Figs. 2d?i, the IDEPM indicated the state-of-theart scores for all datasets. Finally, the computational efficiency of the IDEPM was compared
with that of the truncated DEPM. Figure 3 shows the TDLL as a function of the number
of Gibbs iterations. In keeping with expectations, the IDEPM indicated significantly better
mixing property when compared with that of the DEPM for all datasets. Furthermore,
Fig. 4 shows a comparison of the convergence speed of the IDEPM and DEPM with several
truncation levels (T = {2, 4, 8, 16, 32, 64, 128}). As clearly shown in the figure, the convergence of the IDEPM was significantly faster than that of the DEPM with all truncation
levels. Therefore, we confirmed that the IDEPM indicated a state-of-the-art performance
in generalization ability, link prediction accuracy, mixing efficiency, and convergence speed.
6
Conclusions
In this paper, we analysed the model shrinkage effect of the EPM, which is a Bayesian
nonparametric model for extracting overlapping structure with an optimal dimension from
binary matrices. We derived the expectation of the intensity function of the EPM, and
showed that the redundancy of the EPM?s intensity function disturbs its model shrinkage
effect. According to this finding, we proposed two novel generative construction for the
EPM (i.e., CEPM and DEPM) to ensure that its model shrinkage effect works appropriately. Furthermore, we derived a truly infinite version of the DEPM (i.e, IDEPM), which
can be inferred using collapsed Gibbs sampler without any approximation for the ?P. We experimentally showed that the model shrinkage mechanism of the CEPM and DEPM worked
appropriately. Furthermore, we confirmed that the proposed IDEPM indicated a state-ofthe-art performance in generalization ability, link prediction accuracy, mixing efficiency, and
convergence speed. It is of interest to further investigate whether the truly infinite construction of the IDEPM can be applied to more complex and modern machine learning models,
including deep brief networks [19], and tensor factorization models [20].
8
References
[1]
[2]
[3]
[4]
[5]
[6]
[7]
[8]
[9]
[10]
[11]
[12]
[13]
[14]
[15]
[16]
[17]
[18]
[19]
[20]
Thomas S. Ferguson. ?A Bayesian Analysis of Some Nonparametric Problems?. In:
The Annals of Statistics 1.2 (1973), pp. 209?230.
Yee Whye Teh, Michael I. Jordan, Matthew J. Beal, and David M. Blei. ?Hierarchical
Dirichlet Processes?. In: J. Am. Stat. Assoc. 101.476 (2006), pp. 1566?1581.
Charles Kemp, Joshua B. Tenenbaum, Thomas L. Griffiths, Takeshi Yamada, and
Naonori Ueda. ?Learning Systems of Concepts with an Infinite Relational Model?. In:
Proc. AAAI. Vol. 1. 2006, pp. 381?388.
Edoardo M. Airoldi, David M. Blei, Stephen E. Fienberg, and Eric P. Xing. ?Mixed
Membership Stochastic Blockmodels?. In: J. Mach. Learn. Res. 9 (2008), pp. 1981?
2014.
Thomas L. Griffiths and Zoubin Ghahramani. ?Infinite Latent Feature Models and
the Indian Buffet Process?. In: Proc. NIPS. 2005, pp. 475?482.
Morten M?rup, Mikkel N. Schmidt, and Lars Kai Hansen. ?Infinite Multiple Membership Relational Modeling for Complex Networks?. In: Proc. MLSP. 2011, pp. 1?
6.
Konstantina Palla, David A. Knowles, and Zoubin Ghahramani. ?An Infinite Latent
Attribute Model for Network Data?. In: Proc. ICML. 2012, pp. 1607?1614.
Mingyuan Zhou. ?Infinite Edge Partition Models for Overlapping Community Detection and Link Prediction?. In: Proc. AISTATS. Vol. 38. 2015, pp. 1135?1143.
Jun S. Liu. ?The Collapsed Gibbs Sampler in Bayesian Computations with Applications to a Gene Regulation Problem?. In: J. Am. Stat. Assoc. 89.427 (1994), pp. 958?
966.
Charles J. Geyer. Lower-Truncated Poisson and Negative Binomial Distributions.
Tech. rep. Working Paper Written for the Software R. University of Minnesota, MN
(available: http://cran.r-project.org/web/packages/aster/vignettes/trunc.pdf), 2007.
David Newman, Arthur U. Asuncion, Padhraic Smyth, and Max Welling. ?Distributed
Algorithms for Topic Models?. In: J. Mach. Learn. Res. 10 (2009), pp. 1801?1828.
Michael D. Escobar and Mike West. ?Bayesian Density Estimation and Inference Using
Mixtures?. In: J. Am. Stat. Assoc. 90 (1994), pp. 577?588.
Mingyuan Zhou. ?Beta-Negative Binomial Process and Exchangeable Random Partitions for Mixed-Membership Modeling?. In: Proc. NIPS. 2014, pp. 3455?3463.
Thomas L. Griffiths and Zoubin Ghahramani. ?The Indian Buffet Process: An Introduction and Review?. In: J. Mach. Learn. Res. 12 (2011), pp. 1185?1224.
David Blackwell and James B. MacQueen. ?Ferguson distributions via Polya urn
schemes?. In: The Annals of Statistics 1 (1973), pp. 353?355.
Bryan Klimat and Yiming Yang. ?The Enron Corpus: A New Dataset for Email Classification Research?. In: Proc. ECML. 2004, pp. 217?226.
MovieLens
dataset,
http://www.grouplens.org/.
as
of
2003.
url:
http://www.grouplens.org/.
Jesse Davis and Mark Goadrich. ?The Relationship Between Precision-Recall and
ROC Curves?. In: Proc. ICML. 2006, pp. 233?240.
Mingyuan Zhou, Yulai Cong, and Bo Chen. ?The Poisson Gamma Belief Network?.
In: Proc. NIPS. 2015, pp. 3043?3051.
Changwei Hu, Piyush Rai, and Lawrence Carin. ?Zero-Truncated Poisson Tensor Factorization for Massive Binary Tensors?. In: Proc. UAI. 2015, pp. 375?384.
9
| 6643 |@word version:2 stronger:1 c0:20 hu:1 takuya:1 liu:1 score:1 hereafter:1 com:1 analysed:2 written:1 additive:2 partition:9 j1:1 shape:1 designed:2 update:2 generative:13 discovering:2 selected:1 parameterization:1 geyer:1 yamada:1 colored:1 blei:2 org:3 five:2 c2:4 constructed:1 beta:2 issei:1 introduce:1 manner:6 theoretically:1 expected:3 globally:1 palla:1 automatically:4 actual:1 provided:1 project:1 underlying:2 factorized:1 mass:1 finding:3 unobserved:1 guarantee:1 assoc:3 partitioning:1 exchangeable:1 overestimate:4 positive:3 tends:1 mach:3 solely:1 black:1 might:1 factorization:9 averaged:3 unique:1 block:1 area:1 significantly:3 griffith:3 zoubin:3 undesirable:2 operator:1 collapsed:8 yee:1 www:2 equivalent:2 customer:3 missing:1 jesse:1 straightforward:2 truncating:2 goadrich:1 enabled:1 annals:2 construction:14 massive:1 user:3 smyth:1 hypothesis:1 overlapped:1 element:1 approximated:1 observed:2 mike:1 capture:4 calculate:2 cong:1 ensures:1 overfitted:2 disturbs:2 ran:1 transforming:1 aster:1 ui:14 rup:1 trunc:1 efficiency:5 eric:1 joint:3 represented:1 derivation:1 univ:2 describe:1 newman:1 hyper:2 widely:1 larger:2 kai:1 otherwise:2 ability:11 statistic:4 analyse:2 noisy:2 superscript:1 final:1 beal:1 analytical:1 propose:4 mixing:5 pthe:1 convergence:6 disturb:2 escobar:1 converges:1 yiming:1 piyush:1 derive:9 ac:2 stat:3 polya:1 b0:1 h3:2 ibp:1 eq:10 involves:3 indicate:2 implies:1 differ:1 drawback:2 tokyo:2 attribute:1 stochastic:2 lars:1 generalization:10 probable:1 c00:4 extension:1 hold:2 considered:1 lawrence:1 matthew:1 substituting:1 major:1 a2:9 purpose:2 uniqueness:4 estimation:1 proc:10 outperformed:3 hansen:1 grouplens:2 largest:1 successfully:2 clearly:1 zhou:3 shrinkage:33 factorizes:1 derived:7 bernoulli:1 likelihood:9 indicates:2 tech:1 contrast:3 baseline:1 am:3 inference:11 membership:3 ferguson:2 typically:2 i1:1 among:1 aforementioned:2 classification:1 denoted:1 constrained:6 art:3 marginal:5 construct:4 beach:1 atom:24 sampling:12 icml:2 fairness:1 theart:1 carin:1 simplex:1 modern:1 gamma:46 individual:2 detection:1 interest:1 investigate:1 truly:7 mixture:1 edge:7 naonori:1 arthur:1 re:5 e0:5 column:2 modeling:2 disadvantage:1 assignment:3 entry:11 subset:1 predicate:1 reported:1 straightforwardly:2 synthetic:4 considerably:1 st:1 thanks:4 fundamental:2 density:2 overestimated:1 michael:2 augmentation:2 aaai:1 padhraic:1 mikkel:1 japan:3 closedform:1 sec:17 b2:8 mlsp:1 depends:5 performed:8 h1:2 multiplicative:1 closed:8 tion:1 overfits:3 analyze:1 xing:1 asuncion:1 contribution:1 accuracy:9 efficiently:2 ofthe:1 bayesian:9 multiplying:1 confirmed:4 ijs:9 movielens100k:7 submitted:1 email:1 pp:18 james:1 proof:7 mi:31 sampled:2 dataset:12 recall:2 color:3 mingyuan:3 higher:1 wherein:2 specify:1 shrink:5 changwei:1 furthermore:12 crp:1 overfit:1 working:1 cran:1 horizontal:2 web:1 overlapping:11 hokkaido:1 reveal:1 indicated:6 usa:1 effect:21 concept:1 true:1 regularization:1 assigned:1 hence:1 white:2 attractive:1 undesired:1 during:1 davis:1 whye:1 pdf:1 complete:1 multiplicativity:2 l1:4 novel:6 recently:1 charles:2 superior:1 multinomial:2 empirically:1 jp:3 belong:1 discussed:1 employee:3 measurement:6 gibbs:23 automatic:1 grid:1 similarly:1 minnesota:1 f0:5 posterior:18 showed:2 termed:4 corp:1 binary:10 rep:1 joshua:1 period:1 monotonically:1 stephen:1 ii:3 multiple:2 faster:1 panasonic:2 cross:1 long:1 a1:9 prediction:11 expectation:22 poisson:12 iteration:7 limt:2 c1:5 proposal:1 appropriately:8 ztp:2 enron:3 sent:2 jordan:1 extracting:3 yang:1 unused:1 enough:1 restaurant:1 zi:6 whether:2 url:1 edoardo:1 cause:1 jj:3 deep:1 detailed:4 clear:1 takeshi:1 nonparametric:6 prepared:1 tenenbaum:1 http:3 outperform:1 exist:1 estimated:2 bryan:1 vol:2 ist:1 redundancy:2 achieving:1 drawn:1 hiroki:1 sum:4 ascertaining:1 package:1 throughout:1 c02:1 ueda:1 knowles:1 appendix:7 fold:1 sato:2 constraint:7 worked:1 software:1 speed:5 urn:1 according:7 rai:1 slightly:1 pr:4 fienberg:1 equation:1 conjugacy:2 count:6 mechanism:10 adopted:3 parametrize:1 available:1 rewritten:1 epm:79 hierarchical:3 appropriate:3 schmidt:1 buffet:3 original:5 thomas:4 denotes:8 dirichlet:12 ensure:7 remaining:2 completed:2 binomial:3 marginalized:4 ghahramani:3 chinese:1 especially:1 tensor:3 degrades:1 concentration:1 strategy:1 diagonal:6 september:1 affinity:1 dp:1 link:12 morten:1 simulated:2 topic:2 mail:3 considers:2 kemp:1 reason:3 assuming:3 hdp:1 relationship:2 equivalently:2 regulation:1 negative:14 motivates:1 perform:1 teh:1 observation:2 datasets:5 macqueen:1 finite:4 parametrizing:1 ecml:1 truncated:9 situation:1 relational:3 incorporated:2 excluding:1 arbitrary:5 community:1 intensity:20 inferred:4 rating:2 introduced:1 david:5 blackwell:1 elapsed:4 nip:4 suggested:1 including:4 sorely:1 max:1 belief:1 critical:2 advanced:1 mn:1 scheme:1 movie:3 brief:1 disappears:1 jun:1 extract:1 naive:2 pusing:1 prior:38 review:2 multiplication:2 marginalizing:3 movielens1m:7 fully:2 mixed:2 remarkable:1 validation:1 h2:2 sufficient:1 pi:1 row:1 truncation:16 transpose:1 keeping:1 taking:1 sparse:1 distributed:1 overcome:1 calculated:3 curve:3 evaluating:3 world:2 dimension:1 adopts:1 welling:1 transaction:1 countably:1 gene:1 active:12 uai:1 b1:8 corpus:1 unnecessary:1 xi:8 latent:12 why:3 table:1 learn:3 ca:1 complex:2 vj:14 aistats:1 pk:3 blockmodels:1 hyperparameters:13 fair:1 fig:10 west:1 roc:1 precision:2 ijc:1 comprises:2 third:1 theorem:17 specific:1 incorporating:7 airoldi:1 konstantina:1 chen:1 unexpected:1 expressed:5 bo:1 corresponds:2 extracted:2 viewed:4 sized:1 consequently:3 experimentally:4 specifically:2 infinite:19 movielens:3 sampler:19 lemma:1 total:4 called:1 experimental:2 internal:2 support:1 mark:1 indian:3 evaluate:2 |
6,239 | 6,644 | Pose Guided Person Image Generation
Liqian Ma1
Xu Jia2? Qianru Sun3? Bernt Schiele3 Tinne Tuytelaars2 Luc Van Gool1,4
KU-Leuven/PSI, TRACE (Toyota Res in Europe) 2 KU-Leuven/PSI, IMEC
3
Max Planck Institute for Informatics, Saarland Informatics Campus
4
ETH Zurich
{liqian.ma, xu.jia, tinne.tuytelaars, luc.vangool}@esat.kuleuven.be
{qsun, schiele}@mpi-inf.mpg.de [email protected]
1
Abstract
This paper proposes the novel Pose Guided Person Generation Network (PG2 )
that allows to synthesize person images in arbitrary poses, based on an image of
that person and a novel pose. Our generation framework PG2 utilizes the pose
information explicitly and consists of two key stages: pose integration and image
refinement. In the first stage the condition image and the target pose are fed into a
U-Net-like network to generate an initial but coarse image of the person with the
target pose. The second stage then refines the initial and blurry result by training a
U-Net-like generator in an adversarial way. Extensive experimental results on both
128?64 re-identification images and 256?256 fashion photos show that our model
generates high-quality person images with convincing details.
1
Introduction
Generating realistic-looking images is of great value for many applications such as face editing,
movie making and image retrieval based on synthesized images. Consequently, a wide range of
methods have been proposed including Variational Autoencoders (VAE) [14], Generative Adversarial
Networks (GANs) [6] and Autoregressive models (e.g., PixelRNN [30]). Recently, GAN models have
been particularly popular due to their principle ability to generate sharp images through adversarial
training. For example in [21, 5, 1], GANs are leveraged to generate faces and natural scene images
and several methods are proposed to stabilize the training process and to improve the quality of
generation.
From an application perspective, users typically have a particular intention in mind such as changing
the background, an object?s category, its color or viewpoint. The key idea of our approach is to
guide the generation process explicitly by an appropriate representation of that intention to enable
direct control over the generation process. More specifically, we propose to generate an image by
conditioning it on both a reference image and a specified pose. With a reference image as condition,
the model has sufficient information about the appearance of the desired object in advance. The
guidance given by the intended pose is both explicit and flexible. So in principle this approach can
manipulate any object to an arbitrary pose. In this work, we focus on transferring a person from a
given pose to an intended pose. There are many interesting applications derived from this task. For
example, in movie making, we can directly manipulate a character?s human body to a desired pose or,
for human pose estimation, we can generate training data for rare but important poses.
Transferring a person from one pose to another is a challenging task. A few examples can be seen in
Figure 1. It is difficult for a complete end-to-end framework to do this because it has to generate both
correct poses and detailed appearance simultaneously. Therefore, we adopt a divide-and-conquer
strategy, dividing the problem into two stages which focus on learning global human body structure
?
Equal contribution.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
Condition
image
Target
pose
Target
image (GT)
Coarse
result
Refined
result
Condition
image
(a) DeepFashion
Condition image
Target
pose
Target
image (GT)
Coarse
result
Refined
result
(b) Market-1501
Target pose sequence
Refined results
(c) Generating from a sequence of poses
Figure 1: Generated samples on DeepFashion dataset [16] (a)(c) and Market-1501 dataset [37] (b).
Our generation sequence (refined results)
and appearance details respectively similar to [35, 9, 3, 19]. At stage-I, we explore different ways to
model pose information. A variant of U-Net is employed to integrate the target pose with the person
image. It outputs a coarse generation result that captures the global structure of the human body in
the target image. A masked L1 loss is proposed to suppress the influence of background change
between condition image and target image. However, it would generate blurry result due to the use of
L1. At stage-II, a variant of Deep Convolutional GAN (DCGAN) model is used to further refine the
initial generation result. The model learns to fill in more appearance details via adversarial training
and generates sharper images. Different from the common use of GANs which directly learns to
generate an image from scratch, in this work we train a GAN to generate a difference map between
the initial generation result and the target person image. The training converges faster since it is an
easier task. Besides, we add a masked L1 loss to regularize the training of the generator such that it
will not generate an image with many artifacts. Experiments on two dataset, a low-resolution person
re-identification dataset and a high-resolution fashion photo dataset, demonstrate the effectiveness of
the proposed method.
Our contribution is three-fold. i) We propose a novel task of conditioning image generation on a
reference image and an intended pose, whose purpose is to manipulate a person in an image to an
arbitrary pose. ii) Several ways are explored to integrate pose information with a person image.
A novel mask loss is proposed to encourage the model to focus on transferring the human body
appearance instead of background information. iii) To address the challenging task of pose transfer,
we divide the problem into two stages, with the stage-I focusing on global structure of the human
body and the stage-II on filling in appearance details based on the first stage result.
2
Related works
Recently there have been a lot of works on generative image modeling with deep learning techniques.
These works fall into two categories. The first line of works follow an unsupervised setting. One
popular method under this setting is variational autoencoders proposed by Kingma and Welling [14]
and Rezende et al. [25], which apply a re-parameterization trick to maximize the lower bound of the
data likelihood. Another branch of methods are autogressive models [28, 30, 29] which compute the
product of conditional distributions of pixels in a pixel-by-pixel manner as the joint distribution of
pixels in an image. The most popular methods are generative adversarial networks (GAN) [6], which
simultaneously learn a generator to generate samples and a discriminator to discriminate generated
samples from real ones. Many works show that GANs can generate sharp images because of using
2
concat
coarse
result
skip connection
...
Condition
image
Target
pose
...
Target
image
Generator at Stage-I (G1)
ResUAE (G1) at Stage-I
Condition
image
Real
or
Fake
Real pair
skip connection
difference
map
Discriminator (D)
at Stage-II
concat
refined
result
Condition
image
Condition
image
Fake pair
Generator at Stage-II (G2)
Figure 2: The overall framework of our Pose Guided Person Generation Network (PG2 ). It contains
two stages. Stage-I focuses on pose integration and generates an initial result that captures the global
structure of the human. Stage-II focuses on refining the initial result via adversarial training and
generates sharper images.
the adversarial loss instead of L1 loss. In this work, we also use the adversarial loss in our framework
in order to generate high-frequency details in images.
The second group of works generate images conditioned on either category or attribute labels, texts or
images. Yan et al. [32] proposed a Conditional Variational Autoencoder (CVAE) to achieve attribute
conditioned image generation. Mirza and Osindero [18] proposed to condition both generator
and discriminator of GAN on side information to perform category conditioned image generation.
Lassner et al. [15] generated full-body people in clothing, by conditioning on the fine-grained body
part segments. Reed et al. proposed to generate bird image conditioned on text descriptions by adding
textual information to both generator and discriminator [24] and further explored the use of additional
location, keypoints or segmentation information to generate images [22, 23]. With only these visual
cues as condition and in contrast to our explicit condition on the intended pose, the control exerted
over the image generation process is still abstract. Several works further conditioned image generation
not only on labels and texts but also on images. Researchers [34, 33, 11, 8] addressed the task of face
image generation conditioned on a reference image and a specific face viewpoint. Chen et al. [4]
tackled the unseen view inference as a tensor completion problem, and use latent factors to impute
the pose in unseen views. Zhao et al. [36] explored generating multi-view cloth images from only a
single view input, which is most similar to our task. However, a wide range of poses is consistent
with any given viewpoint making the conditioning less expressive than in our work. In this work, we
make use of pose information in a more explicit and flexible way, that is, using poses in the format
of keypoints to model diverse human body appearance. It should be noted that instead of doing
expensive pose annotation, we use a state-of-the-art pose estimation approach to obtain the desired
human body keypoints.
3
Method
Our task is to simultaneously transfer the appearance of a person from a given pose to a desired pose
and keep important appearance details of the identity. As it is challenging to implement this as an
end-to-end model, we propose a two-stage approach to address this task, with each stage focusing on
one aspect. For the first stage we propose and analyze several model variants and for the second stage
we use a variant of a conditional DCGAN to fill in more appearance details. The overall framework
of the proposed Pose Guided Person Generation Network (PG2 ) is shown in Figure 2.
3.1
Stage-I: Pose integration
At stage-I, we integrate a conditioning person image IA with a target pose PB to generate a coarse
result I?B that captures the global structure of the human body in the target image IB .
3
FConv-ResUAE (G2) at Stage-II
Pose embedding. To avoid expensive annotation of poses, we apply a state-of-the-art pose estimator [2] to obtain approximate human body poses. The pose estimator generates the coordinates of
18 keypoints. Using those directly as input to our model would require the model to learn to map
each keypoint to a position on the human body. Therefore, we encode pose PB as 18 heatmaps. Each
heatmap is filled with 1 in a radius of 4 pixels around the corresponding keypoints and 0 elsewhere
(see Figure 3, target pose). We concatenate IA and PB as input to our model. In this way, we can
directly use convolutional layers to integrate the two kinds of information.
Generator G1. As generator at stage I, we adopt a U-Net-like architecture [20], i.e., convolutional
autoencoder with skip connections as is shown in Figure 2. Specifically, we first use several stacked
convolutional layers to integrate IA and PB from small local neighborhoods to larger ones so
that appearance information can be integrated and transferred to neighboring body parts. Then, a
fully connected layer is used such that information between distant body parts can also exchange
information. After that, the decoder is composed of a set of stacked convolutional layers which are
symmetric to the encoder to generate an image. The result of the first stage is denoted as I?B1 . In the
U-Net, skip connections between encoder and decoder help propagate image information directly
from input to output. In addition, we find that using residual blocks as basic component improves the
generation performance. In particular we propose to simplify the original residual block [7] and have
only two consecutive conv-relu inside a residual block.
Pose mask loss. To compare the generation I?B1
with the target image IB , we adopt L1 distance
as the generation loss of stage-I. However, since
we only have a condition image and a target
pose as input, it is difficult for the model to
generate what the background would look like if
the target image has a different background from
the condition image. Thus, in order to alleviate
the influence of background changes, we add
another term that adds a pose mask MB to the
L1 loss such that the human body is given more
weight than the background. The formulation of
pose mask loss is given in Eq. 1 with denoting
the pixels-wise multiplication:
Target
image
pose keypoints
estimation
Target
pose
Pose
skeleton
pose keypoints
connection
Pose
mask
morphological
operations
Figure 3: Process of computing the pose mask.
LG1 = k(G1(IA , PB ) ? IB ) (1 + MB )k1 .
(1)
The pose mask MB is set to 1 for foreground and 0 for background and is computed by connecting
human body parts and applying a set of morphological operations such that it is able to approximately
cover the whole human body in the target image, see the example in Figure 3.
The output of G1 is blurry because the L1 loss encourages the result to be an average of all possible
cases [10]. However, G1 does capture the global structural information specified by the target pose,
as shown in Figure 2, as well as other low-frequency information such as the color of clothes. Details
of body appearance, i.e. the high-frequency information, will be refined at the second stage through
adversarial training.
3.2
Stage-II: Image refinement
Since the model at the first stage has already synthesized an image which is coarse but close to
the target image in pose and basic color, at the second stage, we would like the model to focus on
generating more details by correcting what is wrong or missing in the initial result. We use a variant
of conditional DCGAN [21] as our base model and condition it on the stage-I generation result.
Generator G2. Considering that the initial result and the target image are already structurally similar,
we propose that the generator G2 at the second stage aims to generate an appearance difference map
that brings the initial result closer to the target image. The difference map is computed using a U-Net
similar to the first stage but with the initial result I?B1 and condition image IA as input instead. The
difference lies in that the fully-connected layer is removed from the U-Net. This helps to preserve
more details from the input because a fully-connected layer compresses a lot of information contained
in the input. The use of difference maps speeds up the convergence of model training since the model
focuses on learning the missing appearance details instead of synthesizing the target image from
4
scratch. In particular, the training already starts from a reasonable result. The overall architecture of
G2 can be seen in Figure 2.
Discriminator D. In traditional GANs, the discriminator distinguishes between real groundtruth
images and fake generated images (which is generated from random noise). However, in our
conditional network, G2 takes the condition image IA instead of a random noise as input. Therefore,
real images are the ones which not only are natural but also satisfy a specific requirement. Otherwise,
G2 will be mislead to directly output IA which is natural by itself instead of refining the coarse result
of the first stage I?B1 . To address this issue, we pair the G2 output with the condition image to make
the discriminator D to recognize the pairs? fakery, i.e., (I?B2 , IA ) vs (IB , IA ). This is diagrammed in
Figure 2. The pairwise input encourages D to learn the distinction between I?B2 and IB instead of
only the distinction between synthesized and natural images.
Another difference from traditional GANs is that noise is not necessary anymore since the generator is
conditioned on an image IA , which is similar to [17]. Therefore, we have the following loss function
for the discriminator D and the generator G2 respectively,
LD
= Lbce (D(IA , IB ), 1) + Lbce (D(IA , G2(IA , I?B1 )), 0),
(2)
adv
LG
= Lbce (D(IA , G2(IA , I?B1 )), 1),
(3)
adv
where Lbce denotes binary cross-entropy loss. Previous work [10, 17] shows that mixing the adversarial loss with a loss minimizing Lp distance can regularize the image generation process. Here
we use the same masked L1 loss as is used at the first stage such that it pays more attention to the
appearance of targeted human body than background,
LG2 = LG + ?k(G2(IA , I?B1 ) ? IB ) (1 + MB )k1 ,
(4)
adv
where ? is the weight of L1 loss. It controls how close the generation looks like the target image at
low frequencies. When ? is small, the adversarial loss dominates the training and it is more likely
to generate artifacts; when ? is big, the the generator with a basic L1 loss dominates the training,
making the whole model generate blurry results2 .
In the training process of our DCGAN, we alternatively optimize discriminator D and generator G2.
As shown in the left part of Figure 2, generator G2 takes the first stage result and the condition image
as input and aims to refine the image to confuse the discriminator. The discriminator learns to classify
the pair of condition image and the generated image as fake while classifying the pair including the
target image as real.
3.3
Network architecture
We summarize the network architecture of the proposed model PG2 . At stage-I, the encoder of G1
consists of N residual blocks and one fully-connected layer , where N depends on the size of input.
Each residual block consists of two convolution layers with stride=1 followed by one sub-sampling
convolution layer with stride=2 except the last block. At stage-II, the encoder of G2 has a fully
convolutional architecture including N -2 convolution blocks. Each block consists of two convolution
layers with stride=1 and one sub-sampling convolution layer with stride=2. Decoders in both G1
and G2 are symmetric to corresponding encoders. Besides, there are shortcut connections between
decoders and encoders, which can be seen in Figure 2. In G1 and G2, no batch normalization or
dropout are applied. All convolution layers consist of 3?3 filters and the number of filters are
increased linearly with each block. We apply rectified linear unit (ReLU) to each layer except the
fully connected layer and the output convolution layer. For the discriminator, we adopt the same
network architecture as DCGAN [21] except the size of the input convolution layer due to different
image resolutions.
4
Experiments
We evaluate the proposed PG2 network on two person datasets (Market-1501 [37] and DeepFashion [16]), which contain person images with diverse poses. We present quantitative and qualitative
results for three main aspects of PG2 : different pose embeddings; pose mask loss vs. standard L1
loss; and two-stage model vs. one-stage model. We also compare with the most related work [36].
2
The influence of ? on generation quality is analyzed in supplementary materials.
5
4.1
Datasets
The DeepFashion (In-shop Clothes Retrieval Benchmark) dataset [16] consists of 52,712 in-shop
clothes images, and 200,000 cross-pose/scale pairs. All images are in high-resolution of 256?256.
In the train set, we have 146,680 pairs each of which is composed of two images of the same person
but different poses. We randomly select 12,800 pairs from the test set for testing.
We also experiment on a more challenging re-identification dataset Market-1501 [37] containing
32,668 images of 1,501 persons captured from six disjoint surveillance cameras. Persons in this
dataset vary in pose, illumination, viewpoint and background, which makes the person generation
task more challenging. All images have size 128?64 and are split into train/test sets of 12,936/19,732
following [37]. In the train set, we have 439,420 pairs each of which is composed of two images of
the same person but different poses. We randomly select 12,800 pairs from the test set for testing.
Implementation details On both datasets, we use the Adam [13] optimizer with ?1 = 0.5 and
?2 = 0.999. The initial learning rate is set to 2e-5. On DeepFashion, we set the number of
convolution blocks N = 6. Models are trained with a minibatch of size 8 for 30k and 20k iterations
respectively at stage-I and stage-II. On Market-1501, we set the number of convolution blocks N = 5.
Models are trained with a minibatch of size 16 for 22k and 14k iterations respectively at stage-I and
stage-II. For data augmentation, we do left-right flip for both datasets3 .
4.2
Qualitative results
As mentioned above, we investigate three aspects of our proposed PG2 network. Different pose
embeddings and losses are compared within stage-I and then we demonstrate the advantage of our
two-stage model over a one-stage model.
Different pose embeddings. To evaluate our proposed pose embedding method, we implement two
alternative methods. For the first, coordinate embedding (CE), we pass the keypoint coordinates
through two fully connected layers and concatenate the embedded feature vector with the image
embedding vector at the bottleneck fully connected layer. For the second, called heatmap embedding
(HME), we feed the 18 keypoint heatmaps to an independent encoder and extract the fully connected
layer feature to concatenate with image embedding vector at the bottleneck fully connected layer.
Columns 4, 5 and 6 of Figure 4 show qualitative results of the different pose embedding methods
when used in stage-I, that is of G1 with CE (G1-CE-L1), with HME (G1-HME-L1) and our G1
(G1-L1). All three use standard L1 loss. We can see that G1-L1 is able to synthesize reasonable
looking images that capture the global structure of a person, such as pose and color. However, the
other two embedding methods G1-CE-L1 and G1-HME-L1 are quite blurry and the color is wrong.
Moreover, results of G1-CE-L1 all get wrong poses. This can be explained by the additional difficulty
to map the keypoint coordinates to appropriate image locations making training more challenging.
Our proposed pose embedding using 18 channels of pose heatmaps is able to guide the generation
process effectively, leading to correctly generated poses. Interestingly, G1-L1 can even generate
reasonable face details like eyes and mouth, as shown by the DeepFashion samples.
Pose mask loss vs. L1 loss. Comparing the results of G1 trained with L1 loss (G1-L1) and G1
trained with poseMaskLoss (G1-poseMaskLoss) for the Market-1501 dataset, we find that pose mask
loss indeed brings improvement to the performance (columns 6 and 7 in Figure 4). By focusing the
image generation on the human body, the synthesized image gets sharper and the color looks nicer.
We can see that for person ID 164, the person?s upper body generated by G1-L1 is more noisy in
color than the one generated by G1-poseMaskLoss. For person ID 23 and 346, the method with pose
mask loss generates more clear boundaries for shoulder and head. These comparisons validate that
our pose mask loss effectively alleviates the influence of noisy backgrounds and guides the generator
to focus on the pose transfer of the human body. The two losses generate similar results for the
DeepFashion samples because the background is much simpler.
Two-stage vs. one-stage. In addition, we demonstrate the advantage of our two-stage model over
a one-stage model. For this we use G1 as generator but train it in an adversarial way to directly
generate a new image given a condition image and a target pose as input. This one-stage model is
denoted as G1+D and our full model is denoted as G1+G2+D. From Figure 4, we can see that our full
model is able to generate photo-realistic results, which contain more details than the one-stage model.
3
More details about parameters of the network architecture are given in supplementary materials.
6
For example, for DeepFashion samples, more details in the face and the clothes are transferred to the
generated images. For person ID 245, the shorts on the result of G1+D have lighter color and more
blurry boundary than G1+G2+D. For person ID 346, the two-stage model is able to generate both
the right color and textures for the clothes, while the one-stage model is only able to generate the
right color. On Market-1501 samples, the quality of the images generated by both methods decreases
because of the more challenging setting. However, the two-stage model is still able to generate
better results than the one-stage method. We can see that for person ID 53, the stripes on the T-shirt
are retained by our full model while the one-stage model can only generate a blue blob as clothes.
Besides, we can also clearly see the stool in the woman?s hands (person ID 23).
1
2
3
Condition image Target pose Target image(GT)
4
5
6
G1-CE-L1
G1-HME-L1
G1-L1
7
G1-poseMaskLoss
(our coarse result)
8
9
G1+D
G1+G2+D
(our refined result)
ID. 245
ID. 346
ID. 116
ID. 53
ID. 164
ID. 23
Figure 4: Test results on DeepFashion (upper 3 rows, images are cut for the sake of display) and
Market-1501 dataset (lower 3 rows). We test G1 in two aspects: (1) three pose embedding methods,
i.e., coordinate embedding (CE), heatmap embedding (HME) and our pose heatmap concatenation
in G1-L1, and (2) two losses, i.e., the proposed poseMaskLoss and the standard L1 loss. Column
7, 8 and 9 show the differences among our stage-I (G1), one-stage adversarial model (G1+D) and
our two-stage adversarial model (G1+G2+D). Note that all three use poseMaskLoss. The IDs are
assigned randomly when splitting the datasets.
4.3
Quantitative results
We also give quantitative results on both datasets. Structural Similarity (SSIM) [31] and the Inception
Score (IS) [26] are adopted to measure the quality of synthesis. Note that in the Market-1501 dataset,
condition images and target images may have different background. Since there is no information
in the input about the background in the target image, our method is not able to imagine what the
7
Table 1: Quantitative evaluation. For all measures, higher is better.
DeepFashion
Market-1501
Model
SSIM
IS
SSIM
IS
mask-SSIM
mask-IS
G1-CE-L1
G1-HME-L1
G1-L1
G1-poseMaskLoss
G1+D
G1+G2+D
0.694
0.735
0.735
0.779
0.761
0.762
2.395
2.427
2.427
2.668
3.091
3.090
0.219
0.294
0.304
0.340
0.283
0.253
2.568
3.171
3.006
3.326
3.490
3.460
0.771
0.802
0.809
0.817
0.803
0.792
2.455
2.508
2.455
2.682
3.310
3.435
Table 2: User study results from AMT
DeepFashion
4
Market-1501
5
Model
R2G
G2R
R2G
G2R
G1+D
G1+G2+D
7.8%
9.2%
9.3%
14.9%
17.1%
11.2%
11.1%
5.5%
new background looks like. To reduce the influence of background in our evaluation, we propose
a variant of SSIM, called mask-SSIM. A pose mask is added to both the synthesis and the target
image before computing SSIM. In this way we only focus on measuring the synthesis quality of a
person?s appearance. Similarly, we employ mask-IS to eliminate the effect of background. However,
it should be noted that image quality does not always correspond to such image similarity metrics.
For example, in Figure 4, our full model generates sharper and more photo-realistic results than
G1-poseMaskLoss, but the latter one has a higher SSIM. This is also observed in super-resolution
papers [12, 27].
The advantages are also clearly shown in the numerical scores in Table 1. E.g. the proposed pose
embedding (G1-L1) consistently outperforms G1-CE-L1 across all measures and both datasets. G1HME-L1 obtains similar quantitative numbers probably due to the similarity of the two embeddings.
Changing the loss from L1 to the proposed poseMaskLoss (G1-poseMaskLoss) consistently improves
further across all measures and for both datasets. Adding the discriminator during training either
after the first stage (G1+D) or in our full model (G1+G2+D) leads to comparable numbers, even
though we have observed clear differences in the qualitative results as discussed above. This is
explained by the fact that blurry images often get good SSIM despite being less convincing and
photo-realistic [12, 27].
4.4
User study
We perform a user study on Amazon Mechanical Turk (AMT) for both datasets. For each one, we
show 55 real images and 55 generated images in a random order to 30 users. Following [10, 15],
each image is shown for 1 second. The first 10 images are used for practice thus are ignored when
computing scores. From the results reported in Table. 2, we can get some observations that (1)
On DeepFashion our generated images of G1+D and G1+G2+D manage to confuse users on 9.3%
and 14.9% trials respectively (see G2R), showing the advantage of G1+G2+D over G1+D; (2) On
Market-1501, the average score of G2R is lower, because the background is much more cluttered than
DeepFashion; (3) On Market-1501, G1+G2+D gets a lower score than G1+D, because G1+G2+D
transfers more backgrounds from the condition image, which can be figured out in Figure. 4, but
in the meantime it brings extra artifacts on backgrounds which lead users to rate ?Fake?; (4) With
respect to R2G, we notice that Market-1501 gets clearly high scores (>10%) because human users
sometimes get confused when facing low-quality surveillance images.
4
5
R2G means #Real images rated as generated / #Real images
G2R means #Generated images rated as Real / #Generated images
8
Condition
image
Target image
(GT)
Ours
(refined)
VariGAN [36]
(refined)
Condition
image
Target
pose
Target
Ours
image (GT) (coarse)
Ours
(refined)
ID. 215
Figure 5: Comparison examples with [36].
4.5
Figure 6: Our failure cases on DeepFashion.
Further analysis
ID. 2662
Since our task with pose condition is novel, there is no direct comparison work. We only compare
with the most related one6 [36], which did multi-view person image synthesis on the DeepFashion
dataset. It is noted that [36] used the condition image and an additional word vector of the target
view e.g. ?side? as network input. Comparison examples are shown in Figure 5. It is clear that
our refined results are much better than those of [36]. Taking the second row as an example, we
can generate high-quality whole body images conditioned on an upper body while the whole body
synthesis by [36] only has a rough body shape.
Additionally, we give two failure DeepFashion examples by our model in Figure 6. In the top row,
only the upper body is generated consistently. The ?pieces of legs? is caused by the rare training
data for such complicated poses. The bottom row shows inaccurate gender which is caused by the
imbalance of training data for male / female. Besides, the condition person wears a long-sleeve jacket
of similar color to his inner short-sleeve, making the generated cloth look like a mixture of both.
5
Conclusions
In this work, we propose the Pose Guided Person Generation Network (PG2 ) to address a novel task
of synthesizing person images by conditioning it on a reference image and a target pose. A divideand-conquer strategy is employed to divide the generation process into two stages. Stage-I aims to
capture the global structure of a person and generate an initial result. A pose mask loss is further
proposed to alleviate the influence of the background on person image synthesis. Stage-II fills in
more appearance details via adversarial training to generate sharper images. Extensive experimental
results on two person datasets demonstrate that our method is able to generate images that are both
photo-realistic and pose-wise correct. In the future work, we plan to generate more controllable and
diverse person images conditioning on both pose and attribute.
Acknowledgments
We gratefully acknowledge the support of Toyota Motors Europe, FWO Structure from Semantics
project, KU Leuven GOA project CAMETRON, and German Research Foundation (DFG CRC 1223).
We would like to thank Bo Zhao for his helpful discussions.
References
[1] Mart?n Arjovsky, Soumith Chintala, and L?on Bottou. Wasserstein GAN. arXiv, 1701.07875, 2017.
[2] Zhe Cao, Tomas Simon, Shih-En Wei, and Yaser Sheikh. Realtime multi-person 2d pose estimation using
part affinity fields. arXiv, 1611.08050, 2016.
[3] Joao Carreira, Pulkit Agrawal, Katerina Fragkiadaki, and Jitendra Malik. Human pose estimation with iterative error feedback. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,
pages 4733?4742, 2016.
6
Results of VariGAN are provided by the authors.
9
[4] Chao-Yeh Chen and Kristen Grauman. Inferring unseen views of people. In CVPR, pages 2003?2010,
2014.
[5] Xi Chen, Xi Chen, Yan Duan, Rein Houthooft, John Schulman, Ilya Sutskever, and Pieter Abbeel. Infogan:
Interpretable representation learning by information maximizing generative adversarial nets. In NIPS,
pages 2172?2180, 2016.
[6] Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair,
Aaron C. Courville, and Yoshua Bengio. Generative adversarial nets. In NIPS, 2014.
[7] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition.
In CVPR, pages 770?778, 2016.
[8] Rui Huang, Shu Zhang, Tianyu Li, and Ran He. Beyond face rotation: Global and local perception gan for
photorealistic and identity preserving frontal view synthesis. arXiv, 1704.04086, 2017.
[9] Xun Huang, Yixuan Li, Omid Poursaeed, John Hopcroft, and Serge Belongie. Stacked generative
adversarial networks. arXiv preprint arXiv:1612.04357, 2016.
[10] Phillip Isola, Jun-Yan Zhu, Tinghui Zhou, and Alexei A. Efros. Image-to-image translation with conditional
adversarial networks. In CVPR, 2017.
[11] Xu Jia, Amir Ghodrati, Marco Pedersoli, and Tinne Tuytelaars. Towards automatic image editing: Learning
to see another you. In BMVC, 2016.
[12] Justin Johnson, Alexandre Alahi, and Li Fei-Fei. Perceptual losses for real-time style transfer and
super-resolution. In ECCV, 2016.
[13] Diederik P. Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv, 1412.6980, 2014.
[14] Diederik P. Kingma and Max Welling. Auto-encoding variational bayes. arXiv, 1312.6114, 2013.
[15] Christoph Lassner, Gerard Pons-Moll, and Peter V Gehler. A generative model of people in clothing. arXiv
preprint arXiv:1705.04098, 2017.
[16] Ziwei Liu, Ping Luo, Shi Qiu, Xiaogang Wang, and Xiaoou Tang. Deepfashion: Powering robust clothes
recognition and retrieval with rich annotations. In CVPR, pages 1096?1104, 2016.
[17] Micha?l Mathieu, Camille Couprie, and Yann LeCun. Deep multi-scale video prediction beyond mean
square error. In ICLR, 2016.
[18] Mehdi Mirza and Simon Osindero. Conditional generative adversarial nets. arXiv, 1411.1784, 2014.
[19] Alejandro Newell, Kaiyu Yang, and Jia Deng. Stacked hourglass networks for human pose estimation. In
European Conference on Computer Vision, pages 483?499. Springer, 2016.
[20] Tran Minh Quan, David G. C. Hildebrand, and Won-Ki Jeong. Fusionnet: A deep fully residual convolutional neural network for image segmentation in connectomics. arXiv, 1612.05360, 2016.
[21] Alec Radford, Luke Metz, and Soumith Chintala. Unsupervised representation learning with deep
convolutional generative adversarial networks. arXiv, 1511.06434, 2015.
[22] Scott Reed, Zeynep Akata, Santosh Mohan, Samuel Tenka, Bernt Schiele, and Honglak Lee. Learning
what and where to draw. In NIPS, 2016.
[23] Scott Reed, A?ron van den Oord, Nal Kalchbrenner, Victor Bapst, Matt Botvinick, and Nando de Freitas.
Generating interpretable images with controllable structure. Technical report, 2016.
[24] Scott E. Reed, Zeynep Akata, Xinchen Yan, Lajanugen Logeswaran, Bernt Schiele, and Honglak Lee.
Generative adversarial text to image synthesis. In ICML, 2016.
[25] Danilo Jimenez Rezende, Shakir Mohamed, and Daan Wierstra. Stochastic backpropagation and approximate inference in deep generative models. In ICML, 2014.
[26] Tim Salimans, Ian J. Goodfellow, Wojciech Zaremba, Vicki Cheung, Alec Radford, and Xi Chen. Improved
techniques for training gans. In NIPS, pages 2226?2234, 2016.
[27] Wenzhe Shi, Jose Caballero, Ferenc Huszar, Johannes Totz, Andrew P. Aitken, Rob Bishop, Daniel
Rueckert, and Zehan Wang. Real-time single image and video super-resolution using an efficient sub-pixel
convolutional neural network.
10
[28] Benigno Uria, Marc-Alexandre C?t?, Karol Gregor, Iain Murray, and Hugo Larochelle. Neural autoregressive distribution estimation. arXiv, 1605.02226, 2016.
[29] A?ron van den Oord, Nal Kalchbrenner, Lasse Espeholt, Koray Kavukcuoglu, Oriol Vinyals, and Alex
Graves. Conditional image generation with pixelcnn decoders. In NIPS, pages 4790?4798, 2016.
[30] A?ron van den Oord, Nal Kalchbrenner, and Koray Kavukcuoglu. Pixel recurrent neural networks. In
ICML, pages 1747?1756, 2016.
[31] Zhou Wang, Alan C. Bovik, Hamid R. Sheikh, and Eero P. Simoncelli. Image quality assessment: from
error visibility to structural similarity. IEEE Trans. Image Processing, 13(4):600?612, 2004.
[32] Xinchen Yan, Jimei Yang, Kihyuk Sohn, and Honglak Lee. Attribute2image: Conditional image generation
from visual attributes. In ECCV, pages 776?791, 2016.
[33] Jimei Yang, Scott Reed, Ming-Hsuan Yang, and Honglak Lee. Weakly-supervised disentangling with
recurrent transformations for 3d view synthesis. In NIPS, 2015.
[34] Junho Yim, Heechul Jung, ByungIn Yoo, Changkyu Choi, Du-Sik Park, and Junmo Kim. Rotating your
face using multi-task deep neural network. In CVPR, 2015.
[35] Han Zhang, Tao Xu, Hongsheng Li, Shaoting Zhang, Xiaolei Huang, Xiaogang Wang, and Dimitris
Metaxas. Stackgan: Text to photo-realistic image synthesis with stacked generative adversarial networks.
arXiv:, 1612.03242, 2016.
[36] Bo Zhao, Xiao Wu, Zhi-Qi Cheng, Hao Liu, and Jiashi Feng. Multi-view image generation from a
single-view. arXiv, 1704.04886, 2017.
[37] Liang Zheng, Liyue Shen, Lu Tian, Shengjin Wang, Jingdong Wang, and Qi Tian. Scalable person
re-identification: A benchmark. In ICCV, pages 1116?1124, 2015.
11
| 6644 |@word trial:1 pieter:1 propagate:1 tenka:1 jingdong:1 ld:1 initial:12 liu:2 contains:1 score:6 jimenez:1 daniel:1 denoting:1 ours:3 interestingly:1 outperforms:1 freitas:1 cvae:1 comparing:1 luo:1 diederik:2 connectomics:1 john:2 refines:1 realistic:6 concatenate:3 distant:1 numerical:1 shape:1 uria:1 motor:1 visibility:1 hourglass:1 interpretable:2 v:5 generative:12 cue:1 alec:2 parameterization:1 concat:2 amir:1 short:2 coarse:10 location:2 ron:3 simpler:1 zhang:4 saarland:1 wierstra:1 direct:2 qualitative:4 consists:5 inside:1 manner:1 pairwise:1 aitken:1 mask:18 indeed:1 market:14 mpg:1 multi:6 shirt:1 ming:1 duan:1 soumith:2 zhi:1 considering:1 conv:1 confused:1 project:2 campus:1 moreover:1 joao:1 provided:1 what:4 kind:1 clothes:7 transformation:1 quantitative:5 jimei:2 pg2:9 sleeve:2 alahi:1 zaremba:1 grauman:1 wrong:3 botvinick:1 sherjil:1 control:3 unit:1 planck:1 before:1 local:2 despite:1 encoding:1 id:15 approximately:1 bird:1 challenging:7 christoph:1 luke:1 micha:1 jacket:1 logeswaran:1 range:2 tian:2 acknowledgment:1 camera:1 lecun:1 testing:2 practice:1 block:11 implement:2 backpropagation:1 lg2:1 yan:5 eth:1 intention:2 pixelrnn:1 word:1 get:7 close:2 zehan:1 influence:6 applying:1 optimize:1 gool1:1 map:7 missing:2 maximizing:1 shi:2 attention:1 cluttered:1 jimmy:1 resolution:7 mislead:1 hsuan:1 splitting:1 amazon:1 correcting:1 tomas:1 pouget:1 shen:1 estimator:2 iain:1 fill:3 regularize:2 his:2 embedding:13 coordinate:5 target:40 fconv:1 imagine:1 user:8 xinchen:2 lighter:1 goodfellow:2 trick:1 synthesize:2 expensive:2 particularly:1 recognition:3 stripe:1 cut:1 gehler:1 observed:2 bottom:1 preprint:2 wang:6 capture:6 connected:9 adv:3 sun:1 morphological:2 decrease:1 removed:1 ran:1 mentioned:1 schiele:3 skeleton:1 warde:1 diagrammed:1 trained:4 weakly:1 ferenc:1 segment:1 tinne:3 joint:1 hopcroft:1 xiaoou:1 train:5 stacked:5 vicki:1 neighborhood:1 refined:11 rein:1 kalchbrenner:3 whose:1 bernt:3 larger:1 supplementary:2 quite:1 cvpr:5 jean:1 otherwise:1 divideand:1 encoder:5 ability:1 g1:62 tuytelaars:2 unseen:3 itself:1 noisy:2 shakir:1 sequence:3 advantage:4 blob:1 net:10 agrawal:1 propose:8 tran:1 product:1 mb:4 neighboring:1 cao:1 mixing:1 alleviates:1 achieve:1 description:1 validate:1 xun:1 sutskever:1 convergence:1 requirement:1 gerard:1 generating:5 adam:2 converges:1 karol:1 object:3 help:2 tim:1 andrew:1 completion:1 pose:101 recurrent:2 eq:1 dividing:1 skip:4 larochelle:1 tianyu:1 guided:5 radius:1 correct:2 attribute:4 filter:2 stochastic:2 human:21 nando:1 enable:1 material:2 crc:1 require:1 exchange:1 espeholt:1 abbeel:1 benigno:1 alleviate:2 kristen:1 hamid:1 heatmaps:3 clothing:2 marco:1 around:1 caballero:1 great:1 efros:1 vary:1 adopt:4 consecutive:1 optimizer:1 purpose:1 estimation:7 label:2 datasets3:1 rough:1 clearly:3 bapst:1 always:1 powering:1 aim:3 super:3 avoid:1 zhou:2 surveillance:2 vae:1 encode:1 derived:1 focus:9 rezende:2 refining:2 improvement:1 consistently:3 likelihood:1 contrast:1 adversarial:23 kim:1 helpful:1 inference:2 cloth:2 inaccurate:1 typically:1 transferring:3 lbce:4 integrated:1 eliminate:1 semantics:1 tao:1 pixel:8 overall:3 issue:1 flexible:2 among:1 denoted:3 proposes:1 heatmap:4 art:2 integration:3 plan:1 g2r:5 santosh:1 equal:1 field:1 exerted:1 beach:1 sampling:2 koray:2 park:1 look:5 unsupervised:2 filling:1 icml:3 foreground:1 future:1 mirza:3 yoshua:1 simplify:1 report:1 few:1 distinguishes:1 employ:1 randomly:3 composed:3 simultaneously:3 preserve:1 recognize:1 dfg:1 intended:4 hongsheng:1 investigate:1 alexei:1 yixuan:1 evaluation:2 zheng:1 male:1 analyzed:1 mixture:1 farley:1 encourage:1 closer:1 necessary:1 filled:1 pulkit:1 divide:3 rotating:1 re:6 desired:4 guidance:1 increased:1 classify:1 modeling:1 column:3 cover:1 measuring:1 pons:1 rare:2 junmo:1 masked:3 jiashi:1 johnson:1 osindero:2 reported:1 encoders:2 person:43 st:1 oord:3 lee:4 informatics:2 connecting:1 synthesis:10 ilya:1 gans:7 nicer:1 augmentation:1 manage:1 containing:1 leveraged:1 huang:3 woman:1 zhao:3 leading:1 style:1 wojciech:1 li:4 ma1:1 hme:7 de:2 stride:4 b2:2 stabilize:1 rueckert:1 satisfy:1 jitendra:1 explicitly:2 caused:2 depends:1 piece:1 view:11 lot:2 doing:1 analyze:1 start:1 bayes:1 metz:1 complicated:1 annotation:3 simon:2 jia:3 contribution:2 square:1 convolutional:9 correspond:1 serge:1 identification:4 metaxas:1 kavukcuoglu:2 ren:1 lu:1 researcher:1 rectified:1 ping:1 failure:2 frequency:4 turk:1 mohamed:1 chintala:2 psi:2 photorealistic:1 dataset:12 popular:3 color:11 improves:2 segmentation:2 akata:2 focusing:3 feed:1 autogressive:1 higher:2 alexandre:2 danilo:1 follow:1 totz:1 supervised:1 wei:1 bmvc:1 editing:2 formulation:1 improved:1 though:1 inception:1 stage:67 autoencoders:2 hand:1 expressive:1 mehdi:2 assessment:1 minibatch:2 brings:3 artifact:3 quality:10 usa:1 effect:1 phillip:1 contain:2 matt:1 assigned:1 symmetric:2 changkyu:1 impute:1 during:1 encourages:2 noted:3 mpi:1 won:1 samuel:1 complete:1 demonstrate:4 l1:36 goa:1 image:152 variational:4 wise:2 novel:6 recently:2 common:1 rotation:1 hugo:1 conditioning:7 discussed:1 he:2 synthesized:4 honglak:4 stool:1 leuven:3 automatic:1 similarly:1 gratefully:1 wear:1 pixelcnn:1 europe:2 similarity:4 han:1 alejandro:1 gt:5 add:3 base:1 female:1 perspective:1 inf:1 poursaeed:1 binary:1 victor:1 seen:3 captured:1 additional:3 arjovsky:1 wasserstein:1 preserving:1 employed:2 isola:1 deng:1 xiangyu:1 maximize:1 ii:12 branch:1 full:6 simoncelli:1 keypoints:7 alan:1 technical:1 faster:1 cross:2 long:2 retrieval:3 manipulate:3 qi:2 prediction:1 variant:6 basic:3 scalable:1 vision:3 metric:1 arxiv:15 iteration:2 normalization:1 sometimes:1 background:21 addition:2 fine:1 addressed:1 imec:1 jian:1 bovik:1 extra:1 lajanugen:1 probably:1 quan:1 effectiveness:1 ee:1 structural:3 yang:4 iii:1 embeddings:4 split:1 results2:1 bengio:1 relu:2 moll:1 architecture:7 reduce:1 idea:1 inner:1 bottleneck:2 six:1 yaser:1 peter:1 shaoqing:1 deep:8 ignored:1 fake:5 detailed:1 clear:3 fragkiadaki:1 johannes:1 fwo:1 sohn:1 category:4 generate:35 notice:1 disjoint:1 correctly:1 blue:1 diverse:3 group:1 key:2 shih:1 stackgan:1 pb:5 changing:2 figured:1 ce:9 nal:3 vangool:2 houthooft:1 jose:1 you:1 reasonable:3 groundtruth:1 yann:1 wu:1 utilizes:1 realtime:1 draw:1 comparable:1 huszar:1 dropout:1 bound:1 layer:20 pay:1 followed:1 ki:1 tackled:1 display:1 fold:1 courville:1 cheng:1 refine:2 xiaogang:2 fei:2 alex:1 scene:1 your:1 sake:1 generates:7 aspect:4 speed:1 format:1 transferred:2 across:2 character:1 lp:1 sheikh:2 rob:1 making:6 leg:1 explained:2 den:3 iccv:1 lg1:1 zurich:1 bing:1 german:1 mind:1 flip:1 kuleuven:1 fed:1 end:4 photo:7 adopted:1 operation:2 apply:3 salimans:1 appropriate:2 blurry:7 yim:1 anymore:1 batch:1 alternative:1 original:1 compress:1 denotes:1 top:1 gan:7 k1:2 murray:1 conquer:2 junho:1 gregor:1 feng:1 tensor:1 malik:1 already:3 added:1 strategy:2 traditional:2 affinity:1 iclr:1 sun3:1 distance:2 thank:1 concatenation:1 decoder:5 zeynep:2 ozair:1 besides:4 byungin:1 retained:1 reed:5 convincing:2 minimizing:1 liang:1 difficult:2 lg:2 disentangling:1 sharper:5 shu:1 trace:1 hao:1 synthesizing:2 suppress:1 ba:1 lassner:2 implementation:1 ziwei:1 perform:2 upper:4 ssim:9 convolution:10 observation:1 datasets:9 imbalance:1 benchmark:2 acknowledge:1 minh:1 daan:1 looking:2 shoulder:1 head:1 arbitrary:3 sharp:2 camille:1 david:2 pair:11 mechanical:1 specified:2 extensive:2 connection:6 discriminator:13 pedersoli:1 jeong:1 distinction:2 textual:1 kingma:3 esat:1 nip:7 address:4 able:9 beyond:2 justin:1 trans:1 pattern:1 perception:1 scott:4 dimitris:1 summarize:1 omid:1 max:2 including:3 video:2 mouth:1 ia:16 natural:4 difficulty:1 meantime:1 residual:7 zhu:1 shop:2 improve:1 movie:2 rated:2 keypoint:4 eye:1 mathieu:1 jun:1 autoencoder:2 extract:1 auto:1 text:5 chao:1 yeh:1 schulman:1 multiplication:1 graf:1 embedded:1 loss:35 fully:11 generation:33 interesting:1 facing:1 generator:18 foundation:1 integrate:5 sufficient:1 consistent:1 heechul:1 principle:2 viewpoint:4 sik:1 xiao:1 classifying:1 translation:1 row:5 eccv:2 elsewhere:1 jung:1 last:1 guide:3 side:2 institute:1 wide:2 fall:1 face:8 taking:1 van:4 boundary:2 feedback:1 rich:1 autoregressive:2 hildebrand:1 author:1 refinement:2 attribute2image:1 welling:2 approximate:2 obtains:1 keep:1 global:9 b1:7 belongie:1 eero:1 xi:3 alternatively:1 zhe:1 latent:1 iterative:1 table:4 additionally:1 ku:3 transfer:5 learn:3 ca:1 channel:1 controllable:2 robust:1 du:1 bottou:1 european:1 marc:1 did:1 main:1 linearly:1 whole:4 noise:3 big:1 wenzhe:1 qiu:1 xu:5 body:27 lasse:1 en:1 fashion:2 structurally:1 position:1 sub:3 explicit:3 inferring:1 lie:1 perceptual:1 ib:7 infogan:1 toyota:2 learns:3 grained:1 ian:2 tang:1 choi:1 specific:2 bishop:1 showing:1 shengjin:1 explored:3 abadie:1 dominates:2 consist:1 adding:2 effectively:2 texture:1 mohan:1 illumination:1 conditioned:8 confuse:2 rui:1 chen:5 easier:1 entropy:1 appearance:17 explore:1 likely:1 visual:2 vinyals:1 dcgan:5 contained:1 kaiming:1 g2:28 bo:2 springer:1 ch:1 gender:1 newell:1 radford:2 amt:2 ma:1 mart:1 tinghui:1 conditional:9 identity:2 targeted:1 cheung:1 consequently:1 towards:1 couprie:1 luc:2 shortcut:1 change:2 carreira:1 specifically:2 except:3 kaiyu:1 called:2 discriminate:1 pas:1 experimental:2 katerina:1 aaron:1 select:2 people:3 support:1 latter:1 kihyuk:1 ethz:1 frontal:1 oriol:1 evaluate:2 yoo:1 scratch:2 |
6,240 | 6,645 | Inference in Graphical Models
via Semidefinite Programming Hierarchies
Murat A. Erdogdu
Microsoft Research
[email protected]
Yash Deshpande
MIT and Microsoft Research
[email protected]
Andrea Montanari
Stanford University
[email protected]
Abstract
Maximum A posteriori Probability (MAP) inference in graphical models amounts
to solving a graph-structured combinatorial optimization problem. Popular inference algorithms such as belief propagation (BP) and generalized belief propagation
(GBP) are intimately related to linear programming (LP) relaxation within the
Sherali-Adams hierarchy. Despite the popularity of these algorithms, it is well
understood that the Sum-of-Squares (SOS) hierarchy based on semidefinite programming (SDP) can provide superior guarantees. Unfortunately, SOS relaxations
for a graph with n vertices require solving an SDP with n?(d) variables where
d is the degree in the hierarchy. In practice, for d
4, this approach does not
scale beyond a few tens of variables. In this paper, we propose binary SDP relaxations for MAP inference using the SOS hierarchy with two innovations focused
on computational efficiency. Firstly, in analogy to BP and its variants, we only
introduce decision variables corresponding to contiguous regions in the graphical
model. Secondly, we solve the resulting SDP using a non-convex Burer-Monteiro
style method, and develop a sequential rounding procedure. We demonstrate that
the resulting algorithm can solve problems with tens of thousands of variables
within minutes, and outperforms BP and GBP on practical problems such as image
denoising and Ising spin glasses. Finally, for specific graph types, we establish a
sufficient condition for the tightness of the proposed partial SOS relaxation.
1
Introduction
Graphical models provide a powerful framework for analyzing systems comprised by a large number
of interacting variables. Inference in graphical models is crucial in scientific methodology with
countless applications in a variety of fields including causal inference, computer vision, statistical
physics, information theory, and genome research [WJ08, KF09, MM09].
In this paper, we propose a class of inference algorithms for pairwise undirected graphical models.
Such models are fully specified by assigning: (i) a finite domain X for the variables; (ii) a finite
graph G = (V, E) for V = [n] ? {1, . . . , n} capturing the interactions of the basic variables;
e
(iii) a collection of functions ? = ({?iv }i2V , {?ij
}(i,j)2E ) that quantify the vertex potentials and
interactions between the variables; whereby for each vertex i 2 V we have ?iv : X ! R and for each
e
edge (i, j) 2 E, we have ?ij
: X ? X ! R (an arbitrary ordering is fixed on the pair of vertices
{i, j}). These parameters can be used to form a probability distribution on X V for the random vector
x = (x1 , x2 , ..., xn ) 2 X V by letting,
X
X
1
e
p(x|?) =
eU (x;?) ,
U (x; ?) =
?ij
(xi , xj ) +
?iv (xi ) ,
(1.1)
Z(?)
(i,j)2E
i2V
where Z(?) is the normalization constant commonly referred to as the partition function. While such
models can encode a rich class of multivariate probability distributions, basic inference tasks are
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
intractable except for very special graph structures such as trees or small treewidth graphs [CD+ 06].
In this paper, we will focus on MAP estimation, which amounts to solving the combinatorial
optimization problem
? (?) ? arg max U (x; ?).
x
(1.2)
x2X V
Intractability plagues other classes of graphical models as well (e.g. Bayesian networks, factor
graphs), and has motivated the development of a wide array of heuristics. One of the simplest such
heuristics is the loopy belief propagation (BP) [WJ08, KF09, MM09]. In its max-product version
(that is well-suited for MAP estimation), BP is intimately related to the linear programming (LP)
relaxation of the combinatorial problem maxx2X V U (x; ?). Denoting the decision variables by
b = ({bi }i2V , {bij }(i,j)2E ), LP relaxation form of BP can be written as
X
X
X X
maximize
?ij (xi , xj )bij (xi , xj ) +
?i (xi )bi (xi ) ,
(1.3)
b
subject to
(i,j)2E xi ,xj 2X
X
i2V xi 2X
bij (xi , xj ) = bi (xi )
xj 2X
8(i, j) 2 E ,
(1.4)
bi 2 X 8i 2 V, bij 2 X ?X 8(i, j) 2 E ,
(1.5)
where S denotes the simplex of probability distributions over set S. The decision variables are
referred to as ?beliefs?, and their feasible set is a relaxation of the polytope of marginals of distributions.
The beliefs satisfy the constraints on marginals involving at most two variables connected by an edge.
Loopy belief propagation is successful on some applications, e.g. sparse locally tree-like graphs
that arise, for instance, decoding modern error correcting codes [RU08] or in random constraint
satisfaction problems [MM09]. However, in more structured instances ? arising for example in
computer vision ? BP can be substantially improved by accounting for local dependencies within
subsets of more than two variables. This is achieved by generalized belief propagation (GBP)
[YFW05] where the decision variables are beliefs bR that are defined on subsets of vertices (a
?region?) R ? [n], and that represent the marginal distributions of the variables
in that region. The
P
basic constraint on the beliefs is the linear marginalization constraint: xR\S bR (xR ) = bS (xS ),
holding whenever S ? R. Hence GBP itself is closely related to LP relaxation of the polytope
of marginals of probability distributions. The relaxation becomes tighter as larger regions are
incorporated. In a prototypical application, G is a two-dimensional grid, and regions are squares
induced by four contiguous vertices (plaquettes), see Figure 1, left frame. Alternatively in the right
frame of the same figure, the regions correspond to triangles.
The LP relaxations that correspond to GBP are closely related to the Sherali-Adams hierarchy
[SA90]. Similar to GBP, the variables within this hierarchy are beliefs
Pover subsets of variables
bR = (bR (xR ))xR 2X R which are consistent under marginalization:
xR\S bR (xR ) = bS (xS ).
However, these two approaches differ in an important point: Sherali-Adams hierarchy uses beliefs
over all subsets of |R| ? d variables, where d is the degree in the hierarchy; this leads to an LP of size
?(nd ). In contrast, GBP only retains regions that are contiguous in G. If G has maximum degree k,
this produces an LP of size O(nk d ), a reduction which is significant for large-scale problems.
Given the broad empirical success of GBP, it is natural to develop better methods for inference in
graphical models using tighter convex relaxations. Within combinatorial optimization, it is well
understood that the semidefinite programming (SDP) relaxations provide superior approximation
guarantees with respect to LP [GW95]. Nevertheless, SDP has found limited applications in inference
tasks for graphical models for at least two reasons. A structural reason: standard SDP relaxations (e.g.
[GW95]) do not account exactly for correlations between neighboring vertices in the graph which is
essential for structured graphical models. As a consequence, BP or GBP often outperforms basic
SDPs. A computational reason: basic SDP relaxations involve ?(n2 ) decision variables, and generic
interior point solvers do not scale well for the large-scale applications. An exception is [WJ04] which
employs the simplest SDP relaxation (degree 2 Sum-Of-Squares, see below) in conjunction with a
relaxation of the entropy and interior point methods ? higher order relaxations are briefly discussed
without implementation as the resulting program suffers from the aforementioned limitations.
In this paper, we revisit MAP inference in graphical models via SDPs, and propose an approach that
carries over the favorable performance guarantees of SDPs into inference tasks. For simplicity, we
focus on models with binary variables, but we believe that many of the ideas developed here can be
naturally extended to other finite domains. We present the following contributions:
2
Region 1
on
gi
Re
Region 2
1
on
gi
Re
2
on
gi
Re
3
on
gi
Re
4
Figure 1: A two dimensional grid, and two typical choices for regions for GBP and PSOS.
Left: Regions are plaquettes comprising four vertices. Right: Regions are triangles.
Partial Sum-Of-Squares relaxations. We use SDP hierarchies, specifically the Sum-Of-Squares
(SOS) hierarchy [Sho87, Las01, Par03] to formulate tighter SDP relaxations for binary MAP inference
that account exactly for the joint distributions of small subsets of variables xR , for R ? V . However,
SOS introduces decision variables for all subsets R ? V with |R| ? d/2 (d is a fixed even integer),
and hence scales poorly for large-scale inference problems. We propose a similar modification as in
GBP. Instead of accounting for all subsets R with |R| ? d/2, we only introduce decision variables to
represent a certain family of such subsets (regions) of vertices in G. The resulting SDP has (for d and
the maximum degree of G bounded) only O(n2 ) decision variables which is suitable for practical
implementations. We refer to these relaxations as Partial Sum-Of-Squares (PSOS), cf. Section 2.
Theoretical analysis. In Section 2.1, we prove that suitable PSOS relaxations are tight for certain
classes of graphs, including planar graphs, with ?v = 0. While this falls short of explaining the
empirical results (which uses simpler relaxations, and ?v 6= 0), it points in the right direction.
Optimization algorithm and rounding. Despite the simplification afforded by PSOS, interior-point
solvers still scale poorly to large instances. In order to overcome this problem, we adopt a non-convex
approach proposed by Burer and Monteiro [BM03]. We constrain the rank of the SDP matrix in
PSOS to be at most r, and solve the resulting non-convex problem using a trust-region coordinate
ascent method, cf. Section 3.1. Further, we develop a rounding procedure called Confidence Lift and
Project (CLAP) which iteratively uses PSOS relaxations to obtain an integer solution, cf. Section 3.2.
Numerical experiments. In Section 4, we present numerical experiments with PSOS by solving
problems of size up to 10, 000 within several minutes. While additional work is required to scale
this approach to massive sizes, we view this as an exciting proof-of-concept. To the best of our
knowledge, no earlier attempt was successful in scaling higher order SOS relaxations beyond tens
of dimensions. More specifically, we carry out experiments with two-dimensional grids ? an image
denoising problem, and Ising spin glasses. We demonstrate through extensive numerical studies that
PSOS significantly outperforms BP and GBP in the inference tasks we consider.
2
Partial Sum-Of-Squares Relaxations
For concreteness, throughout the paper we focus on pairwise models with binary variables. We do not
expect fundamental problems extending the same approach to other domains. For binary variables
x = (x1 , x2 , ..., xn ), MAP estimation amounts to solving the following optimization problem
X
X
e
maximize
?ij
xi xj +
?iv xi ,
(INT)
x
where ? =
e
e
(?ij
)1?i,j?n
i2V
(i,j)2E
subject to xi 2 {+1, 1} ,
and ? =
v
(?iv )1?i?n
8i 2 V ,
are the parameters of the graphical model.
For the reader?s convenience, we recall a few basic facts about SOS relaxations, referring to [BS16]
for further details. For an even integer d, SOS(d) is an SDP relaxation of INT with decision variable
[n]
[n]
X : ?d
! R where ?d
denotes the set of subsets S ? [n] of size |S| ? d; it is given as
X
X
e
maximize
?ij
X({i, j}) +
?iv X({i}) ,
(SOS)
X
i2V
(i,j)2E
subject to X(;) = 1, M(X) < 0 .
The moment matrix M(X) is indexed by sets S, T ? [n], |S|, |T | ? d/2, and has entries M(X)S,T =
X(S4T ) with 4 denoting the symmetric difference of two sets. Note that M(X)S,S = X(;) = 1.
3
Rank
20
10
5
3
2
1e+01
662
1e?01
Duality Gap
Objective Value
600
400
Rank
20
10
5
3
2
200
0
1e?03
1e?05
1e?07
0
50
100
150
Iterations
200
0
50
100
Iterations
150
200
Figure 2: Effect of the rank constraint r on n = 400 square lattice (20 ? 20): Left plot shows the
change in the value of objective at each iteration. Right plot shows the duality gap of the Lagrangian.
We can equivalently represent M(X) as a Gram matrix by letting M(X)S,T = h S , T i for a
[n]
[n]
collection of vectors S 2 Rr indexed by S 2 ?d/2
. The case r = ?d/2
can represent any
semidefinite matrix; however, in what follows it is convenient from a computational perspective to
consider smaller choices of r. The constraint M(X)S,S = 1 is equivalent to k S k = 1, and the
condition M (X)S,T = X(S4T ) can be equivalently written as
h
T1 i
S1 ,
=h
S2 ,
T2 i ,
8S1 4T1 = S2 4T2 .
(2.1)
In the case d = 2, SOS(2) recovers the classical Goemans-Williamson SDP relaxation [GW95].
In the following, we consider the simplest higher-order SDP, namely SOS(4) for which the general
constraints in Eq. (2.1) can be listed explicitly. Fixing a region R ? V , and defining the Gram vectors
; , ( i )i2V , ( ij ){i,j}?V , we list the constraints that involve vectors S for S ? R and |S| = 1, 2:
k ik = 1
h i, j i = h
h i , ij i = h
h
i,
h
ij ,
h
ij ,
jk i
jk i
kl i
=h
=h
;i
ij ,
=h
8i 2 S [ {;},
8i, j 2 S,
8i, j 2 S,
;i
j,
ij i
k,
ik ,
ik ,
8i, j, k 2 S,
;i
8i, j, k 2 S,
jl i
8i, j, k, l 2 S.
(Sphere i )
(Undirected i j)
(Directed i ! j)
(V-shaped ij V k )
i
(Triangle j 4k )
(Loop ik ?jl )
Given an assignment of the Gram vectors = ( ; , ( i )i2V , ( ij ){i,j}?V ), we denote by |R its
restriction to R, namely |R = ( ; , ( i )i2R , ( ij ){i,j}?R ). We denote by ?(R), the set of vectors
|R that satisfy the above constraints. With these notations, the SOS(4) SDP can be written as
X
X
e
maximize
?ij
h i, j i +
?iv h i , ; i ,
(SOS(4))
i2V
(i,j)2E
subject to
2 ?(V ) .
A specific Partial SOS (PSOS) relaxation is defined by a collection of regions R =
{R1 , R2 , . . . , Rm }, Ri ? V . We will require R to be a covering, i.e. [m
i=1 Ri = V and for
each (i, j) 2 E there exists ` 2 [m] such that {i, j} ? R` . Given such a covering, the PSOS(4)
relaxation is
X
X
e
maximize
?ij
h i, j i +
?iv h i , ; i ,
(PSOS(4))
i2V
(i,j)2E
subject to
|Ri 2 ?(Ri )
8i 2 {1, 2, . . . , m} .
Notice that variables ij only enter the above program ifP
{i, j} ? R` for some `. As a consequence,
m
the dimension of the above optimization problem is O(r `=1 |R` |2 ), which is O(nr) if the regions
have bounded size; this will be the case in our implementation. Of course, the specific choice of
regions R is crucial for the quality of this relaxation. A natural heuristic is to choose each region R`
to be a subset of contiguous vertices in G, which is generally the case for GBP algorithms.
4
Algorithm 1: Partial-SOS
Input :G = (V, E), ?e 2 Rn?n , ?v 2 Rn ,
Actives = V [ E \ Reliables, and = 1,
while > tol do
=0
for s 2 Actives do
if s 2 V then
P
e
v
cs = t2@s ?st
t + ?s ;
else
cs = ?se1 s2 ; + ?sv1 s2 + ?sv2
2 Rr?(1+|V |+|E|) , Reliables = ;
/* s 2 V is a vertex */
/* s = (s1 , s2 ) 2 E is an edge */
s1
Form matrix As , vector bs , and the corresponding Lagrange multipliers s (see text).
new
arg max hcs , i + ?2 kAs
bs + s k2
/* sub-problem */
s
s
s
k k=1
new
s
new
s
s + As s
+k
2
sk
+ kAs
s
bs k2
bs
/* update variables */
2.1 Tightness guarantees
Solving exactly INT is NP-hard even if G is a three-dimensional grid [Bar82]. Therefore, we do
not expect PSOS(4) to be tight for general graphs G. On the other hand, in our experiments (cf.
Section 4), PSOS(4) systematically achieves the exact maximum of INT for two-dimensional grids
e
with random edge and vertex parameters (?ij
)(i,j)2E , (?iv )i2V . This finding is quite surprising and
calls for a theoretical explanation. While full understanding remains an open problem, we present
here partial results in that direction.
Recall that a cycle in G is a sequence of distinct vertices (i1 , . . . , i` ) such that, for each j 2 [`] ?
{1, 2, . . . , `}, (ij , ij+1 ) 2 E (where ` + 1 is identified with 1). The cycle is chordless if there is no
j, k 2 [`], with j k 6= ?1 mod ` such that (ij , ik ) 2 E. We say that a collection of regions R
on graph G is circular if for each chordless cycle in G there exists a region in R 2 R such that all
vertices of the cycle belong to R. We also need the following straightforward notion of contractibility.
A contraction of G is a new graph obtained by identifying two vertices connected by an edge in G.
G is contractible to H if there exists a sequence of contractions transforming G into H.
The following theorem is a direct consequence of a result of Barahona and Mahjoub [BM86] (see
Supplement for a proof).
Theorem 1. Consider the problem INT with ?v = 0. If G is not contractible to K5 (the complete
graph over 5 vertices), then PSOS(4) with a circular covering R is tight.
The assumption that ?v = 0 can be made without loss of generality (see Supplement for the reduction
from the general case). Furthermore, INT can be solved in polynomial time if G is planar, and ?v = 0
[Bar82]. Note however, the reduction from ?v 6= 0 to ?v = 0 can transform a planar graph to a
non-planar graph. This theorem implies that (full) SOS(4) is also tight if G is not contractible to K5 .
Notice that planar graphs are not contractible to K5 , and we recover the fact that INT can be solved
in polynomial time if ?v = 0. This result falls short of explaining the empirical findings in Section 4,
for at least two reasons. Firstly the reduction to ?v = 0 induces K5 subhomomorphisms for grids.
Second, the collection of regions R described in the previous section does not include all chordless
cycles. Theoretically understanding the empirical performance of PSOS(4) as stated remains open.
However, similar cycle constraints have proved useful in analyzing LP relaxations [WRS16].
3
3.1
Optimization Algorithm and Rounding
Solving PSOS(4) via Trust-Region Coordinate Ascent
We will approximately solve PSOS(4) while keeping r = O(1). Earlier work implies
p that (under
suitable genericity condition on the SDP) there existspan optimal solution with rank 2 # constraints
[Pat98]. Recent work [BVB16] shows that for r > 2 # constraints, the non-convex optimization
problem has no non-global local maxima. For SOS(2), [MM+ 17] proves that setting r = O(1)
is sufficient for achieving O(1/r) relative error from the global maximum for specific choices of
potentials ?e , ?v . We find that there is little or no improvement beyond r = 10 (cf. Figure 2).
5
Algorithm 2: CLAP: Confidence Lift And Project
Input :G = (V, E), ?e 2 Rn?n , ?v 2 Rn , regions R = {R1 , ..., Rm }
Initialize variable matrix 2 Rr?(1+|V |+|E|) and set Reliables = ;.
while Reliables 6= V [ E do
Run Partial-SOS on inputs G = (V, E), ?e , ?v , , Reliables
/* lift procedure */
Promotions = ; and Confidence = 0.9
while Confidence > 0 and Promotions 6= ; do
for s 2 V [ E \ Reliables do
/* find promotions */
if |h ; , s i| > Confidence then
/* project procedure */
s = sign(h ; , s i) ? ;
Promotions
Promotions [ {sc }
if Promotions = ; then
/* decrease confidence level */
Confidence
Confidence 0.1
Reliables
Reliables [ Promotions
/* update Reliables */
Output :(h
i,
; i)i2V
2 { 1, +1}n
We will assume that R = (R1 , . . . , Rm ) is a covering of G (in the sense introduced in the previous
section), and ?without loss of generality? we will assume that the edge set is
E = (i, j) 2 V ? V :
9` 2 [m] such that {i, j} ? R` .
(3.1)
In other words, E is the maximal set of edges that is compatible with R being a covering. This can
e
always be achieved by adding new edges (i, j) to the original edge set with ?ij
= 0. Hence, the
decision variables s are indexed by s 2 S = {;} [ V [ E. Apart from the norm constraints, all other
consistency constraints take the form h s , r i = h t , p i for some 4-tuple of indices (s, r, t, p). We
denote the set of all such 4-tuples by C, and construct the augmented Lagrangian of PSOS(4) as
?2
X
X
? X ?
e
L( , ) =
?iv h i , ; i +
?ij
h i, j i +
h s , r i h t , p i + s,r,t,p .
2
i2V
(i,j)2E
(s,r,t,p)2C
At each step, our algorithm execute two operations: (i) maximize the cost function with respect to one
of the vectors s ; (ii) perform one step of gradient descent with respect to the corresponding subset
of Lagrangian parameters, to be denoted by s . More precisely, fixing s 2 S \ {;} (by rotational
invariance, it is not necessary to update ; ), we note that s appears in the constraints linearly (or
it does not appear). Hence, we can write these constraints in the form As s = bs where As , bs
depend on ( r )r6=s but not on s . We stack the corresponding Lagrangian parameters in a vector s ;
therefore the Lagrangian term involving s reads (?/2)kAs s bs + s k2 . On the other hand, the
graphical model contribution is that the first two terms in L( , ) are linear in s , and hence they
can be written as hcs , s i. Summarizing, we have
L( , ) =hcs ,
si
+ kAs
bs +
s
2
sk
+ Le (
r )r6=s ,
.
(3.2)
It is straightforward to compute As , bs , cs ; in particular, for (s, r, t, p) 2 C, the rows of As and bs
are indexed by r such that the vectors r form the rows of As , and h t , p i form
P the corresponding
e
v
entry of bs . Further, if s is a vertex and @s are its neighbors, we set cs = t2@s ?st
t + ?s ;
e
v
v
while if s = (s1 , s2 ) is an edge, we set cs = ?s1 s2 ; + ?s1 s2 + ?s2 s1 . Note that we are using the
equivalent representations h i , j i = h ij , ; i, h ij , j i = h i , ; i, and h ij , i i = h j , ; i.
Finally, we maximize Eq. (3.2) with respect to
s
by a Mor?-Sorenson style method [MS83].
3.2 Rounding via Confidence Lift and Project
After Algorithm 1 generates an approximate optimizer for PSOS(4), we reduce its rank to produce
a solution of the original combinatorial optimization problem INT. To this end, we interpret h i , ; i
as our belief about the value of xi in the optimal solution of INT, and h ij , ; i as our belief about
the value of xi xj . This intuition can be formalized using the notion of pseudo-probability [BS16].
We then recursively round the variables about which we have strong beliefs; we fix rounded variables
in the next iteration, and solve the induced PSOS(4) on the remaining ones.
More precisely, we set a confidence threshold Confidence. For any variable s such that |h s , ; i| >
Confidence, we let xs = sign(h s , ; i) and fix s = xs ; . These variables s are no longer
6
Noisy
BP-SP
BP-MP
GBP
PSOS(2)
PSOS(4)
Bernoulli p = 0.2
True
Blockwise p=0.006
U(x) : 25815 19237 26165 26134 26161 26015 26194
Time: - - 2826s 2150s
7894s
454s
5059s
U(x) : 27010 26808 27230 27012 27232 26942 27252
Time: - - 1674s 729s
8844s
248s
4457s
Figure 3: Denoising a binary image by maximizing the objective function Eq. (4.1). Top row: i.i.d.
Bernoulli error with flip probability p = 0.2 with ?0 = 1.26. Bottom row: blockwise noise where
each pixel is the center of a 3 ? 3 error block independently with probability p = 0.006 and ?0 = 1.
updated, and instead the reduced SDP is solved. If no variable satisfies the confidence condition, the
threshold is reduced until variables are found that satisfy it. After the first iteration, most variables
yield strong beliefs and are fixed; hence the consequent iterations have fewer variables and are faster.
4
Numerical Experiments
In this section, we validate the performance of the Partial SOS relaxation and the CLAP rounding
scheme on models defined on two-dimensional grids. Grid-like graphical models are common in a
variety of fields such as computer vision [SSZ02], and statistical physics [MM09]. In Section 4.1, we
study an image denoising example and in Section 4.2 we consider the Ising spin glass ? a model in
statistical mechanics that has been used as a benchmark for inference in graphical models.
Our main objective is to demonstrate that Partial SOS can be used successfully on large-scale
graphical models, and is competitive with the following popular inference methods:
? Belief Propagation - Sum Product (BP-SP): Pearl?s belief propagation computes exact marginal
distributions on trees [Pea86]. Given a graph structured objective function U (x), we apply BP-SP
to the Gibbs-Boltzmann distribution p(x) = exp{U (x)}/Z using the standard sum-product update
rules with an inertia of 0.5 to help convergence [YFW05], and threshold the marginals at 0.5.
? Belief Propagation - Max Product (BP-MP): By replacing the marginal probabilities in the sumproduct updates with max-marginals, we obtain BP-MP, which can be used for exact inference
on trees [MM09]. For general graphs, BP-MP is closely related to an LP relaxation of the
combinatorial problem INT [YFW05, WF01]. Similar to BP-SP, we use an inertia of 0.5. Note
that the Max-Product updates can be equivalently written as Min-Sum updates [MM09].
? Generalized Belief Propagation (GBP): The decision variables in GBP are beliefs (joint probability distributions) over larger subsets of variables in the graph G, and they are updated in a
message passing fashion [YFW00, YFW05]. We use plaquettes in the grid (contiguous groups of
four vertices) as the largest regions, and apply message passing with inertia 0.1 [WF01].
? Partial SOS - Degree 2 (PSOS(2)): By defining regions as single vertices and enforcing only
the sphere constraints, we recover the classical Goemans-Williamson SDP relaxation [GW95].
Non-convex Burer-Monteiro approach is extremely efficient in this case [BM03]. We round the
SDP solution by x
?i = sign(h i , ; i) which is closely related to the classical approach of [GW95].
? Partial SOS - Degree 4 (PSOS(4)): This is the algorithm developed p
in the present
paper. We
p
take
R` to be triangles, cf. Figure 1, right frame. In an n ? n grid, we have
p the regions
2( n 1)2 such regions resulting in O(n) constraints. In Figures 3 and 4, PSOS(4) refers to the
CLAP rounding scheme applied together with PSOS(4) in the lift procedure.
4.1 Image Denoising via Markov Random Fields
p
p
Given a n ? n binary image x0 2 {+1, 1}n , we generate a corrupted version of the same
image y 2 {+1, 1}n . We then try to denoise y by maximizing the following objective function:
X
X
U (x) =
xi xj + ?0
yi x i ,
(4.1)
i2V
(i,j)2E
7
PS
O
PS S(4
O )
GB S(2
P )
BP
-S
BP P
-M
P
PS
O
PS S(4
O )
GB S(2
P )
BP
-S
BP P
-M
P
PS
O
PS S(4
O )
GB S(2
P )
BP
-S
BP P
-M
P
PS
O
PS S(4
O )
GB S(2
P )
BP
-S
BP P
-M
P
PS
O
PS S(4
O )
GB S(2
P )
BP
-S
BP P
-M
P
PS
O
PS S(4
O )
GB S(2
P )
BP
-S
BP P
-M
P
PS
O
PS S(4
O )
GB S(2
P )
BP
-S
BP P
-M
P
PS
O
PS S(4
O )
GB S(2
P )
BP
-S
BP P
-M
P
Ratio to the best algorithm
Figure 4: Solving the MAP inference problem INT for Ising spin glasses on two-dimensional grids.
U and N represent uniform and normal distributions. Each bar contains 100 independent realizations.
We plot the ratio between the objective value achieved by that algorithm and the exact optimum for
n 2 {16, 25}, or the best value achieved by any of the 5 algorithms for n 2 {100, 400, 900}.
p
p
p
where the graph G is the n ? n grid, i.e., V = {i = (i1 , i2 ) : i1 , i2 2 {1, . . . , n}} and
E = {(i, j) : ki jk1 = 1}. In applying Algorithm 1, we add diagonals to the grid (see right plot
e
in Figure 1) in order to satisfy the condition (3.1) with corresponding weight ?ij
= 0.
In Figure 3, we report the output of various algorithms for a 100 ? 100 binary image. We are
not aware of any earlier implementation of SOS(4) beyond tens of variables, while PSOS(4) is
applied here to n = 10, 000 variables. Running times for CLAP rounding scheme (which requires
several runs of PSOS(4)) are of order an hour, and are reported in Figure 3. We consider two noise
models: i.i.d. Bernoulli noise and blockwise noise. The model parameter ?0 is chosen in each
case as to approximately optimize the performances under BP denoising. In these (as well as in 4
other experiments of the same type reported in the supplement), PSOS(4) gives consistently the best
reconstruction (often tied with GBP), in reasonable time. Also, it consistently achieves the largest
value of the objective function among all algorithms.
4.2
Ising Spin Glass
The Ising spin glass (also known as Edwards-Anderson model [EA75]) is one of the most studied
models in statistical physics. It is given by an objective function of the form INT with G a de
dimensional grid, and i.i.d. parameters {?ij
}(i,j)2E , {?iv }i2V . Following earlier work [YFW05],
we use Ising spin glasses as a testing ground for our algorithm. Denoting the uniform and normal
distributions by U and N respectively, we consider two-dimensional grids (i.e. d = 2), and the
e
e
following parameter distributions: (i) ?ij
? U ({+1, 1}) and ?iv ? U ({+1, 1}), (ii) ?ij
?
v
e
v
2
U ({+1, 1}) and ?i ? U ({+1/2, 1/2}), (iii) ?ij ? N(0, 1) and ?i ? N(0, ) with = 0.1
e
(this is the setting considered in [YFW05]), and (iv) ?ij
? N(0, 1) and ?iv ? N(0, 2 ) with = 1.
For each of these settings, we considered grids of size n 2 {16, 25, 100, 400, 900}.
In Figure 4, we report the results of 8 experiments as a box plot. We ran the five inference algorithms
described above on 100 realizations; a total of 800 experiments are reported in Figure 4. For each of
the realizations, we record the ratio of the achieved value of an algorithm to the exact maximum (for
n 2 {16, 25}), or to the best value achieved among these algorithms (for n 2 {100, 400, 900}). This
is because for lattices of size 16 and 25, we are able to run an exhaustive search to determine the true
maximizer of the integer program. Further details are reported in the supplement.
In every single instance of 800 experiments, PSOS(4) achieved the largest objective value, and
whenever this could be verified by exhaustive search (i.e. for n 2 {16, 25}) it achieved an exact
maximizer of the integer program.
8
References
[Bar82] Francisco Barahona. On the computational complexity of Ising spin glass models. Journal of Physics
A: Mathematical and General, 15(10):3241, 1982.
[BM86] Francisco Barahona and Ali Ridha Mahjoub. On the cut polytope. Mathematical programming,
36(2):157?173, 1986.
[BM03] Samuel Burer and Renato DC Monteiro. A nonlinear programming algorithm for solving semidefinite
programs via low-rank factorization. Mathematical Programming, 95(2):329?357, 2003.
[BS16] Boaz Barak and David Steurer. Proofs, beliefs, and algorithms through the lens of sum-of-squares.
Course notes: http://www. sumofsquares. org/public/index. html, 2016.
[BVB16] Nicolas Boumal, Vlad Voroninski, and Afonso Bandeira. The non-convex Burer-Monteiro approach
works on smooth semidefinite programs. In Advances in Neural Information Processing Systems,
pages 2757?2765, 2016.
[CD+ 06] Robert G Cowell, Philip Dawid, Steffen L Lauritzen, and David J Spiegelhalter. Probabilistic
networks and expert systems: Exact computational methods for Bayesian networks. Springer Science
& Business Media, 2006.
[EA75] Samuel Frederick Edwards and Phil W Anderson. Theory of spin glasses. Journal of Physics F:
Metal Physics, 5(5):965, 1975.
[EM15] Murat A Erdogdu and Andrea Montanari. Convergence rates of sub-sampled newton methods. In
Advances in Neural Information Processing Systems, pages 3052?3060, 2015.
[GW95] Michel X Goemans and David P Williamson. Improved approximation algorithms for maximum
cut and satisfiability problems using semidefinite programming. Journal of the ACM (JACM),
42(6):1115?1145, 1995.
[KF09] Daphne Koller and Nir Friedman. Probabilistic graphical models. MIT press, 2009.
[Las01] Jean B Lasserre. An explicit exact SDP relaxation for nonlinear 0-1 programs. In International
Conference on Integer Programming and Combinatorial Optimization, pages 293?303, 2001.
[MM09] Marc M?zard and Andrea Montanari. Information, physics, and computation. Oxford Press, 2009.
[MM+ 17] Song Mei, Theodor Misiakiewicz, Andrea Montanari, and Roberto I Oliveira. Solving SDPs
for synchronization and MaxCut problems via the Grothendieck inequality. arXiv preprint
arXiv:1703.08729, 2017.
[MS83] Jorge J Mor? and Danny C Sorensen. Computing a trust region step. SIAM Journal on Scientific and
Statistical Computing, 4(3):553?572, 1983.
[Par03] Pablo A Parrilo. Semidefinite programming relaxations for semialgebraic problems. Mathematical
programming, 96(2):293?320, 2003.
[Pat98] G?bor Pataki. On the rank of extreme matrices in semidefinite programs and the multiplicity of
optimal eigenvalues. Mathematics of operations research, 23(2):339?358, 1998.
[Pea86] Judea Pearl. Fusion, propagation, and structuring in belief networks. Artificial intelligence, 29(3):241?
288, 1986.
[RU08] Tom Richardson and Ruediger Urbanke. Modern coding theory. Cambridge Press, 2008.
[SA90] Hanif D Sherali and Warren P Adams. A hierarchy of relaxations between the continuous and convex
hull representations for zero-one programming problems. SIAM Journal on Discrete Mathematics,
3(3):411?430, 1990.
[Sho87] Naum Z Shor. Class of global minimum bounds of polynomial functions. Cybernetics and Systems
Analysis, 23(6):731?734, 1987.
[SSZ02] Jian Sun, Heung-Yeung Shum, and Nan-Ning Zheng. Stereo matching using belief propagation. In
European Conference on Computer Vision, pages 510?524. Springer, 2002.
[WF01] Yair Weiss and William T Freeman. On the optimality of solutions of the max-product beliefpropagation algorithm in arbitrary graphs. IEEE Trans. on Info. Theory, 47(2):736?744, 2001.
[WJ04] Martin J Wainwright and Michael I Jordan. Semidefinite relaxations for approximate inference on
graphs with cycles. In Advances in Neural Information Processing Systems, pages 369?376, 2004.
[WJ08] Martin J Wainwright and Michael I Jordan. Graphical models, exponential families, and variational
inference. Foundations and Trends in Machine Learning, 1(1?2):1?305, 2008.
[WRS16] Adrian Weller, Mark Rowland, and David Sontag. Tightness of lp relaxations for almost balanced
models. In Artificial Intelligence and Statistics, pages 47?55, 2016.
[YFW00] Jonathan S Yedidia, William T Freeman, and Yair Weiss. Generalized belief propagation. In
Advances in Neural Information Processing Systems, pages 689?695, 2000.
[YFW05] Jonathan S Yedidia, William T Freeman, and Yair Weiss. Constructing free-energy approximations and generalized belief propagation algorithms. IEEE Transactions on Information Theory,
51(7):2282?2312, 2005.
9
| 6645 |@word briefly:1 version:2 polynomial:3 norm:1 nd:1 open:2 barahona:3 adrian:1 accounting:2 contraction:2 recursively:1 carry:2 moment:1 reduction:4 contains:1 sherali:4 shum:1 denoting:3 outperforms:3 ka:4 surprising:1 si:1 assigning:1 danny:1 written:5 numerical:4 partition:1 plot:5 update:7 intelligence:2 fewer:1 short:2 record:1 toronto:1 firstly:2 simpler:1 org:1 five:1 daphne:1 mathematical:4 direct:1 ik:5 prove:1 introduce:2 theoretically:1 x0:1 pairwise:2 andrea:4 sdp:22 mechanic:1 steffen:1 freeman:3 little:1 solver:2 becomes:1 project:4 bounded:2 notation:1 medium:1 what:1 mahjoub:2 substantially:1 developed:2 finding:2 guarantee:4 pseudo:1 every:1 exactly:3 rm:3 k2:3 wj08:3 appear:1 t1:2 understood:2 local:2 consequence:3 despite:2 analyzing:2 oxford:1 approximately:2 studied:1 limited:1 factorization:1 bi:4 directed:1 practical:2 testing:1 practice:1 block:1 xr:7 procedure:5 mei:1 empirical:4 significantly:1 convenient:1 matching:1 confidence:13 word:1 refers:1 convenience:1 interior:3 applying:1 restriction:1 equivalent:2 map:8 lagrangian:5 center:1 maximizing:2 optimize:1 straightforward:2 www:1 phil:1 independently:1 convex:8 focused:1 formulate:1 simplicity:1 identifying:1 formalized:1 correcting:1 rule:1 array:1 notion:2 coordinate:2 updated:2 hierarchy:12 massive:1 exact:8 programming:13 us:3 dawid:1 trend:1 jk:2 gw95:6 ising:8 cut:2 bottom:1 preprint:1 solved:3 thousand:1 region:29 connected:2 cycle:7 sun:1 ordering:1 eu:1 decrease:1 ran:1 balanced:1 intuition:1 transforming:1 complexity:1 depend:1 solving:10 tight:4 ali:1 efficiency:1 triangle:4 joint:2 yash:2 s4t:2 various:1 distinct:1 kf09:3 artificial:2 sc:1 lift:5 exhaustive:2 quite:1 heuristic:3 stanford:2 solve:5 larger:2 say:1 tightness:3 jean:1 statistic:1 gi:4 richardson:1 transform:1 itself:1 noisy:1 sequence:2 rr:3 eigenvalue:1 propose:4 reconstruction:1 interaction:2 product:6 maximal:1 neighboring:1 loop:1 realization:3 poorly:2 validate:1 convergence:2 p:16 extending:1 r1:3 produce:2 optimum:1 adam:4 help:1 develop:3 fixing:2 ij:35 lauritzen:1 eq:3 edward:2 strong:2 c:6 bs16:3 treewidth:1 implies:2 quantify:1 differ:1 direction:2 ning:1 closely:4 hull:1 public:1 require:2 fix:2 tighter:3 secondly:1 mm:2 considered:2 ground:1 normal:2 exp:1 achieves:2 adopt:1 optimizer:1 estimation:3 favorable:1 combinatorial:7 largest:3 successfully:1 mit:3 promotion:7 always:1 conjunction:1 encode:1 jk1:1 focus:3 structuring:1 improvement:1 consistently:2 rank:8 bernoulli:3 contrast:1 sense:1 glass:9 posteriori:1 inference:22 summarizing:1 koller:1 comprising:1 i1:3 voroninski:1 pixel:1 monteiro:5 arg:2 aforementioned:1 among:2 i2v:15 denoted:1 html:1 development:1 special:1 initialize:1 marginal:3 field:3 construct:1 aware:1 shaped:1 beach:1 broad:1 simplex:1 t2:4 np:1 report:2 few:2 employ:1 modern:2 microsoft:2 william:3 attempt:1 friedman:1 message:2 circular:2 zheng:1 introduces:1 extreme:1 semidefinite:10 sorensen:1 edge:10 tuple:1 partial:12 necessary:1 sorenson:1 tree:4 iv:14 indexed:4 urbanke:1 re:4 causal:1 theoretical:2 instance:4 se1:1 earlier:4 contiguous:5 retains:1 assignment:1 lattice:2 loopy:2 cost:1 vertex:19 subset:12 entry:2 uniform:2 comprised:1 successful:2 rounding:8 weller:1 reported:4 dependency:1 corrupted:1 referring:1 st:3 sv1:1 fundamental:1 international:1 siam:2 probabilistic:2 physic:7 decoding:1 rounded:1 michael:2 together:1 theodor:1 choose:1 expert:1 style:2 michel:1 account:2 potential:2 parrilo:1 de:1 coding:1 int:12 satisfy:4 explicitly:1 mp:4 view:1 try:1 competitive:1 recover:2 contractibility:1 contribution:2 square:9 spin:9 correspond:2 yield:1 bayesian:2 sdps:4 bor:1 cybernetics:1 suffers:1 afonso:1 whenever:2 energy:1 deshpande:1 naturally:1 proof:3 recovers:1 judea:1 sampled:1 proved:1 popular:2 vlad:1 recall:2 knowledge:1 satisfiability:1 appears:1 higher:3 methodology:1 planar:5 improved:2 tom:1 wei:3 execute:1 hcs:3 box:1 generality:2 furthermore:1 anderson:2 correlation:1 until:1 hand:2 trust:3 replacing:1 nonlinear:2 maximizer:2 propagation:13 quality:1 scientific:2 believe:1 usa:1 bm03:3 effect:1 concept:1 multiplier:1 true:2 hence:6 read:1 symmetric:1 iteratively:1 i2:2 em15:1 round:2 covering:5 whereby:1 samuel:2 generalized:5 complete:1 demonstrate:3 image:8 variational:1 superior:2 ifp:1 common:1 jl:2 discussed:1 belong:1 mor:2 marginals:5 interpret:1 significant:1 refer:1 cambridge:1 gibbs:1 enter:1 grid:16 consistency:1 mathematics:2 maxcut:1 longer:1 add:1 multivariate:1 recent:1 perspective:1 apart:1 certain:2 contractible:4 bandeira:1 binary:8 success:1 inequality:1 jorge:1 yi:1 pover:1 minimum:1 additional:1 determine:1 maximize:7 ii:3 full:2 pataki:1 smooth:1 faster:1 burer:5 long:1 sphere:2 variant:1 basic:6 involving:2 vision:4 arxiv:2 iteration:6 normalization:1 represent:5 yeung:1 achieved:8 x2x:1 else:1 jian:1 crucial:2 ascent:2 subject:5 induced:2 undirected:2 mod:1 jordan:2 integer:6 call:1 structural:1 iii:2 variety:2 xj:9 marginalization:2 shor:1 identified:1 reduce:1 idea:1 br:5 motivated:1 gb:8 song:1 stereo:1 sontag:1 passing:2 tol:1 generally:1 useful:1 involve:2 listed:1 amount:3 oliveira:1 ten:4 locally:1 induces:1 simplest:3 reduced:2 generate:1 http:1 revisit:1 notice:2 sign:3 arising:1 popularity:1 write:1 discrete:1 group:1 four:3 nevertheless:1 threshold:3 achieving:1 verified:1 heung:1 graph:24 relaxation:40 concreteness:1 sum:10 run:3 powerful:1 family:2 throughout:1 reader:1 reasonable:1 almost:1 decision:11 scaling:1 capturing:1 ki:1 renato:1 bound:1 nan:1 simplification:1 zard:1 constraint:18 precisely:2 constrain:1 bp:34 x2:2 afforded:1 ri:4 generates:1 min:1 extremely:1 optimality:1 martin:2 structured:4 smaller:1 intimately:2 lp:11 b:13 modification:1 s1:8 multiplicity:1 remains:2 letting:2 flip:1 end:1 operation:2 k5:4 apply:2 yedidia:2 generic:1 yair:3 original:2 denotes:2 remaining:1 cf:6 include:1 top:1 graphical:18 running:1 newton:1 prof:1 establish:1 classical:3 misiakiewicz:1 objective:10 beliefpropagation:1 chordless:3 nr:1 diagonal:1 gradient:1 philip:1 polytope:3 reason:4 enforcing:1 code:1 index:2 rotational:1 ratio:3 innovation:1 equivalently:3 unfortunately:1 robert:1 holding:1 blockwise:3 info:1 stated:1 implementation:4 steurer:1 murat:2 boltzmann:1 perform:1 ruediger:1 markov:1 benchmark:1 finite:3 descent:1 defining:2 extended:1 incorporated:1 frame:3 interacting:1 rn:4 stack:1 dc:1 arbitrary:2 sumproduct:1 introduced:1 david:4 pair:1 required:1 specified:1 extensive:1 namely:2 kl:1 gbp:17 pablo:1 plague:1 pearl:2 hour:1 nip:1 trans:1 beyond:4 plaquettes:3 bar:1 below:1 able:1 frederick:1 program:8 including:2 max:7 explanation:1 belief:25 wainwright:2 suitable:3 satisfaction:1 natural:2 business:1 scheme:3 spiegelhalter:1 roberto:1 nir:1 grothendieck:1 text:1 understanding:2 countless:1 relative:1 synchronization:1 fully:1 expect:2 loss:2 prototypical:1 limitation:1 analogy:1 semialgebraic:1 foundation:1 degree:7 sufficient:2 consistent:1 metal:1 exciting:1 intractability:1 systematically:1 cd:2 row:4 course:2 clap:5 compatible:1 keeping:1 free:1 warren:1 barak:1 wide:1 fall:2 erdogdu:3 explaining:2 neighbor:1 boumal:1 sparse:1 overcome:1 dimension:2 xn:2 gram:3 genome:1 rich:1 computes:1 inertia:3 collection:5 commonly:1 made:1 mm09:7 rowland:1 transaction:1 approximate:2 boaz:1 sv2:1 global:3 active:2 francisco:2 tuples:1 xi:16 alternatively:1 search:2 continuous:1 sk:2 lasserre:1 ca:1 nicolas:1 williamson:3 european:1 constructing:1 domain:3 marc:1 sp:4 main:1 montanari:5 linearly:1 s2:9 noise:4 arise:1 n2:2 denoise:1 x1:2 augmented:1 referred:2 fashion:1 sub:2 explicit:1 exponential:1 r6:2 tied:1 bij:4 minute:2 theorem:3 specific:4 list:1 x:4 r2:1 consequent:1 fusion:1 intractable:1 essential:1 exists:3 sequential:1 adding:1 supplement:4 genericity:1 nk:1 gap:2 suited:1 entropy:1 jacm:1 lagrange:1 cowell:1 springer:2 satisfies:1 acm:1 maxx2x:1 feasible:1 change:1 hard:1 typical:1 except:1 specifically:2 denoising:6 i2r:1 called:1 total:1 goemans:3 duality:2 invariance:1 lens:1 exception:1 mark:1 jonathan:2 |
6,241 | 6,646 | Variable Importance using Decision Trees
Jalil Kazemitabar
UCLA
[email protected]
Arash A. Amini
UCLA
[email protected]
Adam Bloniarz
UC Berkeley?
[email protected]
Ameet Talwalkar
CMU
[email protected]
Abstract
Decision trees and random forests are well established models that not only offer
good predictive performance, but also provide rich feature importance information.
While practitioners often employ variable importance methods that rely on this
impurity-based information, these methods remain poorly characterized from a
theoretical perspective. We provide novel insights into the performance of these
methods by deriving finite sample performance guarantees in a high-dimensional
setting under various modeling assumptions. We further demonstrate the effectiveness of these impurity-based methods via an extensive set of simulations.
1
Introduction
Known for their accuracy and robustness, decision trees and random forests have long been a
workhorse in machine learning [1]. In addition to their strong predictive accuracy, they are equipped
with measures of variable importance that are widely used in applications where model interpretability
is paramount. Importance scores are used for model selection: predictors with high-ranking scores
may be chosen for further investigation, or for building a more parsimonious model.
One common approach naturally couples the model training process with feature selection [2, 5].
This approach, which we call T REE W EIGHT, calculates the feature importance score for a variable
by summing the impurity reductions over all nodes in the tree where a split was made on that
variable, with impurity reductions weighted to account for the size of the node. For ensembles, these
quantities are averaged over constituent trees. T REE W EIGHT is particularly attractive because it can
be calculated without any additional computational expense above the standard training procedure.
However, as the training procedure in random forests combines several complex ingredients?bagging,
random selection of predictor subsets at nodes, line search for optimal impurity reduction, recursive
partitioning?theoretical investigation into T REE W EIGHT is extremely challenging. We propose a
new method called DS TUMP that is inspired by T REE W EIGHT but is more amenable to analysis.
DS TUMP assigns variable importance as the impurity reduction at the root node of a single tree.
In this work we characterize the finite sample performance of DS TUMP under an additive regression
model, which also yields novel results for variable selection under a linear model, both with correlated
and uncorrelated design. We corroborate our theoretical analyses with extensive simulations in which
we evaluate DS TUMP and T REE W EIGHT on the task of feature selection under various modeling
assumptions. We also compare the performance of these techniques against established methods
whose behaviors have been theoretically characterized, including Lasso, SIS, and SpAM [12, 3, 9].
?
Now at Google
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
Our work provides the first finite-sample high-dimensional analyses of tree-based variable selection
techniques, which are commonly used in practice but lacking in theoretical grounding. Although
we focus on DS TUMP, which is a relatively simple tree-based variable selection approach, our
novel proof techniques are highly non-trivial and suggest a path forward for studying more general
multi-level tree-based techniques such as T REE W EIGHT. Moreover, our simulations demonstrate
that such algorithmic generalizations exhibit impressive performance relative to competing methods
under more realistic models, e.g., non-linear models with interaction terms and correlated design.
2
Related Work
Our analysis is distinct from existing work in analyzing variable importance measures of trees and
forests in several ways. To our knowledge, ours is the first analysis to consider the high-dimensional
setting, where the number of variables, p, and the size of the active set s, grow with the sample size
n, and potentially p n.
The closest related work is the analysis of [8], which considers a fixed set of variables, in the limit of
infinite data (n = ?). Unlike DS TUMP?s use of the root node only, [8] does consider importance
scores derived from the full set of splits in a tree as in T REE W EIGHT. However, they make crucial
simplifying (and unrealistic) assumptions that are distinct from those of our analysis: (1) each variable
is split on only once in any given path from the root to a leaf of the tree; (2) at each node a variable is
picked uniformly at random among those not yet used at the parent nodes, i.e., the splits themselves
are not driven by impurity reduction; and (3) all predictors are categorical, with splits being made on
all possible levels of a variable, i.e., the number of child nodes equals the cardinality of the variable
being split. Our analysis instead considers continuous-valued predictors, the split is based on actual
impurity reduction, and our results are nonasymptotic, i.e. they give high-probability bounds on
impurity measures for active and inactive variables that hold in finite samples.
A second line of related work is motivated by a permutation-based importance method [1] for feature
selection. In practice, this method is computationally expensive as it determines variable importance
by comparing the predictive accuracy of a forest before and after random permutation of a predictor.
Additionally, due to the algorithmic complexity of the procedure, it is not immediately amenable to
theoretical analysis, though the asymptotic properties of a simplified variant of the procedure have
been studied in [6].
While our work is the first investigation of finite-sample model selection performance of tree-based
regression methods, alternative methods performing both linear and nonparametric regression in high
dimensions have been studied in the literature. Considering model selection consistency results, most
of the attention has been focused on the linear setting, whereas the nonparametric (nonlinear) setup
has been mostly studied in terms of the prediction consistency. Under a high-dimensional linear
regression model, L ASSO has be extensively studied and is shown to be minimax optimal for variable
selection under appropriate regularity conditions, including the uncorrelated design with a moderate
?min condition. Remarkably, while not tailored to the linear setting, we show that DS TUMP is nearly
minimax optimal for variable selection in the same uncorrelated design setting (cf. Corollary 1). In
fact, DS TUMP can be considered a nonlinear version of SIS [4], itself a simplified form of the L ASSO
when one ignores correlation among features (cf. Section 3 for more details).
The Rodeo framework [7] performs automatic bandwidth selection and variable selection for local
linear smoothers, and is tailored to a more general nonparametric model with arbitrary interactions.
It was shown to possess model selection consistency in high dimensions; however, the results are
asymptotic and focus on achieving optimal prediction rate. In particular, there is no clear ?min
threshold as a function of n, s, and p. RODEO is also computationally burdensome for even modestsized problems (we thus omit it our experimental results in Section 4).
Among the nonlinear methods, S PAM is perhaps the most well-understood in terms of model
selection properties. Under a general high-dimensional sparse additive model, S PAM possesses the
sparsistency property (a term for model selection consistency); the analysis is reduced to a linear
setting by considering expansions in basis functions, and selection consistency is proved under an
irrepresentible condition on the coefficients in those bases. We show that DS TUMP is model selection
consistent in the sparse additive model with uncorrelated design. Compared to S PAM results, our
conditions are stated directly in terms of underlying functions and are not tied to a particular basis;
2
hence our proof technique is quite different. There is no implicit reduction to a linear setting via basis
expansions. Empirically, we show that DS TUMP indeed succeeds in the settings our theory predicts.
3
Selection consistency
The general model selection problem for non-parametric regression can be stated as follows: we
observe noisy samples yi = f (xi1 , . . . , xip ) + wi , i = 1, . . . , n where {wi } is an i.i.d. noise
sequence. Here, p is the total number of features (or covariates) and n is the total number of
observations (or the sample size). In general, f belongs to a class F of functions from Rp ? R.
One further assumes that the functions in F depend on at most s of the features, usually with s p.
That is, for every f ? F, there is some f0 : Rs ? R and a subset S ? [p] with |S| ? s such
that f (z1 , . . . , zp ) = f0 (zS ) where zS = (zi , i ? S). The subset S, i.e., the set of active features,
is unknown in advance and the goal of model selection is to recover it given {(yi , xi )}ni=1 . The
problem is especially challenging in the high-dimensional setting where p n. We will consider
various special cases of this general model when we analyze DS TUMP. For theoretical analysis it is
common to assume s to be known and we will make this assumption throughout. In practice, one
often considers s to be a tunable parameter that can be selected, e.g., via cross-validation or greedy
forward selection.
We characterize the model selection performance of DS TUMP by establishing its sample complexity:
that is, the scaling of n, p, and s that is sufficient to guarantee that DS TUMP identifies the active
set of features with probability converging to 1. Our general results, proved in the technical report,
allow for a correlated design matrix and additive nonlinearities in the true regression function. Our
results for the linear case, derived as a special case of the general theory, allow us to compare the
performance of DS TUMP to the information theoretic limits for sample complexity established in
[11], and to the performance of existing methods more tailored to this setting, such as the Lasso [12].
Given a generative model and the restriction of DS TUMP to using root-level impurity reduction, the
general thrust of our result is straightforward: impurity reduction due to active variables concentrates
at a significantly higher level than that of inactive variables. However, there are significant technical
challenges in establishing this result, mainly deriving from the fact that the splitting procedure
renders the data in the child nodes non-i.i.d., and hence standard concentration inequalities do not
immediately apply. We leverage the fact that the DS TUMP procedure considers splits at the median
of a predictor. Given this median point, the data in each child node is i.i.d., and hence we can
apply standard concentration inequalities in this conditional distribution. Removing this conditioning
presents an additional technical subtlety. For ease of exposition, we first present our results for the
linear setting in Section 3.1, and subsequently summarize our general results in Section 3.2. We
provide a proof of our result in the linear setting in Section 3.3, and defer the proof of our general
result to the supplementary material.
Algorithm 1 DS TUMP
k=p
input {xk ? Rn }k=1
, y ? Rn , # top features s
n
m= 2
for k = 1, . . . , p do
I(xk ) = SortFeatureValues(xk )
y k = SortLabelByFeature(y, I(xk ))
k
k
y[m]
, y[n]\[m]
= SplitAtMidpoint(y k )
k
k
ik = ComputeImpurity(y[m]
, y[n]\[m]
)
end for
S = FindTopImpurityReductions({ik }, s)
output top s features sorted by impurity reduction
The DS TUMP algorithm. In order to describe DS TUMP more precisely, let us introduce
some notation. We write [n] := {1, . . . , n}.
Throughout, y = (yi , i ? [n]) ? Rn will be the
response vector observed for a sample of size
n. For an ordered index set I = (i1 , i2 , . . . , ir ),
we set yI = (yi1 , yi2 . . . , yir ). A similar notation is used for unordered index sets. We write
xj = (x1j , x2j , . . . , xnj ) ? Rn for the vector
collecting values of the jth feature; xj forms the
jth column of the design matrix X ? Rn?p .
Let I(xj ) := (i1 , i2 . . . , in ) be an ordering of
[n] such that xi1 j ? xi2 j ? ? ? ? ? xin j and let
sor(y, xj ) := yI(xj ) ? Rn ; this is an operator
that sorts y relative
to xj . DS TUMP proceeds as
P
follows: Evaluate y k := sor(y, xk ) = sor
?
x
+
w,
x
,
for
k
= 1, . . . , p. Let m := n/2.
j
j
k
j?S
k
k
k
For each k, consider the midpoint split of y into y[m] and y[n]\[m] and evaluate the impurity of the
3
left-half, using empirical variance as impurity:
k
imp(y[m]
) :=
1
m
2
X
1?i<j?m
1 k
(y ? yjk )2 .
2 i
(1)
k
Let imp(y[m]
) be the score of feature k, and output the s features with the smallest scores (corresponding to maximal reduction in impurity). If the generative model is linear, the choice of the
midpoint is justified by our assumption of the uniform distribution for the features (Zi ), and we
further show that this simple choice is effective even under a nonlinear model. The choice of the
left-half in our analysis is for convenience; a similar analysis applies if we take the impurity to be that
of the sum of both halves (or their maximum). DS TUMP is summarized in Algorithm 1. Impurity
k
reduction imp(y[m] ) ? imp(y[m]
) can be considered a form of nonlinear correlation between y and
feature xk . The SIS algorithm is equivalent to replacing this nonlinear correlation with the (absolute)
linear correlation | n1 xTk y|. That is, both procedures assign a score to each feature by considering it
against the response separately, ignoring other features. In the uncorrelated (i.e. orthogonal design)
setting, this is more or less optimal, and as is the case with SIS, we show that DS TUMP also retains
some model selection performance even under correlated designs. In contrast to SIS, we show that
DS TUMP also enjoys performance guarantees in non-linear settings.
The models. We present our consistency results for models of various complexity. We start with the
well-known and extensively studied setting of a linear model with IID design. This basic setup serves
as the benchmark for comparison of model selection procedures. As will become clear in the course
of the proof, analyzing DS TUMP (or impurity-based feature selection in general) is challenging even
in this case, in contrast to linear model based approaches such as SIS or Lasso. Once we have a good
understanding of DS TUMP under the baseline model, we extend the analysis to correlated design and
nonlinear additive models. The structure of our proof is also most clearly seen in this simple case, as
outlined in Section 3.3. We now introduce our general models:
Model 1 (Sparse linear model with ICA-type design). A linear regression model y = X? + w with
e ? Rn?p
e where X
ICA-type (random) design X ? Rn?p has the following properties: (i) X = XM
e is an independent draw from a (column) vector Z = (Z1 , . . . , Zp ) with IID
and each row of X
entries drawn uniformly from [0, 1]. (ii) The noise vector w = (w1 , . . . , wn ) has IID sub-Gaussian
2
entries with variance with variance var(wi ) = vw
and sub-Gaussian norm kwi k?2 ? ?w , . (iii) The
p
? ? R is s-sparse, namely, ?j 6= 0 for j ? S = {1, . . . , s} and zero otherwise.
Model 1 serves both the correlated and uncorrelated design cases. Each row of the design matrix
X is a draw from the vector M T Z, which has covariance c M T M for some constant c. Thus, the
choice of M = I leads to an uncorrelated design. The choice of the interval [0, 1] for covariates
is for convenience; it can be replaced with any other compact interval, in the linear setting, since
variance impurity is invariant to a shift. Similarly the choice of the (active) support indices, S, is for
2
convenience. For simplicity, we often assume vw
= ? 2 and ?w ? C? (only ?w would affect the
results as examining of our proofs shows).
Model
2 (Sparse additive model with uncorrelated design). An additive regression model yi =
Pp
j=1 fj (xij ) + wi , is one with random design X = (xij ) and the noise (wi ) as in Model 1, with
M = I (uncorrelated design). We assume (fk ) to be s-sparse, namely, fj 6= 0 for j ? S = {1, . . . , s}
and zero otherwise.
3.1
Linear Setting
Uncorrelated design. Our baseline result is the following feature selection consistency guarantee
for DS TUMP, for the case M = I of Model 1. Throughout, we let p? := p?s, and C, C1 , . . . , c, c1 , . . .
are absolute positive constants which can be different in each occurrence. For any vector x, let
|x|min := mini |xi |, the minimum absolute value of its entries. The quantity |?S |2min = mink ? S ?k2
appearing in Theorem 1 is a well-known parameter controlling hardness of subset recovery. All our
results are stated in terms of constants ?, ? and ? that are related as:
? ? (0, 1/8),
? = log(1/(8?)),
4
? = 1 ? (1 ? ?)2 .
(2)
Theorem 1. Assume Model 1 with M = I, and (2). The DS TUMP algorithm, which selects
the ?s? least impure features at the root, succeeds in feature selection, with probability at least
1 ? p??c ? 2e??n/2 if log p?/n ? C1 and
r
C
log p?
2
2
2
|?S |min ?
(k?k2 + ? )
(3)
?
n
The result can be read by setting, e.g., ? = 1/16 leading to numerical constants for ? and ?. The
current form allows the flexibility to trade-off the constant (?) in the probability bound with the
constant (?) in the gap condition (3). Although Theorem 1 applies to a general ?, it is worthwhile to
see its consequence in a special regime of interest where |?S |2min 1/s, corresponding to k?k2 1.
We get the following immediate corollary:
Corollary 1. Assume |?S |2min 1/s, ? 2 1 and log p?/n = O(1). Then DS TUMP succeeds with
high probability if n & s2 log p?.
The minimax optimal threshold for support recovery in the regime of Corollary 1 is known to be
n s log p? [11], and achieved by LASSO [12]. Although this result is obtained for Gaussian design,
the same argument goes through for our uniform ensemble. Compared to the optimal threshold, using
DS TUMP we pay a small factor of s in the sample complexity. However, DS TUMP is not tied to the
linear model and as we discuss in Section 3.2, we can generalize the performance of DS TUMP to
nonlinear settings.
Correlated design. We take the following approach to generalize our result to the correlated case:
(1) We show a version of Theorem 1, which holds for an ?approximately sparse? parameter ?e with
uncorrelated design. (2) We derive conditions on M such that the correlated case can be turned into
the uncorrelated case with approximate sparsity. The following theorem details Step 1:
e a general vector in
Theorem 2. Assume Model 1(i)-(ii) with M = I, but instead of (iii) let ? = ?,
p
R . Let S be any subset of [p] of cardinality s. The DS TUMP algorithm, which selects the ?s? least
impure features at the root, recovers S, withp
probability at least 1 ? p??c ? 2e??n/2 if log p?/n ? C1
2
2
2
2
e + ? ) (log p?)/n.
and ?|?eS |min ? k?eS c k? > C(k?k
2
The theorem holds for any ?e and S, but the gap condition required is likely to be violated unless ?e
is approximately sparse w.r.t. S. Going back to Model 1, we see that setting ?e = M ? transforms
the model with correlated design X, and exact sparsity on ?, to the model with uncorrelated design
e The following corollary gives sufficient conditions on M , so
e and approximate sparsity on ?.
X,
that Theorem 2 is applicable. Recall theP
usual (vector) `? norm, kxk? = maxi |xi |, the matrix
`? ? `? operator norm |||A|||? = maxi j |Aij | , and the `2 ? `2 operator norm |||A|||2 .
Corollary 2. Consider a general ICA-type Model 1 with ? and M satisfying
k?S k? ? ?|?S |min ,
|||MSS ? I|||? ?
1??
,
?
|||MS c S |||? ?
?p
?(1 ? ?)
?
(4)
for some ?, ? ? (0, 1] and ? ? 1. Then, the conclusion of Theorem 1 holds, for DS TUMP applied to
e under the gap condition (3) with C/? replaced with C|||MSS |||2 /(? ? ?2 ).
input (y, X),
2
e is reasonable in cases where one can perform consistent ICA.
Access to decorrelated features, X,
This assumption is practically plausible, especially in the low-dimensional regimes, though it would
be desirable if this assumption can be removed theoretically. Moreover, we note that the response y
is based on the correlated features.
In this result, C|||MSS |||22 /(? ? ?2 ) plays the role of a new constant. There is a hard bound on how big
? can be, which via (4) controls how much correlation between off-support and on-support features
are
? tolerated. For example, taking ? = 1/9, we have ? = log(9/8) ? 0.1, ? = 17/81 ? 0.2 and
? ? 0.45 and this is about as big as it can get (the maximum we can allow is ? 0.48). ? can be
arbitrarily close to 0, relaxing the assumption (4), at the expense of increasing the constant in the
threshold. ? controls
deviation of |?j |, j ? S from uniform: in case of equal weights on the support,
?
i.e., |?j | = 1/ s for j ? S, we have ? = 1. Theorem 1 for the uncorrelated design is recovered, by
taking ? = ? = 1.
5
3.2
General Additive Model Setting
To prove results in this more general setting, we need some further regularity conditions on (fk ): Fix
some ? ? (0, 1), let U ? unif(0, 1) and assume the following about the underlying functions (fk ):
2
(F1) kfk (?U )k2?2 ? ?f,k
, ?? ? [0, 1]. (F2) var[fk (?U )] ? var[fk ((1 ? ?)U )], ?? ? 1 ? ?.
P
P
p
2
2
2
Next, we define ?f,?
:= k=1 ?f,k
= k?S ?f,k
along with the following key gap quantities:
gf,k (?) := var[fk (U ))] ? var[fk ((1 ? ?)U )].
1
Theorem 3. Assume additive Model 2 with (F1) and (F2). Let ? = log 8?
for ? ? (0, 1/8). The
DS TUMP algorithm, which selects the ?s? least impure features at the root, succeeds in model
selection, with probability at least 1 ? p??c ? 2e??n/2 if log p?/n ? C1 and
r
log p?
2
2
min gf,k (?) ? C(?f,? + ? )
(5)
k?S
n
In the supplementary material, we explore in detail the class of functions that satisfy conditions (F1)
and (F2), as well as the gap condition in (5). (F1) is relatively mild and satisfied if f is Lipschitz
or bounded. (F2) is more stringent and we show that it is satisfied for convex nondecreasing and
concave nonincreasing functions.2 The gap condition is less restrictive than (F2) and is related to the
slope of the function near the endpoint, i.e., x = 1. Notably, we study one such function that satisfies
all of these conditions, i.e., exp(?) on [?1, 1], in our simulations in Section 4.
3.3
Proof of Theorem 1
We provide the high-level proof of Theorem 1. For brevity, the proofs of the lemmas have been
omitted and can be found in the supplement, where we in fact prove them for the more general setup
of Theorem 3. The analysis boils down to understanding the behavior of y k = sor(y, xk ) as defined
k
earlier. Let yek be obtained from y k by random reshuffling of its left half y[m]
(i.e., rearranging the
entries according to a random permutation). This reshuffling has no effect on the impurity, that is,
k
k
imp(e
y[m]
) = imp(y[m]
), and the reason for it becomes clear when we analyze the case k ? S.
Understanding the distribution of y k . If k ?
/ S, the ordering according to which we sort y is
independent of y (since xk is independent of y), hence the sorted version, before and after reshuffling
has the same distribution as y. Thus, each entry of yek is an IID draw from the same distribution as
the pre-sort version:
X
iid
yeik ? W0 :=
?j Zj + w1 , i = 1, . . . , n.
(6)
j?S
On the other hand, if k ? S, then for i = 1, . . . , n
yik = ?k x(i)k + rik ,
iid
where rik ? Wk :=
X
?j Zj + w1 .
j?S\{k}
Here x(i)k is the ith order statistic of xk , that is, x(1)k ? x(2)k ? ? ? ? ? x(n)k . Note that the residual
terms are still IID since they gather the covariates (and the noise) that are independent of the kth one
and hence its ordering. Note also that rik is independent of the first term ?k x(i)k .
k
Recall that we split at the midpoint and focus on the left split, i.e., we look at y[n/2]
=
k k
k
k
k k
k
(y1 , y2 , . . . , yn/2 ), and its reshuffled version ye[n/2] = (e
y1 , ye2 , . . . , yen/2 ). Intuitively, we would
k
like to claim that the ?signal part? of the ye[n/2]
are approximately IID draws from ?k Unif(0, 1/2).
Unfortunately this is not true, in the sense that the distribution cannot be accurately approximated by
Unif(0, 1 ? ?) for any ? (Lemma 1). However, we show that the distribution can be approximated by
an infinite mixture of IID uniforms of reduced range (Lemma 2).
Let U(1) ? U(2) ? ? ? ? ? U(n) be the order statistics obtained by ordering an IID sample Ui ?
e := (U
e1 , U
e2 . . . , U
em ) be obtained from
Unif(0, 1), i = 1, . . . , n. Recall that m := n/2 and let U
2
We also observe that this condition holds for functions beyond these two categories.
6
e has an exchangeable distribution. We can write
(U(1) , . . . , U(m) ) by random permutation. Then, U
for k ? S,
yeik = ?k u
eki + reik ,
e,
u
ek ? U
and
iid
reik ? Wk , i ? [m]
where the m-vectors u
ek = (e
uki , i ? [m]) and rek = (e
rik , i ? [m]) are also independent.
e:
We have the following result regarding the distribution of U
e is a mixture of IID unif(0, ?) m-vectors with mixing variable
Lemma 1. The distribution of U
? ? Beta(m, m + 1).
Note that Beta(m, m + 1) has mean = m/(2m + 1) = (1 + o(1))/2 as m ? ?, and variance
e
= O(m?1 ). Thus, Lemma 1 makes our intuition precise in the sense that the distribution of U
is a ?range mixture? of IID uniform distributions, with the range concentrating around 1/2. We
now provide a reduced range, finite sample approximation in terms of the total variation distance
e, U
b ) between the distributions of random vectors U
e and U
b.
dTV (U
b be distributed according to a mixture of IID Unif(0, ?
b) m-vectors with ?
b distributed
Lemma 2. Let U
??
e
as a Beta(m, m + 1) truncated to (0, 1 ? ?) for ? = e /8 and ? > 0. With U as in Lemma 1, we
e, U
b ) ? 2 exp(??m).
have dTV (U
e by a truncated version, U
b , is an essential technique in
The approximation of the distribution of the U
our proof. As will become clear in the proof of Lemma 3, we will need to condition on the mixing
e , or its truncated approximation U
e , to allow for the use of concentration inequalities for
variable U
independent variables. The resulting bounds should be devoid of randomness so that by taking
expectation, we can get similar bounds for the exchangeable case. The truncation allows us to
maintain a positive gap in impurities (between on and off support features) throughout this process.
We expect the loss due to truncation to be minimal, only impacting the constants.
b described in Lemma 2,
For k ? S, let u
bk = (b
uki , i ? [m]) be drawn from the distribution of U
independently of anything else in the model, and let ?
bk be its corresponding mixing variable, which
has a Beta distribution truncated to (0, 1 ? ?). Let us define ybik = ?k u
bki + reik , i ? [m] where
k
k
k
k
re = (e
ri ) is as before. This construction provides a simple coupling between ye[m]
and yb[m]
giving
k
the same bound on the their total variation distance. Hence, we can safely work with yb[m] instead
k
of ye[m]
, and pay a price of at most 2 exp(??m) in probability bounds. To simplify discussion, let
k
ybi = yeik for k ?
/ S.
k
Concentration of empirical impurity. We will focus on yb[m]
due the discussion above. We would
k
k
k
like to control imp(b
y[m] ), the empirical variance impurity of yb[m] which is defined as in (1) with y[m]
k
k
replaced with yb[m]
. The idea is to analyze E[imp(b
y[m]
)], or proper bounds on it, and then show that
k
imp(b
y[m] ) concentrates around its mean. Let us consider the concentration first. (1) is a U-statistic of
order 2 with kernel h(u, v) = 21 (u ? v)2 . The classical Hoeffding inequality guarantees concentration
if h is uniformly bounded and the underlying variables are IID. Instead, we use a version of Hanson?
Wright concentration inequality derived in [10], which allows us to derive a concentration bound for
the empirical variance, for general sub-Gaussian vectors, avoiding the boundedness assumption:
Corollary 3. Let w = (w1 , . . . , wm ) ? Rm be a random vector with independent components wi
?1 P
2
which satisfy Ewi = ? and kwi ? ?k?2 ? K. Let imp(w) := m
1?i<j?m (wi ? wj ) be the
2
empirical variance of w. Then, for u ? 0,
P imp(w) ? E imp(w) > K 2 u ? 2 exp ?c (m ? 1) min(u, u2 ) .
(7)
We can immediately apply this result when k ?
/ S. However, for k ? S, a more careful application is
k
needed since we can only guarantee an exchangeable distribution for yb[m]
in this case. The following
lemma summarizes the conclusions:
7
k
Lemma 3. Let Ibm,k = imp(b
y[m]
) and recall that ? was introduced in the definition of ybik . Let
1
?21 := 12
be the variance of Unif(0, 1). Recall that p? := p ? s. Let L = k?k2 . There exist absolute
constants C1 , C2 , c such that if log p?/m ? C1 , then with probability at least 1 ? p??c ,
Ibm,k ? Ik1 + ?m , ?k ? S,
and,
Ibm,k ? I 0 ? ?m , ?k ?
/S
where, letting ? := 1 ? (1 ? ?)2 ,
Ik1 := ?21 (???k2 + L2 ) + ? 2 ,
I 0 := ?21 L2 + ? 2 , and ?m := C2 (L2 + ? 2 )
p
log p?/m.
The key outcome of Lemma 3 is that, on average, there is a positive gap I 0 ? Ik1 = ?21 ??k2 in
impurities between a feature on the support and those off of it, and that due to concentration, the
fluctuations in impurities will be less than this gap for large m. Combined with Lemma 2, we can
k
transfer the results to Iem,k := imp(e
y[m]
).
Corollary 4. The conclusion of Lemma 3 holds for Iem,k in place of Ibm,k , with probability at least
1
1 ? p??c ? 2e??m for ? = log 8?
.
Note that for ? < 1/8, the bound holds with high probability. Thus, as long as I 0 ? Ik1 > 2?m , the
selection algorithm correctly favors the kth feature in S, over the inactive ones (recall that lower
impurity is better). We have our main result after substituting n/2 for m.
4
Simulations
(a)
(b)
(c)
(d)
(e)
(f)
Figure 1: Support recovery performance in a linear regression model augmented with possible
nonlinearities for n = 1024. (a) Linear case with uncorrelated design. (b) Linear case with correlated
design. (c) Nonlinear additive model with exponentials of covariates and uncorrelated design. (d)
Nonlinear model with interaction terms and uncorrelated design. (e) Nonlinear additive model with
exponentials of covariates, interaction terms, and uncorrelated design. (f) Nonlinear additive model
with exponentials of covariates, interaction terms, and correlated design.
In order to corroborate the theoretical analysis, we next present various simulation results. We
consider the following model: y = X? + f (XS ) + w, where f (XS ) is a potential nonlinearity, and
e ? Rn?p is a random
e where X
S is the true support of ?. We generate the training data as X = XM
p?p
matrix with IID Unif(?1, 1) entries, and M ? R
is an upper-triangular matrix that determines
whether the design is IID or correlated. In the IID case we set M = I. To achieve a correlated design
8
we randomly assign values from {0, ??, +?} to the upper triangular cells of M , with probabilities
(1 ? 2?, ?, ?). We observed qualitatively similar results for various values of ? and ? and here we
present results with ? = 0.04, and ?
? = 0.1. The noise is generated as w ? N (0, ? 2 In ). We fix
p = 200, ? = 0.1, and let ?i = ?1/ s over its support i ? S, where |S| = s. That is, only s of the
p = 200 variables are predictive of the response. The nonlinearity, f , optionally contains additive
terms in the form of exponentials of on-support covariates. It can also contain interaction terms across
on-support covariates, i.e., terms of the form ?2s xi xj for some randomly selected pairs of i, j ? S.
Notably, the choice of f is unknown to the variable selection methods. We vary s ? [5, 100] and note
that k?k2 = 1 remains fixed.
The plots in Figure 1 show the fraction of the true support recovered3 as a function of s, for various
methods under different modeling setups: f = 0 (linear), f = 2 exp(?) (additive), f = interaction
(interactions), and f = interaction + 2 exp(?) (interactions+additive) with IID or correlated designs.
Each data point is an average over 100 trials (see supplementary material for results with 95%
confidence intervals). In addition to DS TUMP, we evaluate T REE W EIGHT, S PAM, L ASSO, SIS and
random guessing for comparison. SIS refers to picking the indices of the top s largest values of X T y
in absolute value. When X is orthogonal and the generative model is linear, this approach is optimal,
and we use it as a surrogate for the optimal approach in our nearly orthogonal setup (i.e., the IID
linear case), due to its lack of any tuning parameters. Random guessing is used as a benchmark, and
as expected, on average recovers the fraction s/p = s/200 of the support.
The plots show that, in the linear setting, the performance of DS TUMP is comparable to, and only
slightly worse than, that of SIS or Lasso which are considered optimal in this case. Figure 1(b) shows
that under mildly correlated design the gap between DS TUMP and L ASSO widens. In this case, SIS
loses its optimality and performs at the same level as DS TUMP. This matches our intuition as both
SIS and DS TUMP are both greedy methods that consider covariates independently.
DS TUMP is more robust to nonlinearities, as characterized theoretically in Theorem 3 and evidenced
in Figure 1(c). In contrast, in the presence of exponential nonlinearities, SIS and Lasso are effective in
the very sparse regime of s p, but quickly approach random guessing as s grows. In the presence
of interaction terms, T REE W EIGHT and to a lesser extent S PAM outperform all other methods, as
shown in Figure 1(d), 1(e), and 1(f). We also note that the permutation-based importance method [1],
denoted by T REE W EIGHT P ERMUTATION in the plots in Figure 1, performs substantially worse than
T REE W EIGHT across the various modelling settings.
Overall, these simulations illustrate the promise of multi-level tree-based methods like T REE W EIGHT
under more challenging and realistic modeling settings. Future work involves generalizing our
theoretical analyses to extend to these more complex multi-level tree-based approaches.
5
Discussion
We presented a simple model selection algorithm for decision trees, which we called DS TUMP,
and analyzed its finite-sample performance in a variety of settings, including the high-dimensional,
nonlinear additive model setting. Our theoretical and experimental results show that even a simple
tree-based algorithm that selects at the root can achieve high dimensional selection consistency.
We hope these results pave the way for the finite-sample analysis of more refined tree-based model
selection procedures. Inspired by the empirical success of T REE W EIGHT in nonlinear settings, we are
actively looking at extensions of DS TUMP to a multi-stage algorithm capable of handling interactions
with high-dimensional guarantees.
Moreover, while we mainly focused on the regression problem, our proof technique based on
concentration of impurity reductions is quite general. We expect analogous results to hold, for example
for classification. However, aspects of the proof would be different, since impurity measures used for
classification are different than those of regression. One major hurdle involves deriving concentration
inequalities for the empirical versions of these measures, which are currently unavailable, and would
be of independent interest.
3
In the supplementary material we report analogous results using a more stringent performance metric,
namely the probability of exact support recovery. The results are qualitatively similar.
9
References
[1] L. Breiman. Random forests. Machine learning, 45(1):5?32, 2001.
[2] L. Breiman, J. Friedman, C. J. Stone, and R. A. Olshen. Classification and regression trees.
CRC press, 1984.
[3] J. Fan and J. Lv. Sure independence screening for ultrahigh dimensional feature space. Journal
of the Royal Statistical Society: Series B, 70(5), 2008.
[4] J. Fan and J. Lv. Sure independence screening for ultrahigh dimensional feature space. Journal
of the Royal Statistical Society: Series B (Statistical Methodology), 70(5):849?911, 2008.
[5] J. H. Friedman. Greedy function approximation: A gradient boosting machine. The Annals of
Statistics, 29(5):1189?1232, 2001.
[6] H. Ishwaran. Variable importance in binary regression trees and forests. Electronic Journal of
Statistics, 1:519?537, 2007.
[7] J. Lafferty and L. Wasserman. Rodeo: Sparse, Greedy Nonparametric Regression. Annals of
Statistics, 36(1):28?63, 2008.
[8] G. Louppe, L. Wehenkel, A. Sutera, and P. Geurts. Understanding variable importances in
forests of randomized trees. In Advances in Neural Information Processing Systems 26. 2013.
[9] P. Ravikumar, J. Lafferty, H. Liu, and L. Wasserman. Sparse additive models. Journal of the
Royal Statistical Society: Series B, 71(5):1009?1030, 2009.
[10] M. Rudelson and R. Vershynin. Hanson-Wright inequality and sub-gaussian concentration.
Electron. Commun. Probab, pages 1?10, 2013.
[11] M. J. Wainwright. Information-theoretic limits on sparsity recovery in the high-dimensional
and noisy setting. IEEE Transactions on Information Theory, 55(12):5728?5741, 2009.
[12] M. J. Wainwright. Sharp thresholds for high-dimensional and noisy sparsity Recovery `1 constrained quadratic programming (Lasso). IEEE Transactions on Information Theory,
55(5):2183?2202, 2009.
10
| 6646 |@word mild:1 trial:1 version:8 norm:4 unif:8 simulation:7 r:1 simplifying:1 covariance:1 boundedness:1 reduction:13 liu:1 contains:1 score:7 series:3 ours:1 existing:2 xnj:1 current:1 comparing:1 recovered:1 si:12 yet:1 additive:17 realistic:2 thrust:1 numerical:1 plot:3 greedy:4 leaf:1 selected:2 generative:3 half:4 xk:9 yi1:1 ith:1 provides:2 boosting:1 node:10 along:1 c2:2 become:2 beta:4 ik:2 prove:2 sutera:1 combine:1 introduce:2 theoretically:3 notably:2 ica:4 hardness:1 expected:1 themselves:1 indeed:1 behavior:2 multi:4 inspired:2 actual:1 equipped:1 cardinality:2 considering:3 increasing:1 becomes:1 ye2:1 moreover:3 underlying:3 notation:2 bounded:2 substantially:1 z:2 guarantee:7 safely:1 berkeley:2 every:1 collecting:1 concave:1 k2:8 rm:1 exchangeable:3 partitioning:1 control:3 omit:1 yn:1 before:3 positive:3 understood:1 local:1 limit:3 consequence:1 analyzing:2 establishing:2 ree:13 path:2 approximately:3 fluctuation:1 pam:5 studied:5 challenging:4 relaxing:1 ease:1 ms:3 range:4 averaged:1 recursive:1 practice:3 procedure:9 empirical:7 significantly:1 pre:1 confidence:1 refers:1 suggest:1 get:3 convenience:3 close:1 selection:35 operator:3 cannot:1 restriction:1 equivalent:1 straightforward:1 attention:1 go:1 independently:2 convex:1 focused:2 simplicity:1 splitting:1 assigns:1 immediately:3 recovery:6 wasserman:2 insight:1 deriving:3 variation:2 analogous:2 annals:2 controlling:1 play:1 construction:1 exact:2 programming:1 expensive:1 particularly:1 satisfying:1 approximated:2 predicts:1 observed:2 role:1 louppe:1 wj:1 ordering:4 trade:1 removed:1 intuition:2 complexity:5 covariates:9 ui:1 depend:1 impurity:28 predictive:4 f2:5 basis:3 ewi:1 various:8 distinct:2 describe:1 effective:2 outcome:1 refined:1 whose:1 quite:2 widely:1 valued:1 supplementary:4 plausible:1 otherwise:2 triangular:2 favor:1 statistic:6 nondecreasing:1 itself:1 noisy:3 eki:1 sequence:1 propose:1 interaction:12 maximal:1 turned:1 mixing:3 poorly:1 flexibility:1 achieve:2 constituent:1 parent:1 regularity:2 zp:2 adam:2 derive:2 coupling:1 illustrate:1 stat:1 strong:1 involves:2 concentrate:2 subsequently:1 arash:1 stringent:2 material:4 crc:1 sor:4 assign:2 fix:2 generalization:1 f1:4 investigation:3 extension:1 hold:8 practically:1 around:2 considered:3 wright:2 exp:6 algorithmic:2 claim:1 electron:1 substituting:1 major:1 vary:1 smallest:1 omitted:1 applicable:1 currently:1 largest:1 weighted:1 hope:1 clearly:1 reshuffling:3 gaussian:5 breiman:2 corollary:8 derived:3 focus:4 modelling:1 mainly:2 contrast:3 talwalkar:2 baseline:2 burdensome:1 sense:2 xip:1 going:1 i1:2 selects:4 overall:1 among:3 classification:3 denoted:1 impacting:1 constrained:1 special:3 uc:1 equal:2 once:2 beach:1 look:1 nearly:2 imp:14 future:1 report:2 simplify:1 employ:1 randomly:2 sparsistency:1 replaced:3 n1:1 maintain:1 iem:2 friedman:2 interest:2 screening:2 highly:1 withp:1 mixture:4 analyzed:1 bki:1 nonincreasing:1 amenable:2 capable:1 orthogonal:3 unless:1 tree:21 re:1 theoretical:9 minimal:1 yek:2 column:2 modeling:4 earlier:1 corroborate:2 retains:1 deviation:1 subset:5 entry:6 predictor:6 uniform:5 examining:1 characterize:2 tolerated:1 combined:1 vershynin:1 st:1 devoid:1 randomized:1 xi1:2 off:4 picking:1 quickly:1 w1:4 satisfied:2 hoeffding:1 worse:2 ek:2 leading:1 actively:1 account:1 nonasymptotic:1 nonlinearities:4 potential:1 unordered:1 summarized:1 wk:2 coefficient:1 satisfy:2 ranking:1 root:8 picked:1 analyze:3 start:1 recover:1 sort:3 wm:1 defer:1 slope:1 yen:1 ni:1 accuracy:3 ir:1 variance:9 ensemble:2 yield:1 ybi:1 generalize:2 accurately:1 iid:20 randomness:1 decorrelated:1 definition:1 against:2 pp:1 e2:1 naturally:1 proof:14 recovers:2 boil:1 couple:1 proved:2 tunable:1 concentrating:1 recall:6 knowledge:1 x1j:1 back:1 higher:1 methodology:1 response:4 yb:6 though:2 implicit:1 stage:1 correlation:5 d:42 hand:1 replacing:1 nonlinear:14 lack:1 google:1 perhaps:1 grows:1 usa:1 grounding:1 building:1 effect:1 true:4 y2:1 ye:4 contain:1 hence:6 read:1 i2:2 attractive:1 anything:1 m:1 stone:1 theoretic:2 demonstrate:2 workhorse:1 performs:3 geurts:1 fj:2 novel:3 common:2 empirically:1 conditioning:1 endpoint:1 extend:2 significant:1 automatic:1 tuning:1 consistency:9 outlined:1 similarly:1 fk:7 nonlinearity:2 ik1:4 access:1 f0:2 impressive:1 base:1 closest:1 perspective:1 moderate:1 driven:1 belongs:1 commun:1 inequality:7 binary:1 arbitrarily:1 success:1 yi:6 seen:1 minimum:1 additional:2 signal:1 impure:3 smoother:1 full:1 ii:2 desirable:1 asso:4 technical:3 match:1 characterized:3 offer:1 long:3 cross:1 e1:1 ravikumar:1 calculates:1 prediction:2 variant:1 regression:14 converging:1 basic:1 cmu:2 expectation:1 metric:1 kernel:1 tailored:3 achieved:1 cell:1 c1:7 justified:1 addition:2 whereas:1 remarkably:1 separately:1 interval:3 hurdle:1 else:1 grow:1 median:2 reshuffled:1 crucial:1 unlike:1 posse:2 sure:2 kwi:2 lafferty:2 effectiveness:1 practitioner:1 call:1 vw:2 leverage:1 near:1 uki:2 split:11 iii:2 wn:1 presence:2 variety:1 xj:7 affect:1 zi:2 independence:2 dtv:2 lasso:7 competing:1 bandwidth:1 regarding:1 idea:1 lesser:1 shift:1 inactive:3 whether:1 motivated:1 render:1 yik:1 clear:4 transforms:1 nonparametric:4 extensively:2 category:1 reduced:3 generate:1 outperform:1 xij:2 exist:1 zj:2 correctly:1 write:3 promise:1 key:2 threshold:5 achieving:1 drawn:2 fraction:2 sum:1 place:1 throughout:4 reasonable:1 electronic:1 parsimonious:1 draw:4 decision:4 summarizes:1 scaling:1 comparable:1 bound:10 pay:2 fan:2 quadratic:1 paramount:1 precisely:1 ri:1 ermutation:1 ucla:4 aspect:1 argument:1 extremely:1 min:11 optimality:1 performing:1 ameet:1 relatively:2 xtk:1 according:3 remain:1 across:2 em:1 slightly:1 wi:7 intuitively:1 invariant:1 computationally:2 remains:1 discus:1 xi2:1 needed:1 letting:1 end:1 serf:2 studying:1 ishwaran:1 eight:13 observe:2 apply:3 appropriate:1 amini:1 worthwhile:1 occurrence:1 appearing:1 alternative:1 robustness:1 rp:1 bagging:1 assumes:1 top:3 cf:2 rudelson:1 wehenkel:1 widens:1 giving:1 restrictive:1 especially:2 classical:1 society:3 quantity:3 parametric:1 concentration:12 usual:1 pave:1 guessing:3 surrogate:1 exhibit:1 gradient:1 kth:2 distance:2 w0:1 considers:4 extent:1 trivial:1 reason:1 index:4 mini:1 optionally:1 setup:5 olshen:1 mostly:1 yir:1 unfortunately:1 potentially:1 yjk:1 expense:2 stated:3 mink:1 kfk:1 design:36 proper:1 unknown:2 perform:1 upper:2 observation:1 benchmark:2 finite:8 truncated:4 immediate:1 looking:1 precise:1 y1:2 rn:9 arbitrary:1 sharp:1 bk:2 introduced:1 namely:3 required:1 pair:1 extensive:2 z1:2 evidenced:1 hanson:2 established:3 nip:1 beyond:1 proceeds:1 usually:1 xm:2 regime:4 sparsity:5 challenge:1 summarize:1 interpretability:1 including:3 royal:3 wainwright:2 unrealistic:1 rely:1 residual:1 minimax:3 identifies:1 categorical:1 gf:2 probab:1 literature:1 understanding:4 l2:3 relative:2 asymptotic:2 lacking:1 loss:1 expect:2 permutation:5 ultrahigh:2 var:5 ingredient:1 lv:2 validation:1 rik:4 sufficient:2 consistent:2 gather:1 uncorrelated:18 ibm:4 row:2 course:1 truncation:2 jth:2 enjoys:1 aij:1 allow:4 taking:3 midpoint:3 sparse:11 absolute:5 distributed:2 calculated:1 dimension:2 rek:1 rich:1 ignores:1 forward:2 made:2 commonly:1 qualitatively:2 simplified:2 spam:1 transaction:2 approximate:2 compact:1 active:6 summing:1 xi:4 thep:1 search:1 continuous:1 additionally:1 transfer:1 robust:1 rearranging:1 correlated:17 ignoring:1 ca:1 rodeo:3 forest:8 unavailable:1 expansion:2 complex:2 yi2:1 main:1 s2:1 noise:5 big:2 child:3 augmented:1 sub:4 exponential:5 tied:2 removing:1 theorem:15 down:1 maxi:2 x:2 essential:1 importance:14 supplement:1 gap:10 mildly:1 generalizing:1 likely:1 explore:1 kxk:1 ordered:1 subtlety:1 u2:1 applies:2 loses:1 determines:2 satisfies:1 conditional:1 goal:1 sorted:2 exposition:1 careful:1 lipschitz:1 price:1 hard:1 infinite:2 uniformly:3 lemma:14 called:2 total:4 x2j:1 experimental:2 xin:1 succeeds:4 e:2 support:15 brevity:1 violated:1 evaluate:4 avoiding:1 handling:1 |
6,242 | 6,647 | Preventing Gradient Explosions
in Gated Recurrent Units
Sekitoshi Kanai, Yasuhiro Fujiwara, Sotetsu Iwamura
NTT Software Innovation Center
3-9-11, Midori-cho, Musashino-shi, Tokyo
{kanai.sekitoshi, fujiwara.yasuhiro, iwamura.sotetsu}@lab.ntt.co.jp
Abstract
A gated recurrent unit (GRU) is a successful recurrent neural network architecture
for time-series data. The GRU is typically trained using a gradient-based method,
which is subject to the exploding gradient problem in which the gradient increases
significantly. This problem is caused by an abrupt change in the dynamics of the
GRU due to a small variation in the parameters. In this paper, we find a condition
under which the dynamics of the GRU changes drastically and propose a learning
method to address the exploding gradient problem. Our method constrains the
dynamics of the GRU so that it does not drastically change. We evaluated our
method in experiments on language modeling and polyphonic music modeling. Our
experiments showed that our method can prevent the exploding gradient problem
and improve modeling accuracy.
1
Introduction
Recurrent neural networks (RNNs) can handle time-series data in many applications such as speech
recognition [14, 1], natural language processing [26, 30], and hand writing recognition [13]. Unlike
feed-forward neural networks, RNNs have recurrent connections and states to represent the data.
Back propagation through time (BPTT) is a standard approach to train RNNs. BPTT propagates the
gradient of the cost function with respect to the parameters, such as weight matrices, at each layer and
at each time step by unfolding the recurrent connections through time. The parameters are updated
using the gradient in a way that minimizes the cost function. The cost function is selected according
to the task, such as classification or regression.
Although RNNs are used in many applications, they have problems in that the gradient can be
extremely small or large; these problems are called the vanishing gradient and exploding gradient
problems [5, 28]. If the gradient is extremely small, RNNs can not learn data with long-term
dependencies [5]. On the other hand, if the gradient is extremely large, the gradient moves the RNNs
parameters far away and disrupts the learning process. To handle the vanishing gradient problem,
previous studies [18, 8] proposed sophisticated models of RNN architectures. One successful model
is a long short-term memory (LSTM). However, the LSTM has the complex structures and numerous
parameters with which to learn the long-term dependencies. As a way of reducing the number of
parameters while avoiding the vanishing gradient problem, a gated recurrent unit (GRU) was proposed
in [8]; the GRU has only two gate functions that hold or update the state which summarizes the past
information. In addition, Tang et al. [33] show that the GRU is more robust to noise than the LSTM
is, and it outperforms the LSTM in several tasks [9, 20, 33, 10].
Gradient clipping is a popular approach to address the exploding gradient problem [26, 28]. This
method rescales the gradient so that the norm of the gradient is always less than a threshold. Although
gradient clipping is a very simple method and can be used with GRUs, it is heuristic and does not
analytically derive the appropriate threshold. The threshold has to be manually tuned to the data
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
and tasks by trial and error. Therefore, a learning method is required to more effectively address the
exploding gradient problem in training of GRUs.
In this paper, we propose a learning method for GRUs that addresses the exploding gradient problem.
The method is based on an analysis of the dynamics of GRUs. GRUs suffer from gradient explosions
due to their nonlinear dynamics [11, 28, 17, 3] that enable GRUs to represent time-series data. The
dynamics can drastically change when the parameters cross certain values, called bifurcation points
[36], in the learning process. Therefore, the gradient of the state with respect to the parameters can
drastically increase at a bifurcation point. This paper presents an analysis of the dynamics of GRUs
and proposes a learning method to prevent the parameters from crossing the bifurcation point. It
describes evaluations of this method through language modeling and polyphonic music modeling
experiments. The experiments demonstrate that our method can train GRUs without gradient clipping
and that it can improve the accuracy of GRUs.
The rest of this paper is organized as follows: Section 2 briefly explains the GRU, dynamical systems
and the exploding gradient problem. It also outlines related work. The dynamics of GRUs is analyzed
and our training approach is presented in Section 3. The experiments that verified our method
are discussed in Section 4. The paper concludes in Section 5. Proofs of lemmas are given in the
supplementary material.
2
2.1
Preliminaries
Gated Recurrent Unit
Time-series data often have long and short-term dependencies. In order to model long and short-term
behavior, a GRU is designed to properly keep and forget past information. The GRU controls the past
information by having two gates: an update gate and reset gate. The update gate zt ? Rn?1 at a time
step t is expressed as
zt = sigm(Wxz xt + Whz ht?1 ),
(1)
where xt ? Rm?1 is the input vector, and ht ? Rn?1 is the state vector. Wxz ? Rn?m and
Whz ? Rn?n are weight matrices. sigm(?) represents the element-wise logistic sigmoid function.
The reset gate rt ? Rn?1 is expressed as
rt = sigm(Wxr xt + Whr ht?1 ),
(2)
where Wxr ? Rn?m and Whr ? Rn?n are weight matrices. The activation of the state ht is
expressed as
? t,
ht = zt ht?1 + (1 ? zt ) h
(3)
? t is a candidate for a new
where 1 is the vector of all ones, and means the element-wise product. h
state, expressed as
? t = tanh(Wxh xt + Whh (rt ht?1 )),
h
(4)
where tanh(?) is the element-wise hyperbolic tangent, and Wxh ? Rn?m and Whh ? Rn?n are
weight matrices. The initial value of ht is h0 = 0 where 0 represents the vector of all zeros; the
GRU completely forgets the past information when ht becomes 0.
The training of a GRU can be formulated as an optimization problem as follows:
PN
min? N1 j=1 C(x(j) , y (j) ; ?),
(j)
(j)
(j)
(5)
(j)
where ?, x , y , C(x , y ; ?), and N are all parameters of the model (e.g., elements of
Whh ), the j-th training input data, the j-th training output data, the loss function for the j-th data
(e.g., mean squared error or cross entropy), and the number of training data, respectively. This
optimization problem is usually solved through stochastic gradient descent (SGD). SGD iteratively
updates parameters according to the gradient of a mini-batch, which is randomly sampled data from
the training data. The parameter update at step ? is
P
(6)
? (? ) = ? (? ?1) ? ??? |D1? | (x(j) ,y(j) )?D? C(x(j) , y (j) ; ?),
where D? , |D? |, and ? represent the ? -th mini-batch, the size of the
P mini-batch, and the learning rate
of SGD, respectively. In gradient clipping, the norm of ?? |D1? | (x(j) ,y(j) )?D? C(x(j) , y (j) ; ?) is
clipped by the specified threshold. The size of the parameters ? is 3(n2 + mn) + ?, where ? is the
number of parameters except for the GRU, because the sizes of the six weight matrices of W? in eqs.
(1)-(4) are n?n or n?m. Therefore, the computational cost of gradient clipping is O(n2 + mn + ?).
2
2.2
Dynamical System and Gradient Explosion
An RNN is a nonlinear dynamical system that can be represented as follows:
ht = f (ht?1 , ?),
(7)
where ht is a state vector at time step t, ? is a parameter vector, and f is a nonlinear vector function.
The state evolves over time according to eq. (7). If the state ht? at some time step t? satisfies
ht? = f (ht? , ?), i.e., the new state equals the previous state, the state never changes until an external
input is applied to the system. Such a state point is called a fixed point h? . The state converges to or
goes away from the fixed point h? depending on f and ?. This property is important and is called
stability [36]. The fixed point h? is said to be locally stable if there exists a constant ? such that, for
ht whose initial value h0 satisfies |h0 ? h? | < ?, limt?? |ht ? h? | = 0 holds. In this case, a set of
points h0 such that |h0 ? h? | < ? is called a basin of attraction of the fixed point. Conversely, if
h? is not stable, the fixed point is said to be unstable. Stability and the behavior of ht near a fixed
point, e.g., converging or diverging, can be qualitatively changed by a smooth variation in ?. This
phenomenon is called a local bifurcation, and the value of the parameter of a bifurcation is called a
bifurcation point [36].
Doya [11], Pascanu et al. [28] and Baldi and Hornik [3] pointed out that gradient explosions are due
to bifurcations. The training of an RNN involves iteratively updating its parameters. This process
causes a bifurcation: a small change in parameters can result in a drastic change in the behavior of
the state. As a result, the gradient increases at a bifurcation point.
2.3
Related Work
Kuan et al. [23] established a learning method to avoid the exploding gradient problem. This method
restricts the dynamics of an RNN so that the state remains stable. Yu [37] proposed a learning rate for
stable training through Lyapunov functions. However, these methods mainly target Jordan and Elman
networks called simple RNNs which, unlike GRUs, are difficult to train long-term dependencies.
In addition, they suppose that the mean squared error is used as the loss function. By contrast, our
method targets the GRU, a more sophisticated model, and can be used regardless of the loss function.
Doya [11] showed that bifurcations cause gradient explosions and that real-time recurrent learning
(RTRL) can train an RNN without the gradient explosion. However, RTRL has a high computational
cost: O((n + u)4 ) for each update step where u is the number of output units [19]. More recently,
Arjovsky et al. [2] proposed unitary matrix constraints in order to prevent the gradient vanishing
and exploding. Vorontsov et al. [35], however, showed that it can be detrimental to maintain hard
constraints on matrix orthogonality.
Previous studies analyzed the dynamics of simple RNNs [12, 4, 31, 16, 27]. Barabanov and Prokhorov
[4] analyzed the absolute stability of multi-layer simple RNNs. Haschke and Steil [16] presented a
bifurcation analysis of a simple RNN in which inputs are regarded as the bifurcation parameter. Few
studies have analyzed the dynamics of the modern RNN models. Talathi and Vartak [32] analyzed
the nonlinear dynamics of an RNN with a Relu nonlinearity. Laurent and von Brecht [24] empirically
revealed that LSTMs and GRUs can exhibit chaotic behavior and proposed a novel model that has
stable dynamics. To the best of our knowledge, our study is the first to analyze the stability of GRUs.
3
Proposed Method
As mentioned in Section 2, a bifurcation makes the gradient explode. In this section, through an
analysis of the dynamics of GRUs, we devise a training method that avoids a bifurcation and prevents
the gradient from exploding.
3.1
Formulation of Proposed Training
In Section 3.1, we formulate our training approach. For the sake of clarity, we first explain the
formulation for a one-layer GRU; then, we apply the method to a multi-layer GRU.
3.1.1
One-Layer GRU
The training of a GRU is formulated as eq. (5). This training with SGD can be disrupted by a gradient
explosion. To prevent the gradient from exploding, we formulate the training of a one-layer GRU as
3
the following constrained optimization:
PN
min? N1 j=1 C(x(j) , y (j) ; ?), s.t. ?1 (Whh ) < 2,
(8)
where ?i (?) is the i-th largest singular value of a matrix, and ?1 (?) is called the spectral norm. This
constrained optimization problem keeps the one-layer GRU locally stable and prevents the gradient
from exploding due to a bifurcation of the fixed point on the basis of the following theorem:
Theorem 1. When ?1 (Whh ) < 2, a one-layer GRU is locally stable at a fixed point h? = 0.
We show the proof of this theorem later. This theorem indicates that our training approach of eq. (8)
maintains the stability of the fixed point h? = 0. Therefore, our approach prevents the gradient
explosion caused by the bifurcation of the fixed point h? . In order to prove this theorem, we need to
use the following three lemmas:
Lemma 1. A one-layer GRU has a fixed point at h? = 0.
Lemma 2. Let I be an n ? n identity matrix, ?i (?) be the eigenvalue that has the i-th largest absolute
value, and J = 14 Whh + 12 I. When the spectral radius 1 |?1 (J )| < 1, a one-layer GRU without
input can be approximated by the following linearized GRU near ht = 0:
ht = J ht?1 ,
(9)
and the fixed point h? = 0 of a one-layer GRU is locally stable.
Lemma 2 indicates that we can prevent a change in local stability by exploiting the constraint of
|?1 ( 14 Whh + 21 I)| < 1. This constraint can be represented as a bilinear matrix inequality (BMI)
constraint [7]. However, an optimization problem with a BMI constraint is NP-hard [34]. Therefore,
we relax the optimization problem to that of a singular value constraint as in eq. (8) by using the
following lemma:
Lemma 3. When ?1 (Whh ) < 2, we have |?1 ( 14 Whh + 12 I)| < 1.
By exploiting Lemmas 1, 2, and 3, we can prove Theorem 1 as follows:
Proof. From Lemma 1, there exists a fixed point h? = 0 in a one-layer GRU. This fixed point is
locally stable when |?1 ( 14 Whh + 21 I)| < 1 from Lemma 2. From Lemma 3, |?1 ( 14 Whh + 12 I)| < 1
holds when ?1 (Whh ) < 2. Therefore, when ?1 (Whh ) < 2, the one-layer GRU is locally stable at
the fixed point h? = 0
Lemma 1 indicates that a one-layer GRU has a fixed point. Lemma 2 shows the condition under
which this fixed point is kept stable. Lemma 3 shows that we can use a singular value constraint
instead of an eigenvalue constraint. These lemmas prove Theorem 1, and this theorem ensures that
our method prevents the gradient from exploding because of a local bifurcation.
In our method of eq. (8), h? = 0 is a fixed point. This fixed point is important since the initial
value of the state h0 is 0, and the GRU forgets all the past information when the state is reset to 0
as described in Section 2. If h? = 0 is stable, the state vector near 0 asymptotically converges to 0.
This means that the state vector ht can be reset to 0 after a sufficient time in the absence of an input;
i.e., the GRU can forget the past information entirely. On the other hand, when |?1 (J )| becomes
greater than one, the fixed point at 0 becomes unstable. This means that the state vector ht never
resets to 0; i.e., the GRU can not forget all the past information until we manually reset the state. In
this case, the forget gate and reset gate may not work effectively. In addition, Laurent and von Brecht
[24] show that an RNN model with state that asymptotically converges to zero achieves a level of
performance comparable to that of LSTMs and GRUs. Therefore, our constraint that the GRU is
locally stable at h? = 0 is effective for learning.
3.1.2
Multi-Layer GRU
Here, we extend our method in the multi-layer GRU. An L-layer GRU is represented as follows:
h1,t = f1 (h1,t?1 , xt ), h2,t = f2 (h2,t?1 , h1,t ), . . . , hL,t = fL (hL,t?1 , hL?1,t ),
where hl,t ? Rnl ?1 is a state vector with the length of nl at the l-th layer, and fl represents a
GRU that corresponds to eqs. (1)-(4) at the l-th layer. In the same way as the one-layer GRU,
T T
ht = [hT
1,t , . . . , hL,t ] = 0 is a fixed point, and we have the following lemma:
1
The spectral radius is the maximum value of the absolute value of the eigenvalues.
4
Lemma 4. When |?1 ( 14 Wl,hh + 12 I)| < 1 for l = 1, . . . , L, the fixed point h? = 0 of a multi-layer
GRU is locally stable.
From Lemma 3, we have |?1 (Wl,hh + 21 I)| < 1 when ?1 (Wl,hh ) < 2. Thus, we formulated our
training of a multi-layer GRU to prevent gradient explosions as
PN
min? N1 j=1 C(x(j) , y (j) ; ?), s.t. ?1 (Wl,hh ) < 2, ?1 (Wl,xh )? 2 for l = 1, . . . , L.
(10)
We added the constraint ?1 (Wl,xh ) ? 2 in order to prevent the input from pushing the state out of
the basin of attraction of the fixed point h? = 0. This constrained optimization problem keeps a
multi-layer GRU locally stable.
3.2
Algorithm
The optimization method for eq. (8) needs to find the optimal parameters in the feasible set, in which
the parameters satisfy the constraint: {Whh |Whh ? Rn?n , ?1 (Whh ) < 2}. Here, we modify SGD
in order to solve eq. (8). Our method updates the parameters as follows:
(? ?1)
(? )
(? )
(? ?1)
??Whh = ??Whh ? ??? CD? (?), Whh = P? (Whh
? ??Whh CD? (?)),
(11)
P
(?
)
where CD? (?) represents |D1? | (x(j) ,y(j) )?D? C(x(j) , y (j) ; ?), and ??Whh represents the parame(? )
ters except for Whh . In eq. (11), We compute P? (?) by using the following procedure:
(? )
(? ?1)
? := W
Step 1. Decompose W
hh
hh ???WhhCD?(?) by using singular value decomposition (SVD):
? (? ) = U ?V.
W
hh
Step 2. Replace the singular values that are greater than the threshold 2 ? ?:
? = diag(min(?1 , 2 ? ?), . . . min(?n , 2 ? ?)).
?
(12)
(13)
(? )
? in Steps 1 and 2:
Step 3. Reconstruct Whh by using U , V and ?
(? )
?
Whh ? U ?V.
(14)
By using this procedure, Whh is guaranteed to have a spectral norm of less than or equal to 2 ? ?.
When ? is 0 < ? < 2, our method constrains ?1 (Whh ) to be less than 2. P? (?) in our method brings
back the parameters into the feasible set when the parameters go out the feasible set after SGD. Our
procedure P? (?) is an optimal projection into the feasible set as shown by the following lemma:
(? )
Lemma 5. The weight matrix Whh obtained by P? (?) is a solution of the following optimization:
? (? ) ? W (? ) ||2 , s.t. ?1 (W (? ) ) ? 2??, where || ? ||2 represents the Frobenius norm.
minW (? ) ||W
F
hh
hh F
hh
hh
Lemma 5 indicates that our method can bring back the weight matrix into the feasible set with
minimal variations in the parameters. Therefore, our procedure P? (?) has minimal impact on the
minimization of the loss function. Note that our method does not depend on the learning rate schedule,
and an adaptive learning rate method (such as Adam [21]) can be used with it.
3.3
Computational Cost
Let n be the length of a state vector ht ; a naive implementation of SVD needs O(n3 ) time. Here, we
propose an efficient method to reduce the computational cost. First, let us reconsider the computation
of P? (?). Equations (12)-(14) can be represented as follows:
h
i
(? )
T
? (? ) ? Ps
? (? )
Whh = W
(15)
i=1 ?i (Whh ) ? (2 ? ?) ui vi ,
hh
where s is the number of the singular values greater than 2 ? ?, and ui and vi are the i-th left and
right singular vectors, respectively. Eq. (15) shows that our method only needs the singular values
? (? ) ) > 2 ? ?. In order to reduce the computational cost of our method,
and vectors such that ?i (W
hh
we use the truncated SVD [15] to efficiently compute the top s singular values in O(n2 log(s)) time,
where s is the specified number of singular values. Since the truncated SVD requires s to be set
beforehand, we need to efficiently estimate the number of singular values such that must meet the
? (? ) ) > 2 ? ?. Therefore, we compute upper bounds of the singular values that
condition of ?i (W
hh
meet the condition on the basis of the following lemma:
5
? (? ) ) ?
? (? ) are bounded with the following inequality: ?i (W
Lemma 6. The singular values of W
hh
hh
(? ?1)
?i (Whh ) + |?|||?Whh CD? (?)||F .
Using this upper bound, we can estimate s as the number of the singular values with upper bounds of
greater than 2??. This upper bound can be computed in O(n2 ) time since the size of ?Whh CD? (?) is
(? ?1)
n?n and ?i (Whh ) has already been obtained at step ? . If we did not compute the previous singular
? (? ) ) as ?i (W (? ?K?1) )+
values from ? ? K step to ? ? 1 step, we compute the upper bound of ?i (W
hh
hh
PK
(? )
|?|||?
C
(?)||
from
Lemma
6.
Since
our
training
originally
constrains
?
Whh D? ?k
F
1 (Whh ) <
k=0
? (? ) ) > 2,
2 as described in eq. (8), we can redefine s as the number of singular values such that ?i (W
hh
(?
)
?
instead of ?i (W
hh ) > 2 ? ?. This modification can further reduce the computational cost without
disrupting the training. In summary, our method can efficiently estimate the number of singular
values needed in O(n2 ) time, and we compute the truncated SVD in O(n2 log(s)) time only if we
need to compute singular values by using Lemma 6.
4
Experiments
4.1
Experimental Conditions
To evaluate the effectiveness of our method, we conducted experiments on language modeling and
polyphonic music modeling. We trained the GRU and examined the successful training rate, as well
as the average and standard deviation of the loss. We defined successful training as training in which
the validation loss at each epoch is never greater than the initial value. The experimental conditions
of each modeling are explained below.
4.1.1
Language Modeling
Penn Treebank (PTB) [25] is a widely used dataset to evaluate the performance of RNNs. PTB is
split into training, validation, and test sets, and the sets are composed of 930 k, 74 k, 80 k tokens.
This experiment used a 10 k word vocabulary, and all words outside the vocabulary were mapped
to a special token. The experimental conditions were based on the previous paper [38]. Our model
architecture was as follows: The first layer was a 650 ? 10, 000 linear layer without bias to convert the
one-hot vector input into a dense vector, and we multiplied the output of the first layer by 0.01 because
our method assumes small inputs. The second layer was a GRU layer with 650 units, and we used the
softmax function as the output layer. We applied 50 % dropout to the output of each layer except for
the recurrent connection [38]. We unfolded the GRU for 35 time steps in BPTT and set the mini-batch
size to 20. We trained the GRU with SGD for 75 epochs since the performance of the models trained
by Adam and RMSprop were worse than that trained by SGD in the preliminary experiments, and
Zaremba et al. [38] used SGD. The results and conditions of preliminary experiments are in the
supplementary material. We set the learning rate to one in the first 10 epochs, and then, divided
the learning rate by 1.1 after each epoch. In our method, ? was set to [0.2, 0.5, 0.8, 1.1, 1.4]. In
gradient clipping, a heuristic for setting the threshold is to look at the average norm of the gradient
[28]. We evaluated gradient clipping based on the gradient norm by following the study [28]. In
the supplementary material, we evaluated gradient elementwise clipping which is used practically.
Since the average norm of the gradient was about 10, we set the threshold to [5, 10, 15, 20]. We
initialized the weight matrices except for Whh with a normal distribution N (0, 1/650) , and Whh
as an orthogonal matrix composed of the left singular vectors of a random matrix [29, 8]. After
each epoch, we evaluated the validation loss. The model that achieved the least validation loss was
evaluated using the test set.
4.1.2
Polyphonic Music Modeling
In this modeling, we predicted MIDI note numbers at the next time step given the observed notes of
the previous time steps. We used the Nottingham dataset: a MIDI file containing 1200 folk tunes [6].
We represented the notes at each time step as a 93-dimensional binary vector. This dataset is split
into training, validation and test sets [6]. The experimental conditions were based on the previous
study [20]. Our model architecture was as follows: The first layer was a 200 ? 93 linear layer without
bias, and the output of the first layer was multiplied by 0.01. The second and third layers were GRU
6
Table 1: Language modeling results: success rate and perplexity.
Our method
Delta
0.2
0.5
0.8
1.1
1.4
Gradient clipping
Threshold
5
10
15
20
Success Rate
Validation Loss
Test Loss
100 %
102.0?0.3
97.6?0.4
100 %
102.8?0.3
98.4?0.3
100 %
103.7?0.2
99.0?0.4
100 %
105.2?0.2
100.3?0.2
100 %
107.0?0.4
102.1?0.2
Success Rate
Validation Loss
Test Loss
100 %
109.3?0.4
106.9?0.4
40 %
103.1?0.4
100.4?0.5
0%
N/A
N/A
0%
N/A
N/A
Table 2: Music modeling results: success rate and negative log-likelihood.
Our method
Delta
0.2
0.5
0.8
1.1
1.4
Gradient clipping
Threshold
15
30
45
60
Success Rate
Validation Loss
Test Loss
100 %
3.46?0.05
3.53?0.04
100 %
3.47?0.07
3.53?0.04
100 %
3.59?0.1
3.64?0.2
100 %
4.58?0.2
4.56?0.2
100 %
4.64?0.2
4.62?0.2
Success Rate
Validation Loss
Test Loss
100 %
3.57?0.01
3.64?0.04
100 %
3.61?0.2
3.64?0.2
100 %
3.88?0.2
3.89?0.2
100 %
5.26?3
5.36?3
layers with 200 units per layer, and we used the logistic function as the output layer. 50 % dropout
was applied to non-recurrent connections. We unfolded GRUs for 35 time steps and set the size
of the mini-batch to 20. We used SGD with a learning rate of 0.1 and divided the learning rate by
1.25 if we observed no improvement over 10 consecutive epochs. We repeated the same procedure
until the learning rate became smaller than 10?4 . In our method, ? was set to [0.2, 0.5, 0.8, 1.1,
1.4]. In gradient clipping, the threshold was set to [15, 30, 45, 60], since the average norm of the
gradient was about 30. We initialized the weight matrices except for Whh with a normal distribution
N (0, 10?4 /200) , and Whh as an orthogonal matrix. After each epoch, we evaluated the validation
loss, and the model that achieved the least validation loss was evaluated using the test set.
4.2
Success Rate and Accuracy
Tables 1 and 2 list the success rates of language modeling and music modeling, respectively. These
tables also list the averages and standard deviations of the loss in each modeling to show that our
method outperforms gradient clipping. In these tables, ?Threshold? means the threshold of gradient
clipping, and ?Delta? means ? in our method.
As shown in Table 1, in language modeling, gradient clipping failed to train even though its parameter
was set to 10, which is the average norm of the gradient as recommended by Pascanu et al. [28].
Although gradient clipping successfully trained the GRU when its threshold was five, it failed to
effectively learn the model with this setting; a threshold of 10 achieved lower perplexity than a
threshold of five. As shown in Table 2, in music modeling, gradient clipping successfully trained
the GRU. However, the standard deviation of the loss was high when the threshold was set to 60
(double the average norm). On the other hand, our method successfully trained the GRU in both
modelings. Tables 1 and 2 show that our approach achieved lower perplexity and negative loglikelihood compared with gradient clipping, while it constrained the GRU to be locally stable. This is
because our approach of constraining stability improves the performance of the GRU. The previous
study [22] showed that stabilizing the activation of the RNN can improve performance on several
tasks. In addition, Bengio et al. [5] showed that an RNN is robust to noise when the state remains
in the basin of attraction. Using our method, the state of the GRU tends to remain in the basin of
the attraction of h? = 0. Therefore, our method can improve robustness against noise, which is an
advantage of the GRU [33].
As shown in Table 2, when ? was set to 1.1 or 1.4, the performance of the GRU deteriorated. This is
because the convergence speed of the state depends on ?. As mentioned in Section 3.2, the spectral
norm of Whh is less than or equal to 2 ? ?. This spectral norm gives the upper bound of |?1 (J )|.
|?1 (J )| gives the rate of convergence of a linearized GRU (eq. (9)), which approximates GRU near
ht = 0 when |?1 (J )| < 1. Therefore, the state of the GRU near ht = 0 tends to converge quickly if
? is set to close to two. In this case, the GRU becomes robust to noise since the state affected by the
past noise converges to zero quickly, while the GRU loses effectiveness for long-term dependencies.
We can tune ? from the characteristics of the data: if data have the long-term dependencies, we should
set ? small, whereas we should set ? large for noisy data.
The threshold in gradient clipping is unbounded, and hence, it is difficult to tune. Although the
threshold can be heuristically set on the basis of the average norm, this may not be effective in
language modeling using the GRU, as shown in Table 1. In contrast, the hyper-parameter is bounded
in our method, i.e., 0 < ? < 2, and it is easy to understand its effect as mentioned above.
7
1.4
Norm of gradient
Spectral radius
1.2
1.2
Spectral radius
1.0
1.0
0.8
0.8
0.6
0.6
0.4
0.4
0.2
0.2
0.0
0
Norm of gradient
100
200
300
400
No. of Iteration
0.0
0
500
(a) Gradient clipping (threshold of 5).
100
200
300
No. of Iteration
400
500
(b) Our method (delta of 0.2).
Figure 1: Gradient explosion in language modeling.
Table 3: Computation time in the language modeling (delta is 0.2, threshold is 5).
Naive SVD
4
5.02 ? 10
4.3
Computation time (s)
Truncated SVD
4.55 ? 10
4
Gradient clipping
4.96 ? 104
Relation between Gradient and Spectral Radius
Our method of constraining the GRU to be locally stable is based on the hypothesis that a change in
stability causes an exploding gradient problem. To confirm this hypothesis, we examined (i) the norm
of the gradient before clipping and (ii) the spectral radius of J (in Lemma 2), which determines local
stability, versus the number of iterations until the 500th iteration in Fig. 1. Fig. 1(a) and 1(b) show
the results of gradient clipping with a threshold of 5 and our method with ? of 0.2. Each norm of the
gradient was normalized so that its maximum value was one. The norm of the gradient significantly
increased when the spectral radius crossed one, such as at the 63rd, 79th, and 141st iteration (Fig. 1(a)).
In addition, the spectral radius decreased to less than one after the gradient explosion; i.e., when the
gradient explosion occurred, the gradient became in the direction of decreasing spectral radius. In
contrast, our method kept the spectral radius less than one by constraining the spectral norm of Whh
(Fig. 1(b)). Therefore, our method can prevent the gradient from exploding and effectively train the
GRU.
4.4
Computation Time
We evaluated computation time of the language modeling experiment. The detailed experimental
setup is described in the supplementary material. Table 3 lists the computation time of the whole
learning process using gradient clipping and our method with the naive SVD and with truncated
SVD. This table shows the computation time of our method is comparable to gradient clipping. As
mentioned in Section 2.1, the computational cost of gradient clipping is proportional to the number
of parameters including weight matrices of input and output layers. In language modeling, the sizes
of input and output layers tend to be large due to the large vocabulary size. On the other hand, the
computational cost of our method only depends on the length of the state vector, and our method
can be efficiently computed if the number of singular values greater than 2 is small as described
in Section 3.3. As a result, our method could reduce the computation time comparing to gradient
clipping.
5
Conclusion
We analyzed the dynamics of GRUs and devised a learning method that prevents the exploding
gradient problem. Our analysis of stability provides new insight into the behavior of GRUs. Our
method constrains GRUs so that the states near 0 asymptotically converge to 0. Through language
and music modeling experiments, we confirmed that our method can successfully train GRUs and
found that our method can improve their performance.
8
References
[1] Dario Amodei, Rishita Anubhai, Eric Battenberg, Carl Case, Jared Casper, Bryan Catanzaro,
Jingdong Chen, Mike Chrzanowski, Adam Coates, Greg Diamos, Erich Elsen, Jesse Engel, Linxi
Fan, Christopher Fougner, Awni Hannun, Billy Jun, Tony Han, Patrick LeGresley, Xiangang
Li, Libby Lin, Sharan Narang, Andrew Ng, Sherjil Ozair, Ryan Prenger, Sheng Qian, Jonathan
Raiman, Sanjeev Satheesh, David Seetapun, Shubho Sengupta, Chong Wang, Yi Wang, Zhiqian
Wang, Bo Xiao, Yan Xie, Dani Yogatama, Jun Zhan, and Zhenyao Zhu. Deep speech 2:
End-to-end speech recognition in english and mandarin. In Proc. ICML, pages 173?182, 2016.
[2] Martin Arjovsky, Amar Shah, and Yoshua Bengio. Unitary evolution recurrent neural networks.
In Proc. ICML, pages 1120?1128, 2016.
[3] Pierre Baldi and Kurt Hornik. Universal approximation and learning of trajectories using
oscillators. In Proc. NIPS, pages 451?457. 1996.
[4] Nikita E Barabanov and Danil V Prokhorov. Stability analysis of discrete-time recurrent neural
networks. IEEE Transactions on Neural Networks, 13(2):292?303, 2002.
[5] Yoshua Bengio, Patrice Simard, and Paolo Frasconi. Learning long-term dependencies with
gradient descent is difficult. IEEE Transactions on Neural Networks, 5(2):157?166, 1994.
[6] Nicolas Boulanger-Lewandowski, Yoshua Bengio, and Pascal Vincent. Modeling temporal
dependencies in high-dimensional sequences: Application to polyphonic music generation and
transcription. In Proc. ICML, pages 1159?1166, 2012.
[7] Mahmoud Chilali and Pascal Gahinet. H ? design with pole placement constraints: an lmi
approach. IEEE Transactions on automatic control, 41(3):358?367, 1996.
[8] Kyunghyun Cho, Bart Van Merri?nboer, Caglar Gulcehre, Dzmitry Bahdanau, Fethi Bougares,
Holger Schwenk, and Yoshua Bengio. Learning phrase representations using rnn encoder?
decoder for statistical machine translation. In Proc. EMNLP, pages 1724?1734. ACL, 2014.
[9] Junyoung Chung, Caglar Gulcehre, KyungHyun Cho, and Yoshua Bengio. Empirical evaluation
of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:1412.3555,
2014.
[10] Jasmine Collins, Jascha Sohl-Dickstein, and David Sussillo. Capacity and trainability in
recurrent neural networks. In Proc. ICLR, 2017.
[11] Kenji Doya. Bifurcations in the learning of recurrent neural networks. In Proc. ISCAS, volume 6,
pages 2777?2780. IEEE, 1992.
[12] Bernard Doyon, Bruno Cessac, Mathias Quoy, and Manuel Samuelides. Destabilization and
route to chaos in neural networks with random connectivity. In Proc. NIPS, pages 549?555.
1993.
[13] Alex Graves and J?rgen Schmidhuber. Offline handwriting recognition with multidimensional
recurrent neural networks. In Proc. NIPS, pages 545?552, 2009.
[14] Alex Graves, Abdel-rahman Mohamed, and Geoffrey Hinton. Speech recognition with deep
recurrent neural networks. In Proc. ICASSP, pages 6645?6649. IEEE, 2013.
[15] N Halko, PG Martinsson, and JA Tropp. Finding structure with randomness: Stochastic algorithms for constructing approximate matrix decompositions. arXiv preprint arXiv:0909.4061,
2009.
[16] Robert Haschke and Jochen J Steil. Input space bifurcation manifolds of recurrent neural
networks. Neurocomputing, 64:25?38, 2005.
[17] Michiel Hermans and Benjamin Schrauwen. Training and analysing deep recurrent neural
networks. In Proc. NIPS, pages 190?198. 2013.
[18] Sepp Hochreiter and J?rgen Schmidhuber. Long short-term memory. Neural computation, 9(8):
1735?1780, 1997.
[19] Herbert Jaeger. Tutorial on training recurrent neural networks, covering BPPT, RTRL, EKF
and the" echo state network" approach. GMD-Forschungszentrum Informationstechnik, 2002.
[20] Rafal Jozefowicz, Wojciech Zaremba, and Ilya Sutskever. An empirical exploration of recurrent
network architectures. In Proc. ICML, pages 2342?2350, 2015.
9
[21] Diederik Kingma and Jimmy Ba. Adam: A method for stochastic optimization. In Proc. ICLR,
2015.
[22] David Krueger and Roland Memisevic. Regularizing rnns by stabilizing activations. In Proc.
ICLR, 2016.
[23] Chung-Ming Kuan, Kurt Hornik, and Halbert White. A convergence result for learning in
recurrent neural networks. Neural Computation, 6(3):420?440, 1994.
[24] Thomas Laurent and James von Brecht. A recurrent neural network without chaos. In Proc.
ICLR, 2017.
[25] Mitchell P Marcus, Mary Ann Marcinkiewicz, and Beatrice Santorini. Building a large annotated
corpus of english: The penn treebank. Computational linguistics, 19(2):313?330, 1993.
[26] Tomas Mikolov. Statistical language models based on neural networks. PhD thesis, Brno
University of Technology, 2012.
[27] Hiroyuki Nakahara and Kenji Doya. Dynamics of attention as near saddle-node bifurcation
behavior. In Proc. NIPS, pages 38?44. 1996.
[28] Razvan Pascanu, Tomas Mikolov, and Yoshua Bengio. On the difficulty of training recurrent
neural networks. In Proc. ICML, pages 1310?1318, 2013.
[29] Andrew M Saxe, James L McClelland, and Surya Ganguli. Exact solutions to the nonlinear
dynamics of learning in deep linear neural networks. In Proc. ICLR, 2014.
[30] Ilya Sutskever, Oriol Vinyals, and Quoc V Le. Sequence to sequence learning with neural
networks. In Proc. NIPS, pages 3104?3112. 2014.
[31] Johan AK Suykens, Bart De Moor, and Joos Vandewalle. Robust local stability of multilayer
recurrent neural networks. IEEE Transactions on Neural Networks, 11(1):222?229, 2000.
[32] Sachin S Talathi and Aniket Vartak. Improving performance of recurrent neural network with
relu nonlinearity. arXiv preprint arXiv:1511.03771, 2015.
[33] Zhiyuan Tang, Ying Shi, Dong Wang, Yang Feng, and Shiyue Zhang. Memory visualization
for gated recurrent neural networks in speech recognition. In Proc. ICASSP, pages 2736?2740.
IEEE, 2017.
[34] Onur Toker and Hitay Ozbay. On the np-hardness of solving bilinear matrix inequalities and
simultaneous stabilization with static output feedback. In Proc. of American Control Conference,
volume 4, pages 2525?2526. IEEE, 1995.
[35] Eugene Vorontsov, Chiheb Trabelsi, Samuel Kadoury, and Chris Pal. On orthogonality and
learning recurrent networks with long term dependencies. In Proc. ICML, 2017.
[36] Stephen Wiggins. Introduction to applied nonlinear dynamical systems and chaos, volume 2.
Springer Science & Business Media, 2003.
[37] Wen Yu. Nonlinear system identification using discrete-time recurrent neural networks with
stable learning algorithms. Information sciences, 158:131?147, 2004.
[38] Wojciech Zaremba, Ilya Sutskever, and Oriol Vinyals. Recurrent neural network regularization.
arXiv preprint arXiv:1409.2329, 2014.
10
| 6647 |@word trial:1 briefly:1 norm:20 bptt:3 heuristically:1 linearized:2 decomposition:2 jingdong:1 prokhorov:2 pg:1 sgd:10 initial:4 series:4 bppt:1 tuned:1 kurt:2 past:8 outperforms:2 comparing:1 manuel:1 activation:3 diederik:1 must:1 designed:1 update:7 midori:1 polyphonic:5 bart:2 selected:1 vanishing:4 short:4 provides:1 pascanu:3 node:1 zhang:1 five:2 unbounded:1 rnl:1 prove:3 redefine:1 baldi:2 shubho:1 vartak:2 hardness:1 behavior:6 disrupts:1 elman:1 multi:7 ptb:2 ming:1 decreasing:1 unfolded:2 becomes:4 bounded:2 medium:1 minimizes:1 finding:1 temporal:1 multidimensional:1 zaremba:3 rm:1 sherjil:1 control:3 unit:7 penn:2 before:1 local:5 modify:1 tends:2 bilinear:2 ak:1 laurent:3 meet:2 rnns:11 acl:1 examined:2 conversely:1 co:1 catanzaro:1 chilali:1 fujiwara:2 chaotic:1 razvan:1 procedure:5 rnn:12 yan:1 universal:1 significantly:2 hyperbolic:1 projection:1 empirical:2 word:2 close:1 writing:1 wxr:2 iwamura:2 shi:2 center:1 jesse:1 go:2 regardless:1 sepp:1 jimmy:1 attention:1 formulate:2 stabilizing:2 tomas:2 abrupt:1 qian:1 jascha:1 insight:1 attraction:4 d1:3 regarded:1 lewandowski:1 stability:12 handle:2 variation:3 merri:1 updated:1 deteriorated:1 target:2 suppose:1 exact:1 carl:1 hypothesis:2 crossing:1 element:4 recognition:6 approximated:1 updating:1 observed:2 mike:1 preprint:4 solved:1 wang:4 ensures:1 mentioned:4 benjamin:1 ui:2 constrains:4 rmsprop:1 dynamic:17 trained:8 depend:1 solving:1 f2:1 eric:1 completely:1 basis:3 icassp:2 schwenk:1 represented:5 sigm:3 train:7 effective:2 prenger:1 hyper:1 outside:1 h0:6 whose:1 heuristic:2 supplementary:4 solve:1 widely:1 loglikelihood:1 relax:1 reconstruct:1 narang:1 encoder:1 amar:1 noisy:1 echo:1 kuan:2 patrice:1 advantage:1 eigenvalue:3 sequence:4 propose:3 product:1 reset:7 doyon:1 frobenius:1 exploiting:2 convergence:3 double:1 p:1 sutskever:3 jaeger:1 adam:4 converges:4 derive:1 recurrent:31 depending:1 mandarin:1 andrew:2 sussillo:1 eq:13 predicted:1 involves:1 kenji:2 lyapunov:1 direction:1 radius:10 tokyo:1 annotated:1 stochastic:3 exploration:1 stabilization:1 saxe:1 enable:1 material:4 explains:1 wxz:2 ja:1 beatrice:1 f1:1 marcinkiewicz:1 preliminary:3 decompose:1 ryan:1 awni:1 hold:3 practically:1 normal:2 rgen:2 achieves:1 consecutive:1 proc:22 tanh:2 largest:2 talathi:2 wl:6 engel:1 successfully:4 moor:1 unfolding:1 minimization:1 dani:1 always:1 ekf:1 pn:3 avoid:1 mahmoud:1 haschke:2 properly:1 improvement:1 indicates:4 mainly:1 likelihood:1 contrast:3 sharan:1 linxi:1 ganguli:1 typically:1 relation:1 quoy:1 classification:1 pascal:2 proposes:1 sengupta:1 constrained:4 special:1 bifurcation:20 softmax:1 equal:3 never:3 frasconi:1 having:1 beach:1 manually:2 ng:1 represents:6 holger:1 yu:2 look:1 icml:6 jochen:1 np:2 yoshua:6 few:1 wen:1 modern:1 randomly:1 composed:2 neurocomputing:1 iscas:1 n1:3 maintain:1 evaluation:2 chong:1 analyzed:6 nl:1 beforehand:1 explosion:12 folk:1 minw:1 orthogonal:2 initialized:2 halbert:1 battenberg:1 minimal:2 elsen:1 aniket:1 increased:1 modeling:27 clipping:26 cost:11 whh:42 deviation:3 pole:1 phrase:1 successful:4 vandewalle:1 conducted:1 pal:1 kanai:2 dependency:9 cho:3 raiman:1 st:2 grus:20 lstm:4 disrupted:1 memisevic:1 dong:1 quickly:2 ilya:3 sanjeev:1 connectivity:1 thesis:1 squared:2 von:3 schrauwen:1 containing:1 rafal:1 emnlp:1 lmi:1 worse:1 external:1 american:1 simard:1 danil:1 chung:2 wojciech:2 li:1 de:1 rescales:1 samuelides:1 satisfy:1 caused:2 vi:2 depends:2 crossed:1 later:1 h1:3 lab:1 analyze:1 maintains:1 gahinet:1 accuracy:3 became:2 greg:1 characteristic:1 efficiently:4 whz:2 identification:1 vincent:1 cessac:1 trajectory:1 wxh:2 confirmed:1 randomness:1 explain:1 simultaneous:1 joos:1 against:1 chrzanowski:1 mohamed:1 james:2 proof:3 handwriting:1 static:1 sampled:1 dataset:3 popular:1 mitchell:1 knowledge:1 improves:1 organized:1 schedule:1 hiroyuki:1 sophisticated:2 back:3 feed:1 originally:1 xie:1 formulation:2 evaluated:8 though:1 nottingham:1 until:4 rahman:1 hand:5 sheng:1 lstms:2 christopher:1 tropp:1 nonlinear:7 propagation:1 logistic:2 brings:1 mary:1 building:1 usa:1 effect:1 normalized:1 dario:1 evolution:1 analytically:1 hence:1 kyunghyun:2 regularization:1 iteratively:2 white:1 fethi:1 covering:1 samuel:1 outline:1 disrupting:1 demonstrate:1 bring:1 wise:3 chaos:3 novel:1 recently:1 steil:2 krueger:1 sigmoid:1 empirically:1 jp:1 volume:3 discussed:1 extend:1 approximates:1 elementwise:1 occurred:1 martinsson:1 bougares:1 jozefowicz:1 destabilization:1 rd:1 automatic:1 erich:1 pointed:1 nonlinearity:2 bruno:1 language:15 stable:18 han:1 patrick:1 showed:5 perplexity:3 schmidhuber:2 route:1 certain:1 inequality:3 binary:1 success:8 yi:1 devise:1 herbert:1 arjovsky:2 greater:6 converge:2 recommended:1 exploding:17 ii:1 stephen:1 ntt:2 smooth:1 cross:2 long:12 lin:1 michiel:1 divided:2 devised:1 roland:1 impact:1 converging:1 regression:1 multilayer:1 arxiv:8 iteration:5 represent:3 limt:1 achieved:4 hochreiter:1 suykens:1 addition:5 whereas:1 decreased:1 singular:20 rest:1 unlike:2 file:1 subject:1 tend:1 bahdanau:1 effectiveness:2 jordan:1 unitary:2 near:7 yang:1 revealed:1 split:2 constraining:3 bengio:7 easy:1 relu:2 architecture:5 brecht:3 reduce:4 six:1 suffer:1 speech:5 cause:3 deep:4 detailed:1 tune:3 locally:11 mcclelland:1 gmd:1 sachin:1 restricts:1 coates:1 tutorial:1 delta:5 per:1 bryan:1 discrete:2 affected:1 rishita:1 paolo:1 dickstein:1 threshold:21 clarity:1 prevent:8 verified:1 ht:28 kept:2 asymptotically:3 convert:1 clipped:1 doya:4 summarizes:1 comparable:2 zhan:1 entirely:1 layer:39 fl:2 bound:6 guaranteed:1 dropout:2 fan:1 placement:1 constraint:13 orthogonality:2 alex:2 n3:1 software:1 anubhai:1 sake:1 explode:1 speed:1 min:5 extremely:3 nboer:1 mikolov:2 martin:1 according:3 amodei:1 describes:1 smaller:1 remain:1 rtrl:3 brno:1 evolves:1 modification:1 quoc:1 hl:5 explained:1 yogatama:1 zhiyuan:1 equation:1 visualization:1 remains:2 hannun:1 hh:20 needed:1 jared:1 drastic:1 end:2 gulcehre:2 multiplied:2 apply:1 away:2 appropriate:1 spectral:15 pierre:1 batch:5 robustness:1 shah:1 gate:8 thomas:1 top:1 assumes:1 tony:1 linguistics:1 music:9 pushing:1 boulanger:1 feng:1 move:1 added:1 already:1 rt:3 said:2 exhibit:1 gradient:85 detrimental:1 iclr:5 onur:1 mapped:1 capacity:1 decoder:1 parame:1 chris:1 manifold:1 unstable:2 dzmitry:1 marcus:1 ozair:1 length:3 mini:5 innovation:1 ying:1 difficult:3 setup:1 robert:1 negative:2 reconsider:1 ba:1 implementation:1 design:1 zt:4 satheesh:1 gated:6 upper:6 caglar:2 descent:2 truncated:5 hinton:1 santorini:1 rn:10 wiggins:1 david:3 gru:62 required:1 specified:2 connection:4 established:1 kingma:1 nip:7 address:4 dynamical:4 usually:1 below:1 herman:1 including:1 memory:3 hot:1 natural:1 difficulty:1 business:1 mn:2 zhu:1 improve:5 technology:1 numerous:1 concludes:1 jun:2 naive:3 epoch:7 eugene:1 tangent:1 graf:2 loss:20 generation:1 proportional:1 versus:1 billy:1 geoffrey:1 validation:11 h2:2 abdel:1 basin:4 sufficient:1 propagates:1 xiao:1 treebank:2 cd:5 casper:1 translation:1 changed:1 summary:1 token:2 english:2 drastically:4 bias:2 offline:1 understand:1 absolute:3 van:1 feedback:1 whr:2 vocabulary:3 avoids:1 preventing:1 forward:1 qualitatively:1 adaptive:1 far:1 transaction:4 approximate:1 midi:2 transcription:1 keep:3 confirm:1 corpus:1 forschungszentrum:1 surya:1 table:13 learn:3 johan:1 robust:4 ca:1 nicolas:1 hornik:3 improving:1 complex:1 constructing:1 diag:1 did:1 pk:1 dense:1 bmi:2 whole:1 noise:5 n2:6 repeated:1 fig:4 junyoung:1 xh:2 candidate:1 vorontsov:2 forgets:2 third:1 tang:2 theorem:8 xt:5 list:3 exists:2 sohl:1 effectively:4 diamos:1 phd:1 chen:1 entropy:1 forget:4 halko:1 saddle:1 informationstechnik:1 failed:2 prevents:5 expressed:4 vinyals:2 bo:1 ters:1 springer:1 corresponds:1 loses:1 satisfies:2 determines:1 identity:1 formulated:3 nakahara:1 ann:1 oscillator:1 replace:1 absence:1 feasible:5 change:9 hard:2 analysing:1 except:5 reducing:1 lemma:26 called:9 bernard:1 mathias:1 svd:9 diverging:1 experimental:5 trainability:1 yasuhiro:2 jonathan:1 collins:1 avoiding:1 oriol:2 evaluate:2 regularizing:1 phenomenon:1 |
6,243 | 6,648 | On the Power of Truncated SVD for General
High-rank Matrix Estimation Problems
Simon S. Du
Carnegie Mellon University
[email protected]
Yining Wang
Carnegie Mellon University
[email protected]
Aarti Singh
Carnegie Mellon University
[email protected]
Abstract
? that is close to a general high-rank positive semiWe show that given an estimate A
?
de?nite (PSD) matrix A in spectral norm (i.e., ?A?A?
2 ? ?), the simple truncated
? produces a multiplicative approximation of A
Singular Value Decomposition of A
in Frobenius norm. This observation leads to many interesting results on general
high-rank matrix estimation problems:
1. High-rank matrix completion: we show that it is possible to recover a general
high-rank matrix A up to (1 + ?) relative error in Frobenius norm from partial
observations, with sample complexity independent of the spectral gap of A.
2. High-rank matrix denoising: we design an algorithm that recovers a matrix A
with error in Frobenius norm from its noise-perturbed observations, without
assuming A is exactly low-rank.
3. Low-dimensional approximation of high-dimensional covariance: given N
i.i.d. samples of dimension n from Nn (0, A), we show that it is possible to
approximate the covariance matrix A with relative error in Frobenius norm with
N ? n, improving over classical covariance estimation results which requires
N ? n2 .
1
Introduction
Let A be an unknown general high-rank n ? n PSD data matrix that one wishes to estimate. In many
machine learning applications, though A is unknown, it is relatively easy to obtain a crude estimate
? ? A?2 ? ?). For example, in matrix completion
? that is close to A in spectral norm (i.e., ?A
A
a simple procedure that ?lls all unobserved entries with 0 and re-scales observed entries produces
an estimate that is consistent in spectral norm (assuming the matrix satis?es a spikeness condition,
standard assumption in matrix completion literature). In matrix de-noising, an observation that is
corrupted by Gaussian noise is close to the underlying signal, because Gaussian noise is isotropic and
has small spectral norm. In covariance estimation, the sample covariance in low-dimensional settings
is close to the population covariance in spectral norm under mild conditions [Bunea and Xiao, 2015].
However, in most such applications it is not suf?cient to settle for a spectral norm approximation. For
example, in recommendation systems (an application of matrix completion) the zero-?lled re-scaled
rating matrix is close to the ground truth in spectral norm, but it is an absurd estimator because
most of the estimated ratings are zero. It is hence mandatory to require a more stringent measure of
? ? A?F ,
performance. One commonly used measure is the Frobenius norm of the estimation error ?A
which ensures that (on average) the estimate is close to the ground truth in an element-wise sense. A
? is in general not a good estimate under Frobenius norm, because in
spectral norm approximation A
?
?
? ? A?2 .
high-rank scenarios ?A ? A?F can be n times larger than ?A
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
In this paper, we show that in many cases a powerful multiplicative low-rank approximation in
Frobenius norm can be obtained by applying a simple truncated SVD procedure on a crude, easyto-?nd spectral norm approximate. In particular, given the spectral norm approximation condition
? ? A?2 ? ?, the top-k SVD of A
? k of A
? multiplicatively approximates A in Frobenius norm; that
?A
?
is, ?Ak ? A?F ? C(k, ?, ?k+1 (A))?A ? Ak ?F , where Ak is the best rank-k approximation of A
in Frobenius and spectral norm. To our knowledge, the best existing result under the assumption
? ? A?2 ? ? is due to Achlioptas and McSherry [2007], who showed that ?A
? k ? A?F ?
?A
?
?
1/4
??Ak ?F , which depends on ?Ak ?F and is not multiplicative in
?A ? Ak ?F + k? + 2k
?A ? Ak ?F .
Below we summarize applications in several matrix estimation problems.
High-rank matrix completion Matrix completion is the problem of (approximately) recovering a
data matrix from very few observed entries. It has wide applications in machine learning, especially
in online recommendation systems. Most existing work on matrix completion assumes the data matrix
is exactly low-rank [Candes and Recht, 2012, Sun and Luo, 2016, Jain et al., 2013]. Candes and Plan
[2010], Keshavan et al. [2010] studied the problem of recovering a low-rank matrix corrupted by
stochastic noise; Chen et al. [2016] considered sparse column corruption. All of the aforementioned
work assumes that the ground-truth data matrix is exactly low-rank, which is rarely true in practice.
Negahban and Wainwright [2012] derived minimax rates of estimation error when the spectrum
of the data matrix lies in an ?q ball. Zhang et al. [2015], Koltchinskii et al. [2011] derived oracle
?
inequalities for general matrix completion; however their error bound has an additional O( n)
multiplicative factor. These results also require solving computationally expensive nuclear-norm
penalized optimization problems whereas our method only requires solving a single truncated
singular value decomposition. Chatterjee et al. [2015] also used the truncated SVD estimator for
matrix completion.
However, his bound depends on the nuclear norm of the underlying matrix
?
which may be n times larger than our result. Hardt and Wootters [2014] used a ?soft-de?ation?
technique to remove condition number dependency in the sample complexity; however, their error
bound for general high-rank matrix completion is additive and depends on the ?consecutive? spectral
gap (?k (A) ? ?k+1 (A)), which can be small in practical settings [Balcan et al., 2016, Anderson
et al., 2015]. Eriksson et al. [2012] considered high-rank matrix completion with additional union-ofsubspace structures.
In this paper, we show that if the n ? n data matrix A satis?es ?0 -spikeness condition, 1 then for
? k ? A?F ? (1 + O(?))?A ?
? k satis?es ?A
any ? ? (0, 1), the truncated SVD of zero-?lled matrix A
n max{??4 ,k2 }?20 ?A?2F log n
) ,which can be
Ak ?F if the sample complexity is lower bounded by ?(
?k+1 (A)2
further simpli?ed to ?(?20 max{??4 , k 2 }?k (A)2 ? nrs (A) log n), where ?k (A) = ?1 (A)/?k+1 (A)
is the kth-order condition number and rs (A) = ?A?2F /?A?22 ? rank(A) is the stable rank of
A. Compared to existing work, our error bound is multiplicative, gap-free, and the estimator is
computationally ef?cient. 2
? = A + E be a noisy observation of A, where E is a PSD
High-rank matrix de-noising Let A
Gaussian noise matrix with zero mean and ? 2 /n variance on each entry. By simple concentration
? ? A?2 = ? with high probability; however, A
? is in general not a good estimator
results we have ?A
?
?
of A in Frobenius norm when A is high-rank. Speci?cally, ?A ? A?F can be as large as n?.
?
Applying our main result, we show that
? if ? < ?k+1 (A) for some k ??n, then the top-k SVD Ak of
?
?
A satis?es ?Ak ? A?F ? (1 + O( ?/?k+1 (A)))?A ? Ak ?F + k?. This suggests a form of
bias-variance decomposition as larger rank threshold k induces smaller bias ?A ? Ak ?F but larger
variance k? 2 . Our results generalize existing work on matrix de-noising [Donoho and Gavish, 2014,
Donoho et al., 2013, Gavish and Donoho, 2014], which focus primarily on exact low-rank A.
1
n?A?max ? ?0 ?A?F ; see also De?nition 2.1.
We remark that our relative-error analysis does not, however, apply to exact rank-k matrix where ?k+1 = 0.
This is because for exact rank-k matrix a bound of the form (1 + O(?))?A ? Ak ?F requires exact recovery of
A, which truncated SVD cannot achieve. On the other hand, in the case of ?k+1 = 0 a weaker additive-error
bound is always applicable, as we show in Theorem 2.3.
2
2
Low-rank estimation of high-dimensional covariance The (Gaussian) covariance estimation
problem asks to estimate an n ? n PSD covariance matrix A, either in spectral or Frobenius
norm, from N i.i.d. samples X1 , ? ? ? , XN ? N (0, A). The high-dimensional regime of covariance
estimation, in which N ? n or even N ? n, has attracted enormous interest in the mathematical
statistics literature [Cai et al., 2010, Cai and Zhou, 2012, Cai et al., 2013, 2016]. While most existing
work focus on sparse or banded covariance matrices, the setting where A has certain low-rank
structure has seen rising interest recently [Bunea and Xiao, 2015, Kneip and Sarda, 2011]. In
particular, Bunea and Xiao [2015] shows that if n = O(N ? ) for some ? ? 0 then the sample
? = 1 ?N Xi X ? satis?es
covariance estimator A
i
i=1
N
?
?
?
log
N
? ? A?F = OP ?A?2 re (A)
,
(1)
?A
N
where re (A) = tr(A)/?A?2 ? rank(A) is the effective rank of A. For high-rank matrices where
re (A) ? n, Eq. (1) requires N = ?(n2 log n) to approximate A consistently in Frobenius norm.
? k and show that, if
consider a reduced-rank estimator A
? k ? A?F admits
? c for some small universal constant c > 0, then ?A
a relative Frobenius-norm error bound (1+O(?))?A?Ak ?F with high probability. Our result allows
reasonable approximation of A in Frobenius norm under the regime of N = ?(npoly(k) log n)
if ?k = O (poly (k)), which is signi?cantly more ?exible than N = ?(n2 log n), though the
dependency of ? is worse than [Bunea and Xiao, 2015]. The error bound is also agnostic in nature,
making no assumption on the actual or effective rank of A.
In
this
paper
we
re (A) max{??4 ,k2 }?k (A)2 log N
N
Notations For an n?n PSD matrix A, denote A = U?U? as its eigenvalue decomposition, where
U is an orthogonal matrix and ? = diag(?1 , ? ? ? , ?n ) is a diagonal matrix, with eigenvalues sorted in
descending order ?1 ? ?2 ??
? ? ? ? ?n ? 0. The spectral norm and Frobenius norm of A are de?ned
as ?A?2 = ?1 and ?A?F = ?12 + ? ? ? + ?n2 , respectively. Suppose u1 , ? ? ? , un are eigenvectors as?k
?n
sociated with ?1 , ? ? ? , ?n . De?ne Ak = i=1 ?i ui u?
= Uk ? k U ?
, An?k = i=k+1 ?i ui u?
i
i =
k
?m 2
?
?
and
A
=
?
u
u
=
U
?
U
.
For
a
tall
matrix
Un?k ?n?k U?
m1 :m2
m1 :m2 m1 :m2 m1 :m2
n?k
i=m1 +1 i i i
U ? Rn?k , we use U = Range(U) to denote the linear subspace spanned by the columns of U. For
two linear subspaces U and V, we write W = U ?V if U ?V = {0} and W = {u+v : u ? U , v ? V}.
For a sequence of random variables {Xn }?
n=1 and real-valued function f : N ? R, we say
Xn = OP (f (n)) if for any ? > 0, there exists N ? N and C > 0 such that Pr[|Xn | ? C ? |f (n)|] ? ?
for all n ? N .
2
Multiplicative Frobenius-norm Approximation and Applications
We ?rst state our main result, which shows that truncated SVD on a weak estimator with small
approximation error in spectral norm leads to a strong estimator with multiplicative Frobenius-norm
error bound. We remark that truncated SVD in general has time complexity
?
?
? ?
??
? + npoly (k) ,
O min n2 k, nnz A
? is the number of non-zero entries in A,
? and the time complexity is at most linear in
where nnz(A)
matrix sizes when k is small. We refer readers to [Allen-Zhu and Li, 2016] for details.
Theorem 2.1. Suppose A is an n ? n PSD matrix with eigenvalues ?1 (A) ? ? ? ? ? ?n (A) ? 0,
? ? A?2 ? ? = ?2 ?k+1 (A) for some ? ? (0, 1/4]. Let Ak and
? satis?es ?A
and a symmetric matrix A
? Then
?
Ak be the best rank-k approximations of A and A.
?
? k ? A?F ? (1 + 32?)?A ? Ak ?F + 102 2k?2 ?A ? Ak ?2 .
?A
(2)
?
Remark 2.1. Note when ? = O(1/ k) we obtain an (1 + O (?)) error bound.
Remark 2.2. This theorem only studies PSD matrices. Using similar arguments in the proof, we
believe similar results for general asymmetric matrices can be obtained as well.
3
? k ? A?F assuming ?A
? ? A?2 ? ? is due to
To our knowledge, the best existing bound for ?A
Achlioptas and McSherry [2007], who showed that
?
? k ? A?F ? ?A ? Ak ?F + ?(A
? ? A)k ?F + 2 ?(A
? ? A)k ?F ?Ak ?F
?A
? ?
?
? ?A ? Ak ?F + k??A ? Ak ?2 + 2k 1/4 ? ?Ak ?F .
(3)
?
Compared to Theorem 2.1, Eq. (3) is not relative because the third term 2k 1/4 ?Ak ?F depends on
the k largest eigenvalues of A, which could be much larger than the remainder term ?A ? Ak ?F . In
? k ? A?F could be upper bounded
contrast, Theorem 2.1, together with Remark 2.1, shows that ?A
by a small factor multiplied with the remainder term ?A ? Ak ?F .
We also provide a gap-dependent version.
Theorem 2.2. Suppose A is an n ? n PSD matrix with eigenvalues ?1 (A) ? ? ? ? ? ?n (A) ? 0,
? ? A?2 ? ? = ? (?k (A) ? ?k+1 (A)) for some ? ? (0, 1/4].
? satis?es ?A
and a symmetric matrix A
?
? Then
Let Ak and Ak be the best rank-k approximations of A and A.
?
? k ? A?F ? ?A ? Ak ?F + 102 2k? (?k (A) ? ?k+1 (A)) .
?A
(4)
?
If A is an exact rank-k matrix, Theorem 2.2 implies that truncated SVD gives an ? 2k?k error
approximation in Frobenius norm, which has been established by many previous works [Yi et al.,
2016, Tu et al., 2015, Wang et al., 2016].
Before we proceed to the applications and proof of Theorem 2.1, we ?rst list several examples of A
with classical distribution of eigenvalues and discuss how Theorem 2.1 could be applied to obatin
good Frobenius-norm approximations of A. We begin with the case where eigenvalues of A have a
polynomial decay rate (i.e., power law). Such matrices are ubiquitous in practice [Liu et al., 2015].
? ? A?2 ? ? for some ? ? (0, 1/2] and
Corollary 2.1 (Power-law spectral decay). Suppose ?A
??
?1/?
?j (A) = j for some ? > 1/2. Set k = ?min{C1 ?
, n} ? 1?. If k ? 1 then
? 2??1
?
2??1
? k ? A?F ? C ? ? max ? 2? , n? 2? ,
?A
1
where C1 , C1? > 0 are constants that only depend on ?.
We remark that the
?j (A) = j ?? implies that the eigenvalues lie in an ?q ball for
?nassumption
q
q = 1/?; that is, j=1 ?j (A) = O(1). The error bound in Corollary 2.1 matches the minimax
rate (derived by Negahban and Wainwright ?
[2012]) for matrix completion when the spectrum is
constrained in an ?q ball, by replacing ? with n/N where N is the number of observed entries.
Next, we consider the case where eigenvalues satisfy a faster decay rate.
? ? A?2 ? ? for some ? ? (0, e?16 ) and
Corollary 2.2 (Exponential spectral decay). Suppose ?A
?j (A) = exp{?cj} for some c > 0. Set k = ?min{c?1 log(1/?) ? c?1 log log(1/?), n} ? 1?. If
k ? 1 then
? ?
?
? k ? A?F ? C ? ? max ? log(1/?)3 , n1/2 exp(?cn) ,
?A
2
where C2? > 0 is a constant that only depends on c.
Both corollaries are proved in the appendix. The error bounds in both Corollaries 2.1 and 2.2 are
? ? A?F ? n1/2 ?. We also remark
? which satis?es ?A
signi?cantly better than the trivial estimate A,
that the bound in Corollary 2.1 cannot be obtained by a direct application of the weaker bound Eq. (3),
?
which yields a ? 2??1 bound.
We next state results that are consequences of Theorem 2.1 in several matrix estimation problems.
2.1
High-rank Matrix Completion
Suppose A is a high-rank n ? n PSD matrix that satis?es ?0 -spikeness condition de?ned as follows:
4
De?nition 2.1 (Spikeness condition). An n ? n PSD matrix A satis?es ?0 -spikeness condition if
n?A?max ? ?0 ?A?F , where ?A?max = max1?i,j?n |Aij | is the max-norm of A.
Spikeness condition makes uniform sampling of matrix entries powerful in matrix completion
problems. If A is exactly low rank, the spikeness condition is implied by an upper bound on
max1?i?n ?e?
i Uk ?2 , which is the standard incoherence assumption on the top-k space of A [Candes
and Recht, 2012]. For general high-rank A, the spikeness condition is implied by a more restrictive
incoherence condition that imposes an upper bound on max1?i?n ?e?
i Un?k ?2 and ?An?k ?max ,
which are assumptions adopted in [Hardt and Wootters, 2014].
? is a symmetric re-scaled zero-?lled matrix of observed entries. That is,
Suppose A
?
Aij /p, with probability p;
? ij =
?1 ? i ? j ? n.
[A]
0,
with probability 1 ? p;
(5)
Here p ? (0, 1) is a parameter that controls the probability of observing a particular entry in A,
? and A are symmetric so we only
corresponding to a sample complexity of O(n2 p). Note that both A
?
specify the upper triangle of A. By a simple application of matrix Bernstein inequality [Mackey
? is close to A in spectral norm when A satis?es ?0 -spikeness. Here we
et al., 2014], one can show A
cite a lemma from [Hardt, 2014] to formally establish this observation:
Lemma 2.1 (Corollary of [Hardt, 2014], Lemma A.3). Under the model of Eq. (5) and ?0 -spikeness
condition of A, for t ? (0, 1) it holds with probability at least 1 ? t that
??
??
?
? ?2 ?A?2 log(n/t) ? ?A? log(n/t) ?
0
F
0
F
? ? A?2 ? O ?max
?.
?A
,
?
?
np
np
? k be the best rank-k approximation of A
? in Frobenius/spectral norm. Applying Theorem 2.1
Let A
and 2.2 we obatin the following result:
Theorem 2.3. Fix t ? (0, 1). Then with probability 1 ? t we have
? 2
?
?
?0 ?A?2F log(n/t)
? k ? A?F ? O( k) ? ?A ? Ak ?F
?A
if p = ?
.
n?k+1 (A)2
Furthermore, for ?xed ? ? (0, 1/4], with probability 1 ? t we have
?
? 2
?4 2
2
? k ? A?F ? (1 + O(?)) ?A ? Ak ?F if p = ? ?0 max{? , k }?A?F log(n/t)
?A
n?k+1 (A)2
? k ? A?F ? ?A ? Ak ?F + ? (?k (A) ? ?k+1 (A)) if p = ?
?A
?
?20 k?A?2F log(n/t)
n?2 (?k (A) ? ?k+1 (A))
2
?
.
?
As a remark, because ?0 ? 1 and ?A?F /?k+1 (A) ? k always hold, the sample complexity is
lower bounded by ?(nk log n), the typical sample complexity in noiseless matrix completion. In
the case of high rank A, the results in Theorem 2.3 are the strongest when A has small stable rank
rs (A) = ?A?2F /?A?22 and the top-k condition number ?k (A) = ?1 (A)/?k+1 (A) is not too large.
?
? k ? A?F has an O( k) multiplicative error
For example, if A has stable rank rs (A) = r then ?A
bound with sample complexity ?(?20 ?k (A)2 ? nr log n); or an (1 + O(?)) relative error bound with
sample complexity ?(?20 max{??4 , k 2 }?k (A)2 ? nr log n). Finally, when ?k+1 (A) is very small
and the ?gap? ?k (A) ? ?k+1 (A) is large, a weaker additive-error bound is applicable with sample
complexity independent of ?k+1 (A)?1 .
Comparing with previous works, if? the gap (1 ? ?k+1 /?k ) is of order ?, then sample complexity
of[Hardt, 2014] Theorem 1.2 and [Hardt and Wootters, 2014] Theorem 1 scale with 1/?7 . Our result
improves their results to the scaling of 1/?4 with a much simpler algorithm (truncated SVD).
5
2.2
High-rank matrix de-noising
Let A be an n ? n PSD signal matrix and E a symmetric random Gaussian matrix with zero mean and
i.i.d.
? = A + E.
? 2 /n variance. That is, Eij ? N (0, ? 2 /n) for 1 ? i ? j ? n and Eij = Eji . De?ne A
?
The matrix de-noising problem is then to recover the signal matrix A from noisy observations A.
We refer the readers to [Gavish and Donoho, 2014] for a list of references that shows the ubiquitous
application of matrix de-noising in scienti?c ?elds.
? ? A?2 = ?E?2 =
It is well-known by concentration results of Gaussian random matrices, that ?A
? k be the best rank-k approximation of A
? in Frobenius/spectral norm. Applying
OP (?). Let A
Theorem 2.1, we immediately have the following result:
Theorem 2.4. There exists an absolute constant c > 0 such that, if ? < c ? ?k+1 (A) for some
1 ? k < n, then with probability at least 0.8 we have that
??
?
??
?
?
? k ? A?F ? 1 + O
?A ? Ak ?F + O( k?).
(6)
?A
?k+1 (A)
Eq.?(6) can be understood from a classical bias-variance tradeoff perspective: the ?rst (1 +
cut-off rank k,
O( ?/?k+1 (A)))?A ? Ak ?F acts as a bias term, which decreases as we increase
?
corresponding to a more complicated model; on the other hand, the second O( k?) term acts as the
(square root of) variance, which does not depend on the signal A and increases with k.
2.3
Low-rank estimation of high-dimensional covariance
Suppose A is an n ? n PSD matrix and X1 , ? ? ? , XN are i.i.d. samples drawn from the multivariate
Gaussian distribution Nn (0, A). The question is to estimate A from samples X1 , ? ? ? , XN . A
? = 1 ?N Xi X ? . While in low-dimensional
common estimator is the sample covariance A
i
i=1
N
? is obvious (cf. [Van der Vaart,
regimes (i.e., n ?xed and N ? ?) the asymptotic ef?ciency of A
2000]), its statistical power in high-dimensional regimes where n and N are comparable are highly
? ? A?? , ? = 2/F
non-trivial. Below we cite results by Bunea and Xiao [2015] for estimation error ?A
when n is not too large compared to N :
Lemma 2.2 (Bunea and Xiao [2015]). Suppose n = O(N ? ) for some ? ? 0 and let re (A) =
? =
tr(A)/?A?2 denote the effective rank of the covariance A. Then the sample covariance A
?N
1
?
i=1 Xi Xi satis?es
N
?
?
?
log
N
? ? A?F = OP ?A?2 re (A)
(7)
?A
N
and
? ? A?2 = OP
?A
?
?A?2 max
??
re (A) log(N n) re (A) log(N n)
,
N
N
??
.
(8)
? k be the best rank-k approximation of A
? in Frobenius/spectral norm. Applying Theorem 2.1
Let A
and 2.2 together with Eq. (8), we immediately arrive at the following theorem.
Theorem 2.5. Fix ? ? (0, 1/4] and 1 ? k < n. Recall that re (A) = tr(A)/?A?2 and ?k (A) =
?1 (A)/?k+1 (A). There exists a universal constant c > 0 such that, if
re (A) max{??4 , k 2 }?k (A)2 log(N )
?c
N
then with probability at least 0.8,
and if
? k ? A?F ? (1 + O(?)) ?A ? Ak ?F
?A
re (A)k?A?22 log(N )
N ?2 (?k (A) ? ?k+1 (A))
6
2
?c
then with probability at least 0.8,
? k ? A?F ? ?A ? Ak ?F + ? (?k (A) ? ?k+1 (A)) .
?A
?
Theorem 2.5 shows that it is possible to obtain a reasonable Frobenius-norm approximation of A
by truncated SVD in the asymptotic regime of N = ?(re (A)poly(k) log N ), which is much more
?exible than Eq. (7) that requires N = ?(re (A)2 log N ).
3
Proof Sketch of Theorem 2.1
In this section we give a proof sketch of Theorem 2.1. The proof of Theorem 2.2 is similar and less
challenging so we defer it to appendix. We defer proofs of technical lemmas to Section A.
?
? k ? Ak ?F is upper bounded by an O( k) factor of
? k and Ak are low-rank, ?A
Because both A
? k ? Ak ?2 . From the condition that ?A
? ? A?2 ? ?, a straightforward approach to upper
?A
? k ? Ak ?2 is to consider the decomposition ?A
? k ? A k ? 2 ? ?A
? ? A?2 + 2?Uk U? ?
bound ?A
k
? kU
? k ?2 , where Uk U? and U
? ? ?2 ?A
? kU
? ? are projection operators onto the top-k eigenspaces
U
k
k
k
? respectively. Such a naive approach, however, has two major disadvantages. First,
of A and A,
? k ?2 , which is additive and may be much larger than ?A
? ? A?2 .
the upper bound depends on ?A
? kU
? ? ?2 depends on the ?consecutive? sepctral
Perhaps more importantly, the quantity ?Uk U?
?
U
k
k
gap (?k (A) ? ?k+1 (A)), which could be very small for large matrices.
The key idea in the proof of Theorem 2.1 is to ?nd an ?envelope? m1 ? k ? m2 in the spectrum of
A surrounding k, such that the eigenvalues within the envelope are relatively close. De?ne
m1
=
m2
=
argmax0?j?k {?j (A) ? (1 + 2?)?k+1 (A)};
argmaxk?j?n {?j (A) ? ?k (A) ? 2??k+1 (A)},
where we let ?0 (A) = ? for convenience. Let Um , U?m be basis of the top m-dimensional linear
? respectively. Also denote Un?m and U?n?m as basis of the orthogonal
subspaces of A and A,
complement of Um and U?m . By asymmetric Davis-Kahan inequality (Lemma C.1) and Wely?s
inequality we can obtain the following result.
? ? A?2 ? ?2 ?k+1 (A) for ? ? (0, 1) then ?U
? ? Um ?2 , ?U
? ? Un?m ?2 ? ?.
Lemma 3.1. If ?A
n?k
1
k
2
Let Um1 :m2 be the linear subspace of A associated with eigenvalues ?m1 +1 (A), ? ? ? , ?m2 (A).
Intuitively, we choose a (k ? m1 )-dimensional linear subspace in Um1 :m2 that is ?most aligned? with
? Formally, de?ne
the top-k subspace U?k of A.
?
?
?k .
W = argmaxdim(W)=k?m1 ,W?Um1 :m2 ?k?m1 W? U
W is then a d ? (k ? m1 ) matrix with orthonormal columns that corresponds to a basis of W. W is
carefully constructed so that it is closely aligned with U?k , yet still lies in Uk . In particular, Lemma
? ? W?2 is upper bounded by ?.
3.2 shows that sin ?(W, U?k ) = ?U
n?k
? ? W?2 ? ?.
? ? A?2 ? ?2 ?k+1 (A) for ? ? (0, 1) then ?U
Lemma 3.2. If ?A
n?k
Now de?ne
? = Am + WW? AWW? .
A
1
? as the ?reference matrix" because we can decompose ?A
? k ? A?F as
We use A
?
? k ? A?F ? ?A ? A?
? F + ?A
? k ? A?
? F ? ?A ? A?
? F + 2k?A
? k ? A?
? 2
?A
(9)
and bound each term on the right hand side separately. Here the last inequality holds because both
? have rank at most k. The following lemma bounds the ?rst term.
? k and A
A
? ? A?2 ? ?2 ?k+1 (A)2 for ? ? (0, 1/4] then ?A ? A?
? F ? (1 + 32?)?A ? Ak ?F .
Lemma 3.3. If ?A
7
The proof of this lemma relies Pythagorean theorem and Poincar? separation theorem. Let Um1 :m2
be the (m2 ? m1 )-dimensional linear subspace such that Um2 = Um1 ? Um1 :m2 . De?ne Am1 :m2 =
Um1 :m2 ?m1 :m2 U?
m1 :m2 , where ?m1 :m2 = diag(?m1 +1 (A), ? ? ? , ?m2 (A)) and Um1 :m2 is an orthonormal basis associated with Um1 :m2 . Applying Pythagorean theorem (Lemma C.2), we can
decompose
? 2 = ?A ? Am ?2 + ?Am :m ?2 ? ?WW? Am :m WW? ?2 .
?A ? A?
F
F
2 F
1
2 F
1
2
Applying Poincar? separation theorem (Lemma C.3) where X = ?m1 :m2 and P = U?
m1 :m2 W, we
?m2 ?m1
?m 2
2
2
have ?W? Am1 :m2 W?2F ? j=m
?
(A
)
=
?
(A)
. With some
j
m1 :m2
j=m1 +m2 ?k+1 j
2 ?k+1
routine algebra we can prove Lemma 3.3.
To bound the second term of Eq. (9) we use the following lemma.
? ? A?2 ? ?2 ?k+1 (A) for ? ? (0, 1/4] then ?A
? k ? A?
? 2 ? 102?2 ?A ? Ak ?2 .
Lemma 3.4. If ?A
? k and A.
? Recall the de?nition that
The proof of Lemma 3.4 relies on the low-rankness of A
? and U?? = Null(A).
? Consider ?v?2 = 1 such that v ? (A
? k ? A)v
? = ?A
?
? 2.
U? = Range(A)
? k ? A??
? ?
?
?
?
Because v maximizes v (Ak ?A)v over all unit-length vectors, it must lie in the range of Ak ? A
because otherwise the component outside the range will not contribute. Therefore, we can choose v
? k ) = U?k and v 2 ? Range(A)
? = U?. Subsequently, we have
that v = v 1 + v 2 where v 1 ? Range(A
that
?U
? ?U
? ?v + U
? n?k U
?? v
? kU
(10)
v = U
k
=
n?k
? kU
?U
? ?v + U
? ?U
? ?U
? ? v.
U
k
?
(11)
Consider the following decomposition:
? ?
? ?
? ?
?
?
? ? ?
? ?? ? ??v ? (A
? ? A)v ?? + ??v ? (A
? k ? A)v
? ?? + ??v ? (A ? A)v
? ?? .
?v (Ak ? A)v
? ? A)v| is trivially upper bounded by ?A
? ? A?2 ? ?2 ?k+1 (A). The second
The ?rst term |v ? (A
?
and the third term can be bounded by Wely?s inequality (Lemma C.4) and basic properties of A
(Lemma A.3). See Section A for details.
4
Discussion
We mention two potential directions to further extend results of this paper.
4.1
Model selection for general high-rank matrices
? ? A?2 ? ?2 ?k+1 (A), which could be
The validity of Theorem 2.1 depends on the condition ?A
hard to verify if ?k+1 (A) is unknown and dif?cult to estimate. Furthermore, for general high-rank
matrices, the model selection problem of determining an appropriate (or even optimal) cut-off rank k
requires knowledge of the distribution of the entire spectrum of an unknown data matrix, which is
even more challenging to obtain.
One potential approach is to impose a parametric pattern of decay of the eigenvalues (e.g., polynomial
and exponential decay), and to estimate a small set of parameters (e.g., degree of polynomial) from
? Afterwards, the optimal cut-off rank k could be determined by a theoretical
the noisy observations A.
analysis, similar to the examples in Corollaries 2.1 and 2.2. Another possibility is to use repeated
sampling techniques such as boostrap in a stochastic problem (e.g.,
? matrix de-noising) to estimate the
?bias? term ?A ? Ak ?F for different k, as the variance term k? is known or easy to estimate.
4.2
Minimax rates for polynomial spectral decay
Consider the class of PSD matrices whose eigenvalues follow a polynomial (power-law) decay:
?(?, n) = {A ? Rn?n : A ? 0, ?j (A) = j ?? }. We are interested in the following minimax rates
for completing or de-noising matrices in ?(?, n):
8
Question 1 (Completion of ?(?, n)). Fix n ? N, p ? (0, 1) and de?ne N = pn2 . For M ? ?(?, n),
? ij = Mij with probability p and A
? ij = 0 with probability 1 ? p. Also let ?(?0 , n) = {M ?
let A
Rn?n : n?M?max ? ?0 ?M?F } be the class of all non-spiky matrices. Determine
R1 (?0 , ?, n, N ) := inf
sup
? ?M
? M??(?,n)??(?0 ,n)
A?
? ? M?2 .
E?M
F
? = M + ?/?nZ, where Z is a
Question 2 (De-noising of ?(?, n)). Fix n ? N, ? > 0 and let A
symmetric matrices with i.i.d. standard Normal random variables on its upper triangle. Determine
R2 (?, ?, n) := inf
sup
? ?M
? M??(?,n)
A?
? ? M?2 .
E?M
F
Compared to existing settings on matrix completion and de-noising, we believe ?(?, n) is a more
natural matrix class which allows for general high-rank matrices, but also imposes suf?cient spectral
decay conditions so that spectrum truncation algorithms result in signi?cant bene?ts. Based on
Corollary 2.1 and its matching lower bounds for a larger ?p class [Negahban and Wainwright, 2012],
we make the following conjecture:
Conjecture 4.1. For ? > 1/2 and ? not too small, we conjecture that
R1 (?0 , ?, n, N ) ? C(?0 ) ?
? n ? 2??1
2?
N
and
where C(?0 ) > 0 is a constant that depends only on ?0 .
5
? ? 2??1
R2 (?, ?, n) ? ? 2 2? ,
Acknowledgements
S.S.D. was supported by ARPA-E Terra program. Y.W. and A.S. were supported by the NSF CAREER
grant IIS-1252412.
References
Dimitris Achlioptas and Frank McSherry. Fast computation of low-rank matrix approximations.
Journal of the ACM, 54(2):9, 2007.
Zeyuan Allen-Zhu and Yuanzhi Li. Even faster svd decomposition yet without agonizing pain. In
Advances in Neural Information Processing Systems, pages 974?982, 2016.
David Anderson, Simon Du, Michael Mahoney, Christopher Melgaard, Kunming Wu, and Ming Gu.
Spectral gap error bounds for improving cur matrix decomposition and the nystr?m method. In
Arti?cial Intelligence and Statistics, pages 19?27, 2015.
Maria Florina Balcan, Simon S Du, Yining Wang, and Adams Wei Yu. An improved gap-dependency
analysis of the noisy power method. In 29th Annual Conference on Learning Theory, pages
284?309, 2016.
Florentina Bunea and Luo Xiao. On the sample covariance matrix estimator of reduced effective rank
population matrices, with applications to fpca. Bernoulli, 21(2):1200?1230, 2015.
T Tony Cai and Harrison H Zhou. Optimal rates of convergence for sparse covariance matrix
estimation. The Annals of Statistics, 40(5):2389?2420, 2012.
T Tony Cai, Cun-Hui Zhang, and Harrison H Zhou. Optimal rates of convergence for covariance
matrix estimation. The Annals of Statistics, 38(4):2118?2144, 2010.
T Tony Cai, Zhao Ren, and Harrison H Zhou. Optimal rates of convergence for estimating toeplitz
covariance matrices. Probability Theory and Related Fields, 156(1-2):101?143, 2013.
T Tony Cai, Zhao Ren, and Harrison H Zhou. Estimating structured high-dimensional covariance and
precision matrices: Optimal rates and adaptive estimation. Electronic Journal of Statistics, 10(1):
1?59, 2016.
9
Emmanuel Candes and Benjamin Recht. Exact matrix completion via convex optimization. Communications of the ACM, 55(6):111?119, 2012.
Emmanuel J Candes and Yaniv Plan. Matrix completion with noise. Proceedings of the IEEE, 98(6):
925?936, 2010.
Sourav Chatterjee et al. Matrix estimation by universal singular value thresholding. The Annals of
Statistics, 43(1):177?214, 2015.
Yudong Chen, Huan Xu, Constantine Caramanis, and Sujay Sanghavi. Matrix completion with
column manipulation: Near-optimal sample-robustness-rank tradeoffs. IEEE Transactions on
Information Theory, 62(1):503?526, 2016.
David Donoho and Matan Gavish. Minimax risk of matrix denoising by singular value thresholding.
The Annals of Statistics, 42(6):2413?2440, 2014.
David L Donoho, Matan Gavish, and Andrea Montanari. The phase transition of matrix recovery
from gaussian measurements matches the minimax mse of matrix denoising. Proceedings of the
National Academy of Sciences, 110(21):8405?8410, 2013.
Brian Eriksson, Laura Balzano, and Robert D Nowak. High-rank matrix completion. In AISTATS,
pages 373?381, 2012.
?
Matan Gavish and David L Donoho. The optimal hard threshold for singular values is 4/ 3. IEEE
Transactions on Information Theory, 60(8):5040?5053, 2014.
Moritz Hardt. Understanding alternating minimization for matrix completion. In Foundations of
Computer Science (FOCS), 2014 IEEE 55th Annual Symposium on, pages 651?660. IEEE, 2014.
Moritz Hardt and Mary Wootters. Fast matrix completion without the condition number. In COLT,
pages 638?678, 2014.
Prateek Jain, Praneeth Netrapalli, and Sujay Sanghavi. Low-rank matrix completion using alternating
minimization. In Proceedings of the forty-?fth annual ACM symposium on Theory of computing,
pages 665?674. ACM, 2013.
Raghunandan H Keshavan, Andrea Montanari, and Sewoong Oh. Matrix completion from a few
entries. Information Theory, IEEE Transactions on, 56(6):2980?2998, 2010.
Alois Kneip and Pascal Sarda. Factor models and variable selection in high-dimensional regression
analysis. The Annals of Statistics, pages 2410?2447, 2011.
Vladimir Koltchinskii, Karim Lounici, and Alexandre B Tsybakov. Nuclear-norm penalization and
optimal rates for noisy low-rank matrix completion. The Annals of Statistics, pages 2302?2329,
2011.
Ziqi Liu, Yu-Xiang Wang, and Alexander Smola. Fast differentially private matrix factorization. In
Proceedings of the 9th ACM Conference on Recommender Systems, pages 171?178. ACM, 2015.
Lester Mackey, Michael I Jordan, Richard Y Chen, Brendan Farrell, and Joel A Tropp. Matrix
concentration inequalities via the method of exchangeable pairs. The Annals of Probability, 42(3):
906?945, 2014.
Sahand Negahban and Martin J Wainwright. Restricted strong convexity and weighted matrix
completion: Optimal bounds with noise. The Journal of Machine Learning Research, 13(1):
1665?1697, 2012.
Ruoyu Sun and Zhi-Quan Luo. Guaranteed matrix completion via non-convex factorization. IEEE
Transactions on Information Theory, 62(11):6535?6579, 2016.
Stephen Tu, Ross Boczar, Max Simchowitz, Mahdi Soltanolkotabi, and Benjamin Recht. Low-rank
solutions of linear matrix equations via procrustes ?ow. arXiv preprint arXiv:1507.03566, 2015.
Aad W Van der Vaart. Asymptotic statistics, volume 3. Cambridge university press, 2000.
10
Lingxiao Wang, Xiao Zhang, and Quanquan Gu. A uni?ed computational and statistical framework
for nonconvex low-rank matrix estimation. arXiv preprint arXiv:1610.05275, 2016.
Xinyang Yi, Dohyung Park, Yudong Chen, and Constantine Caramanis. Fast algorithms for robust pca
via gradient descent. In Advances in Neural Information Processing Systems, pages 4152?4160,
2016.
Lijun Zhang, Tianbao Yang, Rong Jin, and Zhi-Hua Zhou. Analysis of nuclear norm regularization
for full-rank matrix completion. arXiv preprint arXiv:1504.06817, 2015.
11
| 6648 |@word mild:1 private:1 version:1 rising:1 polynomial:5 norm:41 nd:2 r:3 decomposition:8 covariance:21 arti:1 eld:1 asks:1 mention:1 tr:3 nystr:1 liu:2 xinyang:1 existing:7 comparing:1 luo:3 yet:2 attracted:1 must:1 additive:4 cant:1 remove:1 mackey:2 intelligence:1 cult:1 isotropic:1 contribute:1 simpler:1 zhang:4 mathematical:1 c2:1 direct:1 constructed:1 symposium:2 focs:1 prove:1 fth:1 andrea:2 ming:1 zhi:2 actual:1 begin:1 estimating:2 underlying:2 bounded:7 npoly:2 agnostic:1 notation:1 null:1 prateek:1 xed:2 maximizes:1 unobserved:1 wely:2 cial:1 absurd:1 act:2 exactly:4 um:3 scaled:2 k2:2 uk:6 exchangeable:1 control:1 unit:1 grant:1 fpca:1 lester:1 positive:1 before:1 understood:1 consequence:1 ak:47 incoherence:2 approximately:1 koltchinskii:2 studied:1 nz:1 suggests:1 challenging:2 dif:1 factorization:2 range:6 practical:1 practice:2 union:1 procedure:2 nite:1 poincar:2 nnz:2 universal:3 projection:1 matching:1 eriksson:2 close:8 cannot:2 operator:1 noising:10 onto:1 convenience:1 applying:7 selection:3 risk:1 descending:1 lijun:1 pn2:1 straightforward:1 tianbao:1 convex:2 recovery:2 immediately:2 m2:27 estimator:10 importantly:1 nuclear:4 spanned:1 orthonormal:2 his:1 oh:1 population:2 annals:7 suppose:9 exact:6 boczar:1 element:1 expensive:1 asymmetric:2 cut:3 observed:4 preprint:3 wang:5 ensures:1 sun:2 decrease:1 benjamin:2 convexity:1 complexity:12 ui:2 singh:1 solving:2 depend:2 algebra:1 max1:3 basis:4 triangle:2 gu:2 caramanis:2 surrounding:1 jain:2 fast:4 effective:4 outside:1 sociated:1 matan:3 whose:1 balzano:1 larger:7 valued:1 say:1 otherwise:1 toeplitz:1 statistic:10 vaart:2 yiningwa:1 kahan:1 noisy:5 online:1 sequence:1 eigenvalue:13 cai:7 simchowitz:1 remainder:2 tu:2 aligned:2 achieve:1 academy:1 frobenius:23 differentially:1 rst:5 convergence:3 yaniv:1 r1:2 produce:2 adam:1 tall:1 completion:29 ij:3 op:5 eq:8 strong:2 netrapalli:1 recovering:2 c:2 signi:3 implies:2 direction:1 closely:1 stochastic:2 subsequently:1 stringent:1 settle:1 require:2 fix:4 decompose:2 brian:1 rong:1 hold:3 aartisingh:1 considered:2 ground:3 normal:1 exp:2 major:1 consecutive:2 aarti:1 gavish:6 estimation:18 applicable:2 ross:1 largest:1 quanquan:1 bunea:7 weighted:1 minimization:2 gaussian:8 always:2 zhou:6 agonizing:1 corollary:9 derived:3 focus:2 maria:1 consistently:1 rank:65 bernoulli:1 contrast:1 brendan:1 sense:1 am:4 dependent:1 nn:2 kneip:2 entire:1 interested:1 aforementioned:1 colt:1 pascal:1 plan:2 constrained:1 field:1 beach:1 sampling:2 park:1 yu:2 sanghavi:2 np:2 lls:1 primarily:1 richard:1 few:2 national:1 phase:1 raghunandan:1 n1:2 psd:13 interest:2 satis:12 highly:1 possibility:1 joel:1 mahoney:1 yining:2 scienti:1 mcsherry:3 nowak:1 partial:1 huan:1 eigenspaces:1 orthogonal:2 re:16 theoretical:1 arpa:1 column:4 soft:1 disadvantage:1 entry:10 uniform:1 too:3 dependency:3 perturbed:1 corrupted:2 st:1 recht:4 negahban:4 terra:1 ssdu:1 cantly:2 off:3 michael:2 together:2 choose:2 worse:1 laura:1 zhao:2 li:2 potential:2 de:24 satisfy:1 farrell:1 depends:9 multiplicative:8 root:1 observing:1 um2:1 sup:2 recover:2 complicated:1 candes:5 simon:3 defer:2 square:1 variance:7 who:2 yield:1 generalize:1 weak:1 ren:2 corruption:1 banded:1 strongest:1 ed:2 obvious:1 proof:9 associated:2 recovers:1 cur:1 proved:1 hardt:8 recall:2 knowledge:3 improves:1 ubiquitous:2 cj:1 routine:1 carefully:1 alexandre:1 nrs:1 follow:1 specify:1 wei:1 improved:1 lounici:1 though:2 anderson:2 furthermore:2 smola:1 achlioptas:3 spiky:1 am1:2 hand:3 sketch:2 tropp:1 replacing:1 keshavan:2 christopher:1 um1:9 perhaps:1 believe:2 mary:1 usa:1 validity:1 verify:1 true:1 hence:1 regularization:1 moritz:2 symmetric:6 alternating:2 karim:1 sin:1 davis:1 dohyung:1 allen:2 balcan:2 wise:1 ef:2 recently:1 common:1 volume:1 extend:1 approximates:1 m1:22 mellon:3 refer:2 measurement:1 cambridge:1 sujay:2 trivially:1 soltanolkotabi:1 stable:3 multivariate:1 showed:2 perspective:1 constantine:2 inf:2 mandatory:1 scenario:1 certain:1 obatin:2 manipulation:1 inequality:7 nonconvex:1 yi:2 der:2 nition:3 seen:1 ruoyu:1 additional:2 simpli:1 impose:1 speci:1 zeyuan:1 determine:2 forty:1 signal:4 ii:1 stephen:1 afterwards:1 full:1 eji:1 match:2 faster:2 technical:1 long:1 basic:1 florina:1 regression:1 noiseless:1 cmu:3 arxiv:6 c1:3 whereas:1 separately:1 harrison:4 singular:5 envelope:2 quan:1 jordan:1 near:1 yang:1 bernstein:1 easy:2 sarda:2 idea:1 cn:1 praneeth:1 tradeoff:2 pca:1 sahand:1 proceed:1 remark:8 wootters:4 eigenvectors:1 procrustes:1 tsybakov:1 induces:1 reduced:2 nsf:1 estimated:1 carnegie:3 write:1 key:1 threshold:2 enormous:1 drawn:1 powerful:2 arrive:1 reasonable:2 reader:2 wu:1 lled:3 separation:2 electronic:1 florentina:1 appendix:2 scaling:1 comparable:1 bound:29 completing:1 guaranteed:1 oracle:1 annual:3 u1:1 argument:1 min:3 relatively:2 conjecture:3 ned:2 martin:1 structured:1 ball:3 smaller:1 cun:1 making:1 intuitively:1 restricted:1 pr:1 computationally:2 equation:1 discus:1 adopted:1 multiplied:1 apply:1 yuanzhi:1 spectral:25 appropriate:1 robustness:1 top:7 assumes:2 cf:1 tony:4 cally:1 restrictive:1 emmanuel:2 especially:1 establish:1 classical:3 implied:2 question:3 quantity:1 parametric:1 concentration:3 diagonal:1 nr:2 pain:1 kth:1 subspace:7 ow:1 gradient:1 trivial:2 assuming:3 length:1 multiplicatively:1 vladimir:1 robert:1 frank:1 design:1 unknown:4 upper:10 recommender:1 observation:8 descent:1 t:1 jin:1 truncated:12 communication:1 rn:3 ww:3 rating:2 david:4 complement:1 pair:1 bene:1 established:1 nip:1 lingxiao:1 below:2 pattern:1 dimitris:1 regime:5 summarize:1 program:1 max:17 wainwright:4 power:6 ation:1 natural:1 zhu:2 minimax:6 ne:7 argmaxk:1 naive:1 literature:2 acknowledgement:1 understanding:1 determining:1 relative:6 law:3 asymptotic:3 xiang:1 interesting:1 suf:2 penalization:1 foundation:1 degree:1 consistent:1 imposes:2 xiao:8 thresholding:2 sewoong:1 penalized:1 supported:2 last:1 free:1 truncation:1 aij:2 bias:5 weaker:3 side:1 aad:1 wide:1 absolute:1 sparse:3 van:2 yudong:2 dimension:1 xn:6 transition:1 commonly:1 adaptive:1 sourav:1 transaction:4 approximate:3 uni:1 xi:4 spectrum:5 un:5 nature:1 ku:5 robust:1 ca:1 career:1 improving:2 du:3 mse:1 poly:2 diag:2 aistats:1 main:2 montanari:2 noise:7 n2:6 repeated:1 x1:3 xu:1 cient:3 precision:1 wish:1 ciency:1 exponential:2 lie:4 crude:2 mahdi:1 third:2 theorem:30 exible:2 list:2 decay:9 admits:1 r2:2 exists:3 hui:1 chatterjee:2 nk:1 gap:9 chen:4 rankness:1 eij:2 recommendation:2 hua:1 mij:1 cite:2 truth:3 corresponds:1 relies:2 acm:6 sorted:1 donoho:7 hard:2 typical:1 determined:1 denoising:3 lemma:20 svd:13 e:12 rarely:1 formally:2 alexander:1 pythagorean:2 |
6,244 | 6,649 | f -GANs in an Information Geometric Nutshell
Richard Nock?,?,?
Zac Cranko?,?
Aditya Krishna Menon?,?
?,?
Lizhen Qu
Robert C. Williamson?,?
?
?
Data61, the Australian National University and ? the University of Sydney
{firstname.lastname, aditya.menon, bob.williamson}@data61.csiro.au
Abstract
Nowozin et al showed last year how to extend the GAN principle to all f divergences. The approach is elegant but falls short of a full description of the
supervised game, and says little about the key player, the generator: for example,
what does the generator actually converge to if solving the GAN game means
convergence in some space of parameters? How does that provide hints on the generator?s design and compare to the flourishing but almost exclusively experimental
literature on the subject? In this paper, we unveil a broad class of distributions for
which such convergence happens ? namely, deformed exponential families, a wide
superset of exponential families ?. We show that current deep architectures are
able to factorize a very large number of such densities using an especially compact
design, hence displaying the power of deep architectures and their concinnity in the
f -GAN game. This result holds given a sufficient condition on activation functions
? which turns out to be satisfied by popular choices. The key to our results is a
variational generalization of an old theorem that relates the KL divergence between
regular exponential families and divergences between their natural parameters. We
complete this picture with additional results and experimental insights on how
these results may be used to ground further improvements of GAN architectures,
via (i) a principled design of the activation functions in the generator and (ii) an
explicit integration of proper composite losses? link function in the discriminator.
1
Introduction
In a recent paper, Nowozin et al. [30] showed that the GAN principle [15] can be extended to the
variational formulation of all f -divergences. In the GAN game, there is an unknown distribution P
which we want to approximate using a parameterised distribution Q. Q is learned by a generator
by finding a saddle point of a function which we summarize for now as f -GAN(P, Q), where f is
a convex function (see eq. (7) below for its formal expression). A part of the generator?s training
involves as a subroutine a supervised adversary ? hence, the saddle point formulation ? called
discriminator, which tries to guess whether randomly generated observations come from P or Q.
Ideally, at the end of this supervised game, we want Q to be close to P, and a good measure of this is
the f -divergence If (PkQ), also known as Ali-Silvey distance [1, 12]. Initially, one choice of f was
considered [15]. Nowozin et al. significantly grounded the game and expanded its scope by showing
that for any f convex and suitably defined, then [30, Eq. 4]:
f -GAN(P, Q) ? If (PkQ) .
(1)
The inequality is an equality if the discriminator is powerful enough. So, solving the f -GAN game
can give guarantees on how P and Q are distant to each other in terms of f -divergence. This elegant
characterization of the supervised game unfortunately falls short of justifying or elucidating all
parameters of the supervised game [30, Section 2.4], and the paper is also silent regarding a key
part of the game: the link between distributions in the variational formulation and the generator, the
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
main player which learns a parametric model of a density. In doing so, the f -GAN approach and
its members remain within an information theoretic framework that relies on divergences between
distributions only [30]. In the GAN world at large, this position contrasts with other prominent
approaches that explicitly optimize geometric distortions between the parameters or support of
distributions [6, 14, 16, 21, 22], and raises the problem of connecting the f -GAN approach to any
sort of information geometric optimization. One such information-theoretic/information-geometric
identity is well known: The Kullback-Leibler (KL) divergence between two distributions of the
same (regular) exponential family equals a Bregman divergence D between their natural parameters
[2, 4, 7, 9, 35], which we can summarize as:
IfKL (PkQ) = D(?k?) .
(2)
Here, ? and ? are respectively the natural parameters of P and Q. Hence, distributions are points on
a manifold on the right-hand side, a powerful geometric statement [4]; however, being restricted to
KL divergence or "just" exponential families, it certainly falls short of the power to explain the GAN
game. To our knowledge, the only generalizations known fall short of the f -divergence formulation
and are not amenable to the variational GAN formulation [5, Theorem 9], [13, Theorem 3].
Our first contribution is such an identity that connects the general If -divergence formulation in eq.
(1) to the general D (Bregman) divergence formulation in eq. (2). We now briefly state it, postponing
the details to Section 3:
f -GAN(P, escort(Q)) = D(?k?) + Penalty(Q) ,
(3)
for P and Q (with respective parameters ? and ?) which happen to lie in a superset of exponential
families called deformed exponential families, that have received extensive treatment in statistical
physics and differential information geometry over the last decade [3, 25]. The right-hand side of
eq. (3) is the information geometric part [4], in which D is a Bregman divergence. Therefore, the
f -GAN problem can be equivalent to a geometric optimization problem [4], like for the Wasserstein
GAN and its variants [6]. Notice also that Q appears in the game in the form of an escort [5]. The
difference vanish only for exponential families (escort(Q) = Q, Penalty(Q) = 0 and f = KL).
Our second contribution drills down into the information-theoretic and information-geometric
parts of (3). In particular, from the former standpoint, we completely specify the parameters of the
supervised game, unveiling a key parameter left arbitrary in [30] (explicitly incorporating the link
function of proper composite losses [32]). From the latter standpoint, we show that the standard deep
generator architecture is powerful at modelling complex escorts of any deformed exponential family,
factorising a number of escorts in order of the total inner layers? dimensions, and this factorization
happens for an especially compact design. This hints on a simple sufficient condition on the activation
function to guarantee the escort modelling, and it turns out that this condition is satisfied, exactly or
in a limit sense, by most popular activation functions (ELU, ReLU, Softplus, ...). We also provide
experiments1 that display the uplift that can be obtained through a principled design of the activation
function (generator), or tuning of the link function (discriminator).
Due to the lack of space, a supplement (SM) provides the proof of the results in the main file and
additional experiments. A longer version with a more exhaustive treatment of related results is
available [27]. The rest of this paper is as follows. Section ? 2 presents definition, ? 3 formally
presents eq. (3), ? 4 derives consequences for deep learning, ? 5 completes the supervised game
picture of [30], Section ? 6 presents experiments and a last Section concludes.
2
Definitions
Throughout this paper, the domain X of observations is a measurable set. We begin with two
important classes of distortion measures, f -divergences and Bregman divergences.
Definition 1 For any two distributions P and Q having respective densities P and Q absolutely
continuous with respect to a base measure ?, the f -divergence between P and Q, where f : R+ ? R
is convex with f (1) = 0, is
Z
P (X)
P (x)
.
If (PkQ) = EX?Q f
=
Q(x) ? f
d?(x) .
(4)
Q(X)
Q(x)
X
1
The code used for our experiments is available through https://github.com/qulizhen/fgan_info_geometric
2
For any convex differentiable ? : Rd ? R, the (?-)Bregman divergence between ? and % is:
.
D? (?k%) = ?(?) ? ?(%) ? (? ? %)> ??(%) ,
where ? is called the generator of the Bregman divergence.
(5)
f -divergences are the key distortion measure of information theory, while Bregman divergences are
the key distortion measure of information geometry. A distribution P from a (regular) exponential
.
family with cumulant C : ? ? R and sufficient statistics ? : X ? Rd has density PC (x|?, ?) =
>
exp(?(x) ? ? C(?)), where ? is a convex open set, C is convex and ensures normalization on the
simplex (we leave implicit the associated dominating measure [3]). A fundamental Theorem ties
Bregman divergences and f -divergences: when P and Q belong to the same exponential family, and
denoting their respective densities PC (x|?, ?) and QC (x|?, ?), it holds that IKL (PkQ) = DC (?k?).
.
Here, IKL is Kullback-Leibler (KL) f -divergence (f = x 7? x log x). Remark that the arguments in
the Bregman divergence are permuted with respect to those in eq. (2) in the introduction. This also
holds if we consider fKL in eq. (2) to be the Csisz?r dual of f [8], namely fKL : x 7? ? log x, since
in this case IfKL (PkQ) = IKL (QkP) = DC (?k?). We made this choice in the introduction for the
sake of readability in presenting eqs. (1 ? 3). We now define generalizations of exponential families,
following [5, 13]. Let
? : R+ ? R+ be non-decreasing [25, Chapter 10].
R z We define the ?-logarithm,
. Rz 1
.
log? , as log? (z) = 1 ?(t)
dt. The ?-exponential is exp? (z) = 1 + 0 ?(t)dt, where ? is defined
.
by ?(log? (z)) = ?(z). In the case where the integrals are improper, we consider the corresponding
limit in the argument / integrand.
Definition 2 [5] A distribution P from a ?-exponential family (or deformed exponential family, ?
being implicit) with convex cumulant C : ? ? R and sufficient statistics ? : X ? Rd has density
.
given by P?,C (x|?, ?) = exp? (?(x)> ? ? C(?)), with respect to a dominating measure ?. Here, ?
is a convex open set and ? is called the coordinate of P. The escort density (or ?-escort) of P?,C is
Z
1
.
.
? ?(P?,C ) , Z =
?(P?,C (x|?, ?))d?(x) .
(6)
P??,C =
Z
X
Z is the escort?s normalization constant.
? the escort distribution of P whose density
We leaving implicit the dominating measure and denote P
is given by eq. (6). We shall name ? the signature of the deformed (or ?-)exponential family, and
sometimes drop indexes to save readability without ambiguity, noting e.g. P? for P??,C . Notice that
normalization in the escort is ensured by a simple integration [5, Eq. 7]. For the escort to exist, we
require that Z < ? and therefore ?(P ) is finite almost everywhere. Such a requirement would be
satisfied in the GAN game. There is another generalization of regular exponential families, known as
generalized exponential families [13, 27]. The starting point of our result is the following Theorem,
in which the information-theoretic part is not amenable to the variational GAN formulation.
Theorem 3 [5][36] for any two ?-exponential distributions P and Q with respective densities
P?,C , Q?,C and coordinates ?, ?, DC (?k?) = EX?Q? [log? (Q?,C (X)) ? log? (P?,C (X))].
We now briefly frame the now popular (f -)GAN adversarial learning [15, 30]. We have a true
unknown distribution P over a set of objects, e.g. 3D pictures, which we want to learn. In the GAN
setting, this is the objective of a generator, who learns a distribution Q? parameterized by vector
?. Q? works by passing (the support of) a simple, uninformed distribution, e.g. standard Gaussian,
through a possibly complex function, e.g. a deep net whose parameters are ? and maps to the support
of the objects of interest. Fitting Q. involves an adversary (the discriminator) as subroutine, which
fits classifiers, e.g. deep nets, parameterized by ?. The generator?s objective is to come up with
arg min? Lf (?) with Lf (?) the discriminator?s objective:
.
Lf (?) = sup{EX?P [T? (X)] ? EX?Q? [f ? (T? (X))]} ,
(7)
?
where ? is Legendre conjugate [10] and T? : X ? R integrates the classifier of the discriminator
and is therefore parameterized by ?. Lf is a variational approximation to a f -divergence [30]; the
discriminator?s objective is to segregate true (P) from fake (Q. ) data. The original GAN choice, [15]
.
fGAN (z) = z log z ? (z + 1) log(z + 1) + 2 log 2
(8)
(the constant ensures f (1) = 0) can be replaced by any convex f meeting mild assumptions.
3
3
A variational information geometric identity for the f -GAN game
We deliver a series of results that will bring us to formalize eq. (3). First, we define a new set of
distortion measures, that we call KL? divergences.
Definition 4 For any ?-logarithm and distributions P, Q having respective densities P and Q
absolutely continuous with respect
to base measure ?, the KL? divergence between P and Q is
.
defined as KL? (PkQ) = EX?P ? log? (Q(X)/P (X)) .
Since ? is non-decreasing, ? log? is convex and so any KL? divergence is an f -divergence.
.
When ?(z) = z, KL? is the KL divergence. In what follows, base measure ? and absolute continuity are implicit, as well as that P (resp. Q) is the density of P (resp. Q). We
now relate KL? divergences vs f -divergences. Let ?f be the subdifferential of convex f and
.
IP,Q = [inf x P (x)/Q(x), supx P (x)/Q(x)) ? R+ denote the range of density ratios of P over Q.
Our first result states that if there is a selection of the subdifferential which is upperbounded on IP,Q ,
the f -divergence If (PkQ) is equal to a KL? divergence.
Theorem 5 Suppose that P, Q are such that there exists a selection ? ? ?f with sup ?(IP,Q ) < ?.
Then ?? : R+ ? R+ non decreasing such that If (PkQ) = KL? (QkP).
Theorem 5 essentially covers most if not all relevant GAN cases, as the assumption has to be satisfied
in the GAN game for its solution not to be vacuous up to a large extent (eq. (7)). We provide a
more complete treatment in the extended version [27]. The proof of Theorem 5 (in SM, Section
I) is constructive: it shows how to pick ? which satisfies all requirements. It brings the following
interesting corollary: under mild assumptions on f , there exists a ? that fits for all densities P and Q.
A prominent example of f that fits is the original GAN choice for which we can pick
.
?GAN (z) =
1
.
log 1 + z1
(9)
We now define a slight generalization of KL? -divergences and allow for ? to depend on the choice
of the expectation?s X, granted that for any of these choices, it will meet the constraints to be
R+ ? R+ and also increasing,
h and therefore define a ivalid signature. For any f : X ? R+ , we
.
.
denote KL?f (PkQ) = EX?P ? log?f (X) (Q(X)/P (X)) , where for any p ? R+ , ?p (t) = p1 ? ?(tp).
Whenever f = 1, we just write KL? as we already did in Definition 4. We note that for any x ? X,
?f (x) is increasing and non negative because of the properties of ? and f , so ?f (x) (t) defines a
?-logarithm. We are ready to state a Theorem that connects KL? -divergences and Theorem 3.
.
.
Theorem 6 Letting P = P?,C and Q = Q?,C for short in Theorem 3, we have EX?Q? [log? (Q(X)) ?
.
?
?
log? (P (X))] = KL?Q? (QkP)
? J(Q), with J(Q) = KL?Q? (QkQ).
(Proof in SM, Section II) To summarize, we know that under mild assumptions relatively to the GAN
game, f -divergences coincide with KL? divergences (Theorems 5). We also know from Theorem
6 that KL?. divergences quantify the geometric proximity between the coordinates of generalized
exponential families (Theorem 3). Hence, finding a geometric (parameter-based) interpretation of
the variational f -GAN game as described in eq. (7) can be done via a variational formulation of the
KL? divergences appearing in Theorem 6. Since penalty J(Q) does not belong to the GAN game (it
?
does not depend on P), it reduces our focus on KL?Q? (QkP).
?
Theorem 7 KL?Q? (QkP)
admits the variational formulation
n
o
?
KL?Q? (QkP)
=
sup
EX?P [T (X)] ? EX?Q? [(? log?Q? )? (T (X))] ,
T ?R++
(10)
X
.
with R++ = R\R++ . Furthermore, letting Z denoting the normalization constant of the ?-escort of
Q, the optimum T ? : X ? R++ to eq. (10) is T ? (x) = ?(1/Z) ? (?(Q(x))/?(P (x))).
4
(Proof in SM, Section III) Hence, the variational f -GAN formulation can be captured in an
information-geometric framework by the following identity using Theorems 3, 5, 7.
Corollary 8 (the variational information-geometric f -GAN identity) Using notations from Theorems
6, 7 and letting ? (resp. ?) denote the coordinate of P (resp. Q), we have:
sup
T ?R++
X
n
o
EX?P [T (X)] ? EX?Q? [(? log?Q? )? (T (X))] = DC (?k?) + J(Q) .
(11)
We shall also name for short vig-f -GAN the identity in eq. (11). We note that we can drill down
further the identity, expressing in particular the Legendre conjugate (? log?Q? )? with an equivalent
"dual" (negative) ?-logarithm in the variational problem [27]. The left hand-side of Eq. (11) has the
exact same overall shape as the variational objective of [30, Eqs 2, 6]. However, it tells the formal
story of GANs in significantly greater details, in particular for what concerns the generator. For
example, eq. (11) yields a new characterization of the generators? convergence: because DC is a
Bregman divergence, it satisfies the identity of the indiscernibles. So, solving the f -GAN game [30]
can guarantees convergence in the parameter space (? vs ?). In the realm of GAN applications, it
makes sense to consider that P (the true distribution) can be extremely complex. Therefore, even when
deformed exponential families are significantly more expressive than regular exponential families
[25], extra care should be put before arguing that complex applications comply with such a geometric
convergence in the parameter space. One way to circumvent this problem is to build distributions in
Q that factorize many deformed exponential families. This is one strong point of deep architectures
that we shall prove next.
4
Deep architectures in the vig-f -GAN game
In the GAN game, distribution Q in eq. (11) is built by the generator (call it Qg ), by passing the
support of a simple distribution (e.g. uniform, standard Gaussian), Qin , through a series of non-linear
transformations. Letting Qin denote the corresponding density, we now compute Qg . Our generator
g : X ? Rd consists of two parts: a deep part and a last layer. The deep part is, given some L ? N,
the computation of a non-linear transformation ?L : X ? RdL as
.
Rdl 3 ?l (x) =
.
?0 (x) =
v(Wl ?l?1 (x) + bl ) , ?l ? {1, 2, ..., L} ,
x?X .
(12)
(13)
v is a function computed coordinate-wise, such as (leaky) ReLUs, ELUs [11, 17, 23, 24], Wl ?
.
Rdl ?dl?1 , bl ? Rdl . The last layer computes the generator?s output from ?L : g(x) = vOUT (??L (x)+
d?dL
d
?), with ? ? R
, ? ? R ; in general, vOUT 6= v and vOUT fits the output to the domain at hand,
ranging from linear [6, 20] to non-linear functions like tanh [30]. Our generator captures the
high-level features of some state of the art generative approaches [31, 37].
To carry our analysis, we make the assumption that the network is reversible, which is going to reguire
that vOUT , ?, Wl (l ? {1, 2, ..., L}) are invertible. We note that popular examples can be invertible (e.g.
DCGAN, if we use ?-ReLU, dimensions match and fractional-strided convolutions are invertible).
At this reasonable price, we get in closed form the generator?s density and it shows the following:
for any continuous signature ?net , there exists an activation function v such that the deep part in the
network factors as escorts for the ?net -exponential family. Let 1i denote the ith canonical basis vector.
Theorem 9 ?vOUT , ?, Wl invertible (l ? {1, 2, ..., L}), for any continuous signature ?net , there exists
.
activation v and bl ? Rd (?l ? {1, 2, ..., L}) such that for any output z, letting x = g ?1 (z),
? deep (x)) ? 1/(Hout (x) ? Znet ), with Znet > 0 a constant,
Qg (z) factorizes as Qg (z) = (Qin (x)/Q
. Qd
.
.
0
>
Hout (x) = i=1 |vOUT (?i ?L (x) + ?i )|, ?i = ?> 1i , and (letting wl,i = W>
l 1i ):
.
? deep (x) =
Q
L Y
d
Y
P??net ,bl,i (x|wl,i , ?l?1 ) .
l=1 i=1
5
(14)
Name
ReLU(?)
Leaky-ReLU(?)
(?, ?)-ELU(?)
prop-? (?)
Softplus(?)
?-ReLU(?)
LSU
v(z)
max{0,
z}
z if z > 0
z if z ? 0
?z
if z > 0
?(exp(z) ? 1) if z ? 0
1
1
?
z
? 0?1 ?(? ? )?1 (? ? (0)z)
? ? (0)
1
?
1 ? 2?z
log 2
?
k + ?? ? (z)
(0)
k + log2 (1 + exp(z))
?
z+ (1??)2 +z 2
k+
2
?
0
if z < ?1
?
k+
(1 + z)2 if z ? [?1, 1]
?
4z
if z > 1
?(z)
1z>0
if z > ??
if z ? ??
if z > ?
if z ? ?
4z 2
(1??)2 +4z 2
?
2 z
4
if
if
z<4
z>4
Table 1: Some strongly/weakly admissible couples (v, ?). (?) : 1. is the indicator function; (?) :
? ? 0, 0 < ? 1 and dom(v) = [?/, +?). (?) : ? ? ? > 0; (?) : ? is Legendre conjugate; (?) :
? ? [0, 1). Shaded: prop-? activations; k is a constant (e.g. such that v(0) = 0) (see text).
(Proof in SM, Section IV) The relationship between the inner layers of a deep net and deformed
exponential families (Definition 2) follows from the theorem: rows in Wl s define coordinates, ?l
define "deep" sufficient statistics, bl are cumulants and the crucial part, the ?-family, is given by the
activation function v. Notice also that the bl s are learned, and so the deformed exponential families?
? deep factors escorts, and in number.
normalization is in fact learned and not specified. We see that Q
What is notable is the compactness achieved by the deep representation: the total dimension of all
? deep (eq. (14)) is L ? d. To handle this, a shallow net with a single inner
deep sufficient statistics in Q
layer would require a matrix W of space ?(L2 ? d2 ). The deep net g requires only O(L ? d2 ) space to
store all Wl s. The proof of Theorem 9 is constructive: it builds v as a function of ?. In fact, the proof
? deep factors ?-escorts.
also shows how to build ? from the activation function v in such a way that Q
The following Lemma essentially says that this is possible for all strongly admissible activations v.
Definition 10 Activation function v is strongly admissible iff dom(v) ? R+ 6= ? and v is C 1 ,
lowerbounded, strictly increasing and convex. v is weakly admissible
iff for any > 0, there exists
. R
v strongly admissible such that ||v ? v ||L1 < , where ||f ||L1 = |f (t)|dt.
Lemma 11 The following holds: (i) for any strongly admissible v, there exists signature ? such that
Theorem 9 holds; (ii) (?,?)-ELU (for any ? > 0), Softplus are strongly admissible. ReLU is weakly
admissible.
(proof in SM, Section V) The proof uses a trick for ReLU which can easily be repeated for (?, ?)ELU, and for leaky-ReLU, with the constraint that the domain has to be lowerbounded. Table 1
provides some examples of strongly / weakly admissible activations. It includes a wide class of
so-called "prop-? activations", where ? is negative a concave entropy, defined on [0, 1] and symmetric
around 1/2 [29]. This concludes our treatment of the information geometric part of the vig-f -GAN
identity. We now complete it with a treatment of its information-theoretic part.
5
A complete proper loss picture of the supervised GAN game
In their generalization of the GAN objective, Nowozin et al. [30] leave untold a key part of the
supervised game: they split in eq. (7) the discriminator?s contribution in two, T? = gf ? V? , where
V? : X ? R is the actual discriminator, and gf is essentially a technical constraint to ensure that
V? (.) is in the domain of f ? . They leave the choice of gf "somewhat arbitrary" [30, Section 2.4]. We
now show that if one wants the supervised loss to have the desirable property to be proper composite
[32]2 , then gf is not arbitrary. We proceed in three steps, first unveiling a broad class of proper
f -GANs that deal with this property. The initial motivation of eq. (7) was that the inner maximisation
may be seen as the f -divergence between P and Q? [26], Lf (?) = If (PkQ? ). In fact, this variational
2
informally, Bayes rule realizes the optimum and the loss accommodates for any real valued predictor.
6
representation of an f -divergence holds more generally: by [33, Theorem 9], we know that for any
convex f , and invertible link function ? : (0, 1) ? R, we have:
inf
E
T : X?R (X,Y)?D
1
[`? (Y, T (X))] = ? ? If (P k Q)
2
(15)
where D is the distribution over (observations ? {fake, real}) and the loss function `? is defined by:
??1 (z)
??1 (z)
.
.
?
0
;
`
(?1,
z)
=
f
f
, (16)
`? (+1, z) = ?f 0
?
1 ? ??1 (z)
1 ? ??1 (z)
.
assuming f differentiable. Note now that picking ?(z) = f 0 (z/(1 ? z)) with z = T (x) and
simplifying eq. (15) with P[Y = fake] = P[Y = real] = 1/2 in the GAN game yields eq. (7). For
other link functions, however, we get an equally valid class of losses whose optimisation will yield
a meaningful estimate of the f -divergence. The losses of eq. (16) belong to the class of proper
composite losses with link function ? [32]. Thus (omitting parameters ?, ?), we rephrase eq. (7) and
refer to the proper f -GAN formulation as inf Q L? (Q) with (` is as per eq. (16)):
.
L? (Q) = sup
E [?`? (+1, T (X))] + E [?`? (?1, T (X))] .
(17)
T : X?R
X?P
X?Q
Note also that it is trivial to start from a suitable proper composite loss, and derive the corresponding
generator f for the f -divergence as per eq. (15). Finally, our proper composite loss view of the
f -GAN game allows us to elicitate gf in [30]: it is the composition of f 0 and ? in eq. (16).
The use of proper composite losses as part of the supervised GAN formulation sheds further light
on another aspect the game: the connection between the value of the optimal discriminator, and
the density ratio between the generator and discriminator distributions. Instead of the optimal
T ? (x) = f 0 (P (x)/Q(x)) for eq. (7) [30, Eq. 5], we now have with the more general eq. (17) the
result T ? (x) = ?((1 + Q(x)/P (x))?1 ). We now show that proper f -GANs can easily be adapted
.
to eq. (11). We let ?? (t) = 1/??1 (1/t).
.
Theorem 12 For any ?, define `x (?1, z) = ? log(?? )
.
1
?
Q(x)
(?z), and let `(+1, z) = ?z. Then
L? (Q) in eq. (17) equals eq. (11). Its link in eq. (17) is ?x (z) = ?1/?Q(x)
(z/(1 ? z)).
?
(Proof in SM, Section VI) Hence, in the proper composite view of the vig-f -GAN identity, the
generator rules over the supervised game: it tempers with both the link function and the loss
? but only for fake examples. Notice also that when z = ?1, the fake examples loss satisfies
`x (?1, ?1) = 0 regardless of x by definition of the ?-logarithm.
6
Experiments
Two of our theoretical contributions are (A) the fact that on the generator?s side, there exists numerous
activation functions v that comply with the design of its density as factoring escorts (Lemma 11),
and (B) the fact that on the discriminator?s side, the so-called output activation function gf of [30]
aggregates in fact two components of proper composite losses, one of which, the link function ?,
should be a fine knob to operate (Theorem 12). We have tested these two possibilities with the idea
that an experimental validation should provide substantial ground to be competitive with mainstream
approaches, leaving space for a finer tuning in specific applications. Also, in order not to mix their
effects, we have treated (A) and (B) separately.
Architectures and datasets ? We provide in SM (Section VI) the detail of all experiments. To summarize, we consider two architectures in our experiments: DCGAN [31] and the multilayer feedforward
network (MLP) used in [30]. Our datasets are MNIST [19] and LSUN tower category [38].
Comparison of varying activations in the generator (A) ? We have compared ?-ReLUs with varying
? in [0, 0.1, ..., 1] (hence, we include ReLU as a baseline for ? = 1), the Softplus and the Least
Square Unit (LSU, Table 1) activation (Figure 1). For each choice of the activation function, all
inner layers of the generator use the same activation function. We evaluate the activation functions
by using both DCGAN and the MLP used in [30] as the architectures. As training divergence, we
adopt both GAN [15] and Wasserstein GAN (WGAN, [6]). Results are shown in Figure 1 (left).
7
Softplus
LSU
ReLU
?
?-ReLU
Softplus / LS / ReLU
Discriminator: varying link
Figure 1: Summary of our results on MNIST, on experiment A (left+center) and B (right). Left:
comparison of different values of ? for the ?-ReLU activation in the generator (ReLU = 1-ReLU, see
text). Thicker horizontal dashed lines present the ReLU average baseline: for each color, points above
the baselines represent values of ? for which ReLU is beaten on average. Center: comparison of
different activations in the generator, for the same architectures as in the left plot. Right: comparison
of different link function in the discriminator (see text, best viewed in color).
Three behaviours emerge when varying ?: either it is globally equivalent to ReLU (GAN DCGAN)
but with local variations that can be better (? = 0.7) or worse (? = 0), or it is almost consistently
better than ReLU (WGAN MLP) or worse (GAN MLP). The best results were obtained for GAN
DCGAN, and we note that the ReLU baseline was essentially beaten for values of ? yielding smaller
variance, and hence yielding smaller uncertainty in the results. The comparison between different
activation functions (Figure 1, center) reveals that (?-)ReLU performs overall the best, yet with some
variations among architectures. We note in particular that, in the same way as for the comparisons
intra ?-ReLU (Figure 1, left), ReLU performs relatively worse than the other criteria for WGAN
MLP, indicating that there may be different best fit activations for different architectures, which is
good news. Visual results on LSUN (SM, Table A5) also display the quality of results when changing
the ?-ReLU activation.
Comparison of varying link functions in the discriminator (B) ? We have compared the replacement
of the sigmoid function by a link which corresponds to the entropy which is theoretically optimal
in
?
.
boosting algorithms, Matsushita entropy [18, 28], for which ?MAT (z) = (1/2) ? (1 + z/ 1 + z 2 ).
Figure 1 (right) displays the comparison Matsushita vs "standard" (more specifically, we use sigmoid
in the case of GAN [30], and none in the case of WGAN to follow current implementations [6]). We
evaluate with both DCGAN and MLP on MNIST (same hyperparameters as for generators, ReLU
activation for all hidden layer activation of generators). Experiments tend to display that tuning the
link may indeed bring additional uplift: for GANs, Matsushita is indeed better than the sigmoid link
for both DCGAN and MLP, while it remains very competitive with the no-link (or equivalently an
identity link) of WGAN, at least for DCGAN.
7
Conclusion
It is hard to exaggerate the success of GAN approaches in modelling complex domains, and with
their success comes an increasing need for a rigorous theoretical understanding [34]. In this paper,
we complete the supervised understanding of the generalization of GANs introduced in [30], and
provide a theoretical background to understand its unsupervised part, showing in particular how deep
architectures can be powerful at tackling the generative part of the game. Experiments display that
the tools we develop may help to improve further the state of the art.
8
Acknowledgments
The authors thank the reviewers, Shun-ichi Amari, Giorgio Patrini and Frank Nielsen for numerous
comments.
References
[1] S.-M. Ali and S.-D.-S. Silvey. A general class of coefficients of divergence of one distribution from another.
Journal of the Royal Statistical Society B, 28:131?142, 1966.
8
[2]
[3]
[4]
[5]
[6]
[7]
[8]
[9]
[10]
[11]
[12]
[13]
[14]
[15]
[16]
[17]
[18]
[19]
[20]
[21]
[22]
[23]
[24]
[25]
[26]
[27]
[28]
[29]
[30]
[31]
[32]
[33]
[34]
[35]
[36]
[37]
[38]
S.-I. Amari. Differential-Geometrical Methods in Statistics. Springer-Verlag, Berlin, 1985.
S.-I. Amari. Information Geometry and Its Applications. Springer-Verlag, Berlin, 2016.
S.-I. Amari and H. Nagaoka. Methods of Information Geometry. Oxford University Press, 2000.
S.-I. Amari, A. Ohara, and H. Matsuzoe. Geometry of deformed exponential families: Invariant, dually-flat
and conformal geometries. Physica A: Statistical Mechanics and its Applications, 391:4308?4319, 2012.
M. Arjovsky, S. Chintala, and L. Bottou. Wasserstein GAN. CoRR, abs/1701.07875, 2017.
K. S. Azoury and M. K. Warmuth. Relative loss bounds for on-line density estimation with the exponential
family of distributions. MLJ, 43(3):211?246, 2001.
A. Ben-Tal, A. Ben-Israel, and M. Teboulle. Certainty equivalents and information measures: Duality and
extremal principles. J. of Math. Anal. Appl., pages 211?236, 1991.
J.-D. Boissonnat, F. Nielsen, and R. Nock. Bregman voronoi diagrams. DCG, 44(2):281?307, 2010.
S. Boyd and L. Vandenberghe. Convex optimization. Cambridge University Press, 2004.
D.-A. Clevert, T. Unterthiner, and S. Hochreiter. Fast and accurate deep network learning by exponential
linear units (ELUs). In 4th ICLR, 2016.
I. Csisz?r. Information-type measures of difference of probability distributions and indirect observation.
Studia Scientiarum Mathematicarum Hungarica, 2:299?318, 1967.
R.-M. Frongillo and M.-D. Reid. Convex foundations for generalized maxent models. In 33rd MaxEnt,
pages 11?16, 2014.
A. Genevay, G. Peyr?, and M. Cuturi. Sinkhorn-autodiff: Tractable Wasserstein learning of generative
models. CoRR, abs/1706.00292, 2017.
I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio.
Generative adversarial nets. In NIPS*27, pages 2672?2680, 2014.
I. Gulrajani, F. Ahmed, M. Arjovsky, V. Dumoulin, and A.-C. Courville. Improved training of wasserstein
GANs. CoRR, abs/1704.00028, 2017.
R.-H.-R. Hahnloser, R. Sarpeshkar, M.-A. Mahowald, R.-J. Douglas, and H.-S. Seung. Digital selection
and analogue amplification coexist in a cortex-inspired silicon circuit. Nature, 405:947?951, 2000.
M.J. Kearns and Y. Mansour. On the boosting ability of top-down decision tree learning algorithms. J.
Comp. Syst. Sc., 58:109?128, 1999.
Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-based learning applied to document recognition.
Proceedings of the IEEE, 86(11):2278?2324, 1998.
H. Lee, R. Ge, T. Ma, A. Risteski, and S. Arora. On the ability of neural nets to express distributions.
CoRR, abs/1702.07028, 2017.
Y. Li, K. Swersky, and R.-S. Zemel. Generative moment matching networks. In 32nd ICML, pages
1718?1727, 2015.
S. Liu, O. Bousquet, and K. Chaudhuri. Approximation and convergence properties of generative adversarial
learning. CoRR, abs/1705.08991, 2017.
A.-L. Maas, A.-Y. Hannun, and A.-Y. Ng. Rectifier nonlinearities improve neural network acoustic models.
In 30th ICML, 2013.
V. Nair and G. Hinton. Rectified linear units improve restricted Boltzmann machines. In 27th ICML, pages
807?814, 2010.
J. Naudts. Generalized thermostatistics. Springer, 2011.
X. Nguyen, M. J. Wainwright, and M. I. Jordan. Estimating divergence functionals and the likelihood ratio
by convex risk minimization. IEEE Transactions on Information Theory, 56(11):5847?5861, Nov 2010.
R. Nock, Z. Cranko, A.-K. Menon, L. Qu, and R.-C. Williamson. f -GANs in an information geometric
nutshell. CoRR, abs/1707.04385, 2017.
R. Nock and F. Nielsen. On the efficient minimization of classification-calibrated surrogates. In NIPS*21,
pages 1201?1208, 2008.
R. Nock and F. Nielsen. Bregman divergences and surrogates for learning. IEEE Trans.PAMI, 31:2048?
2059, 2009.
S. Nowozin, B. Cseke, and R. Tomioka. f -GAN: training generative neural samplers using variational
divergence minimization. In NIPS*29, pages 271?279, 2016.
A. Radford, L. Metz, and S. Chintala. unsupervised representation learning with deep convolutional
generative adversarial networks. In 4th ICLR, 2016.
M.-D. Reid and R.-C. Williamson. Composite binary losses. JMLR, 11, 2010.
M.-D. Reid and R.-C. Williamson. Information, divergence and risk for binary experiments. JMLR,
12:731?817, 2011.
T. Salimans, I.-J. Goodfellow, W. Zaremba, V. Cheung, A. Radford, and X. Chen. Improved techniques for
training gans. In NIPS*29, pages 2226?2234, 2016.
M. Telgarsky and S. Dasgupta. Agglomerative Bregman clustering. In 29 th ICML, 2012.
R.-F. Vigelis and C.-C. Cavalcante. On ?-families of probability distributions. J. Theor. Probab., 21:1?15,
2011.
L. Wolf, Y. Taigman, and A. Polyak. Unsupervised creation of parameterized avatars. CoRR,
abs/1704.05693, 2017.
F. Yu, Y. Zhang, S. Song, A. Seff, and J. Xiao. Lsun: Construction of a large-scale image dataset using
deep learning with humans in the loop. arXiv preprint arXiv:1506.03365, 2015.
9
| 6649 |@word deformed:10 mild:3 version:2 briefly:2 nd:1 suitably:1 open:2 d2:2 simplifying:1 pick:2 naudts:1 carry:1 moment:1 initial:1 liu:1 series:2 exclusively:1 denoting:2 document:1 current:2 com:1 activation:28 yet:1 tackling:1 distant:1 happen:1 shape:1 drop:1 plot:1 v:3 generative:8 guess:1 warmuth:1 ith:1 short:6 characterization:2 provides:2 boosting:2 readability:2 math:1 zhang:1 differential:2 prove:1 consists:1 fitting:1 theoretically:1 indeed:2 p1:1 mechanic:1 inspired:1 globally:1 decreasing:3 little:1 actual:1 increasing:4 indiscernibles:1 begin:1 boissonnat:1 notation:1 estimating:1 circuit:1 what:4 israel:1 finding:2 transformation:2 guarantee:3 certainty:1 concave:1 thicker:1 shed:1 nutshell:2 tie:1 exactly:1 ensured:1 classifier:2 zaremba:1 unit:3 reid:3 before:1 giorgio:1 local:1 limit:2 consequence:1 vig:4 oxford:1 meet:1 pami:1 au:1 shaded:1 appl:1 factorization:1 range:1 acknowledgment:1 lecun:1 arguing:1 maximisation:1 lf:5 significantly:3 composite:10 boyd:1 matching:1 regular:5 get:2 close:1 selection:3 coexist:1 put:1 risk:2 optimize:1 equivalent:4 measurable:1 map:1 center:3 reviewer:1 regardless:1 starting:1 l:1 convex:16 qc:1 pouget:1 insight:1 rule:2 vandenberghe:1 handle:1 coordinate:6 variation:2 qkp:6 resp:4 suppose:1 avatar:1 construction:1 exact:1 us:1 goodfellow:2 trick:1 recognition:1 preprint:1 capture:1 ensures:2 improper:1 news:1 principled:2 substantial:1 cuturi:1 ideally:1 seung:1 warde:1 dom:2 signature:5 raise:1 solving:3 depend:2 weakly:4 ali:2 creation:1 deliver:1 completely:1 basis:1 easily:2 indirect:1 chapter:1 sarpeshkar:1 fast:1 sc:1 tell:1 aggregate:1 zemel:1 exhaustive:1 whose:3 dominating:3 valued:1 say:2 distortion:5 amari:5 ability:2 statistic:5 nagaoka:1 ip:3 differentiable:2 net:11 hout:2 clevert:1 qin:3 relevant:1 loop:1 iff:2 chaudhuri:1 amplification:1 description:1 csisz:2 convergence:6 requirement:2 optimum:2 telgarsky:1 leave:3 ben:2 object:2 help:1 derive:1 develop:1 uninformed:1 received:1 eq:37 strong:1 sydney:1 involves:2 come:3 australian:1 quantify:1 elus:2 qd:1 nock:5 human:1 shun:1 require:2 behaviour:1 generalization:7 theor:1 mathematicarum:1 strictly:1 physica:1 hold:6 proximity:1 around:1 considered:1 ground:2 exp:5 scope:1 adopt:1 estimation:1 integrates:1 realizes:1 ohara:1 tanh:1 extremal:1 wl:8 tool:1 minimization:3 gaussian:2 frongillo:1 factorizes:1 varying:5 corollary:2 knob:1 cseke:1 focus:1 improvement:1 consistently:1 modelling:3 likelihood:1 contrast:1 adversarial:4 rigorous:1 baseline:4 sense:2 voronoi:1 factoring:1 dcg:1 initially:1 compactness:1 hidden:1 going:1 subroutine:2 unveil:1 arg:1 dual:2 overall:2 among:1 classification:1 temper:1 art:2 integration:2 equal:3 having:2 beach:1 ng:1 broad:2 yu:1 unsupervised:3 icml:4 simplex:1 mirza:1 richard:1 hint:2 strided:1 randomly:1 national:1 divergence:52 wgan:5 replaced:1 geometry:6 connects:2 replacement:1 csiro:1 ab:7 interest:1 mlp:7 a5:1 possibility:1 intra:1 elucidating:1 certainly:1 upperbounded:1 yielding:2 pc:2 light:1 farley:1 silvey:2 amenable:2 accurate:1 bregman:13 integral:1 respective:5 experiments1:1 tree:1 iv:1 old:1 logarithm:5 unterthiner:1 maxent:2 theoretical:3 teboulle:1 cover:1 tp:1 cumulants:1 mahowald:1 uniform:1 predictor:1 lsun:3 peyr:1 supx:1 data61:2 calibrated:1 st:1 density:18 fundamental:1 lee:1 physic:1 invertible:5 picking:1 connecting:1 gans:9 ambiguity:1 satisfied:4 possibly:1 worse:3 li:1 syst:1 nonlinearities:1 includes:1 coefficient:1 notable:1 explicitly:2 vi:2 try:1 view:2 closed:1 dumoulin:1 doing:1 sup:5 start:1 sort:1 relus:2 bayes:1 competitive:2 metz:1 contribution:4 square:1 convolutional:1 variance:1 who:1 yield:3 vout:6 none:1 comp:1 rectified:1 bob:1 finer:1 explain:1 whenever:1 definition:9 chintala:2 proof:10 associated:1 couple:1 dataset:1 treatment:5 popular:4 studia:1 thermostatistics:1 knowledge:1 realm:1 fractional:1 color:2 formalize:1 nielsen:4 actually:1 mlj:1 appears:1 dt:3 supervised:13 follow:1 specify:1 qkq:1 improved:2 formulation:13 done:1 strongly:7 furthermore:1 parameterised:1 just:2 implicit:4 hand:4 horizontal:1 expressive:1 reversible:1 lack:1 ikl:3 scientiarum:1 continuity:1 defines:1 brings:1 quality:1 menon:3 gulrajani:1 usa:1 name:3 omitting:1 effect:1 true:3 former:1 hence:8 equality:1 symmetric:1 leibler:2 deal:1 game:30 lastname:1 seff:1 criterion:1 generalized:4 prominent:2 presenting:1 complete:5 theoretic:5 lowerbounded:2 performs:2 l1:2 bring:2 patrini:1 geometrical:1 ranging:1 variational:16 wise:1 image:1 sigmoid:3 permuted:1 lizhen:1 extend:1 belong:3 slight:1 interpretation:1 expressing:1 refer:1 composition:1 cambridge:1 silicon:1 tuning:3 rd:6 risteski:1 longer:1 mainstream:1 sinkhorn:1 cortex:1 pkq:11 base:3 showed:2 recent:1 inf:3 store:1 verlag:2 inequality:1 binary:2 success:2 meeting:1 krishna:1 captured:1 wasserstein:5 additional:3 greater:1 care:1 somewhat:1 seen:1 arjovsky:2 converge:1 dashed:1 ii:3 relates:1 full:1 desirable:1 mix:1 reduces:1 technical:1 match:1 ahmed:1 long:1 justifying:1 equally:1 qg:4 variant:1 multilayer:1 essentially:4 expectation:1 optimisation:1 arxiv:2 grounded:1 normalization:5 sometimes:1 represent:1 achieved:1 hochreiter:1 subdifferential:2 want:4 fine:1 separately:1 background:1 completes:1 diagram:1 leaving:2 standpoint:2 crucial:1 extra:1 rest:1 operate:1 file:1 comment:1 subject:1 tend:1 elegant:2 member:1 jordan:1 call:2 noting:1 feedforward:1 iii:1 superset:2 enough:1 split:1 bengio:2 relu:25 fit:5 architecture:13 silent:1 inner:5 regarding:1 idea:1 haffner:1 polyak:1 whether:1 expression:1 granted:1 penalty:3 song:1 passing:2 proceed:1 remark:1 deep:25 generally:1 fake:5 informally:1 category:1 http:1 exist:1 canonical:1 notice:4 per:2 write:1 shall:3 unveiling:2 mat:1 express:1 ichi:1 key:7 dasgupta:1 changing:1 douglas:1 year:1 taigman:1 everywhere:1 powerful:4 parameterized:4 uncertainty:1 swersky:1 almost:3 family:28 throughout:1 reasonable:1 decision:1 layer:7 bound:1 matsushita:3 display:5 courville:2 adapted:1 constraint:3 flat:1 sake:1 tal:1 bousquet:1 integrand:1 aspect:1 argument:2 min:1 extremely:1 expanded:1 relatively:2 legendre:3 conjugate:3 remain:1 smaller:2 qu:2 shallow:1 happens:2 restricted:2 invariant:1 remains:1 hannun:1 turn:2 know:3 letting:6 ge:1 tractable:1 end:1 conformal:1 available:2 salimans:1 appearing:1 save:1 rz:1 original:2 top:1 clustering:1 ensure:1 include:1 gan:54 log2:1 especially:2 build:3 rdl:4 society:1 bl:6 objective:6 already:1 parametric:1 surrogate:2 gradient:1 iclr:2 distance:1 link:18 thank:1 berlin:2 accommodates:1 tower:1 manifold:1 agglomerative:1 extent:1 trivial:1 ozair:1 assuming:1 code:1 index:1 relationship:1 ratio:3 equivalently:1 unfortunately:1 postponing:1 robert:1 statement:1 relate:1 znet:2 lsu:3 frank:1 negative:3 drill:2 design:6 implementation:1 proper:13 anal:1 unknown:2 boltzmann:1 observation:4 convolution:1 datasets:2 sm:9 finite:1 extended:2 segregate:1 hinton:1 dc:5 frame:1 mansour:1 dually:1 arbitrary:3 introduced:1 vacuous:1 namely:2 kl:26 extensive:1 z1:1 discriminator:16 specified:1 rephrase:1 connection:1 acoustic:1 learned:3 nip:5 trans:1 able:1 adversary:2 below:1 firstname:1 summarize:4 built:1 max:1 royal:1 analogue:1 power:2 suitable:1 wainwright:1 natural:3 treated:1 circumvent:1 indicator:1 improve:3 github:1 numerous:2 picture:4 arora:1 concludes:2 ready:1 gf:6 hungarica:1 text:3 comply:2 geometric:16 literature:1 l2:1 understanding:2 probab:1 relative:1 loss:17 interesting:1 generator:29 validation:1 foundation:1 digital:1 sufficient:6 xiao:1 principle:3 displaying:1 story:1 nowozin:5 factorising:1 row:1 summary:1 maas:1 last:5 formal:2 side:5 allow:1 understand:1 fall:4 wide:2 emerge:1 absolute:1 leaky:3 dimension:3 world:1 valid:1 computes:1 author:1 made:1 coincide:1 nguyen:1 transaction:1 functionals:1 approximate:1 compact:2 nov:1 kullback:2 elu:4 reveals:1 factorize:2 continuous:4 decade:1 table:4 learn:1 nature:1 ca:1 flourishing:1 genevay:1 williamson:5 bottou:2 complex:5 domain:5 did:1 main:2 azoury:1 motivation:1 hyperparameters:1 repeated:1 xu:1 tomioka:1 position:1 explicit:1 exponential:29 lie:1 vanish:1 jmlr:2 learns:2 admissible:9 theorem:27 down:3 specific:1 rectifier:1 showing:2 abadie:1 admits:1 beaten:2 exaggerate:1 concern:1 derives:1 incorporating:1 exists:7 dl:2 mnist:3 corr:7 supplement:1 autodiff:1 chen:1 entropy:3 saddle:2 visual:1 aditya:2 dcgan:8 escort:17 springer:3 radford:2 corresponds:1 wolf:1 satisfies:3 relies:1 ma:1 prop:3 nair:1 hahnloser:1 identity:11 viewed:1 cheung:1 price:1 fkl:2 hard:1 specifically:1 sampler:1 lemma:3 kearns:1 called:6 total:2 duality:1 experimental:3 player:2 meaningful:1 indicating:1 formally:1 support:4 softplus:6 latter:1 cumulant:2 absolutely:2 constructive:2 evaluate:2 tested:1 ex:11 |
6,245 | 665 | Generalization Abilities of
Cascade Network Architectures
E. Littmann*
H. Ritter
Department of Information Science
Bielefeld University
D-4800 Bielefeld, FRG
[email protected]
Department of Information Science
Bielefeld University
D-4800 Bielefeld, FRG
[email protected] bielefeld.de
Abstract
In [5], a new incremental cascade network architecture has been
presented. This paper discusses the properties of such cascade
networks and investigates their generalization abilities under the
particular constraint of small data sets. The evaluation is done for
cascade networks consisting of local linear maps using the MackeyGlass time series prediction task as a benchmark. Our results indicate that to bring the potential of large networks to bear on the
problem of extracting information from small data sets without running the risk of overjitting, deeply cascaded network architectures
are more favorable than shallow broad architectures that contain
the same number of nodes.
1
Introduction
For many real-world applications, a major constraint for the successful learning
from examples is the limited number of examples available. Thus, methods are
required, that can learn from small data sets. This constraint makes the problem
of generalization particularly hard. If the number of adjustable parameters in a
* to whom correspondence should be sent
188
Generalization Abilities of Cascade Network Architectures
network approaches the number of training examples, the problem of overfitting occurs and generalization becomes very poor. This severely limits the size of networks
applicable to a learning task with a small data set. To achieve good generalization
also in these cases, particular attention must be paid to a proper architecture chosen
for the network. The better the architecture matches the structure of the problem
at hand, the better is the chance to achieve good results even with small data sets
and small numbers of units.
In the present paper, we address this issue for the class of so called Cascade Network
Architectures [5, 6] on the basis of an empirical approach, where we use the MackeyGlass time series prediction as a benchmark problem. In our experiments we want
to exploit the potential of large networks to bear on the problem of extracting
information from small data sets without running the risk of overfitting. Our results
indicate that it is more favorable to use deeply cascaded network architectures than
shallow broad architectures, provided the same number of nodes is used in both
cases. The width of each individual layer is essentially determined by the size of
the training data set. The cascade depth is then matched to the total number of
nodes available.
2
Cascade Architecture
So far, mainly architectures with few layers containing many units have been considered, while there has been very little research on narrow, but deeply cascaded
networks. One of the few exceptions is the work of Fahlman [1], who proposed
networks trained by the cascade-correlation algorithm. In his original approach,
training is strictly feed-forward and the nonlinearity is achieved by incrementally
adding percept ron units trained to maximize the covariance with the residual error.
2.1
Construction Algorithm
In [5] we presented a new incremental cascade network architecture based 011 error
minimization instead of covariance maximization. This leads to all architecture
that differs significantly from Fahlman's proposal and allows an inversion of the
construction process of the network. Thus, at each stage of the construction of the
network all cascaded modules provide an approximation of the target function t(e),
albeit corresponding to different states of convergence (Fig. 1).
The algorithm starts with the training of a neural module with output yeo) to
approximate a target function t(e), yielding
(1)
the superscript (0) indicating the cascade level. After an arbitrary number of training epochs, the weight vector w(O) becomes "frozen". Now we add the output y(O)
of this module as a virtual input unit and train another neural module as new output
189
190
Littmann and Ritter
Neural
Module
Output
Cascade layer2
Neural
Module
(Output)
Cascade layer 1
(Output)
Neural
Module
Xn
Input
{
X2
X,
Bias
1
Figure 1: Cascade Network Architecture
unit
y(1)
with
(2)
=
where x(1)(~)
{x(O)(~),y(O)(~)} denotes the extended input. This procedure can
be iterated arbitrarily and generates a network sttucture as shown in Fig. 1.
2.2
Cascade Modules
The details and advantages of this approach are discussed in [5, 6]. In particular,
this architecture can be applied to any arbitrary nonlinear module. It does not rely
on the availability of a procedure for error backpropagation. Therefore, it is also
applicable to (and has been extensively tested with) pure feed-forward approaches
like simple perceptrons [5] and vector quantization or "Local linear maps" ("LLM
networks") [6, 7].
2.3
Local Linear Maps
LLM networks have been introduced earlier ((Fig. 2); for details, d. [11, 12]) and
are related to the GRBF -approach [10] and the self-organizing maps [2, 3, 11].
They consist of N units r = 1, ... ,N, with an input weight vector w~in) E }RL, an
output weight vector w~out) E }RM and a MxL-matrix Ar for each unit r.
Generalization Abilities of Cascade Network Architectures
Output space
Input space
Figure 2: LLM Network Architecture
The output
y(net)
of a single LLM-network for an input feature vector x E }RL is
(3)
the "winner" node s determined by the minimality condition
(4)
This leads to the learning steps for a training sample (x(o), yeo?):
f
1
(x(o) _
W(in?)
8'
(5)
(6)
(7)
applied for T samples (x(o),y(o?),a = 1,2, ... T, and 0 < fi ? 1, i = 1,2,3
denote learning step sizes. The additional term in (6), not given in [11, 121, leads
to a better decoupling of the effects of (5) and (6,7).
3
Experiments
In order to evaluate the generalization performance of this architecture, we consider the problem of time series prediction based on the Mackey-Glass differential
equation, for which results of other networks already have been reported in the
literature.
191
192
Littmann and Ritter
1.25
]
:=
t
1
0.75
0.50
o
100
200
300
400
Time (t)
Figure 3: Mackey-Glass function
3.1
Time Series Prediction
Lapedes and Farber [4] introduced the prediction of chaotic time series as a benchmark problem. The data is based on the Mackey-Glass differential equation [8]:
x(t)
= -bx(t) + (ax(t -
r))J(l
+ xlO(t -
r)).
(8)
With the parameters a = 0.2, b = 0.1, and r = 17, this equation produces a chaotic
time series with a strange attractor of fractal dimension d ~ 2.1 (Fig. 3). The
input data is a vector x(t) = {x(t),x(t - ~),x(t - 2~),x(t - 3~)}T. The learning
task is defined to predict the value x(t + P). To facilitate comparison, we adopt
the standard choice ~ = 6 and P = 85. Results with these parameters have been
reported in [4, 9, 13].
The data was generated by integration with 30 steps per time unit. We performed
different numbers of training epochs with samples randomly chosen from training
sets consisting of 500 (5000 resp.) samples. The performance was measured on an
independent test set of 5000 samples. All results are averages over ten runs. The
error measure is the normalized root mean squ.are error (NRMSE), i.e. predicting
the average value yields an error value of 1.
4
Results and Discussion
The training of the single LLM networks was performed without extensive parameter tuning. If fine tuning for each cascade unit would be necessary, the training
would be unattractively expensive.
The first results were achieved with cascade networks consisting of LLM units after
30 training epochs per layer on a learning set of 500 samples. Figs. 4 and 5 represent
the performance of such LLM cascade networks on the independent test set for
different numbers of cascaded layers as a function of the number of nodes per layer
Generalization Abilities of Cascade Network Architectures
NRMSE
0.5
0.4
----+--- - - - -
E3 1 layer B
2 layers
B
B
5 layers
I -_?? \?_????--1
4 layers
E3
'~----"'---"---"-'--'--~"--'---r~"--
?.. ., .... ? ??... .?.,?. ? ?r..1 ? ? ?
0.3
0 .5
3 layers
,
.
0 ??
D. l
I
. . ..L...?~~-~~j
0.2
0.1 .....................................
...... ...
~.- _.~?????:i
?? --.?????
.. ,J....
. .
~
?? ?? ???????? ~ ?? ?????????? <t ??? ????????? ; ?.? ?????? ?? J
:
:
,
!
':
!
O~------------------------------~?
10 20 30 40 50 60 70 80 90 100
# nodes per layer
Figure 4: Iso-Layer-Dependence
lO),8I"S
Nodes
90
ID
100
Figure 5: Error Landscape
("iso-layer-curves"). The graphs indicate that there is an optimal number N~!~
of nodes for which the performance of the single layer network has a best value
p!~:. Within the single layer architecture, additional nodes lead to a decrease of
performance due to overfitting. This can only be avoided if the training set is
enlarged, since N~!~ grows with the number of available training examples.
However, Figs. 4 and 5 show that adding more units in the form of an additional,
cascaded layer allows to increase performance significantly beyond p!~:. Similarly,
the optimal performance of the resulting two-layer network cannot be improved
beyond an optimal value p!;~ by arbitrarily increasing the number of nodes in the
two-layer system. However, adding a third cascaded layer again allows to make use
of more nodes to improve performance further, although this time the relative gain
is smaller than for the first cascade step. The same situation repeats for larger
numbers of cascaded layers. This suggests that the cascade architecture is very
suitable to exploit the computational capabilities of large numbers of nodes for the
task of building networks that generalize well from small data sets without running
into the problem of overfitting when many nodes are used .
A second way of comparing the benefits of shallow and broad versus narrow and
deep architectures is to compare the performance achieveable by distributing a fixed
number N of nodes over different numbers L of cascaded layers. Fig. 6 shows the
result for the same benchmark problem as in Fig. 4, each graph belonging to one of
the values N = 40,60,120,240 nodes and representing the NRMSE for distributing
the N nodes among L layers of N / L nodes each t, L ranging from 1 to 10 layers
("iso-nodes-curves" ).
1 rounding
to the nearest integral, whenever N / L is nonintegral.
193
194
Littmann and Ritter
----t-------
E3 240 E3
120 B 60 B 40 Nodes
0.5 ........................\.............. ................. ..... ... ...................... ;.. ....:
NRMSE
-1:------_
G??
G.3
0.4
0.2
0.3
0.2
D.5
,l,- :.'
:....,'..: .... '
?
'
.... ___
~......
.
~.~_~
.
0.1 .......... t.~.~.~.~~.~ ..~
????
'.1
! ???? .,
.:
:
:
~~~:r.'=
. . ??????????1????????????:
???????????,
:
j
o~----~--------------------~
4
6
7
8
9 10
1
2
3
5
I
2
J
,?
l"'....
L = # cascaded layers
Figure 6: Iso-Nodes-Dependence
Figure 7: Nodes-Layer-Dependence
The results show that
(i) the optimal number of layers increases monotonously with - and is roughly
proportional to - the number of nodes to be used.
(ii) if for each number of nodes the optimal number of layers is used, performance increases monotonously with the number of available nodes, and
thus, as a consequence of (i), with the number of cascaded layers.
These results are not restricted to small data sets only. The application of the
cascade algorithm is also useful if larger training sets are available. Fig. 7 represents
the performance of LLM cascade networks on the test set after 300 training epochs
overall on a learning set consisting of 5000 samples. As could be expected, there is
still no sign of overfitting, even using LLM networks with 100 nodes per layer. But
regardless of the size of the single LLM unit, network performance is improved by
the cascade process at least in a zone involving a total of some 300 nodes in the
whole cascade.
5
Conclusions
Summarizing, we find that Cascade Network Architectures allow to use the benefits
of large numbers of nodes even for small training data sets, and still bypass the
problem of overfitting. To achieve this, the "width" of each layer must be matched
to the size of the training set. The "depth" of the cascade then is determined by
the total number of nodes available.
Generalization Abilities of Cascade Network Architectures
Acknowledgements
This work was supported by the German Ministry of Research and Technology
(BMFT), Grant No. ITN9104AO. Any responsibility for the contents of this publication is with the authors.
References
[1] Fahlman, S.E., and Lebiere, C. (1989), "The Cascade-Correlation Learning
Architecture", in Advances in Neural Information Processing Systems II, ed.
D.S. Touretzky, pp. 524-532.
[2] Kohonen, T. (1984), Self-Organization and Associative Memory, Springer Series in Information Sciences 8, Springer, Heidelberg.
[3] Kohonen, T. (1990), "The Self-Organizing Map", in Proc. IEEE 78, pp. 14641480.
[4] Lapedes, A., and Farber, R. (1987), "Nonlinear signal processing using neural
networks; Prediction and system modeling", TR LA-UR-87-2662
[5] Littmann, E., Ritter, H. (1992), "Cascade Network Architectures", in Proc.
Intern. Joint Conference On Neural Networks, pp. II/398-404, Baltimore.
[6] Littmann, E., Ritter, H. (1992), "Cascade LLM Networks", in Artificial Neural Networks II, eds. I. Aleksander, J. Taylor, pp. 253-257, Elsevier Science
Publishers (North Holland).
[7] Littmann, E., Meyering, A., Ritter, H. (1992), "Cascaded and Parallel Neural Network Architectures for Machine Vision - A Case Study", in Proc. 14.
DAGM-Symposium 1992, Dresden, ed. S. Fuchs, pp. 81-87, Springer, Heidelberg.
[8] Mackey, M., and Glass, 1. (1977), "Oscillations and chaos in physiological
control systems", in Science, pp. 287-289.
[9] Moody, J., Darken, C. (1988). "Learning with Localized Receptive Fields", in
Proc. of the 1988 Connectionist Models Summer School, Pittsburg, pp. 133143, Morgan Kaufman Publishers, San Mateo, CA.
[10] Poggio, T., Edelman, S. (1990), "A network that learns to recognize threedimensional objects", in Nature 343, pp. 263-266.
[11] Ritter, H. (1991), "Learning with the Self-organizing Map", in Artificial Neural
Networks 1, eds. T. Kohonen, K. Makisara, O. Simula, J. Kangas, pp. 357-364,
Elsevier Science Publishers (North-Holland).
[12] Ritter, H., Martinetz, T., Schulten, K. (1992). Neural Computation and Selforganizing Maps, Addison-Wesley, Reading, MA.
[13] Walter, J., Ritter, H., Schulten, K. (1990). "Non-linear prediction with selforganizing maps", in Proc. Intern. Joint Conference On Neural Networks, San
Diego, Vol.1, pp. 587-592.
195
| 665 |@word inversion:1 covariance:2 paid:1 tr:1 series:7 lapedes:2 meyering:1 comparing:1 must:2 mackey:4 iso:4 node:28 ron:1 differential:2 symposium:1 edelman:1 expected:1 roughly:1 little:1 increasing:1 becomes:2 provided:1 matched:2 kaufman:1 rm:1 control:1 unit:12 grant:1 local:3 limit:1 severely:1 consequence:1 id:1 dresden:1 mateo:1 suggests:1 limited:1 differs:1 backpropagation:1 chaotic:2 procedure:2 empirical:1 cascade:32 significantly:2 cannot:1 risk:2 map:8 attention:1 regardless:1 pure:1 his:1 techfak:2 construction:3 target:2 resp:1 diego:1 simula:1 expensive:1 particularly:1 module:9 decrease:1 deeply:3 trained:2 basis:1 joint:2 train:1 walter:1 artificial:2 larger:2 ability:6 superscript:1 associative:1 advantage:1 frozen:1 net:1 kohonen:3 organizing:3 achieve:3 convergence:1 produce:1 incremental:2 object:1 measured:1 nearest:1 school:1 indicate:3 farber:2 virtual:1 frg:2 generalization:10 strictly:1 considered:1 predict:1 major:1 adopt:1 favorable:2 proc:5 applicable:2 minimization:1 aleksander:1 publication:1 ax:1 mainly:1 summarizing:1 glass:4 elsevier:2 dagm:1 layer2:1 issue:1 among:1 overall:1 pittsburg:1 integration:1 field:1 makisara:1 represents:1 broad:3 connectionist:1 few:2 randomly:1 recognize:1 individual:1 consisting:4 attractor:1 organization:1 evaluation:1 yielding:1 integral:1 necessary:1 poggio:1 taylor:1 earlier:1 modeling:1 ar:1 maximization:1 successful:1 rounding:1 monotonously:2 reported:2 minimality:1 ritter:10 moody:1 again:1 containing:1 yeo:2 bx:1 potential:2 de:2 availability:1 north:2 performed:2 root:1 responsibility:1 start:1 capability:1 parallel:1 who:1 percept:1 yield:1 landscape:1 generalize:1 iterated:1 touretzky:1 whenever:1 ed:4 pp:10 lebiere:1 gain:1 feed:2 wesley:1 improved:2 done:1 stage:1 correlation:2 hand:1 grbf:1 nonlinear:2 incrementally:1 grows:1 building:1 effect:1 facilitate:1 contain:1 normalized:1 width:2 self:4 bring:1 ranging:1 chaos:1 fi:1 rl:2 winner:1 discussed:1 tuning:2 similarly:1 nonlinearity:1 add:1 arbitrarily:2 morgan:1 ministry:1 additional:3 maximize:1 signal:1 ii:4 match:1 prediction:7 involving:1 essentially:1 vision:1 represent:1 achieved:2 proposal:1 want:1 fine:1 baltimore:1 publisher:3 martinetz:1 sent:1 extracting:2 architecture:28 fuchs:1 distributing:2 e3:4 fractal:1 deep:1 useful:1 selforganizing:2 extensively:1 ten:1 sign:1 per:5 vol:1 nrmse:4 graph:2 run:1 bielefeld:6 strange:1 oscillation:1 investigates:1 layer:31 summer:1 correspondence:1 constraint:3 x2:1 generates:1 department:2 poor:1 belonging:1 smaller:1 ur:1 helge:1 shallow:3 restricted:1 equation:3 discus:1 german:1 addison:1 available:6 original:1 denotes:1 running:3 littmann:8 exploit:2 threedimensional:1 already:1 occurs:1 receptive:1 dependence:3 whom:1 proper:1 adjustable:1 darken:1 benchmark:4 situation:1 extended:1 kangas:1 arbitrary:2 introduced:2 required:1 extensive:1 narrow:2 address:1 beyond:2 reading:1 memory:1 suitable:1 rely:1 predicting:1 cascaded:12 residual:1 representing:1 improve:1 technology:1 epoch:4 literature:1 acknowledgement:1 relative:1 bear:2 proportional:1 versus:1 localized:1 mackeyglass:2 bypass:1 lo:1 repeat:1 fahlman:3 supported:1 bias:1 allow:1 benefit:2 curve:2 depth:2 xn:1 world:1 dimension:1 forward:2 author:1 san:2 avoided:1 far:1 approximate:1 uni:2 overfitting:6 llm:11 learn:1 nature:1 ca:1 decoupling:1 heidelberg:2 whole:1 enlarged:1 fig:9 schulten:2 third:1 learns:1 achieveable:1 physiological:1 consist:1 quantization:1 albeit:1 adding:3 bmft:1 intern:2 xlo:1 holland:2 springer:3 chance:1 ma:1 content:1 hard:1 determined:3 called:1 total:3 la:1 perceptrons:1 exception:1 indicating:1 zone:1 evaluate:1 tested:1 |
6,246 | 6,650 | Toward Multimodal Image-to-Image Translation
Jun-Yan Zhu
UC Berkeley
Trevor Darrell
UC Berkeley
Richard Zhang
UC Berkeley
Alexei A. Efros
UC Berkeley
Oliver Wang
Adobe Research
Deepak Pathak
UC Berkeley
Eli Shechtman
Adobe Research
Abstract
Many image-to-image translation problems are ambiguous, as a single input image
may correspond to multiple possible outputs. In this work, we aim to model
a distribution of possible outputs in a conditional generative modeling setting.
The ambiguity of the mapping is distilled in a low-dimensional latent vector,
which can be randomly sampled at test time. A generator learns to map the given
input, combined with this latent code, to the output. We explicitly encourage the
connection between output and the latent code to be invertible. This helps prevent
a many-to-one mapping from the latent code to the output during training, also
known as the problem of mode collapse, and produces more diverse results. We
explore several variants of this approach by employing different training objectives,
network architectures, and methods of injecting the latent code. Our proposed
method encourages bijective consistency between the latent encoding and output
modes. We present a systematic comparison of our method and other variants on
both perceptual realism and diversity.
1
Introduction
Deep learning techniques have recently made steep progress in conditional image generation. For
example, networks have been used to inpaint missing image regions [15, 28, 41], create sentence
conditioned generations [43], add color to grayscale images [14, 15, 21, 44], and sketch-to-photo [15,
34]. However, most techniques in this space have focused on generating a single result conditioned
on the input. In this work, our focus is to model a distribution of potential results, as many of these
problems may be multimodal or ambiguous in nature. For example, in a night-to-day translation task
(see Figure 1), an image captured at night may correspond to many possible day images with different
types of lighting, sky and clouds. There are two main goals of the conditional generation problem:
producing results which are (1) perceptually realistic and (2) diverse, while remaining faithful to the
input. This multimodal mapping from a high-dimensional input to a distribution of high-dimensional
outputs makes the conditional generative modeling task challenging. In existing approaches, this
leads to the common problem of mode collapse [10], where the generator learns to generate only
a small number of unique outputs. We systematically study a family of solutions to this problem,
which learn a low-dimensional latent code for aspects of possible outputs which are not contained in
the input image. The generator then produces an output conditioned on both the given input and this
learned latent code.
We start with the pix2pix framework [15] which has previously been shown to produce good-quality
results for a variety of image-to-image translation tasks. The pix2pix method works by training a
generator network, conditioned on the input image, with two losses (1) a regression loss to produce
a similar output to the known paired ground truth image and (2) a learned discriminator loss to
encourage realism. The authors note that trivially appending a randomly drawn latent code did not
help produce diverse results, and using random dropout at test time only helped marginally. Instead,
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
?
(a) Input night image
(b) Diverse day images sampled by our model
Figure 1: Multimodal image-to-image translation using our proposed method: given an input image from one
domain (night image of a scene), we aim to model a distribution of potential outputs (corresponding day images)
in the target domain, producing both realistic and diverse results.
we propose encouraging a bijection between the output and latent space. Not only do we perform
the direct task of mapping from the input and latent code to the output, we also jointly learn an
encoder from the output back to the latent space. This discourages two different latent codes from
generating the same output (non-injective mapping). During training, the learned encoder is trained
to find a latent code vector that corresponds to the ground truth output image, while passing enough
information to the generator to resolve any ambiguities about the mode of output. For example, when
generating a day image from a night one (Figure 1), the latent vector may encode information about
the sky color, lighting effects on the ground and the density and shape of clouds. In both cases,
applying the encoder and generator, in either order, should be self-consistent, and result in either the
same latent code or the same image.
In this work, we instantiate this idea by exploring several objective functions inspired by the literature
in unconditional generative modeling:
? cVAE-GAN (Conditional Variational Autoencoder GAN): One popular approach to model multimodal output distribution is by learning the distribution of latent encoding given the output as
popularized by VAEs [18]. In the conditional setup (similar to unconditional analogue [20]), we
enforce that the latent distribution encoded by the desired output image maps back to the output via
conditional generator. The latent distribution is regularized using KL-divergence to be close to a
standard normal so as to sample random codes during inference. This variational objective is then
optimized jointly with the discriminator loss.
? cLR-GAN (Conditional Latent Regressor GAN): Another approach to enforce mode-capture in latent
encoding is to explicitly model the inverse mapping. Starting from a randomly sampled latent
encoding, the conditional generator should result into an output which when given itself as input to
the encoder should result back into the same latent code, enforcing self-consistency. This method
could be seen as a conditional formulation of the ?latent regressor" model [6, 8] and also related to
InfoGAN [2].
? BiCycleGAN: Finally, we combine both these approaches to enforce the connection between latent
encoding and output in both directions jointly and achieve improved performance. We show that
our method can produce both diverse and visually appealing results across a wide range of imageto-image translation problems, significantly more diverse than other baselines, including naively
adding noise in the pix2pix framework. In addition to the loss function, we study the performance
with respect to several encoder networks, as well as different ways of injecting the latent code into
the generator network.
We perform a systematic evaluation of these variants by using real humans to judge photo-realism
and an automated distance metric to assess output diversity.
2
Related Work
Generative modeling Parametric modeling of the natural image distribution is a challenging
problem. Classically, this problem has been tackled using autoencoders [13, 38] or Restricted
Boltzmann machines [36]. Variational autoencoders [18] provide an effective approach to model
2
Input Image
?
Ground truth output
Network output
?
?"
?
?(?)
Loss
?"
?
?(?)
?* + ?
?
Deep network
Target latent distribution
Sample from distribution
(a) Testing Usage for all models
(b) Training pix2pix+noise
?*
?* + ?
?
?
?
?
?
?"
?
?(?|?)
z
?"
?
??
?(?)
??
?
?(?)
(c) Training cVAE-GAN
?
(d) Training cLR-GAN
(e) Training BiCycleGAN
Figure 2: Overview: (a) Test time usage of all the methods. To produce a sample output, a latent code z is
first randomly sampled from a known distribution (a standard normal distribution). A generator G maps input
? (yellow). (b) pix2pix+noise [15] baseline,
image A (blue) and latent sample z to produce output sample B
with additional input B (blue) that corresponds to A. (c) cVAE-GAN (and cAE-GAN) start from ground truth
target image B and encode it into the latent space. The generator then attempts to map the input image A
along with a sampled z back into original image B. (d) cLR-GAN randomly samples a latent code from a known
? and then tries to reconstruct the latent code from it. (e) Our
distribution, uses it to map A into the output B,
hybrid BiCycleGAN method combines constraints in both directions.
stochasticity within the network by re-parametrization of a latent distribution. Other approaches
to modeling stochasticity include auto-regressive models [26, 27] which can model multimodality
via sequential conditional prediction. These approaches are trained with a pixel-wise independent
loss on samples of natural images using maximum likelihood and stochastic back-propagation. This
is a disadvantage because two images, which are close regarding a pixel-wise independent metric,
may be far apart on the manifold of natural images. Generative adversarial networks [11] overcome
this issue by learning the loss function using a discriminator network, and have recently been very
successful [1, 2, 4, 6, 8, 29, 30, 43, 46, 47]. Our method builds on the conditional version of VAE [18]
and InfoGAN [2] or latent regressor [6] model via an alternating joint optimization to learn diverse
and realistic samples. We revisit this connection in Section 3.4.
Conditional image generation Potentially, all of the methods defined above could be easily
conditioned. While conditional VAEs [37] and conditional auto-regressive models [26, 27] have
shown promise [39, 40], image-to-image conditional GANs have lead to a substantial boost in
the quality of the results. However, the quality has been attained at the expense of multimodality,
as the generator learns to largely ignore the random noise vector when conditioned on a relevant
context [15, 28, 34, 41, 48]. In fact, it has even been shown that ignoring the noise leads to more
stable training [15, 23, 28].
Explicitly-encoded multimodality One way to express multiple modes in the output is to encode
them, conditioned on some mode-related context in addition to the input image. For example, color
and shape scribbles and other interfaces were used as conditioning in iGAN [47], pix2pix [15],
Scribbler [34] and interactive colorization [45]. Another option is to explicitly force diversity amongst
a mixture of models [9]. While there has been some degree of success for generating multimodal
outputs in unconditional setups, and conditional text generation [5, 11, 20, 25, 30], conditional
image-to-image generation is still far from achieving the same results, unless explicitly encoded as
discussed above. In this work, we learn conditional generation models for modeling multiple modes
of output by enforcing tight connections in both image and latent space.
3
3
Multimodal Image-to-Image Translation
Our goal is to learn a multi-modal mapping between two image domains, for example, edges and
photographs, or day and night images, etc. Consider the input domain A ? RH?W?3 which is to be
mapped to an output domain
B ? RH?W?3
. During training, we are given a dataset of paired instances
from these domains, (A ? A, B ? B) , which is representative of a joint distribution p(A, B). It is
important to note that there could be multiple plausible paired instances B that would correspond to
an input instance A, but the training dataset usually contains only one such pair. However, given a
b
new instance A during test time, our model should be able to generate a diverse set of output B?s,
corresponding to different modes in the distribution p(B|A).
While conditional GANs have achieved success in image-to-image translation tasks [15, 48], they
b given the input image A. On the other
are primarily limited to generating deterministic output B
b from true conditional
hand, we would like to learn the mapping that could sample the output B
distribution given A, and produce results which are both diverse and realistic. To do so, we learn a
low-dimensional latent space z ? RZ , which encapsulates the ambiguous aspects of the output mode
which are not present in the input image. For example, a sketch of a shoe could map to a variety of
colors and textures, which could get compressed in this latent code. We then learn a deterministic
mapping G : (A, z) ? B to the output. To enable stochastic sampling, we desire the latent code
vector z to be drawn from some prior distribution p(z); we use a standard Gaussian distribution
N (0, I) in this work.
We first discuss a simple extension of existing methods and discuss its strengths and weakness,
motivating the development of our proposed approach in the subsequent subsections.
3.1
b
Baseline: pix2pix+noise (z ? B)
The recently proposed pix2pix model [15] has shown high quality results in image-to-image
translation setting. It uses conditional adversarial networks [11, 24] to help produce perceptually
realistic results. GANs train a generator G and discriminator D by formulating their objective as an
adversarial game. The discriminator attempts to differentiate between real images from the dataset
and fake samples produced by the generator. Randomly drawn noise z is added to attempt to induce
stochasticity. We illustrate the formulation in Figure 2(b) and describe it below.
LGAN (G, D) = EA,B?p(A,B) [log(D(A, B))] + EA?p(A),z?p(z) [log(1 ? D(A, G(A, z)))] (1)
To encourage the output of the generator to match the input as well as stabilize the GANs training,
we use an `1 loss between the output and the ground truth image.
Limage
(G) = EA,B?p(A,B),z?p(z) ||B ? G(A, z)||1
1
(2)
The final loss function is comprised of both the GANs term and the `1 term, where ? balances the
two terms.
G? = arg min max LGAN (G, D) + ?Limage
(G)
(3)
1
G
D
In this scenario, there is no incentive for the generator to make use of the noise vector which encodes
random information. It was also noted in the preliminary experiments in [15] that the generator simply
ignored the added noise and hence the noise was removed in their final experiments. This observation
is consistent with Mathieu et al. [23], Pathak et al. [28] and the mode collapse phenomenon observed
in unconditional cases [10, 33]. In this paper, we explore different ways to explicitly enforce the
generator to use the latent encoding by making it capture relevant information than being random.
3.2
b
Conditional Variational Autoencoder GAN: cVAE-GAN (B ? z ? B)
One way to force the latent code z to be ?useful" is to directly map the ground truth B to it
using an encoding function E. The generator G then uses both the latent code and the input
b The overall model can be easily understood as the
image A to synthesize the desired output B.
reconstruction of B, with latent encoding b
z concatenated with the paired A in the middle ? similar to
an autoencoder [13]. This interpretation is better shown in Figure 2(c).
4
This approach has been successfully investigated in Variational Auto-Encoders [18] in the unconditional scenario without the adversarial objective. Extending it to conditional scenario, the distribution
Q(z|B) of latent code z using the encoder E with a Gaussian assumption, Q(z|B) , E(B). To
reflect this, Equation 1 is modified to sampling z ? E(B) using the re-parameterization trick,
allowing direct back-propagation [18].
LVAE
GAN = EA,B?p(A,B) [log(D(A, B))] + EA,B?p(A,B),z?E(B) [log(1 ? D(A, G(A, z)))]
(4)
We make the corresponding change in the `1 loss term in Equation 2 as well to obtain LVAE
1 (G) =
EA,B?p(A,B),z?E(B) ||B ? G(A, z)||1 . Further, the latent distribution encoded by E(B) is encouraged to be close to random gaussian so as to sample z at inference (i.e., B is not known).
LKL (E) = EB?p(B) [DKL (E(B)|| N (0, I))],
R
where DKL (p||q) = ? p(z) log
of the VAE-GAN [20] as
p(z)
q(z) dz.
G? = arg min max
G,E
This forms our cVAE-GAN objective, a conditional version
VAE
LVAE
GAN (G, D, E) + ?L1 (G, E) + ?KL LKL (E).
D
(5)
(6)
As a baseline, we also consider the deterministic version of this approach, i.e., dropping KLdivergence and encoding z = E(B). We call it cAE-GAN and show comparison in the experiments.
However, there is no guarantee in cAE-GAN on the distribution of the latent space z, which makes
test-time sampling of z difficult.
3.3
b ?b
Conditional Latent Regressor GAN: cLR-GAN (z ? B
z)
We explore another method of enforcing the generator network to utilize the latent code embedding z,
while staying close to the actual test time distribution p(z), but from the latent code?s perspective.
We start from the latent code z, as shown in Figure 2(d), and then enforce that E(G(A, z)) map back
to the randomly drawn latent code with an `1 loss. Note that the encoder E here is producing a point
estimate for b
z, whereas the encoder in the previous section was predicting a Gaussian distribution.
Llatent
(G, E) = EA?p(A),z?p(z) ||z ? E(G(A, z))||1
1
(7)
We also include the discriminator loss LGAN (G, D) (Equation 1) on to encourage the network to
generate realistic results, and the full loss can be written as:
G? = arg min max
G,E
D
LGAN (G, D) + ?latent Llatent
(G, E)
1
(8)
The `1 loss for the ground truth image B is not used. In this case, since the noise vector is randomly
b to be the ground truth, but rather a realistic result. The above
drawn, we do not want the predicted B
b is
objective bears similarity to the ?latent regressor" model [2, 6, 8], where the generated sample B
encoded to generate a latent vector.
3.4
Our Hybrid Model: BiCycleGAN
To take the advantages of both directions, we present a hybrid model by combining cVAE-GAN and
b and z ? B
b ?b
cLR-GAN objectives, as well as combining the two GANs cycles: B ? z ? B
z
(hence the name BiCycleGAN). For cVAE-GAN, the encoding is learned from real data, but a random
latent code may not yield realistic images at test time due to two reasons. First, the KL loss may not
be well optimized. Secondly, our adversarial classifier D does not have a chance to see randomly
sampled results during training. In cLR-GAN, the latent space is easily sampled from a simple
distribution, but the generator is trained without the benefit of seeing ground truth input-output pairs.
We propose to train with constraints in both directions.
G? = arg min max
G,,E
D
VAE
LVAE
GAN (G, D, E) + ?L1 (G, E)
+LGAN (G, D) + ?latent Llatent
(G, E) + ?KL LKL (E),
1
where the hyper-parameters ?, ?latent , and ?KL control the importance of each term.
5
(9)
+
+
z
z
Figure 3: Alternatives for injecting z into generator. Latent code z is injected by spatial replication and
concatenation into the generator network. We tried two alternatives, injecting into (left) the input layer and
(right) every intermediate layer in the encoder.
In the unconditional GAN setting, it has been observed that using samples both from random z and
encoded vector E(B) help further improves the results [20]. Hence, we also report one variant which
is the full objective shown above (Equation 9), but without the reconstruction loss on the latent space
Llatent
. We call it cVAE-GAN++, as it is based on cVAE-GAN with additional loss LGAN (G, D), that
1
encourages the discriminator to also see the randomly generated samples.
4
Implementation Details
The code and additional results are publicly available on our website. Please refer to supplement for
more details about the datasets, architectures, and training procedures.
Network architecture For generator G, we use the U-Net [31], which contains an encoder-decoder
architecture, with symmetric skip connections. The architecture has been shown to produce strong
results in the unimodal image prediction setting, when there is spatial correspondence between input
and output pairs. For discriminator D, we use two PatchGAN discriminators [15] at different scales,
which aims to predict real vs. fake for 70 ? 70 and 140 ? 140 overlapping image patches respectively.
For the encoder E, we experiment with two networks: (1) E_CNN: CNN with a few convolutional and
downsampling layers and (2) E_ResNet: a classifier with several residual blocks [12].
Training details We build our model on the Least Squares GANs (LSGANs) variant [22], which
uses a least-squares objective instead of a cross entropy loss. LSGANs produce high quality results
with stable training. We also find that not conditioning the discriminator D leads to better results
(also discussed in [28]), and hence choose to do the same for all methods. We set the parameters
?image = 10, ?latent = 0.5 and ?KL = 0.01 in all our experiments. In practice, we tie the weights for
the generators in the cVAE-GAN and cLR-GAN. We observe that using two separate discriminators
yields slightly better visual results compared to discriminators with shared weights. We train our
networks from scratch using Adam [17] with a batch size of 1 and with learning rate of 0.0002. We
choose latent dimension to be 8 across all the datasets.
Injecting the latent code z to generator G. How to propagate the information encoded by latent
code z to the image generation process is critical to our applications. We explore two choices as
shown in Figure 3: (1) add_to_input: We simply extend a Z-dimensional latent code z to an
H ? W ? Z spatial tensor and concatenate it with the H ? W ? 3 input image. (2) add_to_all:
Alternatively, we add z to each intermediate layer of the network G.
5
Experiments
Datasets We test our method on several image-to-image translation problems from prior work,
including edges ? photos [42, 47], Google maps ? satellite [15], labels ? images [3], and outdoor
night ? day images [19]. These problems are all one-to-many mappings. We train all the models on
256 ? 256 images.
Methods We train the following models described in Section 3: pix2pix+noise, cAE-GAN,
cVAE-GAN, cVAE-GAN++, cLR-GAN and the hybrid model BiCycleGAN.
5.1
Qualitative Evaluation
We show qualitative comparison results on Figure 5. We observe that pix2pix+noise typically
produces a single realistic output, but does not produce any meaningful variation. cAE-GAN adds
6
Input
Ground truth
Generated samples
Figure 4: Example Results We show example results of our hybrid model BiCycleGAN. The left column show
cAE-GAN
BiCycleGAN
Ground truth
cLR-GAN
Input
cVAE-GAN++
Pix2pix
+noise
cVAE-GAN
shows the input. The second shows the ground truth output. The final four columns show randomly generated
samples. We show results of our method on night?day, edges?shoes, edges?handbags, and maps?satellites.
Figure 5: Qualitative method comparison We show an example on the label ? facades dataset
across different methods. The BiCycleGAN method produces results which are both realistic and
diverse.
7
Diversity
VGG-16
Distance
3.520?.021
0.338?.002
2.304?.012
1.350?.013
1.425?.014
a
1.374?.022
1.469?.014
a
We found that cLR-GAN resulted in severe mode collapse, resulting in 15% of the images producing the same
result. Those images were omitted from this calculation.
40
Realism (AMT Fooling Rate [%])
Method
Random real images
pix2pix+noise [15]
cAE-GAN
cVAE-GAN
cVAE-GAN++
cLR-GAN
BiCycleGAN
Realism
AMT Fooling
Rate [%]
50.0%
27.93?2.40 %
13.64?1.80 %
24.93?2.27 %
29.19?2.43 %
29.23?2.48 %
34.33?2.69 %
35
30
25
20
15
10
5
0
0.0
pix2pix+noise
cAE-GAN
cVAE-GAN
cVAE-GAN++
cLR-GAN
BiCycleGAN
0.5
1.0
1.5
2.0
Diversity (Average VGG Distance)
2.5
Figure 6: Realism vs Diversity. We measure diversity using average feature distance in the VGG-16 space
using cosine distance summed across five layers, and realism using a real vs. fake Amazon Mechanical Turk
test on the Google maps ? satellites task. The pix2pix+noise baseline produces realistic results, but little
diversity. Using only cAE-GAN method produces diverse results, but large artifacts during sampling. The hybrid
BiCycleGAN method, which combines cVAE-GAN++ and cLR-GAN, produces results which have higher realism
while maintaining diversity.
variation to the output, but typically reduces quality of results, as shown for an example on facades
on Figure 4. We typically observe more variation in the cVAE-GAN, as the latent space is encouraged
to encode information about ground truth outputs. However, the space is not densely populated,
so drawing random samples may cause artifacts in the output. The cLR-GAN typically shows less
variation in the output, and sometimes suffers from mode collapse. When combining these methods,
however, in the hybrid method BiCycleGAN, we observe results which are both diverse and realistic.
We show example results in Figure 4. Please see our website for a full set of results.
5.2
Quantitative Evaluation
We perform a quantitative analysis on the diversity, realism, and latent space distribution on our six
variants and baselines. We quantitatively test the Google maps ? satellites dataset.
Diversity We randomly draw samples from our model and compute average distance in a deep
feature space. In the context of style transfer, image super-resolution [16], and feature inversion [7],
pre-trained networks have been used as a ?perceptual loss" and explicitly optimized over. In the
context of generative modeling, they have been used as a held-out ?validation" score, for example to
assess how semantic samples from a generative model [33] or the semantic accuracy of a grayscale
colorization [44].
In Figure 5, we show the diversity-score using the cosine distance, averaged across spatial dimensions,
and summed across the five conv layers preceding the pool layers on the VGG-16 network [35],
pre-trained for Imagenet classification [32]. The maximum score is 5.0, as all the feature responses
are non-negative. For each method, we compute the average distance between 1900 pairs of randomly
b images (sampled from 100 input A images). Random pairs of ground truth real
generated output B
images in the B ? B domain produce an average variation of 3.520 using cosine distance. As we are
b which correspond to a specific input, a system which stays faithful to the input
measuring samples B
should definitely not exceed this score.
The pix2pix system [15] produces a single point estimate. Adding noise to the system
pix2pix+noise produces a diversity score of 0.338, confirming the finding in [15] that adding
noise does not produce large variation. Using an cAE-GAN model to encode ground truth image
B into latent code z does increase the variation. The cVAE-GAN, cVAE-GAN++, and BiCycleGAN
models all place explicit constraints on the latent space, and the cLR-GAN model places an implicit
constraint through sampling. These four methods all produce similar diversity scores. We note
that high diversity scores may also indicate that non-realistic images are being generated, causing
meaningless variation. Next, we investigate the visual realism of our samples.
Perceptual Realism To judge the visual realism of our results, we use human judgments, as proposed
in [44] and later used in [15, 48]. The test sequentially presents a real and generated image to a human
for 1 second each, in a random order, asks them to identify the fake, and measures the ?fooling"
8
Encoder
E_ResNet
E_ResNet
E_CNN
E_CNN
Injecting z
add_to_all add_to_input add_to_all add_to_input
label?photo
0.292 ? 0.058 0.292 ? 0.054 0.326 ? 0.066 0.339 ? 0.069
map ? satellite 0.268 ? 0.070 0.266 ? 0.068 0.287 ? 0.067 0.272 ? 0.069
Table 1: The encoding performance with respect to the different encoder architectures and methods
of injecting z. Here we report the reconstruction loss ||B ? G(A, E(B))||1
.
Input label
|z| = 2
|z| = 8
|z| = 256
Figure 7: Different label ? facades results trained with varying length of the latent code |z| ? {2, 8, 256}.
rate. Figure 5(left) shows the realism across methods. The pix2pix+noise model achieves high
realism score, but without large diversity, as discussed in the previous section. The cAE-GAN helps
produce diversity, but this comes at a large cost to the visual realism. Because the distribution of
the learned latent space is unclear, random samples may be from unpopulated regions of the space.
Adding the KL-divergence loss in the latent space, used in the model recovers the visual realism.
Furthermore, as expected, checking randomly drawn z vectors in the cVAE-GAN++ model slightly
increases realism. The cLR-GAN, which draws z vectors from the predefined distribution randomly,
produces similar realism and diversity scores. However, the cLR-GAN model resulted in large mode
collapse - approximately 15% of the outputs produced the same result, independent of the input
image. The full hybrid BiCycleGAN gets the best of both worlds, as it does not suffer from mode
collapse and also has the highest realism score by a significant margin.
Encoder architecture The pix2pix framework [15] have conducted extensive ablation studies
on discriminators and generators. Here we focus on the performance of two encoders E_CNN and
E_ResNet for our applications on maps and facades datasets, and we find that E_ResNet can better
encode the output image, regarding the image reconstruction loss ||B ? G(A, E(B))||1 on validation
datasets as shown in Table 1.
Methods of injecting latent code We evaluate two ways of injecting latent code z: add_to_input
and add_to_all (Section 4), regarding the same reconstruction loss ||B ? G(A, E(B))||1 . Table 1
shows that two methods give similar performance. This indicates that the U_Net [31] can already
propagate the information well to the output without the additional skip connections from z.
Latent code length We study the BiCycleGAN model results with respect to the varying number
of dimensions of latent codes {2, 8, 256} in Figure 7. A low-dimensional latent code may limit the
amount of diversity that can be expressed by the model. On the contrary, a high-dimensional latent
code can potentially encode more information about an output image at the cost of making sampling
quite difficult. The optimal length of z largely depends on individual datasets and applications, and
how much ambiguity there is in the output.
6
Conclusions
In conclusion, we have presented a novel method for obtaining and exploring diverse results in a
conditional image generation application. Our approach produces optimal results in terms of diversity
and realism, while being simple to train and requiring no additional supervision. We see many
interesting avenues of future work, including directly enforcing a distribution in the latent space that
encodes semantically meaningful attributes to allow for image-to-image transformations with user
controllable parameters.
Acknowledgments We thank Phillip Isola and Tinghui Zhou for helpful discussions. This work was
supported in part by Adobe Inc., DARPA, AFRL, DoD MURI award N000141110688, NSF awards
IIS-1633310, IIS-1427425, IIS-1212798, the Berkeley Artificial Intelligence Research (BAIR) Lab,
and hardware donations from NVIDIA. JYZ is supported by Facebook Graduate Fellowship, RZ by
Adobe Research Fellowship, and DP by NVIDIA Graduate Fellowship.
9
References
[1] M. Arjovsky and L. Bottou. Towards principled methods for training generative adversarial
networks. In ICLR, 2017.
[2] X. Chen, Y. Duan, R. Houthooft, J. Schulman, I. Sutskever, and P. Abbeel. Infogan: interpretable
representation learning by information maximizing generative adversarial nets. In NIPS, 2016.
[3] M. Cordts, M. Omran, S. Ramos, T. Rehfeld, M. Enzweiler, R. Benenson, U. Franke, S. Roth,
and B. Schiele. The cityscapes dataset for semantic urban scene understanding. In CVPR, 2016.
[4] E. L. Denton, S. Chintala, R. Fergus, et al. Deep generative image models using a laplacian
pyramid of adversarial networks. In NIPS, 2015.
[5] L. Dinh, J. Sohl-Dickstein, and S. Bengio. Density estimation using real nvp. In ICLR, 2017.
[6] J. Donahue, P. Kr?henb?hl, and T. Darrell. Adversarial feature learning. In ICLR, 2016.
[7] A. Dosovitskiy and T. Brox. Generating images with perceptual similarity metrics based on
deep networks. In NIPS, 2016.
[8] V. Dumoulin, I. Belghazi, B. Poole, A. Lamb, M. Arjovsky, O. Mastropietro, and A. Courville.
Adversarially learned inference. In ICLR, 2016.
[9] A. Ghosh, V. Kulharia, V. Namboodiri, P. H. Torr, and P. K. Dokania. Multi-agent diverse
generative adversarial networks. arXiv preprint arXiv:1704.02906, 2017.
[10] I. Goodfellow. NIPS 2016 tutorial: Generative adversarial networks.
arXiv:1701.00160, 2016.
arXiv preprint
[11] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and
Y. Bengio. Generative adversarial nets. In NIPS, 2014.
[12] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. In CVPR,
2016.
[13] G. E. Hinton and R. R. Salakhutdinov. Reducing the dimensionality of data with neural networks.
Science, 313(5786):504?507, 2006.
[14] S. Iizuka, E. Simo-Serra, and H. Ishikawa. Let there be color!: Joint end-to-end learning of
global and local image priors for automatic image colorization with simultaneous classification.
SIGGRAPH, 35(4), 2016.
[15] P. Isola, J.-Y. Zhu, T. Zhou, and A. A. Efros. Image-to-image translation with conditional
adversarial networks. In CVPR, 2017.
[16] J. Johnson, A. Alahi, and L. Fei-Fei. Perceptual losses for real-time style transfer and superresolution. In ECCV, 2016.
[17] D. Kingma and J. Ba. Adam: A method for stochastic optimization. In ICLR, 2015.
[18] D. P. Kingma and M. Welling. Auto-encoding variational bayes. In ICLR, 2014.
[19] P.-Y. Laffont, Z. Ren, X. Tao, C. Qian, and J. Hays. Transient attributes for high-level understanding and editing of outdoor scenes. SIGGRAPH, 2014.
[20] A. B. L. Larsen, S. K. S?nderby, H. Larochelle, and O. Winther. Autoencoding beyond pixels
using a learned similarity metric. In ICML, 2016.
[21] G. Larsson, M. Maire, and G. Shakhnarovich. Learning representations for automatic colorization. In ECCV, 2016.
[22] X. Mao, Q. Li, H. Xie, R. Y. Lau, Z. Wang, and S. P. Smolley. Least squares generative
adversarial networks. In ICCV, 2017.
10
[23] M. Mathieu, C. Couprie, and Y. LeCun. Deep multi-scale video prediction beyond mean square
error. In ICLR, 2016.
[24] M. Mirza and S. Osindero.
arXiv:1411.1784, 2014.
Conditional generative adversarial nets.
arXiv preprint
[25] A. Nguyen, J. Yosinski, Y. Bengio, A. Dosovitskiy, and J. Clune. Plug & play generative
networks: Conditional iterative generation of images in latent space. In CVPR, 2017.
[26] A. v. d. Oord, N. Kalchbrenner, and K. Kavukcuoglu. Pixel recurrent neural networks. PMLR,
2016.
[27] A. v. d. Oord, N. Kalchbrenner, O. Vinyals, L. Espeholt, A. Graves, and K. Kavukcuoglu.
Conditional image generation with pixelcnn decoders. arXiv preprint arXiv:1606.05328, 2016.
[28] D. Pathak, P. Kr?henb?hl, J. Donahue, T. Darrell, and A. Efros. Context encoders: Feature
learning by inpainting. In CVPR, 2016.
[29] A. Radford, L. Metz, and S. Chintala. Unsupervised representation learning with deep convolutional generative adversarial networks. In ICLR, 2016.
[30] S. Reed, Z. Akata, X. Yan, L. Logeswaran, B. Schiele, and H. Lee. Generative adversarial
text-to-image synthesis. In ICML, 2016.
[31] O. Ronneberger, P. Fischer, and T. Brox. U-net: Convolutional networks for biomedical image
segmentation. In MICCAI, pages 234?241. Springer, 2015.
[32] O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy,
A. Khosla, M. Bernstein, et al. Imagenet large scale visual recognition challenge. IJCV, 2015.
[33] T. Salimans, I. Goodfellow, W. Zaremba, V. Cheung, A. Radford, and X. Chen. Improved
techniques for training gans. arXiv preprint arXiv:1606.03498, 2016.
[34] P. Sangkloy, J. Lu, C. Fang, F. Yu, and J. Hays. Scribbler: Controlling deep image synthesis
with sketch and color. In CVPR, 2017.
[35] K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image
recognition. arXiv preprint arXiv:1409.1556, 2014.
[36] P. Smolensky. Information processing in dynamical systems: Foundations of harmony theory.
Technical report, DTIC Document, 1986.
[37] K. Sohn, X. Yan, and H. Lee. Learning structured output representation using deep conditional
generative models. In NIPS, 2015.
[38] P. Vincent, H. Larochelle, Y. Bengio, and P.-A. Manzagol. Extracting and composing robust
features with denoising autoencoders. In ICML, 2008.
[39] J. Walker, C. Doersch, A. Gupta, and M. Hebert. An uncertain future: Forecasting from static
images using variational autoencoders. In ECCV, 2016.
[40] T. Xue, J. Wu, K. Bouman, and B. Freeman. Visual dynamics: Probabilistic future frame
synthesis via cross convolutional networks. In NIPS, 2016.
[41] C. Yang, X. Lu, Z. Lin, E. Shechtman, O. Wang, and H. Li. High-resolution image inpainting
using multi-scale neural patch synthesis. In CVPR, 2017.
[42] A. Yu and K. Grauman. Fine-grained visual comparisons with local learning. In CVPR, 2014.
[43] H. Zhang, T. Xu, H. Li, S. Zhang, X. Huang, X. Wang, and D. Metaxas. Stackgan: Text to
photo-realistic image synthesis with stacked generative adversarial networks. In ICCV, 2017.
[44] R. Zhang, P. Isola, and A. A. Efros. Colorful image colorization. In ECCV, 2016.
[45] R. Zhang, J.-Y. Zhu, P. Isola, X. Geng, A. S. Lin, T. Yu, and A. A. Efros. Real-time user-guided
image colorization with learned deep priors. SIGGRAPH, 2017.
11
[46] J. Zhao, M. Mathieu, and Y. LeCun. Energy-based generative adversarial network. In ICLR,
2017.
[47] J.-Y. Zhu, P. Kr?henb?hl, E. Shechtman, and A. A. Efros. Generative visual manipulation on
the natural image manifold. In ECCV, 2016.
[48] J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros. Unpaired image-to-image translation using
cycle-consistent adversarial networks. In ICCV, 2017.
12
| 6650 |@word cnn:1 middle:1 version:3 inversion:1 unpopulated:1 tried:1 propagate:2 asks:1 inpainting:2 shechtman:3 contains:2 score:10 document:1 existing:2 cvae:23 written:1 subsequent:1 concatenate:1 realistic:14 confirming:1 shape:2 interpretable:1 v:3 generative:22 instantiate:1 website:2 intelligence:1 parameterization:1 parametrization:1 realism:20 regressive:2 bijection:1 zhang:6 five:2 along:1 direct:2 replication:1 qualitative:3 ijcv:1 combine:3 multimodality:3 expected:1 multi:4 inspired:1 salakhutdinov:1 freeman:1 resolve:1 encouraging:1 actual:1 little:1 duan:1 conv:1 superresolution:1 finding:1 transformation:1 ghosh:1 kldivergence:1 guarantee:1 berkeley:6 sky:2 every:1 quantitative:2 alahi:1 interactive:1 tie:1 zaremba:1 grauman:1 classifier:2 control:1 colorful:1 jyz:1 producing:4 understood:1 local:2 limit:1 encoding:12 approximately:1 eb:1 challenging:2 collapse:7 limited:1 logeswaran:1 range:1 graduate:2 averaged:1 faithful:2 unique:1 acknowledgment:1 testing:1 lecun:2 practice:1 block:1 procedure:1 maire:1 yan:3 significantly:1 pre:2 induce:1 ronneberger:1 seeing:1 get:2 close:4 context:5 applying:1 franke:1 map:14 deterministic:3 missing:1 dz:1 maximizing:1 roth:1 starting:1 focused:1 patchgan:1 resolution:2 amazon:1 qian:1 pouget:1 fang:1 embedding:1 variation:8 target:3 play:1 controlling:1 user:2 us:4 goodfellow:3 trick:1 synthesize:1 recognition:3 nderby:1 muri:1 observed:2 cloud:2 preprint:6 wang:4 capture:2 region:2 cycle:2 sun:1 removed:1 highest:1 substantial:1 principled:1 schiele:2 warde:1 dynamic:1 trained:6 tight:1 shakhnarovich:1 multimodal:7 cae:11 joint:3 easily:3 darpa:1 siggraph:3 train:6 stacked:1 effective:1 describe:1 artificial:1 hyper:1 kalchbrenner:2 quite:1 encoded:7 plausible:1 cvpr:8 drawing:1 reconstruct:1 compressed:1 encoder:14 simonyan:1 fischer:1 jointly:3 itself:1 final:3 differentiate:1 advantage:1 autoencoding:1 net:5 propose:2 reconstruction:5 facade:4 relevant:2 combining:3 causing:1 ablation:1 laffont:1 achieve:1 sutskever:1 darrell:3 extending:1 satellite:5 produce:26 generating:6 adam:2 staying:1 help:5 illustrate:1 donation:1 recurrent:1 progress:1 strong:1 predicted:1 skip:2 judge:2 indicate:1 come:1 larochelle:2 direction:4 guided:1 attribute:2 stochastic:3 human:3 enable:1 transient:1 espeholt:1 abbeel:1 preliminary:1 secondly:1 exploring:2 extension:1 lkl:3 ground:16 normal:2 visually:1 mapping:10 predict:1 efros:7 achieves:1 omitted:1 estimation:1 injecting:9 harmony:1 label:5 create:1 successfully:1 gaussian:4 aim:3 modified:1 rather:1 super:1 zhou:2 varying:2 vae:4 encode:7 clune:1 focus:2 likelihood:1 indicates:1 adversarial:20 baseline:6 helpful:1 inference:3 typically:4 tao:1 pixel:4 issue:1 arg:4 overall:1 classification:2 development:1 spatial:4 summed:2 uc:5 brox:2 distilled:1 beach:1 sampling:6 encouraged:2 adversarially:1 ishikawa:1 yu:3 unsupervised:1 denton:1 icml:3 park:1 geng:1 future:3 report:3 mirza:2 quantitatively:1 richard:1 primarily:1 dosovitskiy:2 few:1 randomly:15 divergence:2 resulted:2 densely:1 individual:1 attempt:3 investigate:1 alexei:1 evaluation:3 severe:1 weakness:1 mixture:1 farley:1 unconditional:6 held:1 predefined:1 oliver:1 edge:4 encourage:4 injective:1 simo:1 unless:1 desired:2 re:2 uncertain:1 bouman:1 instance:4 column:2 modeling:8 disadvantage:1 measuring:1 cost:2 comprised:1 dod:1 successful:1 conducted:1 osindero:1 johnson:1 motivating:1 encoders:3 xue:1 combined:1 st:1 density:2 definitely:1 winther:1 oord:2 stay:1 systematic:2 lee:2 probabilistic:1 invertible:1 regressor:5 pool:1 nvp:1 synthesis:5 gans:8 ambiguity:3 reflect:1 choose:2 huang:2 classically:1 zhao:1 style:2 li:3 potential:2 diversity:20 stabilize:1 inc:1 explicitly:7 depends:1 later:1 helped:1 try:1 lab:1 dumoulin:1 start:3 bayes:1 option:1 metz:1 ass:2 square:4 publicly:1 accuracy:1 convolutional:5 largely:2 correspond:4 yield:2 judgment:1 identify:1 yellow:1 metaxas:1 vincent:1 kavukcuoglu:2 produced:2 marginally:1 ren:2 lu:2 lighting:2 russakovsky:1 simultaneous:1 suffers:1 trevor:1 facebook:1 energy:1 turk:1 larsen:1 chintala:2 recovers:1 static:1 sampled:8 dataset:6 popular:1 color:6 subsection:1 improves:1 dimensionality:1 segmentation:1 akata:1 ea:7 back:7 afrl:1 attained:1 higher:1 day:8 xie:1 modal:1 improved:2 response:1 editing:1 formulation:2 zisserman:1 furthermore:1 implicit:1 biomedical:1 miccai:1 autoencoders:4 sketch:3 hand:1 night:8 su:1 overlapping:1 propagation:2 google:3 mode:15 quality:6 artifact:2 usa:1 effect:1 usage:2 name:1 true:1 requiring:1 phillip:1 hence:4 alternating:1 symmetric:1 semantic:3 during:7 self:2 encourages:2 game:1 ambiguous:3 noted:1 please:2 cosine:3 bijective:1 l1:2 interface:1 image:106 variational:7 wise:2 novel:1 recently:3 common:1 discourages:1 overview:1 conditioning:2 enzweiler:1 discussed:3 interpretation:1 extend:1 he:1 yosinski:1 refer:1 significant:1 dinh:1 automatic:2 doersch:1 consistency:2 trivially:1 populated:1 stochasticity:3 pixelcnn:1 stable:2 similarity:3 supervision:1 cityscape:1 etc:1 add:3 larsson:1 perspective:1 apart:1 scenario:3 manipulation:1 nvidia:2 hay:2 success:2 clr:17 captured:1 seen:1 additional:5 tinghui:1 preceding:1 isola:5 arjovsky:2 deng:1 ii:3 multiple:4 full:4 unimodal:1 reduces:1 technical:1 match:1 calculation:1 cross:2 long:1 plug:1 lin:2 award:2 dkl:2 paired:4 laplacian:1 adobe:4 prediction:3 variant:6 regression:1 metric:4 arxiv:12 sometimes:1 pyramid:1 achieved:1 addition:2 whereas:1 want:1 fellowship:3 krause:1 fine:1 walker:1 benenson:1 meaningless:1 contrary:1 call:2 extracting:1 yang:1 intermediate:2 exceed:1 enough:1 bengio:4 automated:1 variety:2 mastropietro:1 bernstein:1 architecture:7 idea:1 regarding:3 avenue:1 vgg:4 six:1 bair:1 forecasting:1 suffer:1 henb:3 dokania:1 passing:1 cause:1 deep:12 ignored:1 useful:1 fake:4 karpathy:1 amount:1 hardware:1 sohn:1 unpaired:1 generate:4 cordts:1 nsf:1 tutorial:1 revisit:1 blue:2 diverse:15 promise:1 incentive:1 dropping:1 express:1 dickstein:1 four:2 stackgan:1 achieving:1 drawn:6 urban:1 prevent:1 utilize:1 fooling:3 houthooft:1 eli:1 inverse:1 injected:1 place:2 family:1 lamb:1 wu:1 patch:2 draw:2 dropout:1 layer:7 tackled:1 correspondence:1 courville:2 strength:1 constraint:4 fei:2 scene:3 encodes:2 aspect:2 min:4 formulating:1 structured:1 popularized:1 across:7 slightly:2 appealing:1 encapsulates:1 making:2 hl:3 lau:1 restricted:1 iccv:3 equation:4 previously:1 discus:2 end:2 photo:5 available:1 observe:4 salimans:1 enforce:5 pmlr:1 inpaint:1 appending:1 alternative:2 batch:1 original:1 rz:2 remaining:1 include:2 gan:60 maintaining:1 concatenated:1 build:2 tensor:1 objective:10 added:2 already:1 parametric:1 unclear:1 amongst:1 dp:1 iclr:9 distance:9 separate:1 mapped:1 thank:1 concatenation:1 decoder:2 manifold:2 toward:1 enforcing:4 reason:1 ozair:1 lgan:6 code:40 length:3 colorization:6 reed:1 manzagol:1 balance:1 downsampling:1 setup:2 steep:1 difficult:2 potentially:2 expense:1 negative:1 ba:1 implementation:1 boltzmann:1 satheesh:1 perform:3 allowing:1 observation:1 datasets:6 hinton:1 frame:1 pair:5 mechanical:1 kl:7 extensive:1 connection:6 sentence:1 discriminator:13 optimized:3 imagenet:2 learned:8 boost:1 kingma:2 nip:8 able:1 beyond:2 poole:1 usually:1 below:1 dynamical:1 smolensky:1 challenge:1 including:3 max:4 video:1 analogue:1 critical:1 pathak:3 natural:4 hybrid:8 regularized:1 force:2 predicting:1 ramos:1 residual:2 rehfeld:1 zhu:5 mathieu:3 jun:1 autoencoder:3 auto:4 omran:1 text:3 prior:4 literature:1 schulman:1 checking:1 understanding:2 graf:1 loss:25 bear:1 generation:11 interesting:1 generator:27 validation:2 foundation:1 degree:1 agent:1 consistent:3 systematically:1 translation:12 eccv:5 supported:2 hebert:1 allow:1 wide:1 deepak:1 serra:1 benefit:1 overcome:1 dimension:3 world:1 author:1 made:1 nguyen:1 employing:1 far:2 welling:1 scribble:1 ignore:1 belghazi:1 global:1 sequentially:1 n000141110688:1 fergus:1 alternatively:1 grayscale:2 iizuka:1 latent:79 iterative:1 khosla:1 table:3 nature:1 learn:8 robust:1 composing:1 ca:1 ignoring:1 transfer:2 obtaining:1 controllable:1 investigated:1 bottou:1 domain:7 did:1 main:1 rh:2 noise:21 xu:2 representative:1 mao:1 explicit:1 outdoor:2 perceptual:5 infogan:3 learns:3 donahue:2 grained:1 specific:1 abadie:1 gupta:1 naively:1 adding:4 sequential:1 importance:1 sohl:1 supplement:1 texture:1 kr:3 perceptually:2 conditioned:7 margin:1 dtic:1 chen:2 entropy:1 photograph:1 simply:2 explore:4 shoe:2 visual:9 vinyals:1 desire:1 contained:1 lvae:4 expressed:1 radford:2 springer:1 corresponds:2 truth:15 chance:1 amt:2 ma:1 conditional:32 goal:2 cheung:1 towards:1 couprie:1 shared:1 change:1 torr:1 reducing:1 semantically:1 denoising:1 meaningful:2 vaes:2 handbag:1 phenomenon:1 evaluate:1 scratch:1 |
6,247 | 6,651 | Mixture-Rank Matrix Approximation
for Collaborative Filtering
Dongsheng Li1
Chao Chen1
Wei Liu2? Tun Lu3,4
Ning Gu3,4
Stephen M. Chu1
1
IBM Research - China
2
Tencent AI Lab, China
3
School of Computer Science, Fudan University, China
4
Shanghai Key Laboratory of Data Science, Fudan University, China
{ldsli, cshchen, schu}@cn.ibm.com, [email protected], {lutun, ninggu}@fudan.edu.cn
Abstract
Low-rank matrix approximation (LRMA) methods have achieved excellent accuracy among today?s collaborative filtering (CF) methods. In existing LRMA
methods, the rank of user/item feature matrices is typically fixed, i.e., the same rank
is adopted to describe all users/items. However, our studies show that submatrices
with different ranks could coexist in the same user-item rating matrix, so that
approximations with fixed ranks cannot perfectly describe the internal structures
of the rating matrix, therefore leading to inferior recommendation accuracy. In
this paper, a mixture-rank matrix approximation (MRMA) method is proposed, in
which user-item ratings can be characterized by a mixture of LRMA models with
different ranks. Meanwhile, a learning algorithm capitalizing on iterated condition
modes is proposed to tackle the non-convex optimization problem pertaining to
MRMA. Experimental studies on MovieLens and Netflix datasets demonstrate that
MRMA can outperform six state-of-the-art LRMA-based CF methods in terms of
recommendation accuracy.
1
Introduction
Low-rank matrix approximation (LRMA) is one of the most popular methods in today?s collaborative
filtering (CF) methods due to high accuracy [11, 12, 13, 17]. Given a targeted user-item rating matrix
R ? Rm?n , the general goal of LRMA is to find two rank-k matrices U ? Rm?k and V ? Rn?k
? = U V T . After obtaining the user and item feature matrices, the recommendation
such that R ? R
score of the i-th user on the j-th item can be obtained by the dot product between their corresponding
feature vectors, i.e., Ui Vj T .
In existing LRMA methods [12, 13, 17], the rank k is considered fixed, i.e., the same rank is adopted
to describe all users and items. However, in many real-world user-item rating matrices, e.g., Movielens
and Netflix, users/items have a significantly varying number of ratings, so that submatrices with
different ranks could coexist. For instance, a submatrix containing users and items with few ratings
should be of a low rank, e.g., 10 or 20, and a submatrix containing users and items with many ratings
may be of a relatively higher rank, e.g., 50 or 100. Adopting a fixed rank for all users and items
cannot perfectly model the internal structures of the rating matrix, which will lead to imperfect
approximations as well as degraded recommendation accuracy.
In this paper, we propose a mixture-rank matrix approximation (MRMA) method, in which user-item
ratings are represented by a mixture of LRMA models with different ranks. For each user/item, a
probability distribution with a Laplacian prior is exploited to describe its relationship with different
?
This work was conducted while the author was with IBM.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
LRMA models, while a joint distribution of user-item pairs is employed to describe the relationship
between the user-item ratings and different LRMA models. To cope with the non-convex optimization
problem associated with MRMA, a learning algorithm capitalizing on iterated condition modes
(ICM) [1] is proposed, which can obtain a local maximum of the joint probability by iteratively
maximizing the probability of each variable conditioned on the rest. Finally, we evaluate the proposed
MRMA method on Movielens and Netflix datasets. The experimental results show that MRMA can
achieve better accuracy compared against state-of-the-art LRMA-based CF methods, further boosting
the performance for recommender systems leveraging matrix approximation.
2
Related Work
Low-rank matrix approximation methods have been leveraged by much recent work to achieve
accurate collaborative filtering, e.g., PMF [17], BPMF [16], APG [19], GSMF [20], SMA [13],
etc. These methods train one user feature matrix and one item feature matrix first and use these
feature matrices for all users and items without any adaptation. However, all these methods adopt
fixed rank values for the targeted user-item rating matrices. Therefore, as analyzed in this paper,
submatrices with different ranks could coexist in the rating matrices and only adopting a fixed rank
cannot achieve optimal matrix approximation. Besides stand-alone matrix approximation methods,
ensemble methods, e.g., DFC [15], LLORMA [12], WEMAREC [5], etc., and mixture models, e.g.,
MPMA [4], etc., have been proposed to improve the recommendation accuracy and/or scalability
by weighing different base models across different users/items. However, the above methods do not
consider using different ranks to derive different base models. In addition, it is desirable to borrow
the idea of mixture-rank matrix approximation (MRMA) to generate more accurate base models in
the above methods and further enhance their accuracy.
In many matrix approximation-based collaborative filtering methods, auxiliary information, e.g.,
implicit feedback [9], social information [14], contextual information [10], etc., is introduced to
improve the recommendation quality of pure matrix approximation methods. The idea of MRMA is
orthogonal to these methods, and can thus be employed by these methods to further improve their
recommendation accuracy. In general low-rank matrix approximation methods, it is non-trivial to
directly determine the maximum rank of a targeted matrix [2, 3]. Cand?s et al. [3] proved that a
non-convex rank minimization problem can be equivalently transformed into a convex nuclear norm
minimization problem. Based on this finding, we can easily determine the range of ranks for MRMA
and choose different K values (the maximum rank in MRMA) for different datasets.
3
Problem Formulation
In this paper, upper case letters such as R, U, V denote matrices, and k denotes the rank for matrix
approximation. For a targeted user-item rating matrix R ? Rm?n , m denotes the number of users, n
? denotes
denotes the number of items, and Ri,j denotes the rating of the i-th user on the j-th item. R
the low-rank approximation of R. The general goal of k-rank matrix approximation is to determine
? = U V T . The rank k is
user and item feature matrices, i.e., U ? Rm?k , V ? Rn?k , such that R ? R
considered low, because k min{m, n} can achieve good performance in many CF applications.
In real-world rating matrices, e.g., Movielens and Netflix, users/items have a varying number of
ratings, so that a lower rank which best describes users/items with less ratings will easily underfit the
users/items with more ratings, and similarly a higher rank will easily overfit the users/items with less
ratings. A case study is conducted on the Movielens (1M) dataset (with 1M ratings from 6,000 users
on 4,000 movies), which confirms that internal submatrices with different ranks indeed coexist in the
rating matrix. Here, we run the probabilistic matrix factorization (PMF) method [17] using k = 5
and k = 50, and then compare the root mean square errors (RMSEs) for the users/items with less
than 10 ratings and more than 50 ratings.
As shown in Table 1, when the rank is 5, the users/items with less than 10 ratings achieve lower
RMSEs than the cases when the rank is 50. This indicates that the PMF model overfits the users/items
with less than 10 ratings when k = 50. Similarly, we can conclude that the PMF model underfits
the users/items with more than 50 ratings when k = 5. Moreover, PMF with k = 50 achieves lower
RMSE (higher accuracy) than PMF with k = 5, but the improvement comes with sacrificed accuracy
for the users and items with a small number of ratings, e.g., less than 10. This study shows that PMF
2
Table 1: The root mean square errors (RMSEs) of PMF [17] for users/items with different numbers
of ratings when rank k = 5 and k = 50.
rank = 5
rank = 50
#user ratings < 10
#user ratings > 50
#item ratings < 10
#item ratings > 50
0.9058
0.8416
0.9338
0.8520
0.9165
0.8352
0.9598
0.8418
All
0.8614
0.8583
with fixed rank values cannot perfectly model the internal mixture-rank structure of the rating matrix.
To this end, it is desirable to model users and items with different ranks.
4
Mixture-Rank Matrix Approximation (MRMA)
2
?U
??
b?
b?
??
Uik
?ik
?jk
Vjk
?V2
k = {1, ..., K}
Ri,j
?2
j = {1, ..., n}
i = {1, ..., m}
Figure 1: The graphical model for the proposed mixture-rank matrix approximation (MRMA) method.
Following the idea of PMF, we exploit a probabilistic model with Gaussian noise to model the
ratings [17]. As shown in Figure 1, the conditional distribution over the observed ratings for the
mixture-rank model can be defined as follows:
m Y
n X
K
Y
T
Pr(R|U, V, ?, ?, ? 2 ) =
[
?ik ?jk N (Ri,j |Uik Vjk , ? 2 )]1i,j ,
(1)
i=1 j=1 k=1
where N (x|?, ? 2 ) denotes the probability density function of a Gaussian distribution with mean ?
and variance ? 2 . K is the maximum rank among all internal structures of the user-item rating matrix.
?k and ? k are the weight vectors of the rank-k matrix approximation model for all users and items,
respectively. Thus, ?ik and ?jk denote the weights of the rank-k model for the i-th user and j-th item,
respectively. U k and V k are the feature matrices of the rank-k matrix approximation model for all
users and items, respectively. Likewise, Uik and Vjk denote the feature vectors of the rank-k model
for the i-th user and j-th item, respectively. 1i,j is an indication function, which will be 1 if Ri,j is
observed and 0 otherwise.
By placing a zero mean isotropic Gaussian prior [6, 17] on the user and item feature vectors, we have
m
n
Y
Y
2
2
Pr(U k |?U
)=
N (Uik |0, ?U
I), Pr(V k |?V2 ) =
N (Vjk |0, ?V2 I).
(2)
i=1
j=1
For ?k and ? k , we choose a Laplacian prior here, because the models with most suitable ranks
for each user/item should be with large weights, i.e., ?k and ? k should be sparse. By placing the
Laplacian prior on the user and item weight vectors, we have
m
n
Y
Y
Pr(?k |?? , b? ) =
L(?ik |?? , b? ), Pr(? k |?? , b? ) =
L(?jk |?? , b? ),
(3)
i=1
j=1
3
where ?? and b? are the location parameter and scale parameter of the Laplacian distribution for ?,
respectively, and accordingly ?? and b? are the location parameter and scale parameter for ?.
The log of the posterior distribution over the user and item features and weights can be given as
follows:
2
l = ln Pr(U, V, ?, ?|R, ? 2 , ?U
, ?V2 , ?? , b? , ?? , b? )
2
? ln Pr(R|U, V, ?, ?, ? 2 ) Pr(U |?U
) Pr(V |?V2 ) Pr(?|?? , b? ) Pr(?|?? , b? )
=
m X
n
X
1i,j ln
i=1 j=1
K
X
?ik ?jk N (Ri,j |Uik (Vjk )T , ? 2 I)
k=1
K m
K
n
1 XX k 2
1
1 XX k 2 1
2
? 2
(Ui ) ? 2
(Vi ) ? Km ln ?U
? Kn ln ?V2
2?U
2?
2
2
V
i=1
j=1
k=1
?
(4)
k=1
K
K
K m
K
n
1 XX k
1X
1 XX k
1X
m ln b2? ?
n ln b2? + C,
|?i ? ?? | ?
|?j ? ?? | ?
b?
b
2
2
?
i=1
j=1
k=1
k=1
k=1
k=1
where C is a constant that does not depend on any parameters. Since the above optimization problem
is difficult to solve directly, we obtain its lower bound using Jensen?s inequality and then optimize
the following lower bound:
l0
=
?
K
m n
m X
n
1X
X
1 XX
k k
k
k T 2
?
?
(R
?
U
(V
)
)
?
1
1i,j ln ?2
i,j
i,j
i j
i
j
2? 2 i=1 j=1
2 i=1 j=1
k=1
?
?
1
2
2?U
K X
m
X
(Uik )2 ?
k=1 i=1
K
n
1 XX k 2 1
1
2
(Vi ) ? Km ln ?U
? Kn ln ?V2
2
2?V
2
2
j=1
(5)
k=1
K m
K
n
1 XX k
1 XX k
1
1
|?i ? ?? | ?
|?j ? ?? | ? Km ln b2? ? Kn ln b2? + C.
b?
b?
2
2
i=1
j=1
k=1
k=1
If we keep the hyperparameters of the prior distributions fixed, then maximizing l0 is similar to the
popular least square error minimization with `2 regularization on U and V and `1 regularization on ?
and ?. However, keeping the hyperparameters fixed may easily lead to overfitting because MRMA
models have many parameters.
5
Learning MRMA Models
The optimization problem defined in Equation 5 is very likely to overfit if we cannot precisely
estimate the hyperparameters, which automatically control the generalization capacity of the MRMA
model. For instance, ?U and ?V will control the regularization of U and V . Therefore, it is more
desirable to estimate the parameters and hyperparameters simultaneously during model training. One
possible way is to estimate each variable by its maximum a priori (MAP) value while conditioned
on the rest variables and then iterate until convergence, which is also known as iterated conditional
modes (ICM) [1].
The ICM procedure for maximizing Equation 5 is presented as follows.
Initialization: Choose initial values for all variables and parameters.
ICM Step: The values of U , V , ? and ? can be updated by solving the following minimization
problems when conditioned on other variables or hyperparameters.
?k ? {1, ..., K}, ?i ? {1, ..., m} :
Uik ? arg min
0
U
?ik ? arg min
0
?
n
K
K
1 X
X
1 X k 2
k k
k
k T 2
1
?
?
(R
?
U
(V
)
)
+
(Ui ) ,
i,j
i,j
i j
i
j
2
2? 2 j=1
2?U
k=1
k=1
1
2? 2
n
X
j=1
1i,j
K
X
k=1
1
?ik ?jk (Ri,j ? Uik (Vjk )T )2 +
b?
4
K
X
k=1
|?ik ? ?? | .
?k ? {1, ..., K}, ?j ? {1, ..., n} :
Vjk ? arg min
0
V
?jk ? arg min
0
?
m
K
K
X
1 X
1 X k 2
k k
k
k T 2
1
(Vj ) ,
?
+
?
(R
?
U
(V
)
)
i,j
i,j
i j
i
j
2? 2 i=1
2?V2
k=1
k=1
1
2? 2
m
X
i=1
1i,j
K
X
k=1
1
?ik ?jk (Ri,j ? Uik (Vjk )T )2 +
b?
K
X
|?jk ? ?? | .
k=1
The hyperparameters can be learned as their maximum likelihood estimates by setting their partial
derivatives on l0 to 0.
?2 ?
m X
n
X
1i,j
i=1 j=1
2
?U
?
?V2 ?
K
X
m X
n
X
?ik ?jk (Ri,j ? Uik (Vjk )T )2 /
1i,j ,
i=1 j=1
k=1
K X
m
X
K X
m
X
k=1 i=1
k=1 i=1
K X
n
X
K X
n
X
k=1 j=1
k=1 j=1
(Uik )2 /Km, ?? ?
(Vjk )2 /Kn, ?? ?
?ik /Km, b? =
?jk /Kn, b? =
K X
m
X
|?ik ? ?? |/Km,
k=1 i=1
K X
n
X
|?jk ? ?? |/Kn.
k=1 j=1
Repeat: until convergence or the maximum number of iterations reached.
Note that ICM is sensitive to initial values. Our empirical studies show that setting the initial values
of U k and V k by solving the classic PMF method ?
can achieve good performance. Regarding ?
and ?, one of the proper initial values should be 1/ K (K denotes the number of sub-models in
the mixture model). To improve generalization performance and enable online learning [7], we can
update U, V, ?, ? using stochastic gradient descent. Meanwhile, the `1 norms in learning ? and ?
can be approximated by the smoothed `1 method [18]. To deal with massive datasets, we can use the
alternating least squares (ALS) method to learn the parameters of the proposed MRMA model, which
is amenable to parallelization.
6
Experiments
This section presents the experimental results of the proposed MRMA method on three well-known
datasets: 1) MovieLens 1M dataset (?1 million ratings from 6,040 users on 3,706 movies); 2)
MovieLens 10M dataset (?10 million ratings from 69,878 users on 10,677 movies); 3) Netflix Prize
dataset (?100 million ratings from 480,189 users on 17,770 movies). For all accuracy comparisons,
we randomly split each dataset into a training set and a test set by the ratio of 9:1. All results are
reported by averaging over 5 different splits. The root mean square error (RMSE) is adopted to
measure the
rating prediction accuracy of different algorithms, which can be computed as follows:
qP
P
?
? 2 P P 1i,j (1i,j indicates that entry (i, j) appears in the
D(R) =
i
j 1i,j (Ri,j ? Ri,j ) /
i
j
test set). The normalized discounted cumulative gain (NDCG) is adopted to measure the item
ranking accuracy of different algorithms, which can be computed as follows: N DCG@N =
PN
DCG@N/IDCG@N (DCG@N = i=1 (2reli ? 1)/ log2 (i + 1), and IDCG is the DCG value
with perfect ranking).
In ICM-based learning, we adopt = 0.00001 as the convergence threshold and T = 300 as the
maximum number of iterations. Considering efficiency, we only choose a subset of ranks, e.g.,
{10, 20, 30, ..., 300} rather than {1, 2, 3, ..., 300}, in MRMA. The parameters of all the compared
algorithms are adopted from their original papers because all of them are evaluated on the same
datasets.
We compare the recommendation accuracy of MRMA with six matrix approximation-based collaborative filtering algorithms as follows: 1) BPMF [16], which extends the PMF method from a Baysian
view and estimates model parameters using a Markov chain Monte Carlo scheme; 2) GSMF [20],
which learns user/item features with group sparsity regularization in matrix approximation; 3) LLORMA [12], which ensembles the approximations from different submatrices using kernel smoothing;
4) WEMAREC [5], which ensembles different biased matrix approximation models to achieve higher
5
PMF
MRMA
RMSE
computation time
0.86
1000
RMSE
RMSE
0.86
0.84
0.82
0.82
0.80
0.84
k=
10
k=
20
k=
50
k=
0.80
k
k
k
k
M
10 =15 =20 =25 =30 RM
0
0
0
0
0
A
se
t1
se
t2
se
t3
se
t4
se
t
computation time (s)
0.88
0
5
rank setting
Model
Figure 2: Root mean square error comparison Figure 3: The accuracy and efficiency tradeoff of
between MRMA and PMF with different ranks. MRMA.
accuracy; 5) MPMA [4], which combines local and global matrix approximations using a mixture
model; 6) SMA [13], which yields a stable matrix approximation that can achieve good generalization
performance.
6.1
Mixture-Rank Matrix Approximation vs. Fixed-Rank Matrix Approximation
Given a fixed rank k, the corresponding rank-k model in MRMA is identical to probabilistic matrix
factorization (PMF) [17]. In this experiment, we compare the recommendation accuracy of MRMA
with ranks in {10, 20, 50, 100, 150, 200, 250, 300} against those of PMF with fixed ranks on the
MovieLens 1M dataset. For PMF, we choose 0.01 as the learning rate, 0.01 as the user feature
regularization coefficient, and 0.001 as the item feature regularization coefficient, respectively. The
convergence condition is the same as MRMA.
As shown in Figure 2, when the rank increases from 10 to 300, PMF can achieve RMSEs between
0.86 and 0.88. However, the RMSE of MRMA is about 0.84 when mixing all these ranks from 10 to
300. Meanwhile, the accuracy of PMF is not stable when k ? 100. For instance, PMF with k = 10
achieves better accuracy than k = 20 but worse accuracy than k = 50. This is because fixed rank
matrix approximation cannot be perfect for all users and items, so that many users and items either
underfit or overfit at a fixed rank less than 100. Yet when k > 100, only overfitting occurs and PMF
achieves consistently better accuracy when k increases, which is because regularization terms can
help improve generalization capacity. Nevertheless, PMF with all ranks achieves lower accuracy
than MRMA, because individual users/items can give the sub-models with the optimal ranks higher
weights in MRMA and thus alleviate underfitting or overfitting.
6.2
Sensitivity of Rank in MRMA
In MRMA, the set of ranks decide the performance of the final model. However, it is neither efficient
nor necessary to choose all the ranks in [1, 2, ..., K]. For instance, a rank-k approximation will
be very similar to rank-(k ? 1) and rank-(k + 1) approximations, i.e., they may have overlapping
structures. Therefore, a subset of ranks will be sufficient. Figure 3 shows 5 different settings of
rank combinations, in which set 1 = {10, 20, 30, ..., 300}, set 2 = {20, 40, ..., 300}, set 3 =
{30, 60, ..., 300}, set 4 = {50, 100, ..., 300}, and set 5 = {100, 200, 300}. As shown in this figure,
RMSE decreases when more ranks are adopted in MRMA, which is intuitive because more ranks will
help users/items better choose the most appropriate components. However, the computation time
also increases when more ranks are adopted in MRMA. If a tradeoff between accuracy and efficiency
is required, then set 2 or set 3 will be desirable because they achieve slightly worse accuracies but
significantly less computation overheads.
MRMA only contains three sub-models with different ranks in set 5 = {100, 200, 300}, but it still
significantly outperforms PMF with ranks ranging from 10 to 300 in recommendation accuracy (as
shown in Figure 2). This further confirms that MRMA can indeed discover the internal mixture-rank
structure of the user-item rating matrix and thus achieve better recommendation accuracy due to
better approximation.
6
Table 2: RMSE comparison between MRMA and six state-of-the-art matrix approximation-based
collaborative filtering algorithms on MovieLens (10M) and Netflix datasets. Note that MRMA
statistically significantly outperforms the other algorithms with 95% confidence level.
MovieLens (10M)
Netflix
0.8197 ? 0.0004
0.8012 ? 0.0011
0.7855 ? 0.0002
0.7775 ? 0.0007
0.7712 ? 0.0002
0.7682 ? 0.0003
0.7634 ? 0.0009
0.8421 ? 0.0002
0.8420 ? 0.0006
0.8275 ? 0.0004
0.8143 ? 0.0001
0.8139 ? 0.0003
0.8036 ? 0.0004
0.7973 ? 0.0002
BPMF [16]
GSMF [20]
LLORMA [12]
WEMAREC [5]
MPMA [4]
SMA [13]
MRMA
Table 3: NDCG comparison between MRMA and six state-of-the-art matrix approximation-based
collaborative filtering algorithms on Movielens (1M) and Movielens (10M) datasets. Note that
MRMA statistically significantly outperforms the other algorithms with 95% confidence level.
Metric
N=1
N=5
N=10
N=20
Movielens 1M
BPMF
GSMF
LLORMA
WEMAREC
MPMA
SMA
MRMA
0.6870 ? 0.0024
0.6909 ? 0.0048
0.7025 ? 0.0027
0.7048 ? 0.0015
0.7020 ? 0.0005
0.7042 ? 0.0033
0.7153 ? 0.0027
0.6981 ? 0.0029
0.7031 ? 0.0023
0.7101 ? 0.0005
0.7089 ? 0.0016
0.7114 ? 0.0018
0.7109 ? 0.0011
0.7182 ? 0.0005
0.7525 ? 0.0009
0.7555 ? 0.0017
0.7626 ? 0.0023
0.7617 ? 0.0041
0.7606 ? 0.0006
0.7607 ? 0.0008
0.7672 ? 0.0013
0.8754 ? 0.0008
0.8769 ? 0.0011
0.8811 ? 0.0010
0.8796 ? 0.0005
0.8805 ? 0.0007
0.8801 ? 0.0004
0.8837 ? 0.0004
Movielens 10M
Data | Method
NDCG@N
BPMF
GSMF
LLORMA
WEMAREC
MPMA
SMA
MRMA
0.6563 ? 0.0005
0.6708 ? 0.0012
0.6829 ? 0.0014
0.7013 ? 0.0003
0.6908 ? 0.0006
0.7002 ? 0.0006
0.7048 ? 0.0006
0.6845 ? 0.0003
0.6995 ? 0.0008
0.7066 ? 0.0005
0.7176 ? 0.0006
0.7133 ? 0.0002
0.7134 ? 0.0004
0.7219 ? 0.0001
0.7467 ? 0.0007
0.7566 ? 0.0017
0.7632 ? 0.0004
0.7703 ? 0.0002
0.7680 ? 0.0001
0.7679 ? 0.0003
0.7743 ? 0.0001
0.8691 ? 0.0002
0.8748 ? 0.0004
0.8782 ? 0.0012
0.8824 ? 0.0006
0.8808 ? 0.0004
0.8809 ? 0.0002
0.8846 ? 0.0001
6.3
6.3.1
Accuracy Comparison
Rating Prediction Comparison
Table 2 compares the rating prediction accuracy between MRMA and six matrix approximationbased collaborative filtering algorithms on MovieLens (10M) and Netflix datasets. Note that among
the compared algorithms, BPMF, GSMF, MPMA and SMA are stand-alone algorithms, while
LLORMA and WEMAREC are ensemble algorithms. In this experiment, we adopt the set of ranks
as {10, 20, 50, 100, 150, 200, 250, 300} due to efficiency reason, which means that the accuracy of
MRMA should not be optimal. However, as shown in Table 2, MRMA statistically significantly
outperforms all the other algorithms with 95% confidence level. The reason is that MRMA can
choose different rank values for different users/items, which can achieve not only globally better
approximation but also better approximation in terms of individual users or items. This further
confirms that mixture-rank structure indeed exists in user-item rating matrices in recommender
systems. Thus, it is desirable to adopt mixture-rank matrix approximations rather than fixed-rank
matrix approximations for recommendation tasks.
6.3.2
Item Ranking Comparison
Table 3 compares the NDCGs of MRMA with the other six state-of-the-art matrix approximationbased collaborative filtering algorithms on Movielens (1M) and Movielens (10M) datasets. Note that
for each dataset, we keep 20 ratings in the test set for each user and remove users with less than 5
7
ratings in the training set. As shown in the results, MRMA can also achieve higher item ranking
accuracy than the other compared algorithms thanks to the capability of better capturing the internal
mixture-rank structures of the user-item rating matrices. This experiment demonstrates that MRMA
can not only provide accurate rating prediction but also achieve accurate item ranking for each user.
6.4
Interpretation of MRMA
Table 4: Top 10 movies with largest ? values for sub-models with rank k = 20 and k = 200 in MRMA.
Here, #ratings stands for the average number of ratings in the training set for the corresponding
movies.
rank=20
rank=200
movie name
?
Smashing Time
Gate of Heavenly Peace
Man of the Century
Mamma Roma
Dry Cleaning
Dear Jesse
Skipped Parts
The Hour of the Pig
Inheritors
Dangerous Game
0.6114
0.6101
0.6079
0.6071
0.6071
0.6063
0.6057
0.6055
0.6042
0.6034
#ratings
movie name
?
#ratings
2.4
American Beauty
Groundhog Day
Fargo
Face/Off
2001: A Space Odyssey
Shakespeare in Love
Saving Private Ryan
The Fugitive
Braveheart
Fight Club
0.9219
0.9146
0.8779
0.8693
0.8608
0.8553
0.8480
0.8404
0.8247
0.8153
1781.4
To better understand how users/items weigh different sub-models in the mixture model of MRMA,
we present the top 10 movies which have largest ? values for sub-models with rank=20 and rank=200,
show their ? values, and compare their average numbers of ratings in the training set in Table 4.
Intuitively, the movies with more ratings (e.g., over 1000 ratings) should weigh higher towards more
complex models, and the movies with less ratings (e.g., under 10 ratings) should weigh higher towards
simpler models in MRMA.
As shown in Table 4, the top 10 movies with largest ? values for the sub-model with rank 20 have
only 2.4 ratings on average in the training set. On the contrary, the top 10 movies with largest ? values
for the sub-model with rank 200 have 1781.4 ratings on average in the training set, and meanwhile
these movies are very popular and most of them are Oscar winners. This confirms our previous claim
that MRMA can indeed weigh more complex models (e.g., rank=200) higher for movies with more
ratings to prevent underfitting, and weigh less complex models (e.g., rank=20) higher for the movies
with less ratings to prevent overfitting. A similar phenomenon has also been observed from users
with different ? values, and we omit the results due to space limit.
7
Conclusion and Future Work
This paper proposes a mixture-rank matrix approximation (MRMA) method, which describes useritem ratings using a mixture of low-rank matrix approximation models with different ranks to achieve
better approximation and thus better recommendation accuracy. An ICM-based learning algorithm is
proposed to handle the non-convex optimization problem pertaining to MRMA. The experimental
results on MovieLens and Netflix datasets demonstrate that MRMA can achieve better accuracy than
six state-of-the-art matrix approximation-based collaborative filtering methods, further pushing the
frontier of recommender systems. One of the possible extensions of this work is to incorporate other
inference methods into learning the MRMA model, e.g., variational inference [8], because ICM may
be trapped in local maxima and therefore cannot achieve global maxima without properly chosen
initial values.
Acknowledgement
This work was supported in part by the National Natural Science Foundation of China under Grant
No. 61332008 and NSAF under Grant No. U1630115.
8
References
[1] J. Besag. On the statistical analysis of dirty pictures. Journal of the Royal Statistical Society. Series B
(Methodological), pages 259?302, 1986.
[2] E. J. Cand?s and B. Recht. Exact matrix completion via convex optimization. Communications of the
ACM, 55(6):111?119, 2012.
[3] E. J. Cand?s and T. Tao. The power of convex relaxation: Near-optimal matrix completion. IEEE
Transactions on Information Theory, 56(5):2053?2080, 2010.
[4] C. Chen, D. Li, Q. Lv, J. Yan, S. M. Chu, and L. Shang. MPMA: mixture probabilistic matrix approximation for collaborative filtering. In Proceedings of the 25th International Joint Conference on Artificial
Intelligence (IJCAI ?16), pages 1382?1388, 2016.
[5] C. Chen, D. Li, Y. Zhao, Q. Lv, and L. Shang. WEMAREC: Accurate and scalable recommendation
through weighted and ensemble matrix approximation. In Proceedings of the 38th International ACM
SIGIR Conference on Research and Development in Information Retrieval (SIGIR ?15), pages 303?312,
2015.
[6] D. Dueck and B. Frey. Probabilistic sparse matrix factorization. University of Toronto technical report
PSI-2004-23, 2004.
[7] M. Hardt, B. Recht, and Y. Singer. Train faster, generalize better: Stability of stochastic gradient descent,
2015. arXiv:1509.01240.
[8] M. I. Jordan, Z. Ghahramani, T. S. Jaakkola, and L. K. Saul. An introduction to variational methods for
graphical models. Machine learning, 37(2):183?233, 1999.
[9] Y. Koren. Factorization meets the neighborhood: a multifaceted collaborative filtering model. In Proceedings of the 14th ACM SIGKDD international conference on Knowledge discovery and data mining (KDD
?14), pages 426?434. ACM, 2008.
[10] Y. Koren. Collaborative filtering with temporal dynamics. In Proceedings of the 15th ACM SIGKDD
International Conference on Knowledge Discovery and Data Mining (KDD ?09), pages 447?456. ACM,
2009.
[11] Y. Koren, R. Bell, and C. Volinsky. Matrix factorization techniques for recommender systems. Computer,
42(8), 2009.
[12] J. Lee, S. Kim, G. Lebanon, and Y. Singer. Local low-rank matrix approximation. In Proceedings of The
30th International Conference on Machine Learning (ICML ?13), pages 82?90, 2013.
[13] D. Li, C. Chen, Q. Lv, J. Yan, L. Shang, and S. Chu. Low-rank matrix approximation with stability. In The
33rd International Conference on Machine Learning (ICML ?16), pages 295?303, 2016.
[14] H. Ma, H. Yang, M. R. Lyu, and I. King. Sorec: social recommendation using probabilistic matrix
factorization. In Proceedings of the 17th ACM conference on Information and knowledge management
(CIKM ?08), pages 931?940. ACM, 2008.
[15] L. W. Mackey, M. I. Jordan, and A. Talwalkar. Divide-and-conquer matrix factorization. In Advances in
Neural Information Processing Systems (NIPS ?11), pages 1134?1142, 2011.
[16] R. Salakhutdinov and A. Mnih. Bayesian probabilistic matrix factorization using markov chain monte carlo.
In Proceedings of the 25th international conference on Machine learning (ICML ?08), pages 880?887.
ACM, 2008.
[17] R. Salakhutdinov and A. Mnih. Probabilistic matrix factorization. In Advances in Neural Information
Processing Systems (NIPS ?08), pages 1257?1264, 2008.
[18] M. Schmidt, G. Fung, and R. Rosales. Fast optimization methods for L1 regularization: A comparative
study and two new approaches. In European Conference on Machine Learning (ECML ?07), pages 286?297.
Springer, 2007.
[19] K.-C. Toh and S. Yun. An accelerated proximal gradient algorithm for nuclear norm regularized linear
least squares problems. Pacific Journal of Optimization, 6(15):615?640, 2010.
[20] T. Yuan, J. Cheng, X. Zhang, S. Qiu, and H. Lu. Recommendation by mining multiple user behaviors with
group sparsity. In Proceedings of the 28th AAAI Conference on Artificial Intelligence (AAAI ?14), pages
222?228, 2014.
9
| 6651 |@word private:1 norm:3 km:6 confirms:4 initial:5 contains:1 score:1 series:1 outperforms:4 existing:2 com:1 contextual:1 toh:1 yet:1 chu:2 shakespeare:1 kdd:2 remove:1 update:1 v:1 alone:2 intelligence:2 mackey:1 weighing:1 item:63 accordingly:1 isotropic:1 prize:1 dear:1 boosting:1 location:2 club:1 toronto:1 simpler:1 zhang:1 vjk:10 ik:12 yuan:1 combine:1 overhead:1 underfitting:2 indeed:4 behavior:1 cand:3 nor:1 love:1 salakhutdinov:2 discounted:1 globally:1 automatically:1 considering:1 xx:8 moreover:1 discover:1 fudan:3 finding:1 dueck:1 temporal:1 tackle:1 rm:5 demonstrates:1 control:2 grant:2 omit:1 t1:1 local:4 frey:1 limit:1 meet:1 ndcg:3 initialization:1 china:5 factorization:9 range:1 statistically:3 procedure:1 empirical:1 yan:2 submatrices:5 significantly:6 bell:1 confidence:3 cannot:7 coexist:4 optimize:1 map:1 maximizing:3 jesse:1 convex:7 sigir:2 pure:1 borrow:1 nuclear:2 classic:1 century:1 bpmf:6 handle:1 stability:2 updated:1 today:2 user:68 massive:1 cleaning:1 exact:1 approximated:1 jk:12 observed:3 lrma:11 decrease:1 weigh:5 ui:3 dynamic:1 depend:1 solving:2 efficiency:4 easily:4 joint:3 represented:1 sacrificed:1 train:2 fast:1 describe:5 monte:2 pertaining:2 artificial:2 neighborhood:1 solve:1 otherwise:1 final:1 online:1 indication:1 propose:1 product:1 adaptation:1 mixing:1 achieve:17 intuitive:1 scalability:1 convergence:4 ijcai:1 comparative:1 perfect:2 help:2 derive:1 completion:2 school:1 auxiliary:1 come:1 rosales:1 ning:1 stochastic:2 enable:1 odyssey:1 generalization:4 alleviate:1 ryan:1 frontier:1 extension:1 considered:2 lyu:1 claim:1 sma:6 achieves:4 adopt:4 sensitive:1 largest:4 weighted:1 minimization:4 gaussian:3 rather:2 pn:1 beauty:1 varying:2 jaakkola:1 l0:3 improvement:1 consistently:1 rank:102 indicates:2 likelihood:1 properly:1 methodological:1 besag:1 skipped:1 sigkdd:2 kim:1 talwalkar:1 inference:2 typically:1 dcg:4 fight:1 transformed:1 tao:1 arg:4 among:3 priori:1 proposes:1 development:1 art:6 smoothing:1 fugitive:1 saving:1 beach:1 identical:1 placing:2 icml:3 future:1 t2:1 report:1 few:1 randomly:1 simultaneously:1 national:1 individual:2 mining:3 mnih:2 mixture:22 analyzed:1 chain:2 amenable:1 accurate:5 partial:1 necessary:1 orthogonal:1 divide:1 pmf:22 instance:4 entry:1 subset:2 conducted:2 reported:1 kn:6 proximal:1 st:1 density:1 thanks:1 sensitivity:1 recht:2 international:7 probabilistic:8 off:1 lee:1 enhance:1 aaai:2 management:1 containing:2 leveraged:1 choose:8 worse:2 american:1 derivative:1 leading:1 zhao:1 li:3 dfc:1 b2:4 coefficient:2 ranking:5 vi:2 root:4 view:1 lab:1 overfits:1 reached:1 netflix:9 capability:1 rmse:8 collaborative:14 square:7 accuracy:33 degraded:1 variance:1 likewise:1 ensemble:5 t3:1 yield:1 dry:1 generalize:1 bayesian:1 iterated:3 lu:1 carlo:2 against:2 volinsky:1 associated:1 psi:1 gain:1 proved:1 dataset:7 popular:3 hardt:1 knowledge:3 underfits:1 appears:1 higher:10 day:1 wei:1 formulation:1 evaluated:1 implicit:1 until:2 overfit:3 overlapping:1 mode:3 quality:1 multifaceted:1 name:2 usa:1 normalized:1 regularization:8 alternating:1 laboratory:1 iteratively:1 deal:1 during:1 game:1 inferior:1 yun:1 demonstrate:2 l1:1 ranging:1 variational:2 qp:1 shanghai:1 winner:1 million:3 interpretation:1 ai:1 rd:1 similarly:2 dot:1 stable:2 etc:4 base:3 posterior:1 recent:1 inequality:1 exploited:1 employed:2 determine:3 stephen:1 multiple:1 desirable:5 technical:1 faster:1 characterized:1 long:1 retrieval:1 peace:1 laplacian:4 prediction:4 scalable:1 metric:1 arxiv:1 iteration:2 kernel:1 adopting:2 achieved:1 liu2:1 addition:1 parallelization:1 rest:2 biased:1 contrary:1 leveraging:1 groundhog:1 jordan:2 ee:1 near:1 yang:1 split:2 iterate:1 li1:1 perfectly:3 imperfect:1 idea:3 cn:2 regarding:1 tradeoff:2 six:7 se:5 generate:1 outperform:1 trapped:1 cikm:1 group:2 key:1 threshold:1 nevertheless:1 prevent:2 neither:1 relaxation:1 run:1 letter:1 oscar:1 extends:1 decide:1 submatrix:2 capturing:1 apg:1 bound:2 koren:3 cheng:1 dangerous:1 precisely:1 ri:10 min:5 relatively:1 pacific:1 fung:1 combination:1 across:1 describes:2 slightly:1 intuitively:1 pr:11 ln:12 equation:2 singer:2 end:1 capitalizing:2 adopted:7 v2:9 appropriate:1 schmidt:1 gate:1 original:1 denotes:7 top:4 cf:5 dirty:1 graphical:2 log2:1 pushing:1 reli:1 exploit:1 ghahramani:1 conquer:1 society:1 occurs:1 gradient:3 capacity:2 trivial:1 reason:2 rom:1 besides:1 relationship:2 ratio:1 equivalently:1 difficult:1 proper:1 recommender:4 upper:1 datasets:11 markov:2 descent:2 ecml:1 communication:1 rn:2 smoothed:1 rating:63 introduced:1 pair:1 required:1 baysian:1 learned:1 chen1:1 hour:1 nip:3 sparsity:2 pig:1 tun:1 royal:1 power:1 suitable:1 natural:1 regularized:1 scheme:1 improve:5 movie:16 picture:1 fargo:1 columbia:1 chao:1 prior:5 acknowledgement:1 discovery:2 filtering:14 lv:3 rmses:4 foundation:1 sufficient:1 ibm:3 repeat:1 supported:1 keeping:1 understand:1 saul:1 face:1 sparse:2 feedback:1 world:2 stand:3 cumulative:1 author:1 cope:1 social:2 transaction:1 lebanon:1 keep:2 global:2 overfitting:4 conclude:1 table:10 learn:1 ca:1 obtaining:1 tencent:1 excellent:1 complex:3 meanwhile:4 european:1 vj:2 underfit:2 noise:1 hyperparameters:6 qiu:1 icm:8 uik:11 sub:8 learns:1 jensen:1 exists:1 conditioned:3 t4:1 chen:3 likely:1 recommendation:16 springer:1 acm:9 ma:1 conditional:2 goal:2 targeted:4 king:1 towards:2 man:1 movielens:18 averaging:1 useritem:1 shang:3 experimental:4 internal:7 accelerated:1 incorporate:1 evaluate:1 phenomenon:1 |
6,248 | 6,652 | Non-monotone Continuous DR-submodular
Maximization: Structure and Algorithms
An Bian
ETH Zurich
[email protected]
Kfir Y. Levy
ETH Zurich
[email protected]
Andreas Krause
ETH Zurich
[email protected]
Joachim M. Buhmann
ETH Zurich
[email protected]
Abstract
DR-submodular continuous functions are important objectives with wide real-world
applications spanning MAP inference in determinantal point processes (DPPs),
and mean-field inference for probabilistic submodular models, amongst others.
DR-submodularity captures a subclass of non-convex functions that enables both
exact minimization and approximate maximization in polynomial time.
In this work we study the problem of maximizing non-monotone DR-submodular
continuous functions under general down-closed convex constraints. We start
by investigating geometric properties that underlie such objectives, e.g., a strong
relation between (approximately) stationary points and global optimum is proved.
These properties are then used to devise two optimization algorithms with provable
guarantees. Concretely, we first devise a ?two-phase? algorithm with 1/4 approximation guarantee. This algorithm allows the use of existing methods for finding
(approximately) stationary points as a subroutine, thus, harnessing recent progress
in non-convex optimization. Then we present a non-monotone F RANK -W OLFE
variant with 1/e approximation guarantee and sublinear convergence rate. Finally,
we extend our approach to a broader class of generalized DR-submodular continuous functions, which captures a wider spectrum of applications. Our theoretical
findings are validated on synthetic and real-world problem instances.
1
Introduction
Submodularity is classically most well known for set function optimization, where it enables efficient
minimization [23] and approximate maximization [31; 25] in polynomial time. Submodularity has
recently been studied on the integer lattice [34; 33] and on continuous domains [3; 4; 36; 21], with
significant theoretical results and practical applications. For set functions, it is well known that
submodularity is equivalent to the diminishing returns (DR) property. However, this does not hold
for integer-lattice functions or continuous functions, where the DR property defines a subclass of
submodular functions, called DR-submodular functions.
In continuous domains, applying convex optimization techniques enables efficient minimization of
submodular continuous functions [3; 36] (despite the non-convex nature of such objectives). In [4]
it is further shown that continuous submodularity enables constant-factor approximation schemes
for constrained monotone DR-submodular maximization and ?box? constrained non-monotone
submodular maximization problems.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
Many real-world non-convex problems, such as maximizing the softmax extension of DPPs, require
maximizing a non-monotone DR-submodular function over a general down-closed convex constraint.
Yet, current theory [3; 4; 36] does not apply to this general problem setting, which motivates us to
develop guaranteed and efficient algorithms for such problems.
Exploring the structure that underlies DR-submodularity is crucial to deriving guaranteed algorithms.
Combined with a notion of non-stationarity for constrained optimization problems and a new notion
of ?strong DR-submodularity?, we find a rich structure in the problem of continuous DR-submodular
maximization. This in turn gives rise to two approximation algorithms with provable guarantees.
Specifically, we make the following contributions:
- We bound the difference between objective values of stationary points and the global optimum.
Our analysis shows that the bound is even tighter if the objective is strongly DR-submodular
(see Definition 3).
- Based on the geometric properties, we present two algorithms: (i) A two-phase
p F RANK W OLFE-style algorithm with 1/4 approximation guarantee converges with a 1/ k rate; (ii) a
non-monotone F RANK -W OLFE variant exhibits a 1/e approximation guarantee and converges
sublinearly. Even though the worst-case guarantee of the first one is worse than the second, it
yields several practical advantages, which we discuss in Section 4.2.
- We investigate a generalized class of submodular functions on ?conic? lattices. This allows
us to model a larger class of non-trivial applications. These include logistic regression with
a non-convex separable regularizer, non-negative PCA, etc. To optimize them, we provide a
reduction that enables to invoke algorithms for continuous submodular optimization problems.
- We experimentally demonstrate the applicability of our methods on both synthetic and realworld problem instances.
1.1
Problem Statement
Notation. We use boldface letters, e.g., x to represent a vector, boldface capital letters, e.g., A
to denote a matrix. xi is the ith entry of x, Aij is the (ij)th entry of A. We use ei to denote the
standard ith basis vector. f (?) is used to denote a continuous function, and F (?) to represent a set
function. [n] := {1, ..., n} for an integer n 1. k ? k means the Euclidean norm by default. Given
two vectors x, y, x ? y means xi ? yi , 8i. x _ y and x ^ y denote coordinate-wise maximum and
coordinate-wise minimum, respectively.
The general setup of constrained non-monotone DR-submodular (see Definition 1 below) maximization is,
max f (x),
(P)
x2P
Qn
where f : X ! R is continuous DR-submodular, X = i=1 Xi , each Xi is an interval [3; 4]. Wlog1 ,
? The set P ? [0, u]
? is assumed to be a
we assume that the lower bound u of X is 0, i.e., X = [0, u].
down-closed convex set, where down-closedness means: x 2 P and 0 ? y ? x implies that y 2 P.
? We use x? to denote the
The diameter of P is D := maxx,y2P kx yk, and it holds that D ? kuk.
global maximum of (P). One can assume f is non-negative over X , since otherwise one just needs to
? (and box-constrained submodular
find a lower bound for the minimum function value of f over [0, u]
minimization can be solved to arbitrary precision in polynomial time [3]). Over continuous domains,
a DR-submodular function [4] is a submodular function with the diminishing returns (DR) property,
Definition 1 (DR-submodular & DR property). A function f : X 7! R is DR-submodular (has the
DR property) if 8a ? b 2 X , 8i 2 [n], 8k 2 R+ s.t. (kei + a) and (kei + b) are still in X , it holds,
f (kei + a)
f (a)
f (kei + b)
f (b).
(1)
If f is differentiable, one can show that Definition 1 is equivalent to rf being an antitone mapping
from Rn to Rn . Furthermore, if f is twice-differentiable, the DR property is equivalent to all of
the entries of its Hessian being non-positive, i.e., r2ij f (x) ? 0, 8x 2 X , i, j 2 [n]. A function
f : X 7! R is DR-supermodular iff f is DR-submodular. We also assume that f has Lipschitz
gradients,
1
Since otherwise one can work on a new function g(x) := f (x + u) that has 0 as the lower bound of its
domain, and all properties of the function are still preserved.
2
Definition 2. A function f has L-Lipschitz gradients if for all x, y 2 X it holds that,
krf (x)
rf (y)k ? Lkx
yk.
(2)
A brief summary of related work appears in Section 6.
2
Motivating Real-world Examples
Many continuous objectives in practice turn out to be DR-submodular. Here we list several of them.
More can be found in Appendix B.
Softmax extension. Determinantal point processes (DPPs) are probabilistic models of repulsion,
that have been used to model diversity in machine learning [26]. The constrained MAP (maximum a
posteriori) inference problem of a DPP is an NP-hard combinatorial problem in general. Currently,
the methods with the best approximation guarantees are based on either maximizing the multilinear
extension [6] or the softmax extension [20], both of which are DR-submodular functions (details in
Appendix F.1). The multilinear extension is given as an expectation over the original set function
values, thus evaluating the objective of this extension requires expensive sampling. In constast, the
softmax extension has a closed form expression, which is much more appealing from a computational
perspective. Let L be the positive semidefinite kernel matrix of a DPP, its softmax extension is:
I) + I) , x 2 [0, 1]n ,
f (x) = log det (diag(x)(L
(3)
where I is the identity matrix, diag(x) is the diagonal matrix with diagonal elements set as x.
The problem of MAP inference in DPPs corresponds to the problem maxx2P f (x), where P is a
down-closed convex constraint, e.g., a matroid polytope or a matching polytope.
Mean-field inference for log-submodular models. Log-submodular models [9] are a class of
probabilistic models over subsets of a ground set V = [n], where the log-densities
P are submodular
set functions F (S): p(S) = Z1 exp(F (S)). The partition function Z =
S?V exp(F (S)) is
typically hard to evaluate.
One
can
use
mean-field
inference
to
approximate
p(S)
by some factorized
Q
Q
distribution qx (S) := i2S xi j 2S
xj ), x 2 [0, 1]n , by minimizing the distance measured
/ (1
P
x (S)
w.r.t. the Kullback-Leibler divergence between qx and p, i.e., S?V qx (S) log qp(S)
. It is,
XY Y
Xn
KL(x) =
xi
(1 xj )F (S) +
[xi log xi + (1 xi ) log(1 xi )] + log Z.
S?V i2S
i=1
j 2S
/
KL(x) is DR-supermodular w.r.t. x (details in Appendix F.1). Minimizing the Kullback-Leibler
divergence KL(x) amounts to maximizing a DR-submodular function.
2.1
Motivating Example Captured by Generalized Submodularity on Conic Lattices
Submodular continuous functions can already model many scenarios. Yet, there are several interesting cases which are in general not (DR-)Submodular, but can still be captured by a generalized
notion. This generalization enables to develop polynomial algorithms with guarantees by using ideas
from continuous submodular optimization. We present one representative objective here (more in
Appendix B). In Appendix A we show the technical details on how they are covered by a class of
submodular continuous functions over conic lattices.
Consider the logistic regression model with a non-convex separable regularizer. This flexibility may
result in better statistical performance (e.g., in recovering discontinuities, [2]) compared to classical
models with convex regularizers. Let z 1 , ..., z m in Rn be m training points with corresponding
binary labels y 2 {?1}m . Assume that the following mild assumption is satisfied: For any fixed
dimension i, all the data points have the same sign, i.e., sign(zij ) is the same for all j 2 [m] (which
can be achieved by easily scaling if not). The task is to solve the following non-convex optimization
problem,
Xm
minn f (x) := m 1
fj (x) + r(x),
(4)
j=1
x2R
where fj (x) = log(1 + exp( yj x> z j )) is the logistic loss; > 0 is the regularization parameter,
and r(x) is some non-convex separable regularizer. Such separable regularizers are popular in
3
statistics, and two notable choices are r(x) =
Pn
x2i
i=1 1+ x2i ,
sign(zij ), i 2
and r(x) =
Pn
i=1 min{
Pm
1
:= m
j=1
x2i , 1} (see
[2]). Let us define a vector ? 2 {?1}n as ?i =
[n] and l(x)
fj (x). One
can show that l(x) is not DR-submodular or DR-supermodular. Yet, in Appendix A we will show
that l(x) is K? -DR-supermodular, where the latter generalizes DR-supermodularity. Usually, one
? Then the problem is an instance of
can assume the optimal solution x? lies in some box [u, u].
constrained non-monotone K? -DR-submodular maximization.
3
Underlying Properties of Constrained DR-submodular Maximization
In this section we present several properties arising in DR-submodular function maximization. First
we show properties related to concavity of the objective along certain directions, then we establish the
relation between locally stationary points and the global optimum (thus called ?local-global relation?).
These properties will be used to derive guarantees for the algorithms in Section 4. All omitted proofs
are in Appendix D.
3.1
Properties Along Non-negative/Non-positive Directions
A DR-submodular function f is concave along any non-negative/non-positive direction [4]. Notice
that DR-submodularity is a stronger condition than concavity along directions v 2 ?Rn+ : for instance,
a concave function is concave along any direction, but it may not be a DR-submodular function.
For a DR-submodular function with L-Lipschitz gradients, one can get the following quadratic lower
bound using standard techniques by combing the concavity and Lipschitz gradients in (2).
Quadratic lower bound. If f is DR-submodular with a L-Lipschitz gradient, then for all x 2 X
and v 2 ?Rn+ , it holds,
L
kvk2 .
(5)
2
It will be used in Section 4.2 for analyzing the non-monotone F RANK -W OLFE variant (Algorithm 2).
f (x + v)
f (x) + hrf (x), vi
Strong DR-submodularity. DR-submodular objectives may be strongly concave along directions
v 2 ?Rn+ , e.g., for DR-submodular quadratic functions. We will show that such additional structure
may be exploited to obtain stronger guarantees for the local-global relation.
Definition 3 (Strongly DR-submodular). A function f is ?-strongly DR-submodular (? 0) if for
all x 2 X and v 2 ?Rn+ , it holds that,
f (x + v) ? f (x) + hrf (x), vi
3.2
?
kvk2 .
2
(6)
Relation Between Approximately Stationary Points and Global Optimum
First of all, we present the following Lemma, which will motivate us to consider a non-stationarity
measure for general constrained optimization problems.
Lemma 1. If f is ?-strongly DR-submodular, then for any two points x, y in X , it holds:
?
(y x)> rf (x) f (x _ y) + f (x ^ y) 2f (x) + kx yk2 .
(7)
2
Lemma 1 implies that if x is stationary (i.e., rf (x) = 0), then 2f (x) f (x _ y) + f (x ^ y) +
?
yk2 , which gives an implicit relation between x and y. While in practice finding an exact
2 kx
stationary point is not easy, usually non-convex solvers will arrive at an approximately stationary
point, thus requiring a proper measure of non-stationarity for the constrained optimization problem.
Non-stationarity measure. Looking at the LHS of (7), it naturally suggests to use maxy2P (y
x)> rf (x) as the non-stationarity measure, which happens to coincide with the measure proposed
by recent work of [27], and it can be calculated for free for Frank-Wolfe-style algorithms (e.g.,
Algorithm 3). In order to adapt it to the local-global relation, we give a slightly more general
definition here: For any constraint set Q ? X , the non-stationarity of a point x 2 Q is,
gQ (x) := maxhv
v2Q
x, rf (x)i
4
(non-stationarity).
(8)
It always holds that gQ (x) 0, and x is a stationary point in Q iff gQ (x) = 0, so (8) is a natural
generalization of the non-stationarity measure krf (x)k for unconstrained optimization. As the next
statement shows, gQ (x) plays an important role in characterizing the local-global relation.
Proposition 1 (Local-Global Relation). Let x be a point in P with non-stationarity gP (x), and
? x}. Let z be a point in Q with non-stationarity gQ (z). It holds that,
Q := {y 2 P | y ? u
max{f (x), f (z)}
where z ? := x _ x?
1
[f (x? )
4
gP (x)
gQ (z)] +
?
kx
8
x? k2 + kz
z ? k2 ,
(9)
x.
Proof sketch of Proposition 1: The proof uses Lemma 1, the non-stationarity in (8) and a key
observation in the following Claim. The detailed proof is in Appendix D.2.
Claim 1. It holds that f (x _ x? ) + f (x ^ x? ) + f (z _ z ? ) + f (z ^ z ? )
f (x? ).
Note that [7; 20] propose a similar relation for the special cases of multilinear/softmax extensions by
mainly proving the same conclusion as in Claim 1. Their relation does not incorporate the properties
of non-stationarity or strong DR-submodularity. They both use the proof idea of constructing a
complicated auxiliary set function tailored to specific DR-submodular functions. We present a
different proof method by directly utilizing the DR property on carefully constructed auxiliary points
(e.g., (x + z) _ x? in the proof of Claim 1).
4
Algorithms for Constrained DR-submodular Maximization
Based on the properties, we present two algorithms for solving (P). The first is based on the localglobal relation, and the second is a F RANK -W OLFE variant adapted for the non-monotone setting.
All the omitted proofs are deferred to Appendix E.
4.1
An Algorithm Based on the Local-Global Relation
Algorithm 1: TWO - PHASE F RANK -W OLFE for non-monotone DR-submodular maximization
Input: maxx2P f (x), stopping tolerance ?1 , ?2 , #iterations K1 , K2
1 x
N ON - CONVEX F RANK -W OLFE(f, P, K1 , ?1 , x(0) ) ;
// x(0) 2 P
n
? x};
2 Q
P \ {y 2 R+ | y ? u
3 z
N ON - CONVEX F RANK -W OLFE(f, Q, K2 , ?2 , z (0) ) ;
// z (0) 2 Q
Output: arg max{f (x), f (z)} ;
We summarize the TWO - PHASE algorithm in Algorithm 1. It is generalized from the ?two-phase?
method in [7; 20]. It invokes some non-convex solver (we use the NON - CONVEX F RANK -W OLFE
by [27]; pseudocode is included in Algorithm 3 of Appendix C) to find approximately stationary
points in P and Q, respectively, then returns the solution with the larger function value. Though
we use N ON - CONVEX F RANK -W OLFE as the subroutine here, it is worth noting that any algorithm
that is guaranteed to find an approximately stationary point can be plugged into Algorithm 1 as
the subroutine. We give an improved approximation bound by considering more properties of
DR-submodular functions. Borrowing the results from [27] for the NON - CONVEX F RANK -W OLFE
subroutine, we get the following,
Theorem 1. The output of Algorithm 1 satisfies,
?
max{f (x), f (z)}
kx x? k2 + kz z ? k2
8
?
?
1
max{2h1 , Cf (P)}
?
p
+
f (x ) min
, ?1
4
K1 + 1
(10)
min
?
max{2h2 , Cf (Q)}
p
, ?2
K2 + 1
,
where h1 := maxx2P f (x) f (x(0) ), h2 := maxz2Q f (z) f (z (0) ) are the initial suboptimalities,
Cf (P) := supx,v2P, 2[0,1],y=x+ (v x) 22 (f (y) f (x) (y x)> rf (x)) is the curvature of f
w.r.t. P, and z ? = x _ x? x.
5
p
Theorem 1 indicates that Algorithm 1 has a 1/4 approximation guarantee and 1/ k rate. However,
it has good empirical performance as demonstrated by the experiments in Section 5. Informally, this
can be partially explained by the term ?8 kx x? k2 + kz z ? k2 in (10): if x is away from x? ,
this term will augment the bound; if x is close to x? , by the smoothness of f , it should be close to
optimal.
4.2
The Non-monotone F RANK -W OLFE Variant
Algorithm 2: Non-monotone F RANK -W OLFE variant for DR-submodular maximization
Input: maxx2P f (x), prespecified step size 2 (0, 1]
(0)
1 x
0, t(0)
0, k
0;
// k : iteration index, t(k) : cumulative step size
(k)
2 while t
< 1 do
3
v (k)
arg maxv2P,v?u? x(k) hv, rf (x(k) )i;
// shrunken LMO
(k)
4
use uniform step size k = ; set k
min{ k , 1 t };
5
x(k+1)
x(k) + k v (k) , t(k+1)
t(k) + k , k
k + 1;
Output: x(K) ;
// assuming there are K iterations in total
Algorithm 2 summarizes the non-monotone F RANK -W OLFE variant, which is inspired by the unified
continuous greedy algorithm in [13] for maximizing the multilinear extension of a submodular set
function. It initializes the solution x(0) to be 0, and maintains t(k) as the cumulative step size. At
iteration k, it maximizes the linearization of f over a ?shrunken? constraint set: {v|v 2 P, v ?
? x(k) }, which is different from the classical LMO of Frank-Wolfe-style algorithms (hence we
u
refer to it as the ?shrunken LMO?). Then it employs an update step in the direction v (k) chosen by
the LMO with a uniform step size k = . The cumulative step size t(k) is used to ensure that the
overall step sizes sum to one, thus the output solution x(K) is a convex combination of the LMO
outputs, hence also lies in P.
The shrunken LMO (Step 3) is the key difference compared to the monotone F RANK -W OLFE variant
? x(k) is added to prevent too aggressive growth of the solution,
in [4]. The extra constraint v ? u
since in the non-monotone setting such aggressive growth may hurt the overall performance. The
next theorem states the guarantees of Algorithm 2.
Theorem 2. Consider Algorithm 2 with uniform step size . For k = 1, ..., K it holds that,
f (x(k) )
t(k) e
t(k)
LD2
k
2
f (x? )
2
O(
2
)f (x? ).
(11)
By observing that t(K) = 1 and applying Theorem 2, we get the following Corollary:
Corollary 1. The output of Algorithm 2 satisfies f (x(K) )
e
1
f (x? )
LD 2
2K
O
1
K2
f (x? ).
Corollary 1 shows that Algorithm 2 enjoys a sublinear convergence rate towards some point x(K)
inside P, with a 1/e approximation guarantee.
Proof sketch of Theorem 2: The proof is by induction. To prepare the building blocks, we first of
all show that the growth of x(k) is indeed bounded,
Lemma 2. Assume x(0) = 0. For k = 0, ..., K
(k)
1, it holds xi
?u
?i [1
)t
(1
(k)
/
], 8i 2 [n].
Then the following Lemma provides a lower bound, which gets the global optimum involved,
? let 0 = mini2[n] u??ii . Then for all
Lemma 3 (Generalized from Lemma 7 in [8]). Given ? 2 (0, u],
1
?
?
x 2 [0, ?], it holds f (x _ x ) (1
0 )f (x ).
Then the key ingredient for induction is the relation between f (x(k+1) ) and f (x(k) ) indicated by:
Claim 2. For k = 0, ..., K
1 it holds f (x(k+1) )
(1
)f (x(k) )+ (1
)t
(k)
/
f (x? )
LD 2
2
which is derived by a combination of the quadratic lower bound in (5), Lemma 2 and Lemma 3.
6
2
,
Remarks on the two algorithms. Notice that though the TWO - PHASE algorithm has a worse
guarantee than the non-monotone F RANK -W OLFE variant, it is still of interest: i) It allows flexibility
in using a wide range of existing solvers for finding an (approximately) stationary point. ii) The
guarantees that we present rely on a worst-case analysis. The empirical performance of the TWO PHASE algorithm is often comparable or better than that of the F RANK -W OLFE variant. This suggests
to explore more properties in concrete problems that may favor the TWO - PHASE algorithm, which we
leave for future work.
5
Experimental Results
We test the performance of the analyzed algorithms, while considering the following baselines: 1)
QUADPROG IP [39], which is a global solver for non-convex quadratic programming; 2) Projected
1
gradient ascent (P ROJ G RAD) with diminishing step sizes ( k+1
, k starts from 0). We run all the
algorithms for 100 iterations. For the subroutine (Algorithm 3) of TWO - PHASE F RANK -W OLFE,
we set ?1 = ?2 = 10 6 , K1 = K2 = 100. All the synthetic results are the average of 20 repeated
experiments. All experiments were implemented using MATLAB. Source code can be found at:
https://github.com/bianan.
5.1
DR-submodular Quadratic Programming
As a state-of-the-art global solver, QUADPROG IP2 [39] can find the global optimum (possibly in
exponential time), which were used to calculate the approximation ratios. Our problem instances are
synthetic DR-submodular quadratic objectives with down-closed polytope constraints, i.e., f (x) =
m?n
1 >
>
n
m
?
2 x Hx + h x + c and P = {x 2 R+ | Ax ? b, x ? u, A 2 R++ , b 2 R+ }. Both objective
and constraints were randomly generated, in the following two manners:
1) Uniform distribution. H 2 Rn?n is a symmetric matrix with uniformly distributed entries in
[ 1, 0]; A 2 Rm?n has uniformly distributed entries in [?, ? + 1], where ? = 0.01 is a small positive
constant in order to make entries of A strictly positive.
0.95
0.9
1
1
0.95
0.95
Approx. ratio
Approx. ratio
Approx. ratio
1
0.9
0.9
0.85
0.85
8
10
12
14
Dimensionality
(a) m = b0.5nc
16
8
10
12
Dimensionality
(b) m = n
14
16
8
10
12
14
16
Dimensionality
(c) m = b1.5nc
Figure 1: Results on DR-submodular quadratic instances with uniform distribution.
2) Exponential distribution. The entries of H and A were sampled from exponential distributions
Exp( ) (For a random variable y 0, its probability density function is e y , and for y < 0, its
density is 0). Specifically, each entry of H was sampled from Exp(1), then the matrix H was
made to be symmetric. Each entry of A was sampled from Exp(0.25) + ?, where ? = 0.01 is a small
positive constant.
? to be the tightest upper bound of P by
In both the above two cases, we set b = 1m , and u
i
? To
u
?j = mini2[m] Abij
, 8j 2 [n]. In order to make f non-monotone, we set h = 0.2 ? H> u.
1 >
>
make sure that f is non-negative, we first of all solve the problem minx2P 2 x Hx + h x using
? then set c = f (x)
? + 0.1 ? |f (x)|.
?
QUADPROG IP, let the solution to be x,
The approximation ratios w.r.t. dimensionalities (n) are plotted in Figures 1 and 2, for the two
manners of data generation. We set the number of constraints to be m = b0.5nc, m = n and
m = b1.5nc in Figures 1a to 1c (and Figures 2a to 2c), respectively.
2
We used the open source code provided by [39], and the IBM CPLEX optimization studio https://www.
ibm.com/jm-en/marketplace/ibm-ilog-cplex as the subroutine.
7
1
0.95
0.9
0.85
0.8
Approx. ratio
1
0.95
Approx. ratio
Approx. ratio
1
0.95
0.9
0.85
0.9
0.85
0.8
0.8
0.75
0.75
8
10
12
14
16
0.75
8
Dimensionality
8
10
12
14
(a) m = b0.5nc
10
12
14
16
Dimensionality
16
Dimensionality
(b) m = n
(c) m = b1.5nc
Figure 2: Results on quadratic instances with exponential distribution.
One can see that TWO - PHASE F RANK -W OLFE usually performs the best, P ROJ G RAD follows, and
non-monotone F RANK -W OLFE variant is the last. The good performance of TWO - PHASE F RANK W OLFE can be partially explained by the strong DR-submodularity of quadratic functions according
to Theorem 1. Performance of the two analyzed algorithms is consistent with the theoretical bounds:
the approximation ratios of F RANK -W OLFE variant are always much higher than 1/e.
5.2
Maximizing Softmax Extensions
0.2
0.2
0.15
0.15
0.1
0.05
0
-0.05
8
10
12
14
Dimensionality
(a) m = b0.5nc
16
Function value
0.2
0.15
Function value
Function value
With some derivation, one can see the derivative of the softmax extension in (3) is: ri f (x) =
tr((diag(x)(L I) + I) 1 (L I)i ), 8i 2 [n], where (L I)i denotes the matrix obtained by
zeroing all entries except the ith row of (L I). Let C := (diag(x)(L I) + I) 1 , D := (L I),
one can see that ri f (x) = D>
i? C?i , which gives an efficient way to calculate the gradient rf (x).
0.1
0.05
0
-0.05
0.1
0.05
0
-0.05
8
10
12
Dimensionality
(b) m = n
14
16
8
10
12
14
16
Dimensionality
(c) m = b1.5nc
Figure 3: Results on softmax instances with polytope constraints generated from uniform distribution.
Results on synthetic data. We generate the softmax objectives (see (3)) in the following way: first
generate the n eigenvalues d 2 Rn+ , each randomly distributed in [0, 1.5], and set D = diag(d).
After generating a random unitary matrix U, we set L = UDU> . One can verify that L is positive
semidefinite and has eigenvalues as the entries of d.
We generate the down-closed polytope constraints in the same form and same way as that for DRsubmodular quadratic functions, except for setting b = 2 ? 1m . Function values returned by different
solvers w.r.t. n are shown in Figure 3, for which the random polytope constraints were generated
with uniform distribution (results for which the random polytope constraints were generated with
exponential distribution are deferred to Appendix G). The number of constraints was set to be
m = b0.5nc, m = n and m = b1.5nc in Figures 3a to 3c, respectively. One can observe that
TWO - PHASE F RANK -W OLFE still has the best performance, the non-monotone F RANK -W OLFE
variant follows, and P ROJ G RAD has the worst performance.
Real-world results on matched summarization. The task of ?matched summarization? is to select
a set of document pairs out of a corpus of documents, such that the two documents within a pair
are similar, and the overall set of pairs is as diverse as possible. The motivation for this task is very
practical: it could be, for example, to compare the opinions of various politicians on a range of
representative topics.
In our experiments, we used a similar setting to the one in [20]. We experimented on the 2012
US Republican debates data, which consists of 8 candidates: Bachman, Gingrich, Huntsman, Paul,
Perry, Romney and Santorum. Each task involves one pair of candidates, so in total there are
8
6
Function value
Function value
28 = 7 ? 8/2 tasks. Figure 4a plots the averaged function values returned by the three solvers over 28
tasks, w.r.t. different values of a hyperparameter reflecting the matching quality (details see [20]).
Figure 4b traces the objectives
w.r.t.
iterations for a spe2.5
cific
candidate
pair (Bachman,
10
2
Romney). For TWO - PHASE
8
F RANK -W OLFE, the objectives
1.5
6
of
the selected phase were plot1
4
ted.
One can see that TWO 0.5
2
PHASE F RANK -W OLFE also
0
0.2
0.4
0.6
0.8
1
0
20
40
60
80
100
achieves the best performance,
Match quality controller
Iteration
while
the performance of non(a) Average on 28 tasks
(b) Objectives w.r.t. iterations
monotone F RANK -W OLFE variant and P ROJ G RAD is comparaFigure 4: Results on 2012 US Republican debates data.
ble.
Related Work
Submodular optimization and, more broadly, non-convex optimization are extensively studied in the
literature, which renders it very difficult comprehensively surveying all previous work. Here we only
briefly summarize some of the most related papers.
Submodular optimization over integer-lattice and continuous domains. Many results from submodular set function optimization have been generalized to the integer-lattice case [34; 33; 12; 24].
Of particular interest is the reduction [12] from an integer-lattice DR-submodular maximization problem to a submodular set function maximization problem. Submodular optimization over continuous
domains has attracted considerable attention recently [3; 4; 36]. Two classes of functions that are
covered by continuous submodularity are the Lovasz extensions [28] and multilinear extensions
[6] of submodular set functions. Particularly, multilinear extensions of submodular set functions
are also continuous DR-submodular [3], but with the special property that they are coordinate-wise
linear. Combined with the rounding technique of contention resolution [7], maximizing multilinear
extensions [38; 19; 13; 8; 11] has become the state-of-the-art method for submodular set function
maximization. Some of the techniques in maximizing multilinear extensions [13; 7; 8] have inspired
this work. However, we are the first to explore the rich properties and devise algorithms for the
general constrained DR-submodular maximization problem over continuous domains.
Non-convex optimization. Non-convex optimization receives a surge of attention in the past years.
One active research topic is to reach a stationary point for unconstrained optimization [35; 32; 1] or
constrained optimization [18; 27]. However, without proper assumptions, a stationary point may not
lead to any global approximation guarantee. The local-global relation (in Proposition 1) provides a
strong relation between (approximately) stationary points and global optimum, thus making it flexible
to incorporate progress in this area.
7
Conclusion
We have studied the problem of constrained non-monotone DR-submodular continuous maximization.
We explored the structural properties of such problems, and established a local-global relation. Based
on these properties, we presented a TWO - PHASE algorithm with a 1/4 approximation guarantee, and a
non-monotone F RANK -W OLFE variant with a 1/e approximation guarantee. We further generalized
submodular continuous function over conic lattices, which enabled us to model a larger class of
applications. Lastly, our theoretical findings were verified by synthetic and real-world experiments.
Acknowledgement. This research was partially supported by ERC StG 307036, by the Max Planck
ETH Center for Learning Systems, and by the ETH Z?rich Postdoctoral Fellowship program.
9
References
[1] Allen-Zhu, Zeyuan and Hazan, Elad. Variance reduction for faster non-convex optimization. In
International Conference on Machine Learning (ICML), pp. 699?707, 2016.
[2] Antoniadis, Anestis, Gijbels, Ir?ne, and Nikolova, Mila. Penalized likelihood regression for
generalized linear models with non-quadratic penalties. Annals of the Institute of Statistical
Mathematics, 63(3):585?615, 2011.
[3] Bach, Francis. Submodular functions: from discrete to continous domains. arXiv preprint
arXiv:1511.00394, 2015.
[4] Bian, Andrew An, Mirzasoleiman, Baharan, Buhmann, Joachim M., and Krause, Andreas.
Guaranteed non-convex optimization: Submodular maximization over continuous domains. In
International Conference on Artificial Intelligence and Statistics (AISTATS), pp. 111?120, 2017.
[5] Boyd, Stephen and Vandenberghe, Lieven. Convex optimization. Cambridge university press,
2004.
[6] Calinescu, Gruia, Chekuri, Chandra, P?l, Martin, and Vondr?k, Jan. Maximizing a submodular set function subject to a matroid constraint. In Integer programming and combinatorial
optimization, pp. 182?196. Springer, 2007.
[7] Chekuri, Chandra, Vondr?k, Jan, and Zenklusen, Rico. Submodular function maximization via
the multilinear relaxation and contention resolution schemes. SIAM Journal on Computing, 43
(6):1831?1879, 2014.
[8] Chekuri, Chandra, Jayram, TS, and Vondr?k, Jan. On multiplicative weight updates for concave
and submodular function maximization. In Proceedings of the 2015 Conference on Innovations
in Theoretical Computer Science, pp. 201?210. ACM, 2015.
[9] Djolonga, Josip and Krause, Andreas. From map to marginals: Variational inference in bayesian
submodular models. In Neural Information Processing Systems (NIPS), pp. 244?252, 2014.
[10] Eghbali, Reza and Fazel, Maryam. Designing smoothing functions for improved worst-case
competitive ratio in online optimization. In Advances in Neural Information Processing Systems
(NIPS), pp. 3279?3287. 2016.
[11] Ene, Alina and Nguyen, Huy L. Constrained submodular maximization: Beyond 1/e. In
Foundations of Computer Science (FOCS), 2016 IEEE 57th Annual Symposium on, pp. 248?257,
2016.
[12] Ene, Alina and Nguyen, Huy L. A reduction for optimizing lattice submodular functions with
diminishing returns. arXiv preprint arXiv:1606.08362, 2016.
[13] Feldman, Moran, Naor, Joseph, and Schwartz, Roy. A unified continuous greedy algorithm
for submodular maximization. In Foundations of Computer Science (FOCS), 2011 IEEE 52nd
Annual Symposium on, pp. 570?579. IEEE, 2011.
[14] Friedland, S and Gaubert, S. Submodular spectral functions of principal submatrices of a
hermitian matrix, extensions and applications. Linear Algebra and its Applications, 438(10):
3872?3884, 2013.
[15] Fuchssteiner, Benno and Lusky, Wolfgang. Convex cones, volume 56. Elsevier, 2011.
[16] Fujishige, Satoru. Submodular functions and optimization, volume 58. Elsevier, 2005.
[17] Garg, Vijay K. Introduction to lattice theory with computer science applications. John Wiley &
Sons, 2015.
[18] Ghadimi, Saeed, Lan, Guanghui, and Zhang, Hongchao. Mini-batch stochastic approximation
methods for nonconvex stochastic composite optimization. Mathematical Programming, 155
(1-2):267?305, 2016.
[19] Gharan, Shayan Oveis and Vondr?k, Jan. Submodular maximization by simulated annealing. In
Proceedings of the twenty-second annual ACM-SIAM symposium on Discrete Algorithms, pp.
1098?1116. Society for Industrial and Applied Mathematics, 2011.
[20] Gillenwater, Jennifer, Kulesza, Alex, and Taskar, Ben. Near-optimal map inference for determinantal point processes. In Advances in Neural Information Processing Systems (NIPS), pp.
2735?2743, 2012.
10
[21] Hassani, Hamed, Soltanolkotabi, Mahdi, and Karbasi, Amin. Gradient methods for submodular
maximization. In Advances in Neural Information Processing Systems (NIPS), pp. 5837?5847,
2017.
[22] Ito, Shinji and Fujimaki, Ryohei. Large-scale price optimization via network flow. In Advances
in Neural Information Processing Systems (NIPS), pp. 3855?3863, 2016.
[23] Iwata, Satoru, Fleischer, Lisa, and Fujishige, Satoru. A combinatorial strongly polynomial
algorithm for minimizing submodular functions. Journal of the ACM, 48(4):761?777, 2001.
[24] Khodabakhsh, Ali and Nikolova, Evdokia. Maximizing non-monotone dr-submodular functions
with cardinality constraints. arXiv preprint arXiv:1611.09474, 2016.
[25] Krause, Andreas and Golovin, Daniel. Submodular function maximization. Tractability:
Practical Approaches to Hard Problems, 3:19, 2012.
[26] Kulesza, Alex, Taskar, Ben, et al. Determinantal point processes for machine learning. Foundations and Trends R in Machine Learning, 5(2?3):123?286, 2012.
[27] Lacoste-Julien, Simon. Convergence rate of frank-wolfe for non-convex objectives. arXiv
preprint arXiv:1607.00345, 2016.
[28] Lov?sz, L?szl?. Submodular functions and convexity. In Mathematical Programming The State
of the Art, pp. 235?257. Springer, 1983.
[29] Montanari, Andrea and Richard, Emile. Non-negative principal component analysis: Message
passing algorithms and sharp asymptotics. IEEE Transactions on Information Theory, 62(3):
1458?1484, 2016.
[30] Motzkin, Theodore S and Straus, Ernst G. Maxima for graphs and a new proof of a theorem of
tur?n. Canad. J. Math, 17(4):533?540, 1965.
[31] Nemhauser, George L, Wolsey, Laurence A, and Fisher, Marshall L. An analysis of approximations for maximizing submodular set functions ? i. Mathematical Programming, 14(1):265?294,
1978.
[32] Reddi, Sashank J., Sra, Suvrit, Poczos, Barnabas, and Smola, Alexander J. Proximal stochastic
methods for nonsmooth nonconvex finite-sum optimization. In Advances in Neural Information
Processing Systems (NIPS), pp. 1145?1153. 2016.
[33] Soma, Tasuku and Yoshida, Yuichi. A generalization of submodular cover via the diminishing
return property on the integer lattice. In Advances in Neural Information Processing Systems
(NIPS), pp. 847?855, 2015.
[34] Soma, Tasuku, Kakimura, Naonori, Inaba, Kazuhiro, and Kawarabayashi, Ken-ichi. Optimal
budget allocation: Theoretical guarantee and efficient algorithm. In International Conference
on Machine Learning (ICML), pp. 351?359, 2014.
[35] Sra, Suvrit. Scalable nonconvex inexact proximal splitting. In Advances in Neural Information
Processing Systems (NIPS), pp. 530?538, 2012.
[36] Staib, Matthew and Jegelka, Stefanie. Robust budget allocation via continuous submodular
functions. In International Conference on Machine Learning (ICML), pp. 3230?3240, 2017.
[37] Topkis, Donald M. Minimizing a submodular function on a lattice. Operations research, 26(2):
305?321, 1978.
[38] Vondr?k, Jan. Optimal approximation for the submodular welfare problem in the value oracle
model. In Proceedings of the 40th Annual ACM Symposium on Theory of Computing, pp. 67?74,
2008.
[39] Xia, Wei, Vera, Juan, and Zuluaga, Luis F. Globally solving non-convex quadratic programs via
linear integer programming techniques. arXiv preprint arXiv:1511.02423, 2015.
[40] Zass, Ron and Shashua, Amnon. Nonnegative sparse pca. Advances in Neural Information
Processing Systems (NIPS), pp. 1561?1568, 2007.
11
| 6652 |@word mild:1 shayan:1 briefly:1 polynomial:5 norm:1 stronger:2 nd:1 laurence:1 open:1 bachman:2 x2p:1 tr:1 ld:2 reduction:4 initial:1 zij:2 daniel:1 document:3 past:1 existing:2 current:1 com:2 yet:3 attracted:1 luis:1 determinantal:4 john:1 partition:1 enables:6 plot:1 update:2 stationary:15 greedy:2 selected:1 intelligence:1 antoniadis:1 ith:3 prespecified:1 provides:2 math:1 ron:1 zhang:1 mathematical:3 along:6 kvk2:2 constructed:1 become:1 symposium:4 ryohei:1 focs:2 consists:1 naor:1 inside:1 hermitian:1 manner:2 lov:1 indeed:1 sublinearly:1 andrea:1 surge:1 inspired:2 globally:1 calculated:1 jm:1 solver:7 considering:2 cardinality:1 provided:1 vera:1 notation:1 underlying:1 maximizes:1 factorized:1 bounded:1 matched:2 surveying:1 unified:2 finding:5 guarantee:20 subclass:2 concave:5 growth:3 k2:11 rm:1 schwartz:1 underlie:1 planck:1 positive:8 local:8 despite:1 analyzing:1 approximately:8 twice:1 garg:1 studied:3 theodore:1 suggests:2 range:2 averaged:1 fazel:1 practical:4 yj:1 yehuda:1 practice:2 block:1 jan:5 antitone:1 asymptotics:1 area:1 empirical:2 maxx:1 eth:6 submatrices:1 matching:2 boyd:1 composite:1 donald:1 get:4 close:2 satoru:3 applying:2 optimize:1 equivalent:3 map:5 demonstrated:1 www:1 maximizing:12 center:1 ghadimi:1 attention:2 yoshida:1 convex:32 resolution:2 splitting:1 utilizing:1 deriving:1 vandenberghe:1 enabled:1 proving:1 notion:3 coordinate:3 hurt:1 annals:1 play:1 exact:2 programming:7 oveis:1 us:1 quadprog:3 designing:1 element:1 wolfe:3 expensive:1 particularly:1 roy:1 trend:1 inaba:1 role:1 taskar:2 preprint:5 solved:1 capture:2 worst:4 hv:1 calculate:2 tur:1 yk:2 convexity:1 barnabas:1 motivate:1 solving:2 algebra:1 ali:1 y2p:1 basis:1 easily:1 various:1 regularizer:3 derivation:1 artificial:1 marketplace:1 harnessing:1 maxx2p:4 larger:3 solve:2 elad:1 supermodularity:1 otherwise:2 favor:1 statistic:2 gp:2 ip:2 online:1 advantage:1 differentiable:2 eigenvalue:2 propose:1 maryam:1 gq:6 shrunken:4 iff:2 flexibility:2 ernst:1 amin:1 convergence:3 optimum:7 generating:1 mirzasoleiman:1 converges:2 ben:2 i2s:2 leave:1 wider:1 derive:1 develop:2 andrew:1 measured:1 ij:1 b0:5 progress:2 strong:6 recovering:1 auxiliary:2 involves:1 implies:2 implemented:1 direction:7 submodularity:13 stochastic:3 opinion:1 require:1 hx:2 generalization:3 proposition:3 tighter:1 multilinear:9 extension:18 exploring:1 strictly:1 hold:14 ground:1 exp:6 welfare:1 mapping:1 claim:5 matthew:1 achieves:1 omitted:2 combinatorial:3 currently:1 label:1 prepare:1 minimization:4 lovasz:1 always:2 eghbali:1 pn:2 broader:1 gharan:1 corollary:3 validated:1 derived:1 ax:1 joachim:2 rank:28 indicates:1 likelihood:1 mainly:1 industrial:1 stg:1 baseline:1 romney:2 posteriori:1 inference:8 elsevier:2 repulsion:1 stopping:1 typically:1 diminishing:5 borrowing:1 relation:17 subroutine:6 arg:2 overall:3 flexible:1 augment:1 constrained:15 softmax:10 special:2 art:3 smoothing:1 field:3 beach:1 sampling:1 ted:1 icml:3 future:1 djolonga:1 others:1 np:1 nonsmooth:1 roj:4 employ:1 richard:1 randomly:2 divergence:2 kakimura:1 phase:16 cplex:2 saeed:1 stationarity:12 interest:2 message:1 investigate:1 fujimaki:1 deferred:2 szl:1 analyzed:2 semidefinite:2 regularizers:2 kfir:1 naonori:1 xy:1 lh:1 euclidean:1 plugged:1 plotted:1 theoretical:6 josip:1 politician:1 instance:8 marshall:1 cover:1 maximization:26 lattice:13 applicability:1 tractability:1 subset:1 entry:11 uniform:7 rounding:1 too:1 closedness:1 motivating:2 supx:1 proximal:2 synthetic:6 combined:2 guanghui:1 st:1 density:3 international:4 siam:2 tasuku:2 probabilistic:3 invoke:1 concrete:1 satisfied:1 possibly:1 classically:1 juan:1 dr:64 worse:2 derivative:1 style:3 return:5 combing:1 aggressive:2 zenklusen:1 diversity:1 notable:1 vi:2 multiplicative:1 h1:2 closed:7 wolfgang:1 observing:1 hazan:1 francis:1 start:2 competitive:1 maintains:1 complicated:1 shashua:1 simon:1 contribution:1 ir:1 variance:1 yield:1 bayesian:1 worth:1 hamed:1 reach:1 definition:7 inexact:1 pp:20 involved:1 straus:1 naturally:1 proof:11 sampled:3 proved:1 popular:1 kawarabayashi:1 dimensionality:10 hassani:1 carefully:1 reflecting:1 appears:1 rico:1 higher:1 supermodular:4 bian:2 improved:2 wei:1 box:3 strongly:6 though:3 furthermore:1 just:1 implicit:1 ld2:1 lastly:1 chekuri:3 smola:1 sketch:2 receives:1 ei:1 perry:1 defines:1 logistic:3 quality:2 indicated:1 usa:1 building:1 requiring:1 verify:1 mini2:2 regularization:1 hence:2 symmetric:2 leibler:2 staib:1 generalized:9 demonstrate:1 performs:1 allen:1 fj:3 wise:3 variational:1 contention:2 recently:2 pseudocode:1 qp:1 reza:1 volume:2 extend:1 x2r:1 lieven:1 marginals:1 significant:1 refer:1 cambridge:1 feldman:1 dpps:4 smoothness:1 unconstrained:2 approx:6 pm:1 mathematics:2 zeroing:1 erc:1 soltanolkotabi:1 submodular:90 gillenwater:1 yk2:2 lkx:1 etc:1 curvature:1 recent:2 perspective:1 optimizing:1 inf:3 scenario:1 certain:1 nonconvex:3 suvrit:2 binary:1 yi:1 devise:3 exploited:1 topkis:1 captured:2 minimum:2 additional:1 george:1 zeyuan:1 ii:3 stephen:1 gruia:1 technical:1 match:1 adapt:1 faster:1 bach:1 long:1 zass:1 variant:15 underlies:1 regression:3 controller:1 scalable:1 expectation:1 chandra:3 arxiv:10 iteration:8 represent:2 kernel:1 tailored:1 achieved:1 preserved:1 gingrich:1 fellowship:1 krause:4 interval:1 annealing:1 source:2 suboptimalities:1 crucial:1 extra:1 ascent:1 sure:1 subject:1 fujishige:2 flow:1 integer:9 reddi:1 unitary:1 structural:1 noting:1 near:1 easy:1 nikolova:2 xj:2 matroid:2 andreas:4 idea:2 det:1 fleischer:1 amnon:1 expression:1 pca:2 cific:1 penalty:1 render:1 returned:2 sashank:1 poczos:1 hessian:1 passing:1 remark:1 matlab:1 covered:2 detailed:1 informally:1 amount:1 locally:1 extensively:1 ken:1 diameter:1 http:2 generate:3 notice:2 sign:3 arising:1 diverse:1 broadly:1 discrete:2 hyperparameter:1 ichi:1 key:3 soma:2 lan:1 olfe:28 capital:1 alina:2 prevent:1 krf:2 kuk:1 verified:1 lacoste:1 graph:1 relaxation:1 monotone:25 sum:2 year:1 gijbels:1 realworld:1 run:1 letter:2 cone:1 arrive:1 ble:1 appendix:11 scaling:1 summarizes:1 comparable:1 bound:13 guaranteed:4 quadratic:13 annual:4 oracle:1 nonnegative:1 adapted:1 constraint:16 alex:2 ri:2 min:4 separable:4 martin:1 according:1 ip2:1 combination:2 slightly:1 son:1 appealing:1 joseph:1 making:1 happens:1 explained:2 ene:2 karbasi:1 zurich:4 jennifer:1 turn:2 discus:1 generalizes:1 tightest:1 operation:1 apply:1 observe:1 away:1 spectral:1 batch:1 original:1 denotes:1 include:1 cf:3 ensure:1 invokes:1 k1:4 establish:1 classical:2 society:1 objective:17 initializes:1 already:1 added:1 canad:1 diagonal:2 exhibit:1 amongst:1 gradient:8 calinescu:1 distance:1 friedland:1 nemhauser:1 simulated:1 topic:2 polytope:7 trivial:1 spanning:1 provable:2 boldface:2 induction:2 assuming:1 code:2 minn:1 index:1 mini:1 ratio:10 minimizing:4 innovation:1 nc:10 setup:1 difficult:1 statement:2 frank:3 debate:2 trace:1 negative:6 rise:1 motivates:1 proper:2 summarization:2 twenty:1 upper:1 observation:1 ilog:1 finite:1 t:1 looking:1 rn:9 arbitrary:1 sharp:1 pair:5 kl:3 z1:1 continous:1 rad:4 established:1 nip:10 discontinuity:1 beyond:1 below:1 usually:3 xm:1 jayram:1 kulesza:2 summarize:2 program:2 rf:9 max:7 baharan:1 natural:1 rely:1 buhmann:2 zhu:1 scheme:2 github:1 x2i:3 brief:1 republican:2 ne:1 julien:1 conic:4 stefanie:1 geometric:2 literature:1 acknowledgement:1 shinji:1 loss:1 sublinear:2 interesting:1 generation:1 wolsey:1 allocation:2 ingredient:1 emile:1 h2:2 foundation:3 krausea:1 jegelka:1 consistent:1 jbuhmann:1 ibm:3 row:1 summary:1 penalized:1 supported:1 last:1 free:1 enjoys:1 aij:1 lisa:1 institute:1 wide:2 comprehensively:1 characterizing:1 sparse:1 tolerance:1 distributed:3 xia:1 dpp:2 default:1 xn:1 world:6 evaluating:1 rich:3 qn:1 dimension:1 concretely:1 concavity:3 coincide:1 kz:3 cumulative:3 kei:4 projected:1 made:1 nguyen:2 qx:3 transaction:1 approximate:3 vondr:5 kullback:2 sz:1 global:19 active:1 investigating:1 b1:5 corpus:1 assumed:1 xi:11 spectrum:1 postdoctoral:1 continuous:29 yuichi:1 nature:1 robust:1 ca:1 golovin:1 sra:2 constructing:1 domain:9 diag:5 aistats:1 montanari:1 motivation:1 paul:1 huy:2 repeated:1 representative:2 mila:1 en:1 wiley:1 precision:1 lmo:6 exponential:5 lie:2 candidate:3 hrf:2 mahdi:1 levy:2 ito:1 down:7 theorem:8 specific:1 gaubert:1 udu:1 moran:1 list:1 experimented:1 explored:1 linearization:1 budget:2 studio:1 kx:6 vijay:1 explore:2 motzkin:1 partially:3 springer:2 ch:4 corresponds:1 iwata:1 satisfies:2 acm:4 identity:1 towards:1 lipschitz:5 price:1 considerable:1 experimentally:1 hard:3 included:1 specifically:2 except:2 uniformly:2 fisher:1 lemma:10 principal:2 called:2 total:2 experimental:1 select:1 latter:1 alexander:1 ethz:4 incorporate:2 evaluate:1 |
6,249 | 6,653 | Learning with Average Top-k Loss
Yanbo Fan3,4,1 , Siwei Lyu1?, Yiming Ying2 , Bao-Gang Hu3,4
1
Department of Computer Science, University at Albany, SUNY
2
Department of Mathematics and Statistics, University at Albany, SUNY
3
National Laboratory of Pattern Recognition, CASIA
4
University of Chinese Academy of Sciences (UCAS)
{yanbo.fan,hubg}@nlpr.ia.ac.cn, [email protected], [email protected]
Abstract
In this work, we introduce the average top-k (ATk ) loss as a new aggregate loss
for supervised learning, which is the average over the k largest individual losses
over a training dataset. We show that the ATk loss is a natural generalization of the
two widely used aggregate losses, namely the average loss and the maximum loss,
but can combine their advantages and mitigate their drawbacks to better adapt to
different data distributions. Furthermore, it remains a convex function over all
individual losses, which can lead to convex optimization problems that can be
solved effectively with conventional gradient-based methods. We provide an intuitive interpretation of the ATk loss based on its equivalent effect on the continuous
individual loss functions, suggesting that it can reduce the penalty on correctly
classified data. We further give a learning theory analysis of MATk learning on
the classification calibration of the ATk loss and the error bounds of ATk -SVM.
We demonstrate the applicability of minimum average top-k learning for binary
classification and regression using synthetic and real datasets.
1
Introduction
Supervised learning concerns the inference of a function f : X 7? Y that predicts a target y ? Y
from data/features x ? X using a set of labeled training examples {(xi , yi )}ni=1 . This is typically
achieved by seeking a function f that minimizes an aggregate loss formed from individual losses
evaluated over all training samples.
To be more specific, the individual loss for a sample (x, y) is given by `(f (x), y), in which ` is
a nonnegative bivariate function that evaluates the quality of the prediction made by function f .
For example, for binary classification (i.e., yi ? {?1}), commonly used forms for individual loss
include the 0-1 loss, Iyf (x)?0 , which is 1 when y and f (x) have different sign and 0 otherwise, the
hinge loss, max(0, 1 ? yf (x)), and the logistic loss, log2 (1 + exp(?yf (x))), all of which can be
further simplified as the so-called margin loss, i.e., `(y, f (x)) = `(yf (x)). For regression, squared
difference (y ? f (x))2 and absolute difference |y ? f (x)| are two most popular forms for individual
loss, which can be simplified as `(y, f (x)) = `(|y ? f (x)|). Usually the individual loss is chosen
to be a convex function of its input, but recent works also propose various types of non-convex
individual losses (e.g., [10, 15, 27, 28]).
The supervised learning problem is then formulated as minf {L(Lz (f )) + ?(f )}, where L(Lz (f ))
is the aggregate loss accumulates all individual losses over training samples, i.e., Lz (f ) =
{`i (f )}ni=1 , with `i (f ) being the shorthand notation for `(f (xi ), yi ), and ?(f ) is the regularizer
on f . However, in contrast to the plethora of the types of individual losses, there are only a few
choices when we consider the aggregate loss:
?
Corresponding author.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
Classification Boundary
3
Misclassification Rate
0.1
3
Misclassification Rate
0.03
0.025
0.08
2
Classification Boundary
2
0.02
0.06
1
0.015
1
0.04
0.01
0
0
0.02
0
-1
-1
0
1
2
3
1
10
Classification Boundary
3
100 200
k
Misclassification Rate
0.025
3
0.02
2
-1
-1
0.005
0
0
1
2
1
3
Classification Boundary
10
k
100 200
Misclassification Rate
0.025
0.02
2
0.015
0.015
1
1
0.01
0.01
0
0
0.005
0
-1
-1
0
1
2
3
1
10
k
100 200
-1
-1
0.005
0
0
1
2
3
1
10
k
100 200
Figure 1: Comparison of different aggregate losses on 2D synthetic datasets with n = 200 samples for binary
classification on a balanced but multi-modal dataset and with outliers (top) and an imbalanced dataset with
outliers (bottom) with logistic loss (left) and hinge loss (right). Outliers in data are shown as an enlarged ?
and the optimal Bayes classifications are shown as shaded areas. The figures in the second and fourth columns
show the misclassification rate of ATk vs. k for each case.
Pn
? the average loss: Lavg (Lz (f )) = n1 i=1 `i (f ), i.e., the mean of all individual losses;
? the maximum loss: Lmax (Lz (f )) = max1?k?n `i (f ), i.e., the largest individual loss;
? the top-k loss [20]: Ltop-k (Lz (f )) = `[k] (f )2 for 1 ? k ? n, i.e., the k-th largest (top-k)
individual loss.
The average loss is unarguably the most widely used aggregate loss, as it is a unbiased approximation
to the expected risk and leads to the empirical risk minimization in learning theory [1, 7, 22, 25, 26].
Further, minimizing the average loss affords simple and efficient stochastic gradient descent algorithms [3, 21]. On the other hand, the work in [20] shows that constructing learning objective based
on the maximum loss may lead to improved performance for data with separate typical and rare subpopulations. The top-k loss [20] generalizes the maximum loss, as Lmax (Lz (f )) = Ltop-1 (Lz (f )),
and can alleviate the sensitivity to outliers of the latter. However, unlike the average loss or the
maximum loss, the top-k loss in general does not lead to a convex learning objective, as it is not
convex of all the individual losses Lz (f ).
In this work, we propose a new type of aggregate loss that we term as the average top-k (ATk ) loss,
which is the average of the largest k individual losses, that is defined as:
Pk
Lavt-k (Lz (f )) = k1 i=1 `[i] (f ).
(1)
We refer to learning objectives based on minimizing the ATk loss as MATk learning.
The ATk loss generalizes the average loss (k = n) and the maximum loss (k = 1), yet it is less
susceptible to their corresponding drawbacks, i.e., it is less sensitive to outliers than the maximum
loss and can adapt to imbalanced and/or multi-modal data distributions better than the average loss.
This is illustrated with two toy examples of synthesized 2D data for binary classification in Fig.1
(see supplementary materials for a complete illustration). As these plots show, the linear classifier
obtained with the maximum loss is not optimal due to the existence of outliers while the linear classifier corresponding to the average loss has to accommodate the requirement to minimize individual
losses across all training data, and sacrifices smaller sub-clusters of data (e.g., the rare population of
+ class in the top row and the smaller dataset of ? class in the bottom row). In contrast, using ATk
loss with k = 10 can better protect such smaller sub-clusters and leads to linear classifiers closer
to the optimal Bayesian linear classifier. This is also corroborated by the plots of corresponding
misclassification rate of ATk vs. k value in Fig.1, which show that minimum misclassification rates
occur at k value other than 1 (maximum loss) or n (average loss).
The ATk loss is a tight upper-bound of the top-k loss, as Lavt-k (Lz (f )) ? Ltop-k (Lz (f )) with
equality holds when k = 1 or `i (f ) = constant, and it is a convex function of the individual
losses (see Section 2). Indeed, we can express `[k] (f ) as the difference of two convex functions
kLavt-k (Lz (f ))?(k ?1)Lavt-(k?1) (Lz (f )), which shows that in general Ltop-k (Lz (f )) is not convex
with regards to the individual losses.
2
We define the top-k element of a set S = {s1 , ? ? ? , sn } as s[k] , such that s[1] ? s[2] ? ? ? ? ? s[n] .
2
In sequel, we will provide a detailed analysis of the ATk loss and MATk learning. First, we establish
a reformulation of the ATk loss as the minimum of the average of the individual losses over all training examples transformed by a hinge function. This reformulation leads to a simple and effective
stochastic gradient-based algorithm for MATk learning, and interprets the effect of the ATk loss as
shifting down and truncating at zero the individual loss to reduce the undesirable penalty on correctly classified data. When combined with the hinge function as individual loss, the ATk aggregate loss
leads to a new variant of SVM algorithm that we term as ATk SVM, which generalizes the C-SVM
and the ?-SVM algorithms [19]. We further study learning theory of MATk learning, focusing on
the classification calibration of the ATk loss function and error bounds of the ATk SVM algorithm.
This provides a theoretical lower-bound for k for reliable classification performance. We demonstrate the applicability of minimum average top-k learning for binary classification and regression
using synthetic and real datasets.
The main contributions of this work can be summarized as follows.
? We introduce the ATk loss for supervised learning, which can balance the pros and cons
of the average and maximum losses, and allows the learning algorithm to better adapt to
imbalanced and multi-modal data distributions.
? We provide algorithm and interpretation of the ATk loss, suggesting that most existing
learning algorithms can take advantage of it without significant increase in computation.
? We further study the theoretical aspects of ATk loss on classification calibration and error
bounds of minimum average top-k learning for ATk -SVM.
? We perform extensive experiments to validate the effectiveness of the MATk learning.
2
Formulation and Interpretation
The original ATk loss, though intuitive, is not convenient to work with because of the sorting procedure involved. This also obscures its connection with the statistical view of supervised learning as
minimizing the expectation of individual loss with regards to the underlying data distribution. Yet,
it affords an equivalent form, which is based on the following result.
Pk
Lemma 1 (Lemma 1, [16]).
of (x1 , ? ? ? , xn ). Furthermore, for
i=1 x[i] is a convex function
Pk
Pn
xi ? 0 and i = 1, ? ? ? , n, we have i=1 x[i] = min??0 k? + i=1 [xi ? ?]+ , where [a]+ =
max{0, a} is the hinge function.
For completeness, we include a proof of Lemma 1 in supplementary materials. Using Lemma 1, we
can reformulate the ATk loss (1) as
( n
)
k
1X
k
1X
`[i] (f ) ? min
[`i (f ) ? ?]+ + ? .
(2)
Lavt-k (Lz (f )) =
??0
k i=1
n i=1
n
In other words, the ATk loss is equivalent to minimum of the average of individual losses that are
shifted and truncated by the hinge function controlled by ?. This sheds more lights on the ATk loss,
which is particularly easy to illustrate in the context of binary classification using the margin losses,
`(f (x), y) = `(yf (x)).
In binary classification, the ?gold standard? of individual loss is the 0-1 loss Iyf (x)?0 , which exerts
a constant penalty 1 to examples that are misclassified by f and no penalty to correctly classified
examples. However, the 0-1 loss is difficult to work as it is neither continuous nor convex. In
practice, it is usually replaced by a surrogate convex loss. Such convex surrogates afford efficient
algorithms, but as continuous and convex upper-bounds of the 0-1 loss, they typically also penalize
correctly classified examples, i.e., for y and x that satisfy yf (x) > 0, `(yf (x)) > 0, whereas
Iyf (x)?0 = 0 (Fig.2). This implies that when the average of individual losses across all training
examples is minimized, correctly classified examples by f that are ?too close? to the classification
boundary may be sacrificed to accommodate reducing the average loss, as is shown in Fig.1.
In contrast, after the individual loss is combined with the hinge function, i.e., [`(yf (x)) ? ?]+ with
? > 0, it has the effect of ?shifting down? the original individual loss function and truncating it at
zero, see Fig.2. The transformation of the individual loss reduces penalties of all examples, and in
particular benefits correctly classified data. In particular, if such examples are ?far enough? from the
decision boundary, like in the 0-1 loss, their penalty becomes zero. This alleviates the likelihood of
misclassification on those rare sub-populations of data that are close to the decision boundary.
3
Loss
3
Algorithm: The reformulation of the ATk loss in Eq.(2) also facilitates development of optimization algorithms for the
2.5
minimum ATk learning. As practical supervised learning prob2
lems usually use a parametric form of f , as f (x; w), where w
is the parameter, the corresponding minimum ATk objective
1.5
becomes
1
( n
)
1X
k
0.5
min
[`(f (xi ; w), yi ) ? ?]+ + ? + ?(w) ,
w,??0
n i=1
n
0
-1.5
-1
-0.5
0
0.5
1
1.5
(3)
It is not hard to see that if `(f (x; w), y) is convex with respect Figure 2: The ATk loss interpreted at
to w, the objective function of in Eq.(3) is a convex function the individual loss level. Shaded area
for w and ? jointly. This leads to an immediate stochastic (pro- corresponds to data/target with correct
jected) gradient descent [3, 21] for solving (3). For instance, classification.
1
kwk2 , where C > 0 is a regularization facwith ?(w) = 2C
tor, at the t-th iteration, the corresponding MATk objective can be minimized by first randomly
sampling (xit , yit ) from the training set and then updating the parameters as
(t)
w(t+1) ? w(t) ? ?t ?w `(f (xit ; w(t) ), yit ) ? I[`(f (xit ;w(t) ),yit )>?(t) ] + wC
i
h
(4)
?(t+1) ? ?(t) ? ?t nk ? I[`(f (xit ;w(t) ,yit )>?(t) ]
+
where ?w `(f (x; w), y) denotes the sub-gradient with respect to w, and ?t ?
1
?
t
is the step size.
ATk -SVM: As a general aggregate loss, the ATk loss can be combined with any functional form
for individual losses. In the case of binary classification, the ATk loss combined with the individual
hinge loss for a prediction function f from a reproducing kernel Hilbert space (RKHS) [18] leads to
the ATk -SVM model. Specifically, we consider function f as a member of RKHS HK with norm
k ? kK , which is induced from a reproducing kernel K : X ? X ? R. Using the individual hinge
loss, [1 ? yi f (xi )]+ , the corresponding MATk learning objective in RKHS becomes
n
1 X
k
1
min
[1 ? yi f (xi )]+ ? ? + + ? +
kf k2K ,
(5)
f ?HK ,??0 n
n
2C
i=1
where C > 0 is the regularization factor. Furthermore, the outer hinge function in (5) can be
removed due to the following result.
Lemma 2. For a ? 0, b ? 0, there holds [a ? `]+ ? b + = [a ? b ? `]+ .
Proof of Lemma 2 can be found in the supplementary materials. In addition, note that for any
minimizer (fz , ?z ) of (5), setting f (x) = 0, ? = 1 in the objective function of (5), we have nk ?z ?
Pn
k
k
1
1
2
i=1 [1 ? yi fz (xi )]+ ? ?z + + n ?z + 2C kfz kK ? n , so we have 0 ? ?z ? 1 which means
n
that the minimization can be restricted to 0 ? ? ? 1. Using these results and introducing ? = 1 ? ?,
Eq.(5) can be rewritten as
n
1X
k
1
min
[? ? yi f (xi )]+ ? ? +
kf k2K .
(6)
f ?HK ,0???1 n
n
2C
i=1
The ATk -SVM objective generalizes many several existing SVM models. For example, when k = n,
it equals to the standard C-SVM [5]. When C = 1 and with conditions K(xi , xi ) ? 1 for any i,
ATk -SVM reduces to ?-SVM [19] with ? = nk . Furthermore, similar to the conventional SVM
model, writing in the dual form of (6) can lead to a convex quadratic programming problem that can
be solved efficiently. See supplementary materials for more detailed explanations.
Choosing k. The number of top individual losses in the ATk loss is a critical parameter that affects
the learning performance. In concept, using ATk loss will not be worse than using average or
maximum losses as they correspond to specific choices of k. In practice, k can be chosen during
training from a validation dataset as the experiments in Section 4. As k is an integer, a simple grid
search usually suffices to find a satisfactory value. Besides, Theorem 1 in Section 3 establishes a
theoretical lower bound for k to guarantee reliable classification based on the Bayes error. If we
have information about the proportion of outliers, we can also narrow searching space of k based on
the fact that ATk loss is the convex upper bound of the top-k loss, which is similar to [20].
4
3
Statistical Analysis
In this section, we address the statistical properties of the ATk objective in the context of binary
classification. Specifically, we investigate the property of classification calibration [1] of the ATk
general objective, and derive bounds for the misclassification error of the ATk -SVM model in the
framework of statistical learning theory (e.g. [1, 7, 23, 26]).
3.1
Classification Calibration under ATk Loss
We assume the training data z = {(xi , yi )}ni=1 are i.i.d. samples from an unknown distribution p on
X ?{?1}. Let pX be the marginal distribution of p on the input space X . Then, the misclassification
error of a classifier f : X ? {?1} is denoted by R(f ) = Pr(y 6= f (x)) = E[Iyf (x)?0 ]. The Bayes
error is given by R? = inf f R(f ), where the infimum is over all measurable functions. No function
can achieve less risk than the Bayes rule fc (x) = sign(?(x) ? 21 ), where ?(x) = Pr(y = 1|x) [8].
In practice, one uses a surrogate loss ` : R ? [0, ?) which is convex and upper-bound the 0-1
loss. The population `-risk (generalization error) is given by E` (f ) = E[`(yf (x))]. Denote the
optimal `-risk by E`? = inf f E` (f ). A very basic requirement for using such a surrogate loss ` is
the so-called classification calibration (point-wise form of Fisher consistency) [1, 14]. Specifically,
a loss ` is classification calibrated with respect to distribution p if, for any x, the minimizer f`? =
inf f E` (f ) should have the same sign as the Bayes rule fc (x), i.e., sign(f`? (x)) = sign(fc (x))
whenever fc (x) 6= 0.
An appealing result concerning the classification calibration of a loss function ` was obtained in [1],
which states that ` is classification calibrated if ` is convex, differentiable at 0 and `0 (0) < 0. In the
same spirit, we investigate the classification calibration property of the ATk loss. Specifically, we
first obtain the population form of the ATk objective using the infinite limit of (2)
n
1X
k nk ??
[`(yi f (xi )) ? ?]+ + ? ??
??? E [[`(yf (x)) ? ?]+ ] + ??.
n i=1
n n??
We then consider the optimization problem
(f ? , ?? ) = arg inf E [[`(yf (x)) ? ?]+ ] + ??,
f,??0
(7)
where the infimum is taken over all measurable function f : X ? R. We say the ATk (aggregate)
loss is classification calibrated with respect to p if f ? has the same sign as the Bayes rule fc . The
following theorem establishes such conditions.
Theorem 1. Suppose the individual loss ` : R ? R+ is convex, differentiable at 0 and `0 (0) < 0.
Without loss of generality, assume that `(0) = 1. Let (f ? , ?? ) be defined in (7),
(i) If ? > E`? then the ATk loss is classification calibrated.
(ii) If, moreover, ` is monotonically
decreasing and the ATk aggregate loss is classification
R
calibrated then ? ? ?(x)6= 1 min(?(x), 1 ? ?(x))dpX (x).
2
The proof of Theorem 1 can be found in the supplementary materials. Part (i) and (ii) of the above
theorem address respectively the sufficient and necessary conditions on ? such that the ATk loss
becomes classification calibrated. Since ` is an upper bound surrogate of the 0-1 loss, the optimal
`-risk E`? is larger than the Bayes error R? , i.e., E`? ? R? . In particular, if the individual loss `
is the hinge loss then E`? = 2R? . Part (ii) of the above theorem indicates that the ATk aggregate
loss is classification calibrated if ? = limn?? k/n is larger than the optimal generalization error E`?
associated with the individual loss. The choice of k > nE`? thus guarantees classification calibration,
which gives a lower bound of k. This result also provides a theoretical underpinning of the sensitivity
to outliers of the
R maximum loss (ATk loss with k =R 1). If the probability of the set {x : ?(x) = 1/2}
is zero, R? = X min(?(x), 1 ? ?(x))dpX (x) = ?(x)6=1/2 min(?(x), 1 ? ?(x))dpX (x). Theorem
1 indicates that in this case, if the maximum loss is calibrated, one must have n1 ? ? ? R? . In
other words, as the number of training data increases, the Bayes error has to be arbitrarily small,
which is consistent with the empirical observation that the maximum loss works well under the
well-separable data setting but are sensitive to outliers and non-separable data.
5
3.2
Error bounds of ATk -SVM
We next study the excess misclassification error of the ATk -SVM model i.e., R(sign(fz )) ? R? .
Let (fz , ?z ) be the minimizer of the ATk -SVM objective (6) in the RKHS setting. Let fH be the
minimizer of the generalization error over the RKHS space HK , i.e., fH = argminf ?HK Eh (f ),
where we use the notation Eh (f ) = E [[1 ? yf (x)]+ ] to denote the `-risk of the hinge loss. In the
finite-dimension case, the existence of fH follows from the direct method in the variational calculus,
as Eh (?) is lower bounded by zero, coercive, and weakly sequentially lower semi-continuous by its
convexity. For an infinite dimensional HK , we assume the existence of fH . We also assume that
Eh (fH ) < 1 since even a na??ve zero classifier can achieve Eh (0) = 1. Denote the approximation
p
error by A(HK ) = inf f ?HK Eh (f ) ? Eh (fc ) = Eh (fH ) ? Eh (fc ), and let ? = supx?X K(x, x).
The main theorem can be stated as follows.
Theorem 2. Consider the ATk -SVM in RKHS (6). For any ? ? (0, 1] and ? ? (0, 1 ? Eh (fH )),
choosing k = dn(Eh (fH ) + ?)e. Then, it holds
1 + C?,H
n?2 ?2
Pr R(sign(fz )) ? R? ? ? + A(H) + ? + ?
,
? 2 exp ?
(1 + C?,H )2
n?
?
where C?,H = ?(2 2C + 4kfH kK ).
The complete proof of Theorem 2 is given in the supplementary materials. The main idea is to
show that ?z is bounded from below by a positive constant with high probability, and then bound the
excess misclassification error R(sign(fz? )) ? R? by Eh (fz /?z ) ? Eh (fc ). If K is a universal kernel
then A(HK ) = 0 [23]. In this case, let ? = ? ? (0, 1 ? Eh (fH )), then from Theorem 2 we have
1 + C?,H
n?4
Pr R(sign(fz )) ? R? ? 2? + ?
? 2 exp ?
,
(1 + C?,H )2
n?
Consequently, choosing C such that limn?? C/n = 0, which is equivalent to
limn?? (1 + C?,H )2 /n = 0, then R(sign(fz )) can be arbitrarily close to the Bayes error R? ,
with high probability, as long as n is sufficiently large.
4
Experiments
We have demonstrated that ATk loss provides a continuum between the average loss and the maximum loss, which can potentially alleviates their drawbacks. A natural question is whether such
an advantage actually benefits practical learning problems. In this section, we demonstrate the
behaviors of MATk learning coupled with different individual losses for binary classification and
regression on synthetic and real datasets, with minimizing the average loss and the maximum loss
treated as special cases for k = n and k = 1, respectively. For simplicity, in all experiments, we use
homogenized linear prediction functions f (x) = wT x with parameters w and the Tikhonov regu1
larizer ?(w) = 2C
||w||2 , and optimize the MATk learning objective with the stochastic gradient
descent method given in (4).
Binary Classification: We conduct experiments on binary classification using eight benchmark
datasets from the UCI3 and KEEL4 data repositories to illustrate the potential effects of using ATk
loss in practical learning to adapt to different underlying data distributions. A detailed description
of the datasets is given in supplementary materials. The standard individual logistic loss and hinge
loss are combined with different aggregate losses. Note that average loss combined with individual
logistic loss corresponds to the logistic regression model and average loss combined with individual
hinge loss leads to the C-SVM algorithm [5].
For each dataset, we randomly sample 50%, 25%, 25% examples as training, validation and testing
sets, respectively. During training, we select parameters C (regularization factor) and k (number of
top losses) on the validation set. Parameter C is searched on grids of log10 scale in the range of
[10?5 , 105 ] (extended when optimal value is on the boundary), and k is searched on grids of log10
scale in the range of [1, n]. We use k ? to denote the optimal k selected from the validation set.
3
4
https://archive.ics.uci.edu/ml/datasets.html
http://sci2s.ugr.es/keel/datasets.php
6
Logistic Loss
Hinge Loss
Maximum
Average
ATk?
Maximum
Average
Monk
22.41(2.95) 20.46(2.02) 16.76(2.29)
22.04(3.08) 18.61(3.16)
Australian 19.88(6.64) 14.27(3.22) 11.70(2.82)
19.82(6.56) 14.74(3.10)
Madelon
47.85(2.51) 40.68(1.43) 39.65(1.72)
48.55(1.97) 40.58(1.86)
Splice
23.57(1.93) 17.25(0.93) 16.12(0.97)
23.40(2.10) 16.25(1.12)
Spambase
21.30(3.05)
8.36(0.97)
8.36(0.97)
21.03(3.26)
7.40(0.72)
German
28.24(1.69) 25.36(1.27) 23.28(1.16)
27.88(1.61) 24.16(0.89)
Titanic
26.50(3.35) 22.77(0.82) 22.44(0.84)
25.45(2.52) 22.82(0.74)
Phoneme
28.67(0.58) 25.50(0.88) 24.17(0.89)
28.81(0.62) 22.88(1.01)
Table 1: Average misclassification rate (%) of different learning objectives over 8 datasets.
are shown in bold with results that are not significant different to the best results underlined.
0.24
0.22
0.2
0.18
Splice
0.28
Misclassification Rate
Misclassification Rate
Misclassification Rate
Australian
0.35
0.26
0.3
0.25
0.2
0.15
Phoneme
0.3
0.26
Misclassification Rate
Monk
0.28
0.24
0.22
0.2
0.18
ATk?
17.04(2.77)
12.51(4.03)
40.18(1.64)
16.23(0.97)
7.40(0.72)
23.80(1.05)
22.02(0.77)
22.88(1.01)
The best results
0.28
0.26
0.24
0.16
0.16
1
50
100
150
k
216
0.1
0.22
1
100
k
200
300 346
1
500
k
1000
1588
1
500
1000
1500
k
2000
2702
Figure 3: Plots of misclassification rate on testing set vs. k on four datasets.
We report the average performance over 10 random splitting of training/validation/testing for each
dataset with MATk learning objectives formed from individual logistic loss and hinge loss. Table 1
gives their experimental results in terms of misclassification rate (results in terms of other classification quality metrics are given in supplementary materials). Note that on these datasets, the average
loss consistently outperforms the maximum loss, but the performance can be further improved with
the ATk loss, which is more adaptable to different data distributions. This advantage of the ATk loss
is particularly conspicuous for datasets Monk and Australian.
To further understand the behavior of MATk learning on individual datasets, we show plots of misclassification rate on testing set vs. k for four representative datasets in Fig.3 (in which C is fixed to
102 and k ? [1, n]). As these plots show, on all four datasets, there is a clear range of k value with
better classification performance than the two extreme cases k = 1 and k = n, corresponding to the
maximum and average loss, respectively. To be more specific, when k = 1, the potential noises and
outliers will have the highest negative effects on the learned classifier and the related classification
performance is very poor. As k increases, the negative effects of noises and outliers will reduce and
the classification performance becomes better, this is more significant on dataset Monk, Australian
and Splice. However, if k keeps increase, the classification performance may decrease (e.g., when
k = n). This may because that as k increases, more and more well classified samples will be included and the non-zero loss on these samples will have negative effects on the learned classifier (see
our analysis in Section 2), especifically for dataset Monk, Australian and Phoneme.
Regression. Next, we report experimental results of linear regression on one synthetic dataset
(Sinc) and three real datasets from [4], with a detailed description of these datasets given in supplementary materials. The standard square loss and absolute loss are adopted as individual losses.
Note that average loss coupled with individual square loss is standard ridge regression model and
average loss coupled with individual absolute loss reduces to ?-SVR [19]. We normalize the target
output to [0, 1] and report their root mean square error (RMSE) in Table 2, with optimal C and k ?
obtained by a grid search as in the case of classification (performance in terms of mean absolute
square error (MAE) is given in supplementary materials). Similar to the classification cases, using
the ATk loss usually improves performance in comparison to the average loss or maximum loss.
5
Related Works
Most work on learning objectives focus on designing individual losses, and only a few are dedicated
to new forms of aggregate losses. Recently, aggregate loss considering the order of training data have
been proposed in curriculum learning [2] and self-paced learning [11, 9], which suggest to organize
the training process in several passes and samples are included from easy to hard gradually. It is
interesting to note that each pass of self-paced learning [11] is equivalent to minimum the average of
7
Square Loss
Absolute Loss
Maximum
Average
ATk?
Maximum
Average
ATk?
Sinc 0.2790(0.0449) 0.1147(0.0060) 0.1139(0.0057) 0.1916(0.0771) 0.1188(0.0067) 0.1161(0.0060)
Housing 0.1531(0.0226) 0.1065(0.0132) 0.1050(0.0132) 0.1498(0.0125) 0.1097(0.0180) 0.1082(0.0189)
Abalone 0.1544(0.1012) 0.0800(0.0026) 0.0797(0.0026) 0.1243(0.0283) 0.0814(0.0029) 0.0811(0.0027)
Cpusmall 0.2895(0.0722) 0.1001(0.0035) 0.0998(0.0037) 0.2041(0.0933) 0.1170(0.0061) 0.1164(0.0062)
Table 2: Average RMSE on four datasets. The best results are shown in bold with results that are not significant
different to the best results underlined.
Pn
the k smallest individual losses, i.e., k1 i=n?k+1 `[i] (f ), which we term it as the average bottom-k
loss in contrast to the average top-k losses in our case. In [20], the pros and cons of the maximum
loss and the average loss are compared, and the top-k loss, i.e., `[k] (f ), is advocated as a remedy to
the problem of both. However, unlike the ATk loss, in general, neither the average bottom-k loss nor
the top-k loss are convex functions with regards to the individual losses.
Minimizing top-k errors has also been used in individual losses. For ranking problems, the work of
[17, 24] describes a form of individual loss that gives more weights to the top examples in a ranked
list. In multi-class classification, the top-1 loss is commonly used which causes penalties when the
top-1 predicted class is not the same as the target class label [6]. This has been further extended
in [12, 13] to the top-k multi-class loss, in which for a class label that can take m different values,
the classifier is only penalized when the correct value does not show up in the top k most confident
predicted values. As an individual loss, these works are complementary to the ATk loss and they
can be combined to improve learning performance.
6
Discussion
In this work, we introduce the average top-k (ATk ) loss as a new aggregate loss for supervised
learning, which is the average over the k largest individual losses over a training dataset. We show
that the ATk loss is a natural generalization of the two widely used aggregate losses, namely the
average loss and the maximum loss, but can combine their advantages and mitigate their drawbacks
to better adapt to different data distributions. We demonstrate that the ATk loss can better protect
small subsets of hard samples from being swamped by a large number of easy ones, especially for
imbalanced problems. Furthermore, it remains a convex function over all individual losses, which
can lead to convex optimization problems that can be solved effectively with conventional gradientbased methods. We provide an intuitive interpretation of the ATk loss based on its equivalent effect
on the continuous individual loss functions, suggesting that it can reduce the penalty on correctly
classified data. We further study the theoretical aspects of ATk loss on classification calibration and
error bounds of minimum average top-k learning for ATk -SVM. We demonstrate the applicability
of minimum average top-k learning for binary classification and regression using synthetic and real
datasets.
There are many interesting questions left unanswered regarding using the ATk loss as learning objectives. Currently, we use conventional gradient-based algorithms for its optimization, but we are
investigating special instantiations of MATk learning for which more efficient optimization methods
can be developed. Furthermore, the ATk loss can also be used for unsupervised learning problems
(e.g., clustering), which is a focus of our subsequent study. It is also of practical importance to
combine ATk loss with other successful learning paradigms such as deep learning, and to apply it to
large scale real life dataset. Lastly, it would be very interesting to derive error bounds of MATk with
general individual loss functions.
7
Acknowledgments
We thank the anonymous reviewers for their constructive comments. This work was completed when
the first author was a visiting student at SUNY Albany, supported by a scholarship from University of
Chinese Academy of Sciences (UCAS). Siwei Lyu is supported by the National Science Foundation
(NSF, Grant IIS-1537257) and Yiming Ying is supported by the Simons Foundation (#422504) and
the 2016-2017 Presidential Innovation Fund for Research and Scholarship (PIFRS) program from
SUNY Albany. This work is also partially supported by the National Science Foundation of China
(NSFC, Grant 61620106003) for Bao-Gang Hu and Yanbo Fan.
8
References
[1] P. L. Bartlett, M. I. Jordan, and J. D. McAuliffe. Convexity, classification, and risk bounds. Journal of the
American Statistical Association, 101(473):138?156, 2006.
[2] Y. Bengio, J. Louradour, R. Collobert, and J. Weston. Curriculum learning. In ICML, pages 41?48, 2009.
[3] O. Bousquet and L. Bottou. The tradeoffs of large scale learning. In NIPS, pages 161?168, 2008.
[4] C.-C. Chang and C.-J. Lin. Libsvm: a library for support vector machines. TIST, 2(3):27, 2011.
[5] C. Cortes and V. Vapnik. Support-vector networks. Machine learning, 20(3):273?297, 1995.
[6] K. Crammer and Y. Singer. On the algorithmic implementation of multiclass kernel-based vector machines. Journal of machine learning research, 2(Dec):265?292, 2001.
[7] E. De Vito, A. Caponnetto, and L. Rosasco. Model selection for regularized least-squares algorithm in
learning theory. Foundations of Computational Mathematics, 5(1):59?85, 2005.
[8] L. Devroye, L. Gy?orfi, and G. Lugosi. A probabilistic theory of pattern recognition, volume 31. Springer
Science & Business Media, 2013.
[9] Y. Fan, R. He, J. Liang, and B.-G. Hu. Self-paced learning: An implicit regularization perspective. In
AAAI, pages 1877?1833, 2017.
[10] R. He, W.-S. Zheng, and B.-G. Hu. Maximum correntropy criterion for robust face recognition. IEEE
Transactions on Pattern Analysis and Machine Intelligence, 33(8):1561?1576, 2011.
[11] M. P. Kumar, B. Packer, and D. Koller. Self-paced learning for latent variable models. In NIPS, pages
1189?1197, 2010.
[12] M. Lapin, M. Hein, and B. Schiele. Top-k multiclass SVM. In NIPS, pages 325?333, 2015.
[13] M. Lapin, M. Hein, and B. Schiele. Loss functions for top-k error: Analysis and insights. In CVPR, pages
1468?1477, 2016.
[14] Y. Lin. A note on margin-based loss functions in classification. Statistics & probability letters, 68(1):73?
82, 2004.
[15] H. Masnadi-Shirazi and N. Vasconcelos. On the design of loss functions for classification: theory, robustness to outliers, and savageboost. In NIPS, pages 1049?1056, 2009.
[16] W. Ogryczak and A. Tamir. Minimizing the sum of the k largest functions in linear time. Information
Processing Letters, 85(3):117?122, 2003.
[17] C. Rudin. The p-norm push: A simple convex ranking algorithm that concentrates at the top of the list.
Journal of Machine Learning Research, 10(Oct):2233?2271, 2009.
[18] B. Sch?olkopf and A. J. Smola. Learning with kernels: support vector machines, regularization, optimization, and beyond. MIT press, 2001.
[19] B. Sch?olkopf, A. J. Smola, R. C. Williamson, and P. L. Bartlett. New support vector algorithms. Neural
computation, 12(5):1207?1245, 2000.
[20] S. Shalev-Shwartz and Y. Wexler. Minimizing the maximal loss: How and why. In ICML, 2016.
[21] N. Srebro and A. Tewari. Stochastic optimization for machine learning. ICML Tutorial, 2010.
[22] I. Steinwart. On the optimal parameter choice for ?-support vector machines. IEEE Transactions on
Pattern Analysis and Machine Intelligence, 25(10):1274?1284, 2003.
[23] I. Steinwart and A. Christmann. Support vector machines. Springer Science & Business Media, 2008.
[24] N. Usunier, D. Buffoni, and P. Gallinari. Ranking with ordered weighted pairwise classification. In ICML,
pages 1057?1064, 2009.
[25] V. Vapnik. Statistical learning theory, volume 1. Wiley New York, 1998.
[26] Q. Wu, Y. Ying, and D.-X. Zhou. Learning rates of least-square regularized regression. Foundations of
Computational Mathematics, 6(2):171?192, 2006.
[27] Y. Wu and Y. Liu. Robust truncated hinge loss support vector machines. Journal of the American Statistical Association, 102(479):974?983, 2007.
[28] Y. Yu, M. Yang, L. Xu, M. White, and D. Schuurmans. Relaxed clipping: A global training method for
robust regression and classification. In NIPS, pages 2532?2540, 2010.
9
| 6653 |@word madelon:1 repository:1 norm:2 proportion:1 calculus:1 hu:3 wexler:1 accommodate:2 liu:1 tist:1 rkhs:6 spambase:1 existing:2 outperforms:1 yet:2 must:1 subsequent:1 plot:5 fund:1 v:4 intelligence:2 selected:1 rudin:1 monk:5 provides:3 uca:2 completeness:1 dn:1 direct:1 shorthand:1 combine:3 introduce:3 pairwise:1 sacrifice:1 indeed:1 expected:1 behavior:2 nor:2 multi:5 decreasing:1 considering:1 becomes:5 notation:2 underlying:2 moreover:1 bounded:2 medium:2 interpreted:1 minimizes:1 developed:1 correntropy:1 coercive:1 transformation:1 guarantee:2 mitigate:2 shed:1 classifier:9 gallinari:1 grant:2 organize:1 mcauliffe:1 positive:1 limit:1 accumulates:1 nsfc:1 lugosi:1 china:1 shaded:2 range:3 practical:4 acknowledgment:1 testing:4 practice:3 dpx:3 procedure:1 area:2 universal:1 empirical:2 orfi:1 convenient:1 word:2 subpopulation:1 suggest:1 svr:1 undesirable:1 close:3 selection:1 risk:8 context:2 writing:1 optimize:1 conventional:4 equivalent:6 measurable:2 demonstrated:1 lyu1:1 reviewer:1 truncating:2 convex:25 simplicity:1 keel:1 splitting:1 prob2:1 rule:3 insight:1 population:4 searching:1 unanswered:1 target:4 suppose:1 nlpr:1 programming:1 us:1 designing:1 element:1 recognition:3 particularly:2 updating:1 corroborated:1 predicts:1 labeled:1 bottom:4 solved:3 decrease:1 removed:1 highest:1 balanced:1 convexity:2 schiele:2 lavg:1 vito:1 weakly:1 tight:1 solving:1 max1:1 various:1 regularizer:1 sacrificed:1 effective:1 aggregate:18 choosing:3 shalev:1 widely:3 supplementary:10 larger:2 say:1 cvpr:1 otherwise:1 presidential:1 statistic:2 jointly:1 housing:1 advantage:5 differentiable:2 propose:2 maximal:1 uci:1 alleviates:2 achieve:2 academy:2 gold:1 intuitive:3 description:2 validate:1 bao:2 normalize:1 olkopf:2 cluster:2 requirement:2 plethora:1 yiming:2 illustrate:2 derive:2 ac:1 advocated:1 eq:3 predicted:2 christmann:1 implies:1 australian:5 concentrate:1 drawback:4 correct:2 stochastic:5 atk:76 material:10 larizer:1 suffices:1 generalization:5 alleviate:1 anonymous:1 hold:3 underpinning:1 sufficiently:1 gradientbased:1 ic:1 exp:3 lyu:1 algorithmic:1 tor:1 continuum:1 smallest:1 fh:9 albany:6 label:2 currently:1 sensitive:2 largest:6 establishes:2 weighted:1 minimization:2 mit:1 pn:4 zhou:1 iyf:4 obscures:1 xit:4 focus:2 consistently:1 likelihood:1 indicates:2 hk:9 contrast:4 inference:1 typically:2 koller:1 transformed:1 misclassified:1 arg:1 classification:52 dual:1 html:1 denoted:1 development:1 special:2 marginal:1 equal:1 vasconcelos:1 beach:1 sampling:1 yu:1 unsupervised:1 icml:4 minf:1 minimized:2 report:3 few:2 masnadi:1 randomly:2 packer:1 national:3 ve:1 individual:56 replaced:1 n1:2 investigate:2 zheng:1 extreme:1 light:1 ogryczak:1 closer:1 necessary:1 conduct:1 hein:2 theoretical:5 instance:1 column:1 clipping:1 applicability:3 introducing:1 subset:1 rare:3 cpusmall:1 successful:1 too:1 supx:1 synthetic:6 combined:8 calibrated:8 st:1 confident:1 sensitivity:2 sequel:1 probabilistic:1 na:1 squared:1 aaai:1 rosasco:1 worse:1 american:2 toy:1 suggesting:3 potential:2 de:1 gy:1 summarized:1 bold:2 student:1 satisfy:1 ranking:3 collobert:1 view:1 root:1 bayes:9 simon:1 rmse:2 contribution:1 minimize:1 square:7 ni:3 formed:2 php:1 phoneme:3 efficiently:1 correspond:1 bayesian:1 classified:8 siwei:2 whenever:1 evaluates:1 involved:1 proof:4 associated:1 con:2 dataset:12 popular:1 improves:1 hilbert:1 actually:1 adaptable:1 focusing:1 supervised:7 modal:3 improved:2 formulation:1 evaluated:1 though:1 generality:1 furthermore:6 implicit:1 lastly:1 smola:2 hand:1 steinwart:2 logistic:7 yf:11 quality:2 infimum:2 shirazi:1 usa:1 effect:8 concept:1 unbiased:1 remedy:1 equality:1 regularization:5 laboratory:1 satisfactory:1 illustrated:1 white:1 during:2 self:4 abalone:1 criterion:1 complete:2 demonstrate:5 ridge:1 dedicated:1 pro:3 wise:1 variational:1 recently:1 functional:1 volume:2 association:2 interpretation:4 he:2 mae:1 synthesized:1 kwk2:1 refer:1 significant:4 grid:4 mathematics:3 consistency:1 calibration:10 imbalanced:4 recent:1 perspective:1 inf:5 tikhonov:1 underlined:2 binary:13 arbitrarily:2 kfh:1 life:1 yi:10 minimum:11 relaxed:1 paradigm:1 monotonically:1 ii:4 semi:1 reduces:3 caponnetto:1 adapt:5 long:2 lin:2 concerning:1 controlled:1 prediction:3 variant:1 regression:11 basic:1 expectation:1 exerts:1 metric:1 iteration:1 kernel:5 achieved:1 dec:1 penalize:1 buffoni:1 whereas:1 addition:1 limn:3 sch:2 unlike:2 archive:1 pass:1 comment:1 induced:1 facilitates:1 member:1 spirit:1 effectiveness:1 jordan:1 integer:1 yang:1 bengio:1 easy:3 enough:1 affect:1 interprets:1 reduce:4 idea:1 cn:1 regarding:1 tradeoff:1 multiclass:2 whether:1 bartlett:2 penalty:8 york:1 afford:1 cause:1 deep:1 tewari:1 detailed:4 clear:1 ugr:1 http:2 fz:9 affords:2 nsf:1 tutorial:1 shifted:1 sign:11 correctly:7 express:1 four:4 reformulation:3 suny:4 yit:4 neither:2 libsvm:1 sum:1 letter:2 fourth:1 wu:2 decision:2 bound:17 paced:4 fan:3 quadratic:1 nonnegative:1 gang:2 occur:1 bousquet:1 wc:1 aspect:2 min:8 kumar:1 separable:2 px:1 department:2 poor:1 across:2 smaller:3 describes:1 conspicuous:1 appealing:1 s1:1 lem:1 swamped:1 outlier:12 restricted:1 pr:4 gradually:1 taken:1 remains:2 german:1 singer:1 jected:1 adopted:1 generalizes:4 usunier:1 rewritten:1 eight:1 apply:1 robustness:1 existence:3 original:2 top:32 denotes:1 include:2 clustering:1 completed:1 log2:1 hinge:17 log10:2 k1:2 chinese:2 establish:1 especially:1 scholarship:2 seeking:1 objective:18 question:2 parametric:1 surrogate:5 visiting:1 gradient:7 separate:1 thank:1 outer:1 besides:1 devroye:1 illustration:1 reformulate:1 minimizing:7 balance:1 kk:3 ying:2 difficult:1 susceptible:1 innovation:1 liang:1 potentially:1 argminf:1 stated:1 negative:3 implementation:1 design:1 hu3:1 perform:1 unknown:1 upper:5 observation:1 datasets:19 benchmark:1 finite:1 descent:3 truncated:2 immediate:1 extended:2 reproducing:2 namely:2 extensive:1 connection:1 learned:2 narrow:1 protect:2 nip:6 address:2 beyond:1 usually:5 pattern:4 below:1 program:1 max:2 reliable:2 explanation:1 shifting:2 ia:1 misclassification:20 critical:1 natural:3 eh:14 treated:1 ranked:1 regularized:2 curriculum:2 business:2 improve:1 titanic:1 ne:1 library:1 coupled:3 ltop:4 sn:1 kf:2 loss:211 interesting:3 srebro:1 validation:5 foundation:5 sufficient:1 consistent:1 row:2 lmax:2 penalized:1 supported:4 understand:1 face:1 absolute:5 benefit:2 regard:3 boundary:8 dimension:1 xn:1 tamir:1 author:2 made:1 commonly:2 simplified:2 lz:16 far:1 transaction:2 excess:2 keep:1 ml:1 global:1 sequentially:1 investigating:1 instantiation:1 xi:13 shwartz:1 continuous:5 search:2 latent:1 why:1 table:4 robust:3 ca:1 schuurmans:1 williamson:1 bottou:1 constructing:1 louradour:1 pk:3 main:3 k2k:2 noise:2 complementary:1 x1:1 enlarged:1 fig:6 representative:1 xu:1 wiley:1 sub:4 splice:3 down:2 theorem:11 specific:3 list:2 svm:23 sinc:2 cortes:1 concern:1 bivariate:1 vapnik:2 effectively:2 importance:1 push:1 margin:3 nk:4 sorting:1 fc:8 ordered:1 partially:1 chang:1 springer:2 corresponds:2 minimizer:4 savageboost:1 weston:1 oct:1 formulated:1 consequently:1 fisher:1 hard:3 included:2 typical:1 specifically:4 reducing:1 infinite:2 wt:1 lemma:6 called:2 pas:1 e:1 experimental:2 select:1 searched:2 support:7 latter:1 crammer:1 constructive:1 lapin:2 |
6,250 | 6,654 | Learning multiple visual domains with residual
adapters
Sylvestre-Alvise Rebuffi1
Hakan Bilen1,2
1
Visual Geometry Group
University of Oxford
{srebuffi,hbilen,vedaldi}@robots.ox.ac.uk
Andrea Vedaldi1
2
School of Informatics
University of Edinburgh
Abstract
There is a growing interest in learning data representations that work well for many
different types of problems and data. In this paper, we look in particular at the
task of learning a single visual representation that can be successfully utilized in
the analysis of very different types of images, from dog breeds to stop signs and
digits. Inspired by recent work on learning networks that predict the parameters of
another, we develop a tunable deep network architecture that, by means of adapter
residual modules, can be steered on the fly to diverse visual domains. Our method
achieves a high degree of parameter sharing while maintaining or even improving
the accuracy of domain-specific representations. We also introduce the Visual
Decathlon Challenge, a benchmark that evaluates the ability of representations to
capture simultaneously ten very different visual domains and measures their ability
to perform well uniformly.
1
Introduction
While research in machine learning is often directed at improving the performance of algorithms on
specific tasks, there is a growing interest in developing methods that can tackle a large variety of
different problems within a single model. In the case of perception, there are two distinct aspects of
this challenge. The first one is to extract from a given image diverse information, such as image-level
labels, semantic segments, object bounding boxes, object contours, occluding boundaries, vanishing
points, etc. The second aspect is to model simultaneously many different visual domains, such as
Internet images, characters, glyph, animal breeds, sketches, galaxies, planktons, etc (fig. 1).
In this work we explore the second challenge and look at how deep learning techniques can be used
to learn universal representations [5], i.e. feature extractors that can work well in several different
image domains. We refer to this problem as multiple-domain learning to distinguish it from the more
generic multiple-task learning.
Multiple-domain learning contains in turn two sub-challenges. The first one is to develop algorithms
that can learn well from many domains. If domains are learned sequentially, but this is not a
requirement, this is reminiscent of domain adaptation. However, there are two important differences.
First, in standard domain adaptation (e.g. [9]) the content of the images (e.g. ?telephone?) remains
the same, and it is only the style of the images that changes (e.g. real life vs gallery image). Instead
in our case a domain shift changes both style and content. Secondly, the difficulty is not just to adapt
the model from one domain to another, but to do so while making sure that it still performs well on
the original domain, i.e. to learn without forgetting [21].
The second challenge of multiple-domain learning, and our main concern in this paper, is to construct
models that can represent compactly all the domains. Intuitively, even though images in different
domains may look quite different (e.g. glyph vs. cats), low and mid-level visual primitives may still
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
(a)
(b)
(c)
(d)
(e)
(f)
(g)
(h)
(i)
(j)
Figure 1: Visual Decathlon. We explore deep architectures that can learn simultaneously different
tasks from very different visual domains. We experiment with ten representative ones: (a) Aircraft, (b)
CIFAR-100, (c) Daimler Pedestrians, (d) Describable Textures, (e) German Traffic Signs, (f) ILSVRC
(ImageNet) 2012, (g) VGG-Flowers, (h) OmniGlot, (i) SVHN, (j) UCF101 Dynamic Images.
be largely shareable. Sharing knowledge between domains should allow to learn compact multivalent
representations. Provided that sufficient synergies between domains exist, multivalent representations
may even work better than models trained individually on each domain (for a given amount of training
data).
The primary contribution of this paper (section 3) is to introduce a design for multivalent neural
network architectures for multiple-domain learning (section 3 fig. 2). The key idea is reconfigure
a deep neural network on the fly to work on different domains as needed. Our construction is based on
recent learning-to-learn methods that showed how the parameters of a deep network can be predicted
from another [2, 16]. We show that these formulations are equivalent to packing the adaptation
parameters in convolutional layers added to the network (section 3). The layers in the resulting
parametric network are either domain-agnostic, hence shared between domains, or domain-specific,
hence parametric. The domain-specific layers are changed based on the ground-truth domain of the
input image, or based on an estimate of the latter obtained from an auxiliary network. In the latter
configuration, our architecture is analogous to the learnet of [2].
Based on such general observations, we introduce in particular a residual adapter module and
use it to parameterize the standard residual network architecture of [13]. The adapters contain a
small fraction of the model parameters (less than 10%) enabling a high-degree of parameter sharing
between domains. A similar architecture was concurrently proposed in [31], which also results in
the possibility of learning new domains sequentially without forgetting. However, we also show a
specific advantage of the residual adapter modules: the ability to modulate adaptation based on the
size of the target dataset.
Our proposed architectures are thoroughly evaluated empirically (section 5). To this end, our second
contribution is to introduce the visual decathlon challenge (fig. 1 and section 4), a new benchmark
for multiple-domain learning in image recognition. The challenge consists in performing well
simultaneously on ten very different visual classification problems, from ImageNet and SVHN to
action classification and describable texture recognition. The evaluation metric, also inspired by the
decathlon discipline, rewards models that perform better than strong baselines on all the domains
simultaneously. A summary of our finding is contained in section 6.
2
Related Work
Our work touches on multi-task learning, learning without forgetting, domain adaptation, and other
areas. However, our multiple-domain setup differs in ways that make most of the existing approaches
not directly applicable to our problem.
Multi-task learning (MTL) looks at developing models that can address different tasks, such as
detecting objects and segmenting images, while sharing information and computation among them.
Earlier examples of this paradigm have focused on kernel methods [10, 1] and deep neural network
(DNN) models [6]. In DNNs, a standard approach [6] is to share earlier layers of the network, training
the tasks jointly by means of back-propagation. Caruana [6] shows that sharing network parameters
between tasks is beneficial also as a form of regularization, putting additional constraints on the
learned representation and thus improving it.
MTL in DNNs has been applied to various problems ranging from natural language processing [8, 22],
speech recognition [14] to computer vision [41, 42, 4]. Collobert et al. [8] show that semi-supervised
learning and multi-task learning can be combined in a DNN model to solve several language
processing prediction tasks such as part-of-speech tags, chunks, named entity tags and semantic
2
roles. Huang et al. [14] propose a shared multilingual DNN which shares hidden layers across
many languages. Liu et al. [22] combine multiple-domain classification and information retrieval for
ranking web search with a DNN. Multi-task DNN models are also reported to achieve performance
gains in computer vision problems such as object tracking [41], facial-landmark detection [42], object
and part detection [4], a collection of low-level and high-level vision tasks [18]. The main focus of
these works is learning a diverse set of tasks in the same visual domain. In contrast, our paper focuses
on learning a representation from a diverse set of domains.
Our investigation is related to the recent paper of [5], which studied the ?size? of the union of different
visual domains measured in terms of the capacity of the model required to learn it. The authors
propose to absorb different domain in a single neural network by tuning certain parameters in batch
and instance normalization layers throughout the architecture; we show that our residual adapter
modules, which include the latter as a special case, lead to far superior results.
Life-long learning. A particularly important aspect of MTL is the ability of learning multiple tasks
sequentially, as in Never Ending Learning [25] and Life-long Learning [38]. Sequential learning
typically suffers in fact from forgetting the older tasks, a phenomenon aptly referred to as ?catastrophic
forgetting? in [11]. Recent work in life-long learning try to address forgetting in two ways. The first
one [37, 33] is to freeze the network parameters for the old tasks and learn a new task by adding
extra parameters. The second one aims at preserving knowledge of the old tasks by retaining the
response of the original network on the new task [21, 30], or by keeping the network parameters of
the new task close to the original ones [17]. Our method can be considered as a hybrid of these two
approaches, as it can be used to retain the knowledge of previous tasks exactly, while adding a small
number of extra parameters for the new tasks.
Transfer learning. Sometimes one is interested in maximizing the performance of a model on a
target domain. In this case, sequential learning can be used as a form of initialization[29]. This is
very common in visual recognition, where most DNN are initialize on the ImageNet dataset and then
fine-tuned on a target domain and task. Note, however, that this typically results in forgetting the
original domain, a fact that we confirm in the experiments.
Domain adaptation. When domains are learned sequentially, our work can be related to domain
adaptation. There is a vast literature in domain adaptation, including recent contributions in deep
learning such as [12, 39] based on the idea of minimizing domain discrepancy. Long et al. [23]
propose a deep network architecture for domain adaptation that can jointly learn adaptive classifiers
and transferable features from labeled data in the source domain and unlabeled data in the target
domain. There are two important differences with our work: First, in these cases different domains
contain the same objects and is only the visual style that changes (e.g. webcam vs. DSLR), whereas
in our case the object themselves change. Secondly, domain adaptation is a form of transfer learning,
and, as the latter, is concerned with maximizing the performance on the target domain reagardless of
potential forgetting.
3
Method
Our primary goal is to develop neural network architectures that can work well in a multiple-domain
setting. Modern neural networks such as residual networks (ResNet [13]) are known to have very high
capacity, and are therefore good candidates to learn from diverse data sources. Furthermore, even
when domains look fairly different, they may still share a significant amount of low and mid-level
visual patterns. Nevertheless, we show in the experiments (section 5) that learning a ResNet (or a
similar model) directly from multiple domains may still not perform well.
In order to address this problem, we consider a compact parametric family of neural networks
?? : X ? V indexed by parameters ?. Concretely, X ? RH?W ?3 can be a space of RGB images
and V = RHv ?Wv ?Cv a space of feature tensors. ?? can then be obtained by taking all but the last
classification layer of a standard ResNet model. The parametric feature extractors ?? is then used to
construct predictors for each domain d as ?d = ?d ? ??d , where ?d are domain-specific parameters
and ?d (v) = softmax(Wd v) is a domain-specific linear classifier V ? Yd mapping features to
image labels.
If ? comprises all the parameters of the feature extractor ?? , this approach reduces to learning
independent models for each domain. On the contrary, our goal is to maximize parameter sharing,
which we do below by introducing certain network parametrizations.
3
w1
BN
?
0
(?1s , ?1b ) ?1w
0
w2
(?1s , ?1b )
?
BN
+
BN
?
[?]+
0
(?2s , ?2n ) ?2w
0
(?2s , ?2b )
?
+
BN
+
[?]+
Figure 2: Residual adapter modules. The figure shows a standard residual module with the inclusion
of adapter modules (in blue). The filter coefficients (w1 , w2 ) are domain-agnostic and contains the
vast majority of the model parameters; (?1 , ?2 ) contain instead a small number of domain-specific
parameters.
3.1
Learning to learn and filter prediction
The problem of adapting a neural network dynamically to variations of the input data is similar to the
one found in recent approaches to learning to learn. A few authors [34, 16, 2], in particular, have
proposed to learn neural networks that predict, in a data-dependent manner, the parameters of another.
Formally, we can write ?d = Aedx where edx is the indicator vector of the domain dx of image x
and A is a matrix whose columns are the parameter vectors ?d . As shown later, it is often easy to
construct an auxiliary network that can predict d from x, so that the parameter ? = ?(x) can also be
expressed as the output of a neural network. If d is known, then ?(x, d) = ?d as before, and if not ?
can be constructed as suggested above or from scratch as done in [2].
The result of this construction is a network ??(x) (x) whose parameters are predicted by a second
network ?(x). As noted in [2], while this construction is conceptually simple, its implementation is
more subtle. Recall that the parameters w of a deep convolutional neural network consist primarily of
the coefficients of the linear filters in the convolutional layers. If w = ?, then ? = ?(x) would need
to predict millions of parameters (or to learn independent models when d is observed). The solution
of [2] is to use a low-rank decomposition of the filters, where w = ?(w0 , ?) is a function of a filter
basis w0 and ? is a small set of tunable parameters.
Here we build on the same idea, with some important extensions. First, we note that linearly
parametrizing a filter bank is the same as introducing a new, intermediate convolutional layer in the
network. Specifically, let Fk ? RHf ?Wf ?Cf be a basis of K filters of size Hf ? Wf operating on
Cf input feature channels. Given parameters [?tk ] ? RT ?K , we can express a bank of T filters as
PK
linear combinations Gt = k=1 ?tk Fk . Applying the bank to a tensor x and using associativity and
PK
linearity of convolution results in G ? x = k=1 ?:k (Fk ? x) = ? ? F ? x where we interpreted ? as
a 1 ? 1 ? T ? K filter bank. While [2] used a slightly different low-rank filter decomposition, their
parametrization can also be seen as introducing additional filtering layers in the network.
An advantage of this parametrization is that it results in a useful decomposition, where part of the
convolutional layers contain the domain-agnostic parameters F and the others contain the domainspecific ones ?d . As discussed in section 5, this is particularly useful to address the forgetting
problem. In the next section we refine these ideas to obtain an effective parametrization of residual
networks.
3.2
Residual adapter modules
As an example of parametric network, we propose to modify a standard residual network. Recall that
a ResNet is a chain gm ? ? ? ? ? g1 of residual modules gt . In the simplest variant of the model, each
residual module g takes as input a tensor RH?W ?C and produces as output a tensor of the same size
using g(x; w) = x + ((w2 ? ?) ? [?]+ ? (w1 ? ?))(x). Here w1 and w2 are the coefficients of banks of
small linear filters, [z]+ = max{0, z} is the ReLU operator, w ? z is the convolution of z by the filter
bank w, and ? denotes function composition. Note that, for the addition to make sense, filters must
be configured such that the dimensions of the output of the last bank are the same as x.
Our goal is to parametrize the ResNet module. As suggested in the previous section, rather than
changing the filter coefficients directly, we introduce additional parametric convolutional layers. In
fact, we go one step beyond and make them small residual modules in their own right and call them
4
residual adapter modules (blue blocks in fig. 2). These modules have the form:
g(x; ?) = x + ? ? x.
In order to limit the number of domain-specific parameters, ? is selected to be a bank of 1 ? 1 filters.
A major advantage of adopting a residual architecture for the adapter modules is that the adapters
reduce to the identity function when their coefficients are zero. When learning the adapters on small
domains, this provides a simple way of controlling over-fitting, resulting in substantially improved
performance in some cases.
Batch normalization and scaling. Batch Normalization (BN) [15] is an important part of very deep
neural networks. This module is usually inserted after convolutional layers in order to normalize
their outputs and facilitate learning (fig. 2). The normalization operation is followed by rescaling and
shift operations s x + b, where (s, b) are learnable parameters. In our architecture, we incorporate
the BN layers into the adapter modules (fig. 2). Furthermore, we add a BN module right before the
adapter convolution layer.1 Note that the BN scale and bias parameters are also dataset-dependent ?
as noted in the experiments, this alone provides a certain degree of model adaptation.
Domain-agnostic vs domain-specific parameters. If the residual module of fig. 2 is configured to
process an input tensor with C feature channels, and if the domain-agnostic filters w1 , w2 are of size
h ? h ? C, then the model has 2(h2 C 2 + hC) domain-agnostic parameters (including biases in the
convolutional layers) and 2(C 2 + 5C) domain-specific parameters.2 Hence, there are approximately
h2 more domain-agnostic parameters than domain specific ones (usually h2 = 9).
3.3
Sequential learning and avoiding forgetting
While in this paper we are not concerned with sequential learning, we have found it to be a good
strategy to bootstrap a model when a large number of domains have to be learned. However, the most
popular approach to sequential learning, fine-tuning (section 2), is often a poor choice for learning
shared representations as it tends to quickly forget the original tasks.
The challenge in learning without forgetting is to maintain information about older tasks as new
ones are learned (section 2). With respect to forgetting, our adapter modules are similar to the
tower model [33] as they preserve the original model exactly: one can pre-train the domain-agnostic
parameters w on a large domain such as ImageNet, and then fine-tune only the domain-specific
parameters ?d for each new domain. Like the tower method, this preserves the original task exactly,
but it is far less expensive as it does not require to introduce new feature channels for each new
domain (a quadratic cost). Furthermore, the residual modules naturally reduce to the identity function
when sufficient shrinking regularization is applied to the adapter weights ?w . This allows the adapter
to be tuned depending on the availability of data for a target domain, sometimes significantly reducing
overfitting.
4
Visual decathlon
In this section we introduce a new benchmark, called visual decathlon, to evaluate the performance
of algorithms in multiple-domain learning. The goal of the benchmark is to assess whether a method
can successfully learn to perform well in several different domains at the same time. We do so by
choosing ten representative visual domains, from Internet images to characters, as well as by selecting
an evaluation metric that rewards performing well on all tasks.
Datasets. The decathlon challenge combines ten well-known datasets from multiple visual domains:
FGVC-Aircraft Benchmark [24] contains 10,000 images of aircraft, with 100 images for each of
100 different aircraft model variants such as Boeing 737-400, Airbus A310. CIFAR100 [19] contains
60,000 32 ? 32 colour images for 100 object categories. Daimler Mono Pedestrian Classification
Benchmark (DPed) [26] consists of 50,000 grayscale pedestrian and non-pedestrian images, cropped
and resized to 18 ? 36 pixels. Describable Texture Dataset (DTD) [7] is a texture database, consisting of 5640 images, organized according to a list of 47 terms (categories) such as bubbly, cracked,
1
While the bias and scale parameters of the latter can be incorporated in the following filter bank, we found
it easier to leave them separated from the latter
2
Including all bias and scaling vectors; 2(C 2 + 3C) if these are absorbed in the filter banks when possible.
5
marbled. The German Traffic Sign Recognition (GTSR) Benchmark [36] contains cropped images for 43 common traffic sign categories in different image resolutions. Flowers102 [28] is a
fine-grained classification task which contains 102 flower categories from the UK, each consisting of
between 40 and 258 images. ILSVRC12 (ImNet) [32] is the largest dataset in our benchmark contains 1000 categories and 1.2 million images. Omniglot [20] consists of 1623 different handwritten
characters from 50 different alphabets. Although the dataset is designed for one-shot learning, we use
the dataset for standard multi-class classification task and include all the character categories in train
and test splits. The Street View House Numbers (SVHN) [27] is a real-world digit recognition
dataset with around 70,000 32 ? 32 images. UCF101 [35] is an action recognition dataset of realistic
human action videos, collected from YouTube. It contains 13,320 videos for 101 action categories.
In order to make this dataset compatible with our benchmark, we convert the videos into images by
using the Dynamic Image encoding of [3] which summarizes each video into an image based on a
ranking principle.
Challenge and evaluation. Each dataset Dd , d = 1, . . . , 10 is formed of pairs (x, y) ? Dd where x
is an image and y ? {1, . . . , Cd } = Yd is a label. For each dataset, we specify a training, validation
and test subsets. The goal is to train the best possible model to address all ten classification tasks
using only the provided training and validation data (no external data is allowed). A model ? is
evaluated on the test data, where, given an image x and its ground-truth domain dx label, it has to
predict the corresponding label y = ?(x, dx ) ? Yd .
Performance is measured in terms of a single scalar score S determined as in the decathlon discipline.
Performing well at this metric requires algorithms to perform well in all tasks, compared to a
minimum level of baseline performance for each. In detail, S is computed as follows:
S=
10
X
?d max{0, Edmax ? Ed }?d ,
Ed =
d=1
1
|Ddtest |
X
1{y6=?(x,d)} .
(1)
(x,y)?Ddtest
where Ed is the average test error for each domain. Edmax the baseline error (section 5), above which
no points are scored. The exponent ?d ? 1 rewards more reductions of the classification error
as this becomes close to zero and is set to ?d = 2 for all domains. The coefficient ?d is set to
1, 000 (Edmax )??d so that a perfect result receives a score of 1,000 (10,000 in total).
Data preprocessing. Different domains contain a different set of image classes as well as a different
number of images. In order to reduce the computational burden, all images have been resized
isotropically to have a shorter side of 72 pixels. For some datasets such as ImageNet, this is a
substantial reduction in resolution which makes training models much faster (but still sufficient to
obtain excellent classification results with baseline models). For the datasets for which there exists
training, validation, and test subsets, we keep the original splits. For the rest, we use 60%, 20% and
20% of the data for training, validation, and test respectively. For the ILSVRC12, since the test labels
are not available, we use the original validation subset as the test subset and randomly sample a new
validation set from their training split. We are planning to make the data and an evaluation server
public soon.
5
Experiments
In this section we evaluate our method quantitatively against several baselines (section 5.1), investigate
the ability of the proposed techniques to learn models for ten very diverse visual domains.
Implementation details. In all experiments we choose to use the powerful ResNets [13] as base
architectures due to their remarkable performance. In particular, as a compromise of accuracy and
speed, we chose the ResNet28 model [40] which consists of three blocks of four residual units. Each
residual unit contains 3 ? 3 convolutional, BN and ReLU modules (fig. 2). The network accepts
64 ? 64 images as input, downscales the spatial dimensions by two at each block and ends with a
global average pooling and a classifier layer followed by a softmax. We set the number of filters to
64, 128, 256 for these blocks respectively. Each network is optimized to minimize its cross-entropy
loss with stochastic gradient descent. The network is run for 80 epochs and the initial learning rate of
0.1 is lowered to 0.01 and then 0.001 gradually.
6
Model
#par. ImNet Airc. C100 DPed DTD GTSR Flwr
1.3m
# images
7k
50k
30k
4k
40k
OGlt SVHN UCF
2k
26k
70k
mean
S
9k
Scratch
Scratch+
10? 59.87 57.10 75.73 91.20 37.77 96.55 56.30 88.74 96.63 43.27 70.32 1625
11? 59.67 59.59 76.08 92.45 39.63 96.90 56.66 88.74 96.78 44.17 71.07 1826
Feature extractor
Finetune
LwF [21]
1? 59.67 23.31 63.11 80.33 45.37 68.16 73.69 58.79 43.54 26.80 54.28 544
10? 59.87 60.34 82.12 92.82 55.53 97.53 81.41 87.69 96.55 51.20 76.51 2500
10? 59.87 61.15 82.23 92.34 58.83 97.57 83.05 88.08 96.10 50.04 76.93 2515
BN adapt. [5]
? 1? 59.87 43.05
2? 59.67 56.68
Res. adapt.
2? 59.67 61.87
Res. adapt. decay
Res. adapt. finetune all
2? 59.23 63.73
Res. adapt. dom-pred
Res. adapt. (large)
78.62
81.20
81.20
81.31
92.07
93.88
93.88
93.30
51.60
50.85
57.13
57.02
95.82
97.05
97.57
97.47
74.14
66.24
81.67
83.43
84.83
89.62
89.62
89.82
94.10
96.13
96.13
96.17
43.51
47.45
50.12
50.28
71.76
73.88
76.89
77.17
1363
2118
2621
2643
2.5? 59.18 63.52 81.12 93.29 54.93 97.20 82.29 89.82 95.99 50.10 76.74 2503
? 12? 67.00 67.69 84.69 94.28 59.41 97.43 84.86 89.92 96.59 52.39 79.43 3131
Table 1: Multiple-domain networks. The figure reports the (top-1) classification accuracy (%) of
different models on the decathlon tasks and final decathlon score (S). ImageNet is used to prime the
network in every case, except for the networks trained from scratch. The model size is the number of
parameters w.r.t. the baseline ResNet. The fully-finetuned model, written blue, is used as a baseline
to compute the decathlon score.
Model
Airc.
C100
DPed
1.1 60.3 3.6 63.1 0.6
Finetune
4.1 61.1 21.0 82.2 23.8
LwF [21] high lr
38.0 50.6 33.0 80.7 53.3
LwF [21] low lr
Res. adapt. finetune all 59.2 63.7 59.2 81.3 59.2
DTD
GTSR
80.3 0.7 45.3 1.4
92.3 36.7 58.8 11.5
92.2 47.0 57.2 23.7
93.3 59.2 57.0 59.2
68.1
97.6
96.6
97.5
Flwr
27.2
34.2
45.7
59.2
OGlt
SVHN
UCF
73.6 13.4 87.7 0.2 96.6 5.4 51.2
83.1 3.0 88.1 0.2 96.1 18.6 50.0
75.7 21.0 86.0 13.3 94.8 29.0 44.6
83.4 59.2 89.8 59.2 96.1 59.2 50.3
Table 2: Pairwise forgetting. Each pair of numbers report the top-1 accuracy (%) on the old task
(ImageNet) and a new target task after the network is fully finetuned on the latter. We also show
the performance of LwF when it is finetuned on the new task with a high and low learning rate,
trading-off forgetting ImageNet and improving the results on the target domain. By comparison, we
show the performance of tuning only the residual adapters, which by construction does not result in
any performance loss in ImageNet while still achieving very good performance on each target task.
5.1
Results
There are two possible extremes. The first one is to learn ten independent models, one for each
dataset, and the second one is to learn a single model where all feature extractor parameters are
shared between the ten domains. We evaluate next different approaches to learn such models.
Pairwise learning. In the first experiment (table 1), we start by learning a ResNet model on ImageNet,
and then use different techniques to extend it to the remaining nine tasks, one at a time. Depending
on the method, this may produce an overall model comprising ten ResNet architectures, or just one
ResNet with a few domain-specific parameters; thus we also report the total number of parameters
used, where 1? is the size of a single ResNet (excluding the last classification layer, which can never
be shared).
As baselines, we evaluate four cases: i) learning an individual ResNet model from scratch for each
task, ii) freezing all the parameters of the pre-trained network, using the network as feature extractor
and only learn a linear classifier, iii) standard finetuning and iv) applying a reimplementation of the
LwF technique of [21] that encourages the fine-tuned network to retain the responses of the original
ImageNet model while learning the new task.
In terms of accuracy, learning from scratch performs poorly on small target datasets and, by learning
10 independent models, requires 10? parameters in total. Freezing the ImageNet feature extraction is
very efficient in terms of parameter sharing (1? parameters in total), preserves the original domain
exactly, but generally performs very poorly on the target domain. Full fine-tuning leads to accurate
results both for large and small datasets; however, it also forgets the ImageNet domain substantially
(table 2), so it still requires learning 10 complete ResNet models for good overall performance.
When LwF is run as intended by the original authors [21], is still leads to a noticeable performance
drop on the original task, even when learning just two domains (table 2), particularly if the target
domain is very different from ImageNet (e.g. Omniglot and SVHN). Still, if one chooses a different
trade-off point and allows the method to forget ImageNet more, it can function as a good regularizer
that slightly outperforms vanilla fine-tuning overall (but still resulting in a 10? model).
7
Next, we evaluate the effect of sharing the majority of parameters between tasks, whereas still
allowing a small number of domain-specific parameters to change. First, we consider specializing
only the BN layer scaling and bias parameters, which is equivalent to the approach of [5]. In this
case, less than the 0.1% of the model parameters are domain-specific (for the ten domains, this results
in a model with 1.01? parameters overall). Hence the model is very similar to the one with the
frozen feature extractor; nevertheless, the performances increase very substantially in most cases (e.g.
23.31% ? 43.05% accuracy on Aircraft).
As the next step, we introduce the residual adapter modules, which increase by 11% the number
of parameters per domain, resulting in a 2? model. In the pre-training phase, we first pretrain on
ImageNet the network with the added modules. Then, we freeze the task agnostic parameters and
train the task specific parameters on the different datasets. Differently from vanilla fine-tuning, there
is no forgetting in this setting. While most of the parameters are shared, our method is either close or
better than full fine-tuning. As a further control, we also train 10 models from scratch with the added
parameters (denoted as Scratch+), but do not observe any noticeable performance gain in average,
demonstrating that parameters sharing is highly beneficial. We also contrast learning the adapter
modules with two values of weight decay (0.002 and 0.005) higher than the default 0.0005. These
parameters are obtained after a coarse grid search using cross-validation for each dataset. Using
higher decay significantly improves the performance on smaller datasets such as Flowers, whereas
the smaller decay is best for larger datasets. This shows both the importance and utility of controlling
overfitting in the adaptation process. In practice, there is an almost direct correspondence between
the size of the data and which one of these values to use. The optimal decay can be selected via
validation, but a rough choice can be performed by simply looking at the dataset size.
We also compare to another baseline where we only finetune the last two convolutional layers and
freeze the others, which may be thought to be generic. This amounts to having a network with twice
the number of total parameters in a vanilla ResNet which is equal to our proposed architecture. This
model obtains 64.7% mean accuracy over ten datasets, which is significantly lower than our 73.9%,
likely due to overfitting (controlling overfitting is one of the advantages of our technique).
Furthermore, we also assess the quality of our adapter without residual connections, which corresponds to the low rank filter parametrization of section 3.1; this approach achieves an accuracy of
70.3%, which is worse than our 73.9%. We also observe that this configuration requires notably more
iterations to converge. Hence, the residual architecture for the adapters results in better performances,
better control of overfitting, and a faster convergence.
End-to-end learning. So far, we have shown that our method, by learning only the adapter modules
for each new domain, does not suffer from forgetting. However, for us sequential learning is just a
scalable learning strategy. Here, we also show (table 1) that we can further improve the results by
fine-tuning all the parameters of the network end-to-end on the ten tasks. We do so by sampling a
batch from each dataset in a round robin fashion, allowing each domain to contribute to the shared
parameters. A final pass is done on the adapter modules to take into account the change in the shared
parameters.
Domain prediction. Up to now we assume that the domain of each image is given during test
time for all the methods. If this is unavailable, it can be predicted on the fly by means of a small
neural-network predictor. We train a light ResNet, which is composed three stacks of two residual
networks, half deep as the original net, obtaining 99.8% accuracy in domain prediction, resulting in a
barely noticeable drop in the overall multiple-domain challenge (see Res. adapt dom-pred in table 1).
Note that similar performance drop would be observed for the other baselines.
Decathlon evaluation: overall performance. While so far we have looked at results on individual
domain, the Decathlon score eq. (1) can be used to compare performance overall. As baseline error
rates in eq. (1), we double the error rates of the fully finetuned networks on each task. In this manner,
this 10? model achieves a score of 2,500 points (over 10,000 possible ones, see eq. (1)). The last
column of table 1 reports the scores achieved by the other architectures. As intended, the decathlon
score favors the methods that perform well overall, emphasizes their consistency rather than just their
average accuracy. For instance, although the Res. adapt. model (trained with single decay coefficient
for all domains) performs well in terms of average accuracy (73.88%), its decathlon score (2118) is
relatively low because the model performs poorly in DTD and Flowers. This also shows that, once
the weight decays are configured properly, our model achieves superior performance (2643 points) to
all the baselines using only 2? the capacity of a single ResNet.
8
Finally we show that using a higher capacity ResNet28 (12?, ResNet adapt. (large) in table 1), which
is comparable to 10 independent networks, significantly improves our results and outperforms the
finetuning baseline by 600 point in decathlon score. As a matter of fact, this model outperforms the
state-of-the-art [40] (81.2%) by 3.5 points in CIFAR100. In other cases, our performances are in
general in line to current state-of-the-art methods. When this is not the case, this is due to reduced
image resolution (ImageNet, Flower) or due to the choice of a specific video representation in UCF
(dynamic image).
6
Conclusions
As machine learning applications become more advanced and pervasive, building data representations
that work well for multiple problems will become increasingly important. In this paper, we have
introduced a simple architectural element, the residual adapter module, that allows compressing many
visual domains in relatively small residual networks, with substantial parameter sharing between
them. We have also shown that they allow addressing the forgetting problem, as well as adapting to
target domain for which different amounts of training data are available. Finally, we have introduced
a new multi-domain learning challenge, the Visual Decathlon, to allow a systematic comparison of
algorithms for multiple-domain learning.
Acknowledgments: This work acknowledges the support of Mathworks/DTA DFR02620 and ERC 677195IDIU.
References
[1] A. Argyriou, T. Evgeniou, and M Pontil. Multi-task feature learning. In Proc. NIPS, volume 19,
page 41. MIT; 1998, 2007.
[2] L. Bertinetto, J. F. Henriques, J. Valmadre, P. Torr, and A. Vedaldi. Learning feed-forward
one-shot learners. In Proc. NIPS, pages 523?531, 2016.
[3] H. Bilen, B. Fernando, E. Gavves, A. Vedaldi, and S. Gould. Dynamic image networks for
action recognition. In Proc. CVPR, 2016.
[4] H. Bilen and A. Vedaldi. Integrated perception with recurrent multi-task neural networks. In
Proc. NIPS, 2016.
[5] H. Bilen and A. Vedaldi. Universal representations: The missing link between faces, text,
planktons, and cat breeds. arXiv preprint arXiv:1701.07275, 2017.
[6] R. Caruana. Multitask learning. Machine Learning, 28, 1997.
[7] M. Cimpoi, S. Maji, I. Kokkinos, S. Mohamed, and A. Vedaldi. Describing textures in the wild.
In Proc. CVPR, 2014.
[8] R. Collobert and J. Weston. A unified architecture for natural language processing: Deep neural
networks with multitask learning. In icml, pages 160?167. ACM, 2008.
[9] H. Daum? III. Frustratingly easy domain adaptation. ACL 2007, page 256, 2007.
[10] T. Evgeniou and M. Pontil. Regularized multi?task learning. In SIGKDD, pages 109?117.
ACM, 2004.
[11] R. M. French. Catastrophic forgetting in connectionist networks. Trends in cognitive sciences,
3(4):128?135, 1999.
[12] Y. Ganin and V. Lempitsky. Unsupervised domain adaptation by backpropagation. Proc. ICML,
2015.
[13] K. He, X. Zhang, S. Ren, and J. Sun. Identity mappings in deep residual networks. In Proc.
ECCV, pages 630?645. Springer, 2016.
[14] J. T. Huang, J. Li, D. Yu, L. Deng, and Y. Gong. Cross-language knowledge transfer using
multilingual deep neural network with shared hidden layers. In ICASSP, pages 7304?7308,
2013.
9
[15] S. Ioffe and C. Szegedy. Batch normalization: Accelerating deep network training by reducing
internal covariate shift. CoRR, 2015.
[16] X. Jia, B. De Brabandere, T. Tuytelaars, and L. Gool. Dynamic filter networks. In Proc. NIPS,
pages 667?675, 2016.
[17] J. Kirkpatrick, E. Pascanu, N. Rabinowitz, J. Veness, G. Desjardins, A. A. Rusu, K. Milan,
J. Quan, T. Ramalho, A. Grabska-Barwinska, et al. Overcoming catastrophic forgetting in
neural networks. National Academy of Sciences, 2017.
[18] I. Kokkinos. Ubernet: Training auniversal?convolutional neural network for low-, mid-, and
high-level vision using diverse datasets and limited memory. Proc. CVPR, 2017.
[19] A. Krizhevsky and G. Hinton. Learning multiple layers of features from tiny images. 2009.
[20] B. M. Lake, R. Salakhutdinov, and J. B. Tenenbaum. Human-level concept learning through
probabilistic program induction. Science, 350(6266):1332?1338, 2015.
[21] Z. Li and D. Hoiem. Learning without forgetting. In Proc. ECCV, pages 614?629, 2016.
[22] X. Liu, J. Gao, X. He, L. Deng, K. Duh, and Y. Wang. Representation learning using multi-task
deep neural networks for semantic classification and information retrieval. In HLT-NAACL,
pages 912?921, 2015.
[23] M. Long, H. Zhu, J. Wang, and M. I. Jordan. Unsupervised Domain Adaptation with Residual
Transfer Networks. In Proc. NIPS, pages 136?144, 2016.
[24] S. Maji, J. Kannala, E. Rahtu, M. Blaschko, and A. Vedaldi. Fine-grained visual classification
of aircraft. Technical report, 2013.
[25] T. Mitchell. Never-ending learning. Technical report, DTIC Document, 2010.
[26] S. Munder and D. M. Gavrila. An experimental study on pedestrian classification. PAMI,
28(11):1863?1868, 2006.
[27] Y. Netzer, T. Wang, A. Coates, A. Bissacco, B. Wu, and A. Y. Ng. Reading digits in natural images with unsupervised feature learning. In NIPS Workshop on Deep Learning and
Unsupervised Feature Learning, 2011.
[28] M-E. Nilsback and A. Zisserman. Automated flower classification over a large number of
classes. In ICCVGIP, Dec 2008.
[29] Ali Sharif Razavian, Hossein Azizpour, Josephine Sullivan, and Stefan Carlsson. CNN features
off-the-shelf: an astounding baseline for recognition. In CVPR DeepVision Workshop, 2014.
[30] S. A. Rebuffi, A. Kolesnikov, G. Sperl, and C. H. Lampert. icarl: Incremental classifier and
representation learning. In Proc. CVPR, 2017.
[31] Amir Rosenfeld and John K Tsotsos. Incremental learning through deep adaptation. arXiv
preprint arXiv:1705.04228, 2017.
[32] O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy,
A. Khosla, M. Bernstein, A. C. Berg, and K. Fei-Fei. Imagenet large scale visual recognition
challenge, 2014.
[33] A. A. Rusu, N. C. Rabinowitz, G. Desjardins, H. Soyer, J. Kirkpatrick, K. Kavukcuoglu,
R. Pascanu, and R. Hadsell. Progressive neural networks. arXiv preprint arXiv:1606.04671,
2016.
[34] J. Schmidhuber. Learning to control fast-weight memories: An alternative to dynamic recurrent
networks. Neural Computation, 4(1):131?139, 1992.
[35] K. Soomro, A. R. Zamir, and M. Shah. Ucf101: A dataset of 101 human actions classes from
videos in the wild. arXiv preprint arXiv:1212.0402, 2012.
[36] J. Stallkamp, M. Schlipsing, J. Salmen, and C. Igel. Man vs. computer: Benchmarking machine
learning algorithms for traffic sign recognition. Neural Networks, 32(0):323?332, 2012.
[37] A. V. Terekhov, G. Montone, and J. K. O?Regan. Knowledge transfer in deep block-modular
neural networks. In Biomimetic and Biohybrid Systems, pages 268?279, 2015.
10
[38] S. Thrun. Lifelong learning algorithms. In Learning to learn, pages 181?209. Springer, 1998.
[39] E. Tzeng, J. Hoffman, T. Darrell, and K. Saenko. Simultaneous deep transfer across domains
and tasks. In Proc. CVPR, pages 4068?4076, 2015.
[40] Sergey Zagoruyko and Nikos Komodakis.
arXiv:1605.07146, 2016.
Wide residual networks.
arXiv preprint
[41] T. Zhang, B. Ghanem, S. Liu, and N. Ahuja. Robust visual tracking via structured multi-task
sparse learning. IJCV, 101(2):367?383, 2013.
[42] Z. Zhang, P. Luo, C. C. Loy, and X. Tang. Facial landmark detection by deep multi-task learning.
In Proc. ECCV, 2014.
11
| 6654 |@word aircraft:6 multitask:2 cnn:1 kokkinos:2 rgb:1 bn:11 decomposition:3 shot:2 reduction:2 initial:1 configuration:2 contains:9 liu:3 selecting:1 score:10 hoiem:1 tuned:3 document:1 outperforms:3 existing:1 current:1 wd:1 luo:1 dx:3 reminiscent:1 must:1 written:1 john:1 gavves:1 realistic:1 designed:1 drop:3 v:5 alone:1 half:1 selected:2 amir:1 parametrization:4 vanishing:1 bissacco:1 lr:2 detecting:1 provides:2 coarse:1 contribute:1 pascanu:2 zhang:3 constructed:1 direct:1 become:2 consists:4 ijcv:1 combine:2 fitting:1 wild:2 manner:2 introduce:8 pairwise:2 notably:1 forgetting:20 andrea:1 learnet:1 themselves:1 growing:2 multi:12 planning:1 inspired:2 salakhutdinov:1 munder:1 becomes:1 provided:2 blaschko:1 linearity:1 agnostic:9 grabska:1 interpreted:1 substantially:3 unified:1 finding:1 every:1 tackle:1 exactly:4 classifier:5 uk:2 control:3 unit:2 segmenting:1 before:2 modify:1 tends:1 limit:1 soyer:1 encoding:1 oxford:1 yd:3 approximately:1 pami:1 chose:1 twice:1 initialization:1 studied:1 acl:1 dynamically:1 limited:1 igel:1 directed:1 acknowledgment:1 union:1 block:5 reimplementation:1 differs:1 practice:1 backpropagation:1 bootstrap:1 digit:3 sullivan:1 pontil:2 area:1 universal:2 adapting:2 vedaldi:7 ucf101:3 significantly:4 pre:3 thought:1 close:3 unlabeled:1 operator:1 applying:2 equivalent:2 missing:1 maximizing:2 primitive:1 go:1 focused:1 resolution:3 hadsell:1 bertinetto:1 variation:1 analogous:1 cifar100:2 construction:4 target:13 gm:1 controlling:3 element:1 trend:1 recognition:11 particularly:3 utilized:1 expensive:1 finetuned:4 imnet:2 labeled:1 database:1 biomimetic:1 observed:2 role:1 module:28 fly:3 inserted:1 preprint:5 capture:1 parameterize:1 wang:3 sharif:1 zamir:1 compressing:1 sun:1 trade:1 substantial:2 reward:3 dynamic:6 dom:2 trained:4 segment:1 compromise:1 ali:1 learner:1 basis:2 compactly:1 packing:1 icassp:1 finetuning:2 differently:1 cat:2 various:1 regularizer:1 maji:2 alphabet:1 train:6 separated:1 distinct:1 ramalho:1 effective:1 fast:1 choosing:1 quite:1 whose:2 larger:1 solve:1 cvpr:6 modular:1 ability:5 favor:1 g1:1 tuytelaars:1 breed:3 jointly:2 rosenfeld:1 final:2 advantage:4 frozen:1 net:1 propose:4 adaptation:16 parametrizations:1 poorly:3 achieve:1 academy:1 normalize:1 milan:1 convergence:1 double:1 requirement:1 rhf:1 darrell:1 produce:2 perfect:1 leave:1 incremental:2 object:8 resnet:16 tk:2 develop:3 ac:1 gong:1 depending:2 recurrent:2 measured:2 ganin:1 school:1 noticeable:3 eq:3 strong:1 auxiliary:2 predicted:3 trading:1 hakan:1 filter:21 stochastic:1 human:3 public:1 require:1 dnns:2 investigation:1 secondly:2 extension:1 around:1 considered:1 ground:2 mapping:2 predict:5 major:1 achieves:4 desjardins:2 proc:14 applicable:1 label:6 individually:1 largest:1 successfully:2 hoffman:1 stefan:1 rough:1 concurrently:1 mit:1 aim:1 rather:2 shelf:1 resized:2 rusu:2 azizpour:1 pervasive:1 focus:2 properly:1 rank:3 pretrain:1 contrast:2 sigkdd:1 baseline:14 sense:1 wf:2 dependent:2 typically:2 integrated:1 associativity:1 hidden:2 dnn:6 interested:1 comprising:1 pixel:2 overall:8 classification:16 among:1 hossein:1 denoted:1 exponent:1 retaining:1 animal:1 spatial:1 special:1 tzeng:1 initialize:1 fairly:1 softmax:2 construct:3 evgeniou:2 never:3 beach:1 extraction:1 having:1 sampling:1 y6:1 equal:1 look:5 unsupervised:4 icml:2 yu:1 progressive:1 discrepancy:1 others:2 report:6 quantitatively:1 connectionist:1 few:2 primarily:1 modern:1 randomly:1 composed:1 simultaneously:5 preserve:3 airbus:1 individual:2 national:1 geometry:1 consisting:2 intended:2 phase:1 astounding:1 maintain:1 detection:3 interest:2 once:1 possibility:1 investigate:1 highly:1 evaluation:5 bilen:3 kirkpatrick:2 extreme:1 light:1 chain:1 accurate:1 netzer:1 shorter:1 facial:2 indexed:1 iv:1 old:3 re:8 instance:2 column:2 earlier:2 schlipsing:1 caruana:2 cost:1 introducing:3 addressing:1 subset:4 predictor:2 krizhevsky:1 reported:1 combined:1 thoroughly:1 st:1 chunk:1 chooses:1 retain:2 systematic:1 off:3 informatics:1 probabilistic:1 discipline:2 quickly:1 w1:5 huang:3 choose:1 worse:1 external:1 cognitive:1 steered:1 style:3 rescaling:1 li:2 szegedy:1 account:1 potential:1 valmadre:1 de:1 availability:1 coefficient:7 pedestrian:5 configured:3 matter:1 ranking:2 collobert:2 later:1 try:1 view:1 performed:1 razavian:1 traffic:4 start:1 hf:1 jia:1 contribution:3 ass:2 formed:1 minimize:1 accuracy:11 convolutional:11 largely:1 conceptually:1 handwritten:1 kavukcuoglu:1 emphasizes:1 ren:1 russakovsky:1 veness:1 simultaneous:1 suffers:1 sharing:10 dslr:1 ed:3 hlt:1 evaluates:1 against:1 mohamed:1 galaxy:1 naturally:1 stop:1 gain:2 tunable:2 dataset:17 popular:1 mitchell:1 recall:2 knowledge:5 improves:2 organized:1 subtle:1 back:1 finetune:5 feed:1 higher:3 supervised:1 mtl:3 response:2 improved:1 specify:1 zisserman:1 formulation:1 evaluated:2 ox:1 box:1 though:1 furthermore:4 just:5 done:2 sketch:1 receives:1 web:1 touch:1 freezing:2 su:1 propagation:1 french:1 quality:1 rabinowitz:2 building:1 glyph:2 usa:1 contain:6 sylvestre:1 facilitate:1 effect:1 concept:1 hence:5 regularization:2 naacl:1 semantic:3 round:1 komodakis:1 during:1 encourages:1 transferable:1 noted:2 complete:1 performs:5 svhn:6 dtd:4 image:43 ranging:1 superior:2 common:2 empirically:1 volume:1 million:2 discussed:1 extend:1 he:2 refer:1 significant:1 freeze:3 composition:1 cv:1 tuning:8 vanilla:3 fk:3 grid:1 consistency:1 inclusion:1 omniglot:3 erc:1 language:5 lowered:1 robot:1 ucf:3 operating:1 etc:2 gt:2 add:1 base:1 own:1 recent:6 showed:1 prime:1 schmidhuber:1 certain:3 server:1 wv:1 life:4 kolesnikov:1 preserving:1 seen:1 additional:3 minimum:1 nikos:1 deng:3 converge:1 paradigm:1 maximize:1 fernando:1 semi:1 ii:1 multiple:19 full:2 reduces:1 barwinska:1 faster:2 adapt:11 technical:2 cross:3 long:6 cifar:1 retrieval:2 specializing:1 prediction:4 variant:2 scalable:1 vision:4 metric:3 nilsback:1 arxiv:10 sergey:1 resnets:1 represent:1 kernel:1 normalization:5 sometimes:2 adopting:1 iteration:1 achieved:1 whereas:3 fine:11 addition:1 cropped:2 dec:1 krause:1 source:2 extra:2 w2:5 rest:1 zagoruyko:1 sure:1 pooling:1 gavrila:1 quan:1 contrary:1 jordan:1 call:1 bernstein:1 intermediate:1 easy:2 concerned:2 split:3 variety:1 adapter:26 relu:2 iii:2 automated:1 architecture:18 reduce:3 idea:4 vgg:1 shift:3 whether:1 icarl:1 utility:1 colour:1 accelerating:1 soomro:1 suffer:1 speech:2 nine:1 action:6 deep:21 useful:2 generally:1 tune:1 karpathy:1 amount:4 mid:3 ten:13 tenenbaum:1 daimler:2 category:7 simplest:1 reduced:1 exist:1 coates:1 sign:5 per:1 blue:3 diverse:7 write:1 express:1 group:1 key:1 putting:1 four:2 nevertheless:2 demonstrating:1 achieving:1 mono:1 changing:1 vast:2 fraction:1 convert:1 tsotsos:1 run:2 ilsvrc12:2 powerful:1 salmen:1 named:1 throughout:1 family:1 almost:1 architectural:1 wu:1 lake:1 summarizes:1 scaling:3 comparable:1 layer:22 internet:2 followed:2 distinguish:1 correspondence:1 quadratic:1 refine:1 constraint:1 fei:2 tag:2 aspect:3 speed:1 performing:3 relatively:2 gould:1 structured:1 developing:2 according:1 combination:1 poor:1 dta:1 beneficial:2 across:2 slightly:2 character:4 smaller:2 increasingly:1 describable:3 making:1 intuitively:1 gradually:1 remains:1 turn:1 german:2 describing:1 mathworks:1 needed:1 end:6 parametrize:1 operation:2 available:2 observe:2 generic:2 batch:5 alternative:1 shah:1 original:14 denotes:1 top:2 include:2 cf:2 remaining:1 maintaining:1 daum:1 build:1 rebuffi:1 webcam:1 tensor:5 added:3 looked:1 parametric:6 primary:2 rt:1 strategy:2 gradient:1 link:1 duh:1 thrun:1 entity:1 landmark:2 capacity:4 aptly:1 majority:2 w0:2 tower:2 street:1 collected:1 gallery:1 barely:1 induction:1 minimizing:1 art:2 loy:1 setup:1 boeing:1 design:1 implementation:2 satheesh:1 perform:6 plankton:2 allowing:2 observation:1 convolution:3 datasets:11 benchmark:9 enabling:1 rhv:1 parametrizing:1 descent:1 hinton:1 incorporated:1 excluding:1 looking:1 stack:1 rahtu:1 hbilen:1 overcoming:1 pred:2 introduced:2 dog:1 required:1 pair:2 optimized:1 imagenet:18 connection:1 learned:5 accepts:1 nip:7 address:5 beyond:1 suggested:2 flower:6 perception:2 pattern:1 below:1 usually:2 reading:1 challenge:13 program:1 including:3 max:2 video:6 memory:2 gool:1 difficulty:1 natural:3 hybrid:1 regularized:1 indicator:1 residual:31 advanced:1 zhu:1 older:2 lwf:6 improve:1 acknowledges:1 extract:1 text:1 epoch:1 literature:1 carlsson:1 loss:2 par:1 fully:3 regan:1 filtering:1 remarkable:1 ghanem:1 validation:8 h2:3 multivalent:3 degree:3 sufficient:3 principle:1 dd:2 bank:10 tiny:1 share:3 cd:1 eccv:3 compatible:1 changed:1 summary:1 last:5 keeping:1 soon:1 henriques:1 bias:5 allow:3 side:1 lifelong:1 wide:1 taking:1 face:1 sparse:1 edinburgh:1 boundary:1 dimension:2 default:1 ending:2 world:1 contour:1 author:3 collection:1 adaptive:1 concretely:1 domainspecific:1 preprocessing:1 forward:1 far:4 josephine:1 compact:2 obtains:1 absorb:1 multilingual:2 synergy:1 confirm:1 keep:1 global:1 sequentially:4 overfitting:5 cimpoi:1 ioffe:1 grayscale:1 search:2 shareable:1 khosla:1 frustratingly:1 robin:1 table:9 learn:21 transfer:6 robust:1 ca:1 channel:3 obtaining:1 improving:4 unavailable:1 hc:1 excellent:1 domain:119 pk:2 main:2 linearly:1 rh:2 bounding:1 scored:1 lampert:1 allowed:1 fig:8 representative:2 referred:1 benchmarking:1 ng:1 fashion:1 ahuja:1 shrinking:1 sub:1 comprises:1 candidate:1 house:1 forgets:1 extractor:7 reconfigure:1 grained:2 tang:1 sperl:1 specific:18 covariate:1 brabandere:1 learnable:1 list:1 decay:7 concern:1 workshop:2 consist:1 burden:1 exists:1 sequential:6 adding:2 importance:1 corr:1 texture:5 dtic:1 easier:1 entropy:1 forget:2 simply:1 explore:2 likely:1 gao:1 absorbed:1 visual:27 expressed:1 contained:1 tracking:2 scalar:1 isotropically:1 springer:2 corresponds:1 truth:2 stallkamp:1 acm:2 ma:1 weston:1 lempitsky:1 modulate:1 goal:5 identity:3 shared:9 man:1 content:2 change:6 youtube:1 telephone:1 specifically:1 uniformly:1 reducing:2 determined:1 except:1 torr:1 called:1 total:5 pas:1 catastrophic:3 experimental:1 saenko:1 occluding:1 formally:1 ilsvrc:1 internal:1 support:1 berg:1 latter:7 avoiding:1 phenomenon:1 incorporate:1 evaluate:5 argyriou:1 scratch:8 |
6,251 | 6,655 | Dykstra?s Algorithm, ADMM, and Coordinate
Descent: Connections, Insights, and Extensions
Ryan J. Tibshirani
Department of Statistics and Machine Learning Department
Carnegie Mellon University
Pittsburgh, PA 15213
[email protected]
Abstract
We study connections between Dykstra?s algorithm for projecting onto an intersection of convex sets, the augmented Lagrangian method of multipliers or ADMM,
and block coordinate descent. We prove that coordinate descent for a regularized
regression problem, in which the penalty is a separable sum of support functions,
is exactly equivalent to Dykstra?s algorithm applied to the dual problem. ADMM
on the dual problem is also seen to be equivalent, in the special case of two sets,
with one being a linear subspace. These connections, aside from being interesting
in their own right, suggest new ways of analyzing and extending coordinate descent. For example, from existing convergence theory on Dykstra?s algorithm over
polyhedra, we discern that coordinate descent for the lasso problem converges at
an (asymptotically) linear rate. We also develop two parallel versions of coordinate
descent, based on the Dykstra and ADMM connections.
1
Introduction
In this paper, we study two seemingly unrelated but closely connected convex optimization problems,
and associated algorithms. The first is the best approximation problem: given closed, convex sets
C1 , . . . , Cd ? Rn and y ? Rn , we seek the point in C1 ? ? ? ? ? Cd (assumed nonempty) closest to y,
and solve
minn ky ? uk22 subject to u ? C1 ? ? ? ? ? Cd .
(1)
u?R
The second problem is the regularized regression problem: given a response y ? Rn and predictors
X ? Rn?p , and a block decomposition Xi ? Rn?pi , i = 1, . . . , d of the columns of X (i.e., these
could be columns, or groups of columns), we build a working linear model by applying blockwise
regularization over the coefficients, and solve
minp
w?R
d
X
1
ky ? Xwk22 +
hi (wi ),
2
i=1
(2)
where hi : Rpi ? R, i = 1, . . . , d are convex functions, and we write wi ? Rpi , P
i = 1, . . . , d for the
d
appropriate block decomposition of a coefficient vector w ? Rp (so that Xw = i=1 Xi wi ).
Two well-studied algorithms for problems (1), (2) are Dykstra?s algorithm (Dykstra, 1983; Boyle and
Dykstra, 1986) and (block) coordinate descent (Warga, 1963; Bertsekas and Tsitsiklis, 1989; Tseng,
1990), respectively. The jumping-off point for our work in this paper is the following fact: these two
algorithms are equivalent for solving (1) and (2). That is, for a particular relationship between the
sets C1 , . . . , Cd and penalty functions h1 , . . . , hd , the problems (1) and (2) are duals of each other,
and Dykstra?s algorithm on the primal problem (1) is exactly the same as coordinate descent on the
dual problem (2). We provide details in Section 2.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
This equivalence between Dykstra?s algorithm and coordinate descent can be essentially found in
the optimization literature, dating back to the late 1980s, and possibly earlier. (We say ?essentially?
here because, to our knowledge, this equivalence has not been stated for a general regression matrix
X, and only in the special case X = I; but, in truth, the extension to a general matrix X is fairly
straightforward.) Though this equivalence has been cited and discussed in various ways over the
years, we feel that it is not as well-known as it should be, especially in light of the recent resurgence
of interest in coordinate descent methods. We revisit the connection between Dykstra?s algorithm
and coordinate descent, and draw further connections to a third method?the augmented Lagrangian
method of multipliers or ADMM (Glowinski and Marroco, 1975; Gabay and Mercier, 1976)?that
has also received a great deal of attention recently. While these basic connections are interesting in
their own right, they also have important implications for analyzing and extending coordinate descent.
Below we give a summary of our contributions.
1. We prove in Section 2 (under a particular relationship between C1 , . . . , Cd and h1 , . . . , hd )
that Dykstra?s algorithm for (1) is equivalent to block coordinate descent for (2). (This is a
mild generalization of the previously known connection when X = I.)
2. We also show in Section 2 that ADMM is closely connected to Dykstra?s algorithm, in that
ADMM for (1), when d = 2 and C1 is a linear subspace, matches Dykstra?s algorithm.
3. Leveraging existing results on the convergence of Dykstra?s algorithm for an intersection of
halfspaces, we establish in Section 3 that coordinate descent for the lasso problem has an
(asymptotically) linear rate of convergence, regardless of the dimensions of X (i.e., without
assumptions about strong convexity of the problem). We derive two different explicit forms
for the error constant, which shed light onto how correlations among the predictor variables
affect the speed of convergence.
4. Appealing to parallel versions of Dykstra?s algorithm and ADMM, we present in Section 4
two parallel versions of coordinate descent (each guaranteed to converge in full generality).
5. We extend in Section 5 the equivalence between coordinate descent and Dykstra?s algorithm
to the case of nonquadratic loss in (2), i.e., non-Euclidean projection in (1). This leads to a
Dykstra-based parallel version of coordinate descent for (separably regularized) problems
with nonquadratic loss, and we also derive an alternative ADMM-based parallel version of
coordinate descent for the same class of problems.
2
Preliminaries and connections
Dykstra?s algorithm. Dykstra?s algorithm was first proposed by Dykstra (1983), and was extended
to Hilbert spaces by Boyle and Dykstra (1986). Since these seminal papers, a number of works have
analyzed and extended Dykstra?s algorithm in various interesting ways. We will reference many of
these works in the coming sections, when we discuss connections between Dykstra?s algorithm and
other methods; for other developments, see the comprehensive books Deutsch (2001); Bauschke and
Combettes (2011) and review article Bauschke and Koch (2013).
Dykstra?s algorithm for the best approximation problem (1) can be described as follows. We initialize
u(0) = y, z (?d+1) = ? ? ? = z (0) = 0, and then repeat, for k = 1, 2, 3, . . .:
u(k) = PC[k] (u(k?1) + z (k?d) ),
z (k) = u(k?1) + z (k?d) ? u(k) ,
(3)
where PC (x) = argminc?C kx ? ck22 denotes the (Euclidean) projection of x onto a closed, convex
set C, and [?] denotes the modulo operator taking values in {1, . . . , d}. What differentiates Dykstra?s
algorithm from the classical alternating projections method of von Neumann (1950); Halperin (1962)
is the sequence of (what we may call) dual variables z (k) , k = 1, 2, 3, . . .. These track, in a cyclic
fashion, the residuals from projecting onto C1 , . . . , Cd . The simpler alternating projections method
will always converge to a feasible point in C1 ? ? ? ? ? Cd , but will not necessarily converge to the
solution in (1) unless C1 , . . . , Cd are subspaces (in which case alternating projections and Dykstra?s
algorithm coincide). Meanwhile, Dykstra?s algorithm converges in general (for any closed, convex
sets C1 , . . . , Cd with nonempty intersection, see, e.g., Boyle and Dykstra (1986); Han (1988); Gaffke
and Mathar (1989)). We note that Dykstra?s algorithm (3) can be rewritten in a different form, which
2
(0)
(0)
(0)
will be helpful for future comparisons. First, we initialize ud = y, z1 = ? ? ? = zd = 0, and then
repeat, for k = 1, 2, 3, . . .:
(k)
(k?1)
u0 = ud
(k)
ui
(k)
zi
,
(k)
(k?1)
= PCi (ui?1 + zi
=
(k)
ui?1
+
(k?1)
zi
?
)
),
(k)
ui ,
for i = 1, . . . , d.
(4)
Coordinate descent. Coordinate descent methods have a long history in optimization, and have
been studied and discussed in early papers and books such as Warga (1963); Ortega and Rheinboldt
(1970); Luenberger (1973); Auslender (1976); Bertsekas and Tsitsiklis (1989), though coordinate
descent was still likely in use much earlier. (Of course, for solving linear systems, coordinate descent
reduces to Gauss-Seidel iterations, which dates back to the 1800s.) Some key papers analyzing the
convergence of coordinate descent methods are Tseng and Bertsekas (1987); Tseng (1990); Luo and
Tseng (1992, 1993); Tseng (2001). In the last 10 or 15 years, a considerable interest in coordinate
descent has developed across the optimization community. With the flurry of recent work, it would be
difficult to give a thorough account of the recent progress on the topic. To give just a few examples,
recent developments include finite-time (nonasymptotic) convergence rates for coordinate descent,
and exciting extensions such as accelerated, parallel, and distributed versions of coordinate descent.
We refer to Wright (2015), an excellent survey that describes this recent progress.
In (block) coordinate descent1 for (2), we initialize say w(0) = 0, and repeat, for k = 1, 2, 3, . . .:
2
X
X
1
(k)
(k)
(k?1)
wi = argmin
y
?
X
w
?
X
w
?
X
w
i = 1, . . . , d. (5)
j j
j j
i i
+ hi (wi ),
wi ?Rpi 2
2
j<i
j>i
We assume here and throughout that Xi ? Rn?pi , i = 1, . . . , d each have full column rank so that
the updates in (5) are uniquely defined (this is used for convenience, and is not a strong assumption;
note that there is no restriction on the dimensionality of the full problem in (2), i.e., we could still
have X ? Rn?p with p n). The precise form of these updates, of course, depends on the penalty
functions. Suppose that each hi is the support function of a closed, convex set Di ? Rpi , i.e.,
hi (v) = max hd, vi,
d?Di
for i = 1, . . . , d.
Suppose also that Ci = (XiT )?1 (Di ) = {v ? Rn : XiT v ? Di }, the inverse image of Di under the
linear map XiT , for i = 1, . . . , d. Then, perhaps surprisingly, it turns out that the coordinate descent
iterations (5) are exactly the same as the Dykstra iterations (4), via a duality argument. We extract
the key relationship as a lemma below, for future reference, and then state the formal equivalence.
Proofs of these results, as with all results in this paper, are given in the supplement.
Lemma 1. Assume that Xi ? Rn?pi has full column rank and hi (v) = maxd?Di hd, vi for a closed,
convex set Di ? Rpi . Then for Ci = (XiT )?1 (Di ) ? Rn and any b ? Rn ,
w
?i = argmin
wi ?Rpi
1
kb ? Xi wi k22 + hi (wi ) ?? Xi w
?i = (Id ? PCi )(b).
2
where Id(?) denotes the identity mapping.
Theorem 1. Assume the setup in Lemma 1, for each i = 1, . . . , d. Then problems (1), (2) are dual to
each other, and their solutions, denoted u
?, w,
? respectively, satisfy u
? = y ? X w.
? Further, Dykstra?s
algorithm (4) and coordinate descent (5) are equivalent, and satisfy at all iterations k = 1, 2, 3, . . .:
X
X
(k)
(k)
(k)
(k)
(k?1)
zi = Xi wi
and ui = y ?
Xj wj ?
Xj wj
, for i = 1, . . . , d.
j>i
j?i
The equivalence between coordinate descent and Dykstra?s algorithm dates back to (at least) Han
(1988); Gaffke and Mathar (1989), under the special case X = I. In fact, Han (1988), presumably
unaware of Dykstra?s algorithm, seems to have reinvented the method and established convergence
1
To be precise, this is cyclic coordinate descent, where exact minimization is performed along each block of
coordinates. Randomized versions of this algorithm have recently become popular, as have inexact or proximal
versions. While these variants are interesting, they are not the focus of our paper.
3
through its relationship to coordinate descent. This work then inspired Tseng (1993) (who must have
also been unaware of Dykstra?s algorithm) to improve the existing analyses of coordinate descent,
which at the time all assumed smoothness of the objective function. (Tseng continued on to become
arguably the single most important contributor to the theory of coordinate descent of the 1990s and
2000s, and his seminal work Tseng (2001) is still one of the most comprehensive analyses to date.)
References to this equivalence can be found speckled throughout the literature on Dykstra?s method,
but given the importance of the regularized problem form (2) for modern statistical and machine
learning estimation tasks, we feel that the connection between Dykstra?s algorithm and coordinate
descent and is not well-known enough and should be better explored. In what follows, we show that
some old work on Dykstra?s algorithm, fed through this equivalence, yields new convergence results
for coordinate descent for the lasso and a new parallel version of coordinate descent.
ADMM. The augmented Lagrangian method of multipliers or ADMM was invented by Glowinski
and Marroco (1975); Gabay and Mercier (1976). ADMM is a member of a class of methods generally
called operator splitting techniques, and is equivalent (via a duality argument) to Douglas-Rachford
splitting (Douglas and Rachford, 1956; Lions and Mercier, 1979). Recently, there has been a strong
revival of interest in ADMM (and operator splitting techniques in general), arguably due (at least in
part) to the popular monograph of Boyd et al. (2011), where it is argued that the ADMM framework
offers an appealing flexibility in algorithm design, which permits parallelization in many nontrivial
situations. As with coordinate descent, it would be difficult thoroughly describe recent developments
on ADMM, given the magnitude and pace of the literature on this topic. To give just a few examples,
recent progress includes finite-time linear convergence rates for ADMM (see Nishihara et al. 2015;
Hong and Luo 2017 and references therein), and accelerated extensions of ADMM (see Goldstein
et al. 2014; Kadkhodaie et al. 2015 and references therein).
To derive an ADMM algorithm for (1), we introduce auxiliary variables and equality constraints to
put the problem in a suitable ADMM form. While different formulations for the auxiliary variables
and constraints give rise to different algorithms, loosely speaking, these algorithms generally take on
similar forms to Dykstra?s algorithm for (1). The same is also true of ADMM for the set intersection
problem, a simpler task than the best approximation problem (1), in which we only seek a point in
the intersection C1 ? ? ? ? ? Cd , and solve
d
X
minn
ICi (ui ),
(6)
u?R
i=1
where IC (?) denotes the indicator function of a set C (equal to 0 on C, and ? otherwise). Consider
the case of d = 2 sets, in which case the translation of (6) into ADMM form is unambiguous. ADMM
for (6), properly initialized, appears highly similar to Dykstra?s algorithm for (1); so similar, in fact,
that Boyd et al. (2011) mistook the two algorithms for being equivalent, which is not generally true,
and was shortly thereafter corrected by Bauschke and Koch (2013).
Below we show that when d = 2, C1 is a linear subspace, and y ? C1 , an ADMM algorithm for (1)
(and not the simpler set intersection problem (6)) is indeed equivalent to Dykstra?s algorithm for (1).
Introducing auxiliary variables, the problem (1) becomes
min n ky ? u1 k22 + IC1 (u1 ) + IC2 (u2 ) subject to u1 = u2 .
u1 ,u2 ?R
The augmented Lagrangian is L(u1 , u2 , z) = ky ? u1 k22 + IC1 (u1 ) + IC2 (u2 ) + ?ku1 ? u2 + zk22
??kzk22 , where ? > 0 is an augmented Lagrangian parameter. ADMM repeats, for k = 1, 2, 3, . . .:
(k?1)
? z (k?1) )
y
?(u2
(k)
u1 = PC1
+
,
1+?
1+?
(7)
(k)
(k)
u2 = PC2 (u1 + z (k?1) ),
(k)
(k)
z (k) = z (k?1) + u1 ? u2 .
Suppose we initialize
= y, z (0) = 0, and set ? = 1. Using linearity of PC1 , the fact that y ? C1 ,
and a simple inductive argument, the above iterations can be rewritten as
(0)
u2
(k)
(k?1)
u1 = PC1 (u2
(k)
u2
=
(k)
PC2 (u1
),
+ z (k?1) ),
(k)
(k)
z (k) = z (k?1) + u1 ? u2 ,
4
(8)
which is precisely the same as Dykstra?s iterations (4), once we realize that, due again to linearity of
(k)
PC1 , the sequence z1 , k = 1, 2, 3, . . . in Dykstra?s iterations plays no role and can be ignored.
Though d = 2 sets in (1) may seem like a rather special case, the strategy for parallelization in both
Dykstra?s algorithm and ADMM stems from rewriting a general d-set problem as a 2-set problem, so
the above connection between Dykstra?s algorithm and ADMM can be relevant even for problems
with d > 2, and will reappear in our later discussion of parallel coordinate descent. As a matter of
conceptual interest only, we note that for general d (and no constraints on the sets being subspaces),
Dykstra?s iterations (4) can be viewed as a limiting version of the ADMM iterations either for (1) or
for (6), as we send the augmented Lagrangian parameters to ? or to 0 at particular scalings. See the
supplement for details.
3
Coordinate descent for the lasso
The lasso problem (Tibshirani, 1996; Chen et al., 1998), defined for a tuning parameter ? ? 0 as
1
min ky ? Xwk22 + ?kwk1 ,
(9)
w?Rp 2
is a special case of (2) where the coordinate blocks are of each size 1, so that Xi ? Rn , i = 1, . . . , p
are just the columns of X, and wi ? R, i = 1, . . . , p are the components of w. This problem fits into
the framework of (2) with hi (wi ) = ?|wi | = maxd?Di dwi for Di = [??, ?], for each i = 1, . . . , d.
Coordinate descent is widely-used for the lasso (9), both because of the simplicity of the coordinatewise updates, which reduce to soft-thresholding, and because careful implementations can achieve
state-of-the-art performance, at the right problem sizes. The use of coordinate descent for the lasso
was popularized by Friedman et al. (2007, 2010), but was studied earlier or concurrently by several
others, e.g., Fu (1998); Sardy et al. (2000); Wu and Lange (2008).
As we know from Theorem 1, the dual of problem (9) is the best approximation problem (1), where
Ci = (XiT )?1 (Di ) = {v ? Rn : |XiT v| ? ?} is an intersection of two halfspaces, for i = 1, . . . , p.
This makes C1 ? ? ? ? ? Cd an intersection of 2p halfspaces, i.e., a (centrally symmetric) polyhedron.
For projecting onto a polyhedron, it is well-known that Dykstra?s algorithm reduces to Hildreth?s
algorithm (Hildreth, 1957), an older method for quadratic programming that itself has an interesting
history in optimization. Theorem 1 hence shows coordinate descent for the lasso (9) is equivalent not
only to Dykstra?s algorithm, but also to Hildreth?s algorithm, for (1).
This equivalence suggests a number of interesting directions to consider. For example, key practical
speedups have been developed for coordinate descent for the lasso that enable this method to attain
state-of-the-art performance at the right problem sizes, such as clever updating rules and screening
rules (e.g., Friedman et al. 2010; El Ghaoui et al. 2012; Tibshirani et al. 2012; Wang et al. 2015).
These implementation tricks can now be used with Dykstra?s (Hildreth?s) algorithm. On the flip side,
as we show next, older results from Iusem and De Pierro (1990); Deutsch and Hundal (1994) on
Dykstra?s algorithm for polyhedra, lead to interesting new results on coordinate descent for the lasso.
Theorem 2 (Adaptation of Iusem and De Pierro 1990). Assume the columns of X ? Rn?p are in
general position, and ? > 0. Then coordinate descent for the lasso (9) has an asymptotically linear
convergence rate, in that for large enough k,
1/2
a2
kw(k+1) ? wk
? ?
?
,
(10)
T X )/ max
2
kw(k) ? wk
? ?
a2 + ?min (XA
A
i?A kXi k2
where w
? is the lasso solution in (9), ? = X T X, and kzk2? = z T ?z for z ? Rp , A = supp(w)
? is
the active set of w,
? a = |A| is its size, XA ? Rn?a denotes the columns of X indexed by A, and
T
T
?min (XA
XA ) denotes the smallest eigenvalue of XA
XA .
Theorem 3 (Adaptation of Deutsch and Hundal 1994). Assume the same conditions and notation
as in Theorem 2. Then for large enough k,
!1/2
?
a?1
Y kP{i
Xij k22
kw(k+1) ? wk
? ?
j+1 ,...,ia }
? 1?
,
(11)
kXij k22
kw(k) ? wk
? ?
j=1
?
where we enumerate A = {i1 , . . . , ia }, i1 < . . . < ia , and we denote by P{i
the projection
j+1 ,...,ia }
onto the orthocomplement of the column span of X{ij+1 ,...,ia } .
5
The results in Theorems 2, 3 both rely on the assumption of general position for the columns of X.
This is only used for convenience and can be removed at the expense of more complicated notation.
Loosely put, the general position condition simply rules out trivial linear dependencies between small
numbers of columns of X, but places no restriction on the dimensions of X (i.e., it still allows for
p n). It implies that the lasso solution w
? is unique, and that XA (where A = supp(w))
? has full
column rank. See Tibshirani (2013) for a precise definition of general position and proofs of these
facts. We note that when XA has full column rank, the bounds in (10), (11) are strictly less than 1.
Remark 1 (Comparing (10) and (11)). Clearly, both the bounds in (10), (11) are adversely affected
by correlations among Xi , i ? A (i.e., stronger correlations will bring each closer to 1). It seems to
us that (11) is usually the smaller of the two bounds, based on simple mathematical and numerical
comparisons. More detailed comparisons would be interesting, but is beyond the scope of this paper.
Remark 2 (Linear convergence without strong convexity). One striking feature of the results in
Theorems 2, 3 is that they guarantee (asymptotically) linear convergence of the coordinate descent
iterates for the lasso, with no assumption about strong convexity of the objective. More precisely,
there are no restrictions on the dimensionality of X, so we enjoy linear convergence even without an
assumption on the smooth part of the objective. This is in line with classical results on coordinate
descent for smooth functions, see, e.g., Luo and Tseng (1992). The modern finite-time convergence
analyses of coordinate descent do not, as far as we understand, replicate this remarkable property.
For example, Beck and Tetruashvili (2013); Li et al. (2016) establish finite-time linear convergence
rates for coordinate descent, but require strong convexity of the entire objective.
Remark 3 (Active set identification). The asymptotics developed in Iusem and De Pierro (1990);
Deutsch and Hundal (1994) are based on a notion of (in)active set identification: the critical value of
k after which (10), (11) hold is based on the (provably finite) iteration number at which Dykstra?s
algorithm identifies the inactive halfspaces, i.e., at which coordinate descent identifies the inactive
set of variables, Ac = supp(w)
? c . This might help explain why in practice coordinate descent for the
lasso performs exceptionally well with warm starts, over a decreasing sequence of tuning parameter
values ? (e.g., Friedman et al. 2007, 2010): here, each coordinate descent run is likely to identify the
(in)active set?and hence enter the linear convergence phase?at an early iteration number.
4
Parallel coordinate descent
Parallel-Dykstra-CD. An important consequence of the connection between Dykstra?s algorithm
and coordinate descent is a new parallel version of the latter, stemming from an old parallel version
of the former. A parallel version of Dykstra?s algorithm is usually credited to Iusem and Pierro (1987)
for polyhedra and Gaffke and Mathar (1989) for general sets, but really the idea dates back to the
product space formalization of Pierra (1984). We rewrite problem (1) as
d
X
min
u=(u1 ,...,ud )?Rnd
?i ky ? ui k22
subject to
u ? C0 ? (C1 ? ? ? ? ? Cd ),
(12)
i=1
where C0 = {(u1 , . . . , ud ) ? Rnd : u1 = ? ? ? = ud }, and ?1 , . . . , ?d > 0 are weights that sum to 1.
After rescaling appropriately to turn (12) into an unweighted best approximation problem, we can
(0)
(0)
(0)
(0)
apply Dykstra?s algorithm, which sets u1 = ? ? ? = ud = y, z1 = ? ? ? = zd = 0, and repeats:
(k)
u0 =
d
X
(k?1)
?i ui
,
i=1
(k)
= PCi (u0 + zi
(k)
= u0 + zi
ui
zi
(k)
(k)
(k?1)
(k?1)
),
(k)
(13)
)
for i = 1, . . . , d,
? ui ,
for k = 1, 2, 3, . . .. The steps enclosed in curly brace above can all be performed in parallel, so that
(13) is a parallel version of Dykstra?s algorithm (4) for (1). Applying Lemma 1, and a straightforward
inductive argument, the above algorithm can be rewritten as follows. We set w(0) = 0, and repeat:
2
1
(k)
(k?1)
wi = argmin
y ? Xw(k?1) + Xi wi
/?i ? Xi wi /?i
+ hi (wi /?i ), i = 1, . . . , d, (14)
2
wi ?Rpi 2
for k = 1, 2, 3, . . ., which we call parallel-Dykstra-CD (with CD being short for coordinate descent).
Again, note that the each of the d coordinate updates in (14) can be performed in parallel, so that
6
(14) is a parallel version of coordinate descent (5) for (2). Also, as (14) is just a reparametrization of
Dykstra?s algorithm (13) for the 2-set problem (12), it is guaranteed to converge in full generality, as
per the standard results on Dykstra?s algorithm (Han, 1988; Gaffke and Mathar, 1989).
Theorem 4. Assume that Xi ? Rn?pi has full column rank and hi (v) = maxd?Di hd, vi for a closed,
convex set Di ? Rpi , for i = 1, . . . , d. If (2) has a unique solution, then the iterates in (14) converge
to this solution. More generally, if the interior of ?di=1 (XiT )?1 (Di ) is nonempty, then the sequence
w(k) , k = 1, 2, 3, . . . from (14) has at least one accumulation point, and any such point solves (2).
Further, Xw(k) , k = 1, 2, 3, . . . converges to X w,
? the optimal fitted value in (2).
There have been many recent exciting contributions to the parallel coordinate descent literature; two
standouts are Jaggi et al. (2014); Richtarik and Takac (2016), and numerous others are described in
Wright (2015). What sets parallel-Dykstra-CD apart, perhaps, is its simplicity: convergence of the
iterations (14), given in Theorem 4, just stems from the connection between coordinate descent and
Dykstra?s algorithm, and the fact that the parallel Dykstra iterations (13) are nothing more than the
usual Dykstra iterations after a product space reformulation. Moreover, parallel-Dykstra-CD for the
lasso enjoys an (asymptotic) linear convergence rate under essentially no assumptions, thanks once
again to an old result on the parallel Dykstra (Hildreth) algorithm from Iusem and De Pierro (1990).
The details can be found in the supplement.
Parallel-ADMM-CD. As an alternative to the parallel method derived using Dykstra?s algorithm,
ADMM can also offer a version of parallel coordinate descent. Since (12) is a best approximation
problem with d = 2 sets, we can refer back to our earlier ADMM algorithm in (7) for this problem.
By passing these ADMM iterations through the connection developed in Lemma 1, we arrive at what
(0)
we call parallel-ADMM-CD, which initializes u0 = y, w(?1) = w(0) = 0, and repeats:
Pd
(k?1)
( i=1 ?i )u0
y ? Xw(k?1)
X(w(k?2) ? w(k?1) )
(k)
u0 =
+
+
,
Pd
Pd
Pd
1 + i=1 ?i
1 + i=1 ?i
1 + i=1 ?i
(15)
2
1
(k)
(k?1)
(k)
/?i ? Xi wi /?i
+ hi (wi /?i ), i = 1, . . . , d,
wi = argmin
u0 + Xi wi
2
wi ?Rpi 2
for k = 1, 2, 3, . . ., where ?1 , . . . , ?d > 0 are augmented Lagrangian parameters. In each iteration,
(k)
the updates to wi , i = 1, . . . , d above can be done in parallel. Just based on their form, it seems
that (15) can be seen as a parallel version of coordinate descent (5) for problem (2). The next result
confirms this, leveraging standard theory for ADMM (Gabay, 1983; Eckstein and Bertsekas, 1992).
Theorem 5. Assume that Xi ? Rn?pi has full column rank and hi (v) = maxd?Di hd, vi for a closed,
convex set Di ? Rpi , for i = 1, . . . , d. Then the sequence w(k) , k = 1, 2, 3, . . . in (15) converges to
a solution in (2).
The parallel-ADMM-CD iterations in (15) and parallel-Dykstra-CD iterations in (14) differ in that,
(k)
where the latter uses a residual y ? Xw(k?1) , the former uses an iterate u0 that seems to have a
(k?1)
more complicated form, being a convex combination of u0
and y ? Xw(k?1) , plus a quantity
that acts like a momentum term. It turns out that when ?1 , . . . , ?d sum to 1, the two methods (14),
(15) are exactly the same. While this may seem like a surprising coincidence, it is in fact nothing
more than a reincarnation of the previously established equivalence between Dykstra?s algorithm (4)
and ADMM (8) for a 2-set best approximation problem, as here C0 is a linear subspace.
Of course, with ADMM we need not choose probability weights for ?1 , . . . , ?d , and the convergence
in Theorem 5 is guaranteed for any fixed values of these parameters. Thus, even though they were
derived from different perspectives, parallel-ADMM-CD subsumes parallel-Dykstra-CD, and it is a
strictly more general approach. It is important to note that larger values of ?1 , . . . , ?d can often lead
to faster convergence in practice, as we show in Figure 1. More detailed study and comparisons to
related parallel methods are worthwhile, but are beyond the scope of this work.
5
Discussion and extensions
We studied connections between Dykstra?s algorithm, ADMM, and coordinate descent. Leveraging
these connections, we established an (asymptotically) linear convergence rate for coordinate descent
for the lasso, as well as two parallel versions of coordinate descent (one based on Dykstra?s algorithm
and the other on ADMM). Some extensions and possibilities for future work are described below.
7
No parallelization
1e+01
1e+04
Coordinate descent
Par?Dykstra?CD
Par?ADMM?CD, rho=10
Par?ADMM?CD, rho=50
Par?ADMM?CD, rho=200
1e?02
1e?08
1e?08
1e?05
1e?02
Suboptimality
1e+01
1e+04
Coordinate descent
Par?Dykstra?CD
Par?ADMM?CD, rho=10
Par?ADMM?CD, rho=50
Par?ADMM?CD, rho=200
1e?05
Suboptimality
10% parallelization
0
500
1000
1500
2000
0
50
Actual iteration number
100
150
Effective iteration number
Figure 1: Suboptimality curves for serial coordinate descent,
Ppparallel-Dykstra-CD, and three tunings
of parallel-ADMM-CD (i.e., three different values of ? = i=1 ?i ), each run over the same 30 lasso
problems with n = 100 and p = 500. For details of the experimental setup, see the supplement.
Nonquadratic loss: Dykstra?s algorithm and coordinate descent. Given a convex function f , a
generalization of (2) is the regularized estimation problem
minp f (Xw) +
w?R
d
X
hi (wi ).
(16)
i=1
Regularized regression (2) is given byP
f (z) = 12 ky ? zk22 , and e.g., regularized classification (under
n
T
the logistic loss) by f (z) = ?y z + i=1 log(1 + ezi ). In (block) coordinate descent for (16), we
(0)
initialize say w = 0, and repeat, for k = 1, 2, 3, . . .:
X
X
(k)
(k)
(k?1)
wi = argmin f
Xj wj +
Xj wj
+ Xi wi + hi (wi ), i = 1, . . . , d.
(17)
wi ?Rpi
j<i
j>i
On the other hand, given a differentiable and strictly convex function g, we can generalize (1) to the
following best Bregman-approximation problem,
min Dg (u, b)
u?Rn
subject to
u ? C1 ? ? ? ? ? Cd .
(18)
where Dg (u, b) = g(u) ? g(b) ? h?g(b), u ? bi is the Bregman divergence between u and b with
respect to g. When g(v) = 12 kvk22 (and b = y), this recovers the best approximation problem (1). As
shown in Censor and Reich (1998); Bauschke and Lewis (2000), Dykstra?s algorithm can be extended
(0)
(0)
(0)
to apply to (18). We initialize ud = b, z1 = ? ? ? = zd = 0, and repeat for k = 1, 2, 3, . . .:
(k)
(k?1)
u0 = ud
(k)
ui
(k)
zi
,
?
(k)
(k?1)
? ?g ? ) ?g(ui?1 ) + zi
, ?
=
(PCg i
=
(k)
?g(ui?1 )
+
(k?1)
zi
?
(k)
?g(ui ),
PCg (x)
for i = 1, . . . , d,
(19)
?
where
= argminc?C Dg (c, x) denotes the Bregman (rather than Euclidean) projection of x
onto a set C, and g ? is the conjugate function of g. Though it may not be immediately obvious, when
g(v) = 12 kvk22 the above iterations (19) reduce to the standard (Euclidean) Dykstra iterations in (4).
Furthermore, Dykstra?s algorithm and coordinate descent are equivalent in the more general setting.
Theorem 6. Let f be a strictly convex, differentiable function that has full domain. Assume that
Xi ? Rn?pi has full column rank and hi (v) = maxd?Di hd, vi for a closed, convex set Di ? Rpi , for
i = 1, . . . , d. Also, let g(v) = f ? (?v), b = ??f (0), and Ci = (XiT )?1 (Di ) ? Rn , i = 1, . . . , d.
8
Then (16), (18) are dual to each other, and their solutions w,
? u
? satisfy u
? = ??f (X w).
? Moreover,
Dykstra?s algorithm (19) and coordinate descent (17) are equivalent, i.e., for k = 1, 2, 3, . . .:
X
X
(k)
(k)
(k)
(k)
(k?1)
zi = Xi wi
and ui = ??f
Xj wj +
Xj wj
, for i = 1, . . . , d.
j>i
j?i
Nonquadratic loss: parallel coordinate descent methods. For a general regularized estimation
problem (16), parallel coordinate descent methods can be derived by applying Dykstra?s algorithm
and ADMM to a product space reformulation of the dual. Interestingly, the subsequent coordinate
descent algorithms are no longer equivalent (for a unity augmented Lagrangian parameter), and they
feature quite different computational structures. Parallel-Dykstra-CD for (16) initializes w(0) = 0,
and repeats:
(k)
(k)
wi = argmin f Xw(k) ? Xi wi /?i + Xi wi /?i + hi (wi /?i ), i = 1, . . . , d,
(20)
wi ?Rpi
for k = 1, 2, 3, . . ., and weights ?1 , . . . , ?d > 0 that sum to 1. In comparison, parallel-ADMM-CD
(0)
for (16) begins with u0 = 0, w(?1) = w(0) = 0, and repeats:
!
X
d
(k)
(k)
(k)
(k?1)
(k?2)
(k?1)
Find u0 such that: u0 = ??f
?i (u0 ? u0
) ? X(w
? 2w
) ,
i=1
(21)
2
1
(k)
(k)
(k?1)
wi = argmin
u0 + Xi wi
/?i ? Xi wi /?i
+ hi (wi /?i ), i = 1, . . . , d,
2
wi ?Rpi 2
for k = 1, 2, 3, . . ., and parameters ?1 , . . . , ?d > 0. Derivation details are given in the supplement.
Notice the stark contrast between the parallel-Dykstra-CD iterations (20) and the parallel-ADMMCD iterations (21). In (20), we perform (in parallel) coordinatewise hi -regularized minimizations
involving f , for i = 1, . . . , d. In (21), we perform a single quadratically-regularized minimization
involving f for the u0 -update, and then for the w-update, we perform (in parallel) coordinatewise
hi -regularized minimizations involving a quadratic loss, for i = 1, . . . , d (these are typically much
cheaper than the analogous minimizations for typical nonquadratic losses f of interest).
We note that the u0 -update in the parallel-ADMM-CD iterations (21) simplifies
Pn for many losses f
of interest; in particular, for separable loss functions of the form f (v) = i=1 fi (vi ), for convex,
univariate functions fi , i = 1, . . . , n, the u0 -update separates into n univariate minimizations. As an
example, consider the logistic lasso problem,
n
X
T
T
minp ?y Xw +
log(1 + exi w ) + ?kwk1 ,
(22)
w?R
i=1
Pp
where xi ? Rp , i = 1, . . . , n denote the rows of X. Abbreviating ? = i=1 ?i , and denoting by
?x
?(x) = 1/(1 + e ) the sigmoid function, and by St (x) = sign(x)(|x| ? t)+ the soft-thresholding
function at a level t > 0, the parallel-ADMM-CD iterations (21) for (22) reduce to:
(k)
(k)
(k?1)
Find u0i such that: (? ? 1)u0i = ?u0i
+ xTi (w(k?2) ? 2w(k?1) ) +
(k)
(k?1)
? ?u0i ? ?u0i
? xTi (w(k?2) ? 2w(k?1) ) , i = 1, . . . , n,
(23)
(k)
(k?1)
T
?i Xi (u0 + Xi wi
/?i )
(k)
, i = 1, . . . , p,
wi = S??i /kXi k22
kXi k22
for k = 1, 2, 3, . . .. Now we see that both the u0 -update and w-update in (23) can be parallelized,
and each coordinate update in the former can be done with, say, a simple bisection search.
Asynchronous parallel algorithms, and coordinate descent in Hilbert spaces. We finish with
some directions for possible future work. Asynchronous variants of parallel coordinate descent are
currently of great interest, e.g., see the review in Wright (2015). Given the link between ADMM and
coordinate descent developed in this paper, it would be interesting to investigate the implications of
the recent exciting progress on asynchronous ADMM, e.g., see Chang et al. (2016a,b) and references
therein, for coordinate descent. In a separate direction, much of the literature on Dykstra?s algorithm
emphasizes that this method works seamlessly in Hilbert spaces. It would be interesting to consider
the connections to (parallel) coordinate descent in infinite-dimensional function spaces, which we
would encounter, e.g., in alternating conditional expectation algorithms or backfitting algorithms in
additive models.
9
References
Alfred Auslender. Optimisation: Methodes Numeriques. Masson, 1976.
Heinz H. Bauschke and Patrick L. Combettes. Convex Analysis and Monotone Operator Theory in
Hilbert Spaces. Springer, 2011.
Heinz H. Bauschke and Valentin R. Koch. Projection methods: Swiss army knives for solving
feasibility and best approximation problems with halfspaces. arXiv: 1301.4506, 2013.
Heinz H. Bauschke and Adrian S. Lewis. Dykstra?s algorithm with Bregman projections: a convergence proof. Optimization, 48:409?427, 2000.
Amir Beck and Luba Tetruashvili. On the convergence of block coordinate descent type methods.
SIAM Journal on Optimization, 23(4):2037?2060, 2013.
Dimitri P. Bertsekas and John N. Tsitsiklis. Parallel and Distributed Computation: Numerical
Methods. Prentice Hall, 1989.
Steve Boyd, Neal Parikh, Eric Chu, Borja Peleato, and Jonathan Eckstein. Distributed optimization
and statistical learning via the alternative direction method of multipliers. Foundations and Trends
in Machine Learning, 3(1):1?122, 2011.
James P. Boyle and Richard L. Dykstra. A method for finding projections onto the intersection of
convex sets in hilbert spaces. Advances in Order Restricted Statistical Inference: Proceedings of
the Symposium on Order Restricted Statistical Inference, pages 28?47, 1986.
Yair Censor and Simeon Reich. The Dykstra algorithm with Bregman projections. Communications
in Applied Analysis, 48:407?419, 1998.
Tsung-Hui Chang, Mingyi Hong, Wei-Cheng Liao, and Xiangfeng Wang. Asynchronous distributed
ADMM for large-scale optimization?part i: Algorithm and convergence analysis. IEEE Transactions on Signal Processing, 64(12):3118?3130, 2016a.
Tsung-Hui Chang, Wei-Cheng Liao, Mingyi Hong, and Xiangfeng Wang. Asynchronous distributed
ADMM for large-scale optimization?part ii: Linear convergence analysis and numerical performance. IEEE Transactions on Signal Processing, 64(12):3131?3144, 2016b.
Scott Chen, David L. Donoho, and Michael Saunders. Atomic decomposition for basis pursuit. SIAM
Journal on Scientific Computing, 20(1):33?61, 1998.
Frank Deutsch. Best Approximation in Inner Product Spaces. Springer, 2001.
Frank Deutsch and Hein Hundal. The rate of convergence of Dykstra?s cyclic projections algorithm:
The polyhedral case. Numerical Functional Analysis and Optimization, 15(5?6):537?565, 1994.
Jim Douglas and H. H. Rachford. On the numerical solution of heat conduction problems in two and
three space variables. Transactions of the American Mathematical Society, 82:421?439, 1956.
Richard L. Dykstra. An algorithm for restricted least squares regression. Journal of the American
Statistical Association, 78(384):837?842, 1983.
Jonathan Eckstein and Dimitri P. Bertsekas. On the Douglas-Rachford splitting method and the
proximal point algorithm for maximal monotone operators. Mathematical Programming, 55(1):
293?318, 1992.
Laurent El Ghaoui, Vivian Viallon, and Tarek Rabbani. Safe feature elimination in sparse supervised
learning. Pacific Journal of Optimization, 8(4):667?698, 2012.
Jerome Friedman, Trevor Hastie, Holger Hoefling, and Robert Tibshirani. Pathwise coordinate
optimization. Annals of Applied Statistics, 1(2):302?332, 2007.
Jerome Friedman, Trevor Hastie, and Robert Tibshirani. Regularization paths for generalized linear
models via coordinate descent. Journal of Statistical Software, 33(1):1?22, 2010.
10
Wenjiang J. Fu. Penalized regressions: The bridge versus the lasso. Journal of Computational and
Graphical Statistics, 7(3):397?416, 1998.
Daniel Gabay. Applications of the method of multipliers to variational inequalities. Studies in
Mathematics and Its Applications, 15:299?331, 1983.
Daniel Gabay and Bertrand Mercier. A dual algorithm for the solution of nonlinear variational
problems via finite element approximation. Computers & Mathematics with Applications, 2(1):
17?40, 1976.
Norbert Gaffke and Rudolf Mathar. A cyclic projection algorithm via duality. Metrika, 36(1):29?54,
1989.
Roland Glowinski and A. Marroco. Sur l?approximation, par elements finis d?ordre un, et la resolution,
par penalisation-dualite d?une classe de problemes de Dirichlet non lineaires. Modelisation
Mathematique et Analyse Numerique, 9(R2):41?76, 1975.
Tom Goldstein, Brendan O?Donoghue, Simon Setzer, and Richard Baraniuk. Fast alternating direction
optimization methods. SIAM Journal on Imaging Sciences, 7(3):1588?1623, 2014.
Israel Halperin. The product of projection operators. Acta Scientiarum Mathematicarum, 23:96?99,
1962.
Shih-Ping Han. A successive projection algorithm. Mathematical Programming, 40(1):1?14, 1988.
Clifford Hildreth. A quadratic programming procedure. Naval Research Logistics Quarterly, 4(1):
79?85, 1957.
Mingyi Hong and Zhi-Quan Luo. On the linear convergence of the alternating direction method of
multipliers. Mathematical Programming, 162(1):165?199, 2017.
Alfredo N. Iusem and Alvaro R. De Pierro. On the convergence properties of Hildreth?s quadratic
programming algorithm. Mathematical Programming, 47(1):37?51, 1990.
Alfredo N. Iusem and Alvaro R. De Pierro. A simultaneous iterative method for computing projections
on polyhedra. SIAM Journal on Control and Optimization, 25(1):231?243, 1987.
Martin Jaggi, Virginia Smith, Martin Takac, Jonathan Terhorst, Sanjay Krishnan, Thomas Hofmann,
and Michael I. Jordan. Communication-efficient distributed dual coordinate ascent. Advances in
Neural Information Processing, 27:3068?3076, 2014.
Mojtaba Kadkhodaie, Konstantina Christakopoulou, Maziar Sanjabi, and Arindam Banerjee. Accelerated alternating direction method of multipliers. International Conference on Knowledge
Discovery and Data Mining, 21:497?506, 2015.
Xingguo Li, Tuo Zhao, Raman Arora, Han Liu, and Mingyi Hong. An improved convergence analysis
of cyclic block coordinate descent-type methods for strongly convex minimization. International
Conference on Artificial Intelligence and Statistics, 19:491?499, 2016.
P. L. Lions and B. Mercier. Splitting algorithms for the sum of two nonlinear operators. SIAM Journal
on Numerical Analysis, 16(6):964?979, 1979.
David Luenberger. Introduction to Linear and Nonlinear Programming. Addison-Wesley, 1973.
Zhi-Quan Luo and Paul Tseng. On the convergence of the coordinate descent method for convex
differentiable minimization. Journal of Optimization Theory and Applications, 72(1):7?35, 1992.
Zhi-Quan Luo and Paul Tseng. On the convergence rate of dual ascent methods for linearly constrained
convex minimization. Mathematics of Operations Research, 18(4):846?867, 1993.
Robert Nishihara, Laurent Lessard, Benjamin Recht, Andrew Packard, and Michael I. Jordan. A
general analysis of the convergence of ADMM. International Conference on Machine Learning,
32:343?352, 2015.
11
James M. Ortega and Werner C. Rheinboldt. Iterative Solution of Nonlinear Equations in Several
Variables. Academic Press, 1970.
G. Pierra. Decomposition through formalization in a product space. Mathematical Programming, 28
(1):96?115, 1984.
Peter Richtarik and Martin Takac. Parallel coordinate descent methods for big data optimization.
Mathematical Programming, 156(1):433?484, 2016.
Sylvain Sardy, Andrew G. Bruce, and Paul Tseng. Block coordinate relaxation methods for nonparametric wavelet denoising. Journal of Computational and Graphical Statistics, 9(2):361?379,
2000.
Robert Tibshirani. Regression shrinkage and selection via the lasso. Journal of the Royal Statistical
Society: Series B, 58(1):267?288, 1996.
Robert Tibshirani, Jacob Bien, Jerome Friedman, Trevor Hastie, Noah Simon, Jonathan Taylor, and
Ryan J. Tibshirani. Strong rules for discarding predictors in lasso-type problems. Journal of the
Royal Statistical Society: Series B, 74(2):245?266, 2012.
Ryan J. Tibshirani. The lasso problem and uniqueness. Electronic Journal of Statistics, 7:1456?1490,
2013.
Paul Tseng. Dual ascent methods for problems with strictly convex costs and linear constraints: A
unified approach. SIAM Journal on Control and Optimization, 28(1):214?29, 1990.
Paul Tseng. Dual coordinate ascent methods for non-strictly convex minimization. Mathematical
Programming, 59(1):231?247, 1993.
Paul Tseng. Convergence of a block coordinate descent method for nondifferentiable minimization.
Journal of Optimization Theory and Applications, 109(3):475?494, 2001.
Paul Tseng and Dimitri P. Bertsekas. Relaxation methods for problems with strictly convex separable
costs and linear constraints. Mathematical Programming, 38(3):303?321, 1987.
John von Neumann. Functional Operators, Volume II: The Geometry of Orthogonal Spaces. Princeton
University Press, 1950.
Jie Wang, Peter Wonka, and Jieping Ye. Lasso screening rules via dual polytope projection. Journal
of Machine Learning Research, 16:1063?1101, 2015.
Jack Warga. Minimizing certain convex functions. Journal of the Society for Industrial and Applied
Mathematics, 11(3):588?593, 1963.
Stephen J. Wright. Coordinate descent algorithms. Mathematical Programming, 151(1):3?34, 2015.
Tong Tong Wu and Kenneth Lange. Coordinate descent algorithms for lasso penalized regression.
The Annals of Applied Statistics, 2(1):224?244, 2008.
12
| 6655 |@word mild:1 version:18 seems:4 replicate:1 stronger:1 c0:3 adrian:1 confirms:1 seek:2 decomposition:4 jacob:1 cyclic:5 liu:1 series:2 daniel:2 denoting:1 interestingly:1 reinvented:1 existing:3 comparing:1 luo:6 surprising:1 rpi:14 chu:1 must:1 john:2 realize:1 stemming:1 numerical:6 subsequent:1 additive:1 wenjiang:1 hofmann:1 update:12 aside:1 intelligence:1 metrika:1 amir:1 une:1 smith:1 short:1 iterates:2 successive:1 simpler:3 rabbani:1 mathematical:11 along:1 become:2 symposium:1 prove:2 backfitting:1 polyhedral:1 introduce:1 xwk22:2 indeed:1 abbreviating:1 heinz:3 inspired:1 bertrand:1 decreasing:1 zhi:3 actual:1 xti:2 becomes:1 begin:1 unrelated:1 linearity:2 notation:2 moreover:2 what:5 israel:1 argmin:7 developed:5 unified:1 finding:1 guarantee:1 thorough:1 act:1 shed:1 exactly:4 k2:1 control:2 enjoy:1 bertsekas:7 arguably:2 consequence:1 analyzing:3 id:2 laurent:2 path:1 credited:1 might:1 plus:1 ryantibs:1 therein:3 studied:4 acta:1 argminc:2 equivalence:10 suggests:1 bi:1 practical:1 unique:2 atomic:1 practice:2 block:13 reappear:1 swiss:1 procedure:1 asymptotics:1 attain:1 projection:17 boyd:3 suggest:1 onto:8 convenience:2 clever:1 operator:8 interior:1 put:2 prentice:1 applying:3 seminal:2 selection:1 restriction:3 equivalent:12 map:1 lagrangian:8 accumulation:1 xiangfeng:2 send:1 straightforward:2 attention:1 regardless:1 masson:1 convex:25 survey:1 resolution:1 simplicity:2 splitting:5 immediately:1 boyle:4 insight:1 continued:1 rule:5 his:1 hd:7 notion:1 coordinate:95 analogous:1 feel:2 limiting:1 annals:2 suppose:3 play:1 modulo:1 exact:1 programming:13 ck22:1 us:2 pa:1 trick:1 curly:1 trend:1 element:2 updating:1 invented:1 role:1 coincidence:1 wang:4 wj:6 connected:2 revival:1 removed:1 halfspaces:5 monograph:1 benjamin:1 pd:4 convexity:4 ui:15 flurry:1 solving:3 rewrite:1 eric:1 basis:1 exi:1 various:2 derivation:1 heat:1 fast:1 describe:1 effective:1 kp:1 artificial:1 pci:3 saunders:1 quite:1 widely:1 solve:3 larger:1 say:4 rho:6 otherwise:1 statistic:7 analyse:1 itself:1 seemingly:1 sequence:5 eigenvalue:1 differentiable:3 coming:1 product:6 adaptation:2 maximal:1 relevant:1 date:4 flexibility:1 achieve:1 ky:7 convergence:33 extending:2 neumann:2 converges:4 help:1 derive:3 develop:1 ac:1 stat:1 andrew:2 hundal:4 ij:1 received:1 progress:4 ic1:2 solves:1 auxiliary:3 strong:7 implies:1 deutsch:6 direction:7 differ:1 safe:1 closely:2 kb:1 duals:1 enable:1 elimination:1 argued:1 require:1 mathematique:1 generalization:2 really:1 preliminary:1 ryan:3 mathematicarum:1 extension:6 strictly:7 hold:1 koch:3 hall:1 wright:4 ic:1 presumably:1 great:2 mapping:1 scope:2 early:2 a2:2 smallest:1 uniqueness:1 estimation:3 currently:1 bridge:1 contributor:1 minimization:11 concurrently:1 clearly:1 always:1 rather:2 pn:1 shrinkage:1 derived:3 xit:8 focus:1 naval:1 properly:1 polyhedron:6 rank:7 sanjabi:1 seamlessly:1 contrast:1 brendan:1 industrial:1 zk22:2 helpful:1 kzk22:1 censor:2 inference:2 el:2 entire:1 typically:1 i1:2 provably:1 dual:14 among:2 classification:1 denoted:1 pcg:2 development:3 art:2 special:5 fairly:1 initialize:6 constrained:1 equal:1 once:2 u0i:5 beach:1 kw:4 holger:1 future:4 others:2 richard:3 few:2 modern:2 dg:3 divergence:1 comprehensive:2 cheaper:1 beck:2 phase:1 geometry:1 friedman:6 interest:7 screening:2 mining:1 possibility:1 highly:1 investigate:1 analyzed:1 light:2 primal:1 pc:2 implication:2 bregman:5 fu:2 closer:1 jumping:1 orthogonal:1 unless:1 indexed:1 euclidean:4 old:3 loosely:2 initialized:1 taylor:1 hein:1 fitted:1 column:16 earlier:4 soft:2 werner:1 cost:2 introducing:1 predictor:3 valentin:1 virginia:1 bauschke:7 byp:1 dependency:1 conduction:1 proximal:2 kxi:3 thoroughly:1 st:2 cited:1 thanks:1 randomized:1 siam:6 alvaro:2 international:3 recht:1 off:1 michael:3 von:2 again:3 clifford:1 choose:1 possibly:1 adversely:1 book:2 american:2 zhao:1 dimitri:3 rescaling:1 stark:1 li:2 supp:3 account:1 nonasymptotic:1 ku1:1 de:8 wk:4 includes:1 coefficient:2 matter:1 subsumes:1 satisfy:3 kzk2:1 depends:1 vi:6 tsung:2 performed:3 h1:2 nishihara:2 closed:8 later:1 start:1 parallel:51 complicated:2 reparametrization:1 simon:2 bruce:1 contribution:2 square:1 who:1 kxij:1 yield:1 identify:1 richtarik:2 generalize:1 identification:2 emphasizes:1 bisection:1 history:2 explain:1 ping:1 simultaneous:1 trevor:3 definition:1 inexact:1 pp:1 james:2 obvious:1 associated:1 di:20 proof:3 recovers:1 popular:2 knowledge:2 dimensionality:2 hilbert:5 goldstein:2 back:5 appears:1 wesley:1 steve:1 supervised:1 tom:1 response:1 wei:2 improved:1 formulation:1 done:2 though:5 strongly:1 generality:2 furthermore:1 just:6 xa:8 hoefling:1 correlation:3 jerome:3 working:1 hand:1 nonlinear:4 banerjee:1 scientiarum:1 hildreth:7 logistic:2 halperin:2 perhaps:2 vivian:1 scientific:1 usa:1 ye:1 k22:8 multiplier:7 true:2 tarek:1 inductive:2 regularization:2 equality:1 hence:2 alternating:7 symmetric:1 former:3 neal:1 deal:1 uniquely:1 unambiguous:1 suboptimality:3 hong:5 generalized:1 ortega:2 alfredo:2 performs:1 bring:1 image:1 variational:2 jack:1 arindam:1 recently:3 fi:2 parikh:1 sigmoid:1 functional:2 volume:1 rachford:4 discussed:2 extend:1 association:1 mellon:1 refer:2 orthocomplement:1 enter:1 smoothness:1 tuning:3 mathematics:4 reich:2 han:6 longer:1 ezi:1 patrick:1 jaggi:2 closest:1 own:2 recent:9 perspective:1 apart:1 certain:1 inequality:1 maxd:5 kwk1:2 seen:2 parallelized:1 converge:5 ud:8 signal:2 u0:22 ii:2 full:11 stephen:1 reduces:2 stem:2 seidel:1 smooth:2 borja:1 match:1 faster:1 academic:1 offer:2 long:2 knife:1 serial:1 roland:1 penalisation:1 feasibility:1 variant:2 regression:8 basic:1 involving:3 essentially:3 cmu:1 expectation:1 optimisation:1 arxiv:1 iteration:26 liao:2 c1:17 pierro:7 appropriately:1 parallelization:4 brace:1 ascent:4 subject:4 quan:3 member:1 leveraging:3 seem:2 jordan:2 call:3 enough:3 krishnan:1 iterate:1 affect:1 xj:6 zi:11 fit:1 finish:1 lasso:25 hastie:3 reduce:3 lange:2 idea:1 simplifies:1 inner:1 donoghue:1 inactive:2 setzer:1 nonquadratic:5 penalty:3 peter:2 speaking:1 passing:1 remark:3 jie:1 ignored:1 generally:4 enumerate:1 detailed:2 descent1:1 nonparametric:1 xij:1 revisit:1 notice:1 sign:1 tibshirani:10 track:1 pace:1 per:1 zd:3 alfred:1 carnegie:1 write:1 affected:1 group:1 key:3 thereafter:1 reformulation:2 shih:1 douglas:4 rewriting:1 kenneth:1 viallon:1 ordre:1 asymptotically:5 imaging:1 monotone:2 relaxation:2 sum:5 year:2 run:2 inverse:1 baraniuk:1 striking:1 discern:1 place:1 throughout:2 arrive:1 wu:2 electronic:1 draw:1 raman:1 pc2:2 scaling:1 bound:3 hi:19 guaranteed:3 centrally:1 cheng:2 quadratic:4 nontrivial:1 noah:1 constraint:5 precisely:2 warga:3 software:1 u1:17 speed:1 argument:4 min:6 span:1 separable:3 martin:3 xingguo:1 speedup:1 department:2 pacific:1 popularized:1 combination:1 conjugate:1 across:1 describes:1 smaller:1 unity:1 wi:42 appealing:2 projecting:3 restricted:3 ghaoui:2 marroco:3 equation:1 previously:2 discus:1 turn:3 nonempty:3 differentiates:1 know:1 flip:1 addison:1 fed:1 mercier:5 luenberger:2 pursuit:1 operation:1 rewritten:3 permit:1 finis:1 apply:2 quarterly:1 worthwhile:1 appropriate:1 alternative:3 encounter:1 shortly:1 tetruashvili:2 rp:4 yair:1 thomas:1 denotes:7 dirichlet:1 include:1 graphical:2 xw:9 build:1 especially:1 establish:2 dykstra:88 classical:2 society:4 objective:4 initializes:2 quantity:1 strategy:1 usual:1 subspace:6 separably:1 separate:2 link:1 uk22:1 jieping:1 nondifferentiable:1 topic:2 polytope:1 tseng:16 trivial:1 sur:1 minn:2 relationship:4 minimizing:1 difficult:2 setup:2 robert:5 blockwise:1 expense:1 frank:2 wonka:1 stated:1 resurgence:1 rise:1 design:1 implementation:2 perform:3 finite:6 descent:86 logistics:1 situation:1 extended:3 communication:2 glowinski:3 precise:3 jim:1 rn:20 pc1:4 community:1 peleato:1 tuo:1 david:2 eckstein:3 connection:18 z1:4 quadratically:1 established:3 nip:1 auslender:2 beyond:2 below:4 lion:2 usually:2 scott:1 sanjay:1 bien:1 rheinboldt:2 max:2 packard:1 royal:2 ia:5 suitable:1 critical:1 rely:1 regularized:11 warm:1 indicator:1 residual:2 older:2 improve:1 kvk22:2 numerous:1 identifies:2 arora:1 extract:1 dating:1 review:2 literature:5 discovery:1 asymptotic:1 loss:9 par:10 interesting:10 enclosed:1 versus:1 remarkable:1 foundation:1 minp:3 article:1 exciting:3 thresholding:2 lessard:1 pi:6 cd:39 translation:1 row:1 course:3 penalized:2 summary:1 repeat:11 last:1 surprisingly:1 asynchronous:5 enjoys:1 tsitsiklis:3 formal:1 side:1 understand:1 taking:1 sparse:1 distributed:6 curve:1 dimension:2 unaware:2 unweighted:1 dwi:1 coincide:1 ici:1 far:1 transaction:3 active:4 conceptual:1 pittsburgh:1 assumed:2 xi:25 search:1 un:1 iterative:2 why:1 simeon:1 ca:1 excellent:1 necessarily:1 meanwhile:1 ic2:2 domain:1 linearly:1 big:1 paul:7 coordinatewise:3 gabay:5 nothing:2 augmented:8 fashion:1 tong:2 combettes:2 formalization:2 position:4 momentum:1 explicit:1 classe:1 late:1 third:1 wavelet:1 theorem:13 discarding:1 explored:1 r2:1 importance:1 ci:4 supplement:5 hui:2 magnitude:1 konstantina:1 terhorst:1 kx:1 chen:2 intersection:9 simply:1 likely:2 univariate:2 army:1 norbert:1 pathwise:1 u2:13 chang:3 rnd:2 springer:2 truth:1 mingyi:4 lewis:2 conditional:1 identity:1 viewed:1 donoho:1 careful:1 admm:55 feasible:1 considerable:1 exceptionally:1 typical:1 infinite:1 corrected:1 sylvain:1 denoising:1 lemma:5 called:1 duality:3 gauss:1 experimental:1 la:1 mojtaba:1 takac:3 rudolf:1 support:2 latter:2 jonathan:4 accelerated:3 princeton:1 |
6,252 | 6,656 | Learning Spherical Convolution
for Fast Features from 360? Imagery
Yu-Chuan Su
Kristen Grauman
The University of Texas at Austin
Abstract
While 360? cameras offer tremendous new possibilities in vision, graphics, and
augmented reality, the spherical images they produce make core feature extraction non-trivial. Convolutional neural networks (CNNs) trained on images from
perspective cameras yield ?flat" filters, yet 360? images cannot be projected to a
single plane without significant distortion. A naive solution that repeatedly projects
the viewing sphere to all tangent planes is accurate, but much too computationally
intensive for real problems. We propose to learn a spherical convolutional network
that translates a planar CNN to process 360? imagery directly in its equirectangular projection. Our approach learns to reproduce the flat filter outputs on 360?
data, sensitive to the varying distortion effects across the viewing sphere. The key
benefits are 1) efficient feature extraction for 360? images and video, and 2) the
ability to leverage powerful pre-trained networks researchers have carefully honed
(together with massive labeled image training sets) for perspective images. We
validate our approach compared to several alternative methods in terms of both raw
CNN output accuracy as well as applying a state-of-the-art ?flat" object detector
to 360? data. Our method yields the most accurate results while saving orders of
magnitude in computation versus the existing exact reprojection solution.
1
Introduction
Unlike a traditional perspective camera, which samples a limited field of view of the 3D scene
projected onto a 2D plane, a 360? camera captures the entire viewing sphere surrounding its optical
center, providing a complete picture of the visual world?an omnidirectional field of view. As such,
viewing 360? imagery provides a more immersive experience of the visual content compared to
traditional media.
360? cameras are gaining popularity as part of the rising trend of virtual reality (VR) and augmented
reality (AR) technologies, and will also be increasingly influential for wearable cameras, autonomous
mobile robots, and video-based security applications. Consumer level 360? cameras are now common
on the market, and media sharing sites such as Facebook and YouTube have enabled support for
360? content. For consumers and artists, 360? cameras free the photographer from making real-time
composition decisions. For VR/AR, 360? data is essential to content creation. As a result of this great
potential, computer vision problems targeting 360? content are capturing the attention of both the
research community and application developer.
Immediately, this raises the question: how to compute features from 360? images and videos?
Arguably the most powerful tools in computer vision today are convolutional neural networks (CNN).
CNNs are responsible for state-of-the-art results across a wide range of vision problems, including
image recognition [14, 39], object detection [11, 27], image and video segmentation [13, 18, 25], and
action detection [10, 29]. Furthermore, significant research effort over the last five years (and really
decades [24]) has led to well-honed CNN architectures that, when trained with massive labeled image
datasets [8], produce ?pre-trained" networks broadly useful as feature extractors for new problems.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
?
Strategy I
Equirectangular
Projection
?
Input: 360? image
?
Np
???
Fully Convolution
?
Sample n
???
?
???
Perspective
Projection
Output
Np
???
Np
???
Np
Strategy II
Figure 1: Two existing strategies for applying CNNs to 360? images. Top: The first strategy unwraps the
360? input into a single planar image using a global projection (most commonly equirectangular projection),
then applies the CNN on the distorted planar image. Bottom: The second strategy samples multiple tangent
planar projections to obtain multiple perspective images, to which the CNN is applied independently to obtain
local results for the original 360? image. Strategy I is fast but inaccurate; Strategy II is accurate but slow. The
proposed approach learns to replicate flat filters on spherical imagery, offering both speed and accuracy.
Indeed such networks are widely adopted as off-the-shelf feature extractors for other algorithms and
applications (c.f., VGG [30], ResNet [14], and AlexNet [22] for images; C3D [33] for video).
However, thus far, powerful CNN features are awkward if not off limits in practice for 360? imagery.
The problem is that the underlying projection models of current CNNs and 360? data are different.
Both the existing CNN filters and the expensive training data that produced them are ?flat", i.e., the
product of perspective projection to a plane. In contrast, a 360? image is projected onto the unit
sphere surrounding the camera?s optical center.
To address this discrepancy, there are two common, though flawed, approaches. In the first, the
spherical image is projected to a planar one,1 then the CNN is applied to the resulting 2D image [16,23]
(see Fig. 1, top). However, any sphere-to-plane projection introduces distortion, making the resulting
convolutions inaccurate. In the second existing strategy, the 360? image is repeatedly projected
to tangent planes around the sphere, each of which is then fed to the CNN [31, 32, 35, 38] (Fig. 1,
bottom). In the extreme of sampling every tangent plane, this solution is exact and therefore accurate.
However, it suffers from very high computational cost. Not only does it incur the cost of rendering
each planar view, but also it prevents amortization of convolutions: the intermediate representation
cannot be shared across perspective images because they are projected to different planes.
We propose a learning-based solution that, unlike the existing strategies, sacrifices neither accuracy
nor efficiency. The main idea is to learn a CNN that processes a 360? image in its equirectangular
projection (fast) but mimics the ?flat" filter responses that an existing network would produce on
all tangent plane projections for the original spherical image (accurate). Because convolutions are
indexed by spherical coordinates, we refer to our method as spherical convolution (S PH C ONV). We
develop a systematic procedure to adjust the network structure in order to account for distortions.
Furthermore, we propose a kernel-wise pre-training procedure which significantly accelerates the
training process.
In addition to providing fast general feature extraction for 360? imagery, our approach provides a
bridge from 360? content to existing heavily supervised datasets dedicated to perspective images.
In particular, training requires no new annotations?only the target CNN model (e.g., VGG [30]
pre-trained on millions of labeled images) and an arbitrary collection of unlabeled 360? images.
We evaluate S PH C ONV on the Pano2Vid [32] and PASCAL VOC [9] datasets, both for raw convolution accuracy as well as impact on an object detection task. We show that it produces more precise
outputs than baseline methods requiring similar computational cost, and similarly precise outputs as
the exact solution while using orders of magnitude less computation. Furthermore, we demonstrate
that S PH C ONV can successfully replicate the widely used Faster-RCNN [27] detector on 360? data
when training with only 1,000 unlabeled 360? images containing unrelated objects. For a similar cost
as the baselines, S PH C ONV generates better object proposals and recognition rates.
1
e.g., with equirectangular projection, where latitudes are mapped to horizontal lines of uniform spacing
2
2
Related Work
360? vision Vision for 360? data is quickly gaining interest in recent years. The SUN360 project
samples multiple perspective images to perform scene viewpoint recognition [35]. PanoContext [38]
parses 360? images using 3D bounding boxes, applying algorithms like line detection on perspective
images then backprojecting results to the sphere. Motivated by the limitations of existing interfaces
for viewing 360? video, several methods study how to automate field-of-view (FOV) control for
display [16, 23, 31, 32], adopting one of the two existing strategies for convolutions (Fig. 1). In these
methods, a noted bottleneck is feature extraction cost, which is hampered by repeated sampling of
perspective images/frames, e.g., to represent the space-time ?glimpses" of [31, 32]. This is exactly
where our work can have positive impact.
Knowledge distillation Our approach relates to knowledge distillation [3, 5, 12, 15, 26, 28, 34],
though we explore it in an entirely novel setting. Distillation aims to learn a new model given existing
model(s). Rather than optimize an objective function on annotated data, it learns the new model
that can reproduce the behavior of the existing model, by minimizing the difference between their
outputs. Most prior work explores distillation for model compression [3, 5, 15, 28]. For example,
a deep network can be distilled into a shallower [3] or thinner [28] one, or an ensemble can be
compressed to a single model [15]. Rather than compress a model in the same domain, our goal is to
learn across domains, namely to link networks on images with different projection models. Limited
work considers distillation for transfer [12, 26]. In particular, unlabeled target-source paired data can
help learn a CNN for a domain lacking labeled instances (e.g., RGB vs. depth images) [12], and
multi-task policies can be learned to simulate action value distributions of expert policies [26]. Our
problem can also be seen as a form of transfer, though for a novel task motivated strongly by image
processing complexity as well as supervision costs. Different from any of the above, we show how to
adapt the network structure to account for geometric transformations caused by different projections.
Also, whereas most prior work uses only the final output for supervision, we use the intermediate
representation of the target network as both input and target output to enable kernel-wise pre-training.
Spherical image projection Projecting a spherical image into a planar image is a long studied
problem. There exists a large number of projection approaches (e.g., equirectangular, Mercator,
etc.) [4]. None is perfect; every projection must introduce some form of distortion. The properties of
different projections are analyzed in the context of displaying panoramic images [37]. In this work,
we unwrap the spherical images using equirectangular projection because 1) this is a very common
format used by camera vendors and researchers [1, 32, 35], and 2) it is equidistant along each row and
column so the convolution kernel does not depend on the azimuthal angle. Our method in principle
could be applied to other projections; their effect on the convolution operation remains to be studied.
CNNs with geometric transformations There is an increasing interest in generalizing convolution in CNNs to handle geometric transformations or deformations. Spatial transformer networks
(STNs) [17] represent a geometric transformation as a sampling layer and predict the transformation
parameters based on input data. STNs assume the transformation is invertible such that the subsequent
convolution can be performed on data without the transformation. This is not possible in spherical images because it requires a projection that introduces no distortion. Active convolution [19]
learns the kernel shape together with the weights for a more general receptive field, and deformable
convolution [7] goes one step further by predicting the receptive field location. These methods are
too restrictive for spherical convolution, because they require a fixed kernel size and weight. In
contrast, our method adapts the kernel size and weight based on the transformation to achieve better
accuracy. Furthermore, our method exploits problem-specific geometric information for efficient
training and testing. Some recent work studies convolution on a sphere [6, 21] using spectral analysis,
but those methods require manually annotated spherical images as training data, whereas our method
can exploit existing models trained on perspective images as supervision. Also, it is unclear whether
CNNs in the spectral domain can reach the same accuracy and efficiency as CNNs on a regular grid.
3
Approach
We describe how to learn spherical convolutions in equirectangular projection given a target network
trained on perspective images. We define the objective in Sec. 3.1. Next, we introduce how to adapt
the structure from the target network in Sec. 3.2. Finally, Sec. 3.3 presents our training process.
3
? = 180?
? = 108?
? = 36?
Figure 2: Inverse perspective projections P ?1 to equirectangular projections at different polar angles ?. The
same square image will distort to different sizes and shapes depending on ?. Because equirectangular projection
unwraps the 180? longitude, a line will be split into two if it passes through the 180? longitude, which causes
the double curve in ? = 36?.
3.1
Problem Definition
Let Is be the input spherical image defined on spherical coordinates (?, ?), and let Ie ? IWe ?He ?3
be the corresponding flat RGB image in equirectangular projection. Ie is defined by pixels on the
image coordinates (x, y) ? De , where each (x, y) is linearly mapped to a unique (?, ?). We define
the perspective projection operator P which projects an ?-degree field of view (FOV) from Is to
W pixels on the the tangent plane n
? = (?, ?). That is, P(Is , n
? ) = Ip ? IW ?W ?3 . The projection
operator is characterized by the pixel size ?p ? = ?/W in Ip , and Ip denotes the resulting perspective
image. Note that we assume ?? = ?? following common digital imagery.
Given a target network2 Np trained on perspective images Ip with receptive field (Rf) R ? R, we
define the output on spherical image Is at n
? = (?, ?) as
Np (Is )[?, ?] = Np (P(Is , (?, ?))),
(1)
where w.l.o.g. we assume W = R for simplicity. Our goal is to learn a spherical convolution network
Ne that takes an equirectangular map Ie as input and, for every image position (x, y), produces as
output the results of applying the perspective projection network to the corresponding tangent plane
for spherical image Is :
Ne (Ie )[x, y] ? Np (Is )[?, ?],
?(x, y) ? De , (?, ?) = (
180? ? y 360? ? x
,
).
He
We
(2)
This can be seen as a domain adaptation problem where we want to transfer the model from the
domain of Ip to that of Ie . However, unlike typical domain adaptation problems, the difference
between Ip and Ie is characterized by a geometric projection transformation rather than a shift
in data distribution. Note that the training data to learn Ne requires no manual annotations: it
consists of arbitrary 360? images coupled with the ?true" Np outputs computed by exhaustive planar
reprojections, i.e., evaluating the rhs of Eq. 1 for every (?, ?). Furthermore, at test time, only a single
equirectangular projection of the entire 360? input will be computed using Ne to obtain the dense
(inferred) Np outputs, which would otherwise require multiple projections and evaluations of Np .
3.2
Network Structure
The main challenge for transferring Np to Ne is the distortion introduced by equirectangular projection. The distortion is location dependent?a k ? k square in perspective projection will not be a
square in the equirectangular projection, and its shape and size will depend on the polar angle ?. See
Fig. 2. The convolution kernel should transform accordingly. Our approach 1) adjusts the shape
of the convolution kernel to account for the distortion, in particular the content expansion, and 2)
reduces the number of max-pooling layers to match the pixel sizes in Ne and Np , as we detail next.
We adapt the architecture of Ne from Np using the following heuristic. The goal is to ensure each
kernel receives enough information from the input in order to compute the target output. First, we
untie the weight of convolution kernels at different ? by learning one kernel Key for each output row
y. Next, we adjust the shape of Key such that it covers the Rf of the original kernel. We consider
Key ? Ne to cover Kp ? Np if more than 95% of pixels in the Rf of Kp are also in the Rf of Ke
in Ie . The Rf of Kp in Ie is obtained by backprojecting the R ? R grid to n
? = (?, 0) using P ?1 ,
where the center of the grid aligns on n
? . Ke should be large enough to cover Kp , but it should also
be as small as possible to avoid overfitting. Therefore, we optimize the shape of Kel,y for layer l as
follows. The shape of Kel,y is initialized as 3 ? 3. We first adjust the height kh and increase kh by 2
2
e.g., Np could be AlexNet [22] or VGG [30] pre-trained for a large-scale recognition task.
4
Kel+1
l+1
Ke
Kel+1
..
....
Kel
Kel
l
Ke
..
....
?
?
Figure 3: Spherical convolution. The kernel weight in spherical convolution is tied only along each row of the
equirectangular image (i.e., ?), and each kernel convolves along the row to generate 1D output. Note that the
kernel size differs at different rows and layers, and it expands near the top and bottom of the image.
until the height of the Rf is larger than that of Kp in Ie . We then adjust the width kw similar to kh .
Furthermore, we restrict the kernel size kh ? kw to be smaller than an upper bound Uk . See Fig. 4.
Because the Rf of Kel depends on Kel?1 , we search for the kernel size starting from the bottom layer.
It is important to relax the kernel from being square to being rectangular, because equirectangular
projection will expand content horizontally near the poles of the sphere (see Fig. 2). If we restrict the
kernel to be square, the Rf of Ke can easily be taller but narrower than that of Kp which leads to
overfitting. It is also important to restrict the kernel size, otherwise the kernel can grow wide rapidly
near the poles and eventually cover the entire row. Although cutting off the kernel size may lead to
information loss, the loss is not significant in practice because pixels in equirectangular projection do
not distribute on the unit sphere uniformly; they are denser near the pole, and the pixels are by nature
redundant in the region where the kernel size expands dramatically.
Besides adjusting the kernel sizes, we also adjust the number of pooling layers to match the pixel
size ?? in Ne and Np . We define ??e = 180?/He and restrict We = 2He to ensure ??e = ??e .
Because max-pooling introduces shift invariance up to kw pixels in the image, which corresponds to
kw ? ?? degrees on the unit sphere, the physical meaning of max-pooling depends on the pixel size.
Since the pixel size is usually larger in Ie and max-pooling increases the pixel size by a factor of kw ,
we remove the pooling layer in Ne if ??e ? ??p .
Fig. 3 illustrates how spherical convolution differs from ordinary CNN. Note that we approximate
one layer in Np by one layer in Ne , so the number of layers and output channels in each layer is
exactly the same as the target network. However, this does not have to be the case. For example,
we could use two or more layers to approximate each layer in Np . Although doing so may improve
accuracy, it would also introduce significant overhead, so we stick with the one-to-one mapping.
3.3
Training Process
Given the goal in Eq. 2 and the architecture described in Sec. 3.2, we would like to learn the network
Ne by minimizing the L2 loss E[(Ne (Ie ) ? Np (Is ))2 ]. However, the network converges slowly,
possibly due to the large number of parameters. Instead, we propose a kernel-wise pre-training
process that disassembles the network and initially learns each kernel independently.
To perform kernel-wise pre-training, we further require Ne to generate the same intermediate representation as Np in all layers l:
Nel (Ie )[x, y] ? Npl (Is )[?, ?] ?l ? Ne .
(3)
Given Eq. 3, every layer l ? Ne is independent of each other. In fact, every kernel is independent and
can be learned separately. We learn each kernel by taking the ?ground truth? value of the previous
layer Npl?1 (Is ) as input and minimizing the L2 loss E[(Nel (Ie ) ? Npl (Is ))2 ], except for the first layer.
Note that Npl refers to the convolution output of layer l before applying any non-linear operation,
e.g. ReLU, max-pooling, etc. It is important to learn the target value before applying ReLU because
it provides more information. We combine the non-linear operation with Kel+1 during kernel-wise
pre-training, and we use dilated convolution [36] to increase the Rf size instead of performing
max-pooling on the input feature map.
For the first convolution layer, we derive the analytic solution directly.
P The projection operator P is
linear in the pixels in equirectangular projection: P(Is , n
? )[x, y] = ij cij Ie [i, j], for coefficients cij
5
Target Kernel Kp
Target Network Np
Receptive Field
Perspective Projection
(Inverse)
Kernel Ke
Yes
>
Network Ne
Receptive Field
No
kw
No
Increase
kh
Output kh
Yes
kw? kh > Uk
kh
Figure 4: Method to select the kernel height kh . We project the receptive field of the target kernel to equirectangular projection Ie and increase kh until it is taller than the target kernel in Ie . The kernel width kw is determined
using the same procedure after kh is set. We restrict the kernel size kw ? kh by an upper bound Uk .
from,
P e.g., bilinear interpolation. Because convolution is a weighted sum of input pixels Kp ? Ip =
xy wxy Ip [x, y], we can combine the weight wxy and interpolation coefficient cij as a single
convolution operator:
X
X
XX
Kp1 ? Is [?, ?] =
wxy
cij Ie [i, j] =
wxy cij Ie [i, j] = Ke1 ? Ie .
(4)
xy
ij
ij
xy
Ne1
The output value of
will be exact and requires no learning. Of course, the same is not possible for
l > 1 because of the non-linear operations between layers.
After kernel-wise pre-training, we can further fine-tune the network jointly across layers and kernels
by minimizing the L2 loss of the final output. Because the pre-trained kernels cannot fully recover
the intermediate representation, fine-tuning can help to adjust the weights to account for residual
errors. We ignore the constraint introduced in Eq. 3 when performing fine-tuning. Although Eq. 3
is necessary for kernel-wise pre-training, it restricts the expressive power of Ne and degrades the
performance if we only care about the final output. Nevertheless, the weights learned by kernel-wise
pre-training are a very good initialization in practice, and we typically only need to fine-tune the
network for a few epochs.
One limitation of S PH C ONV is that it cannot handle very close objects that span a large FOV. Because
the goal of S PH C ONV is to reproduce the behavior of models trained on perspective images, the
capability and performance of the model is bounded by the target model Np . However, perspective
cameras can only capture a small portion of a very close object in the FOV, and very close objects are
usually not available in the training data of the target model Np . Therefore, even though 360? images
offer a much wider FOV, S PH C ONV inherits the limitations of Np , and may not recognize very close
large objects. Another limitation of S PH C ONV is the resulting model size. Because it unties the
kernel weights along ?, the model size grows linearly with the equirectangular image height. The
model size can easily grow to tens of gigabytes as the image resolution increases.
4
Experiments
To evaluate our approach, we consider both the accuracy of its convolutions as well as its applicability
for object detections in 360? data. We use the VGG architecture3 and the Faster-RCNN [27] model as
our target network Np . We learn a network Ne to produce the topmost (conv5_3) convolution output.
Datasets We use two datasets: Pano2Vid for training, and Pano2Vid and PASCAL for testing.
Pano2Vid: We sample frames from the 360? videos in the Pano2Vid dataset [32] for both training
and testing. The dataset consists of 86 videos crawled from YouTube using four keywords: ?Hiking,?
?Mountain Climbing,? ?Parade,? and ?Soccer?. We sample frames at 0.05fps to obtain 1,056 frames
for training and 168 frames for testing. We use ?Mountain Climbing? for testing and others for
training, so the training and testing frames are from disjoint videos. See Supp. for sampling process.
Because the supervision is on a per pixel basis, this corresponds to N ? We ? He ? 250M (non
i.i.d.) samples. Note that most object categories targeted by the Faster-RCNN detector do not appear
in Pano2Vid, meaning that our experiments test the content-independence of our approach.
PASCAL VOC: Because the target model was originally trained and evaluated on PASCAL VOC 2007,
we ?360-ify? it to evaluate the object detector application. We test with the 4,952 PASCAL images,
which contain 12,032 bounding boxes. We transform them to equirectangular images as if they
3
https://github.com/rbgirshick/py-faster-rcnn
6
originated from a 360? camera. In particular, each object bounding box is backprojected to 3 different
scales {0.5R, 1.0R, 1.5R} and 5 different polar angles ??{36?, 72?, 108?, 144?, 180?} on the 360?
image sphere using the inverse perspective projection, where R is the resolution of the target network?s
Rf. Regions outside the bounding box are zero-padded. See Supp. for details. Backprojection allows
us to evaluate the performance at different levels of distortion in the equirectangular projection.
Metrics We generate the output widely used in the literature (conv5_3) and evaluate it with the
following metrics.
Network output error measures the difference between Ne (Ie ) and Np (Is ). In particular, we report
the root-mean-square error (RMSE) over all pixels and channels. For PASCAL, we measure the error
over the Rf of the detector network.
Detector network performance measures the performance of the detector network in Faster-RCNN
using multi-class classification accuracy. We replace the ROI-pooling in Faster-RCNN by pooling
over the bounding box in Ie . Note that the bounding box is backprojected to equirectangular projection
and is no longer a square region.
Proposal network performance evaluates the proposal network in Faster-RCNN using average
Intersection-over-Union (IoU). For each bounding box centered at n
? , we project the conv5_3 output
to the tangent plane n
? using P and apply the proposal network at the center of the bounding box on
the tangent plane. Given the predicted proposals, we compute the IoUs between foreground proposals
and the bounding box and take the maximum. The IoU is set to 0 if there is no foreground proposal.
Finally, we average the IoU over bounding boxes.
We stress that our goal is not to build a new object detector; rather, we aim to reproduce the behavior
of existing 2D models on 360? data with lower computational cost. Thus, the metrics capture how
accurately and how quickly we can replicate the exact solution.
Baselines We compare our method with the following baselines.
? E XACT ? Compute the true target value Np (Is )[?, ?] for every pixel. This serves as an upper
bound in performance and does not consider the computational cost.
? D IRECT ? Apply Np on Ie directly. We replace max-pooling with dilated convolution to produce
a full resolution output. This is Strategy I in Fig. 1 and is used in 360? video analysis [16, 23].
? I NTERP ? Compute Np (Is )[?, ?] every S-pixels and interpolate the values for the others. We set
S such that the computational cost is roughly the same as our S PH C ONV. This is a more efficient
variant of Strategy II in Fig. 1.
? P ERSPECT ? Project Is onto a cube map [2] and then apply Np on each face of the cube, which is
a perspective image with 90? FOV. The result is backprojected to Ie to obtain the feature on Ie .
We use W =960 for the cube map resolution so ?? is roughly the same as Ip . This is a second
variant of Strategy II in Fig. 1 used in PanoContext [38].
S PH C ONV variants We evaluate three variants of our approach:
? O PT S PH C ONV ? To compute the output for each layer l, O PT S PH C ONV computes the exact
output for layer l?1 using Np (Is ) then applies spherical convolution for layer l. O PT S PH C ONV
serves as an upper bound for our approach, where it avoids accumulating any error across layers.
? S PH C ONV-P RE ? Uses the weights from kernel-wise pre-training directly without fine-tuning.
? S PH C ONV ? The full spherical convolution with joint fine-tuning of all layers.
Implementation details We set the resolution of Ie to 640?320. For the projection operator P, we
map ?=65.5? to W =640 pixels following SUN360 [35]. The pixel size is therefore ??e =360?/640
for Ie and ??p =65.5?/640 for Ip . Accordingly, we remove the first three max-pooling layers so
Ne has only one max-pooling layer following conv4_3. The kernel size upper bound Uk =7 ? 7
following the max kernel size in VGG. We insert batch normalization for conv4_1 to conv5_3. See
Supp. for details.
4.1
Network output accuracy and computational cost
Fig. 5a shows the output error of layers conv3_3 and conv5_3 on the Pano2Vid [32] dataset (see
Supp. for similar results on other layers.). The error is normalized by that of the mean predictor. We
evaluate the error at 5 polar angles ? uniformly sampled from the northern hemisphere, since error is
roughly symmetric with the equator.
7
conv3 3 RMSE
conv5 3 RMSE
2
2
1
1
0
18? 36? 54? 72? 90?
?
0
18? 36? 54? 72? 90?
Direct
Interp
Perspective
Exact
OptSphConv
SphConv-Pre
SphConv
0.8
Accuracy
1
0.6
0.5
0.5
0.4
conv5 3 RMSE
1.5
0.7
101
102
103
0
0
2
4
6
Tera-MACs
(a) Network output errors vs. polar angle
(b) Cost vs. accuracy
Figure 5: (a) Network output error on Pano2Vid; lower is better. Note the error of E XACT is 0 by definition.
Our method?s convolutions are much closer to the exact solution than the baselines?. (b) Computational cost
vs. accuracy on PASCAL. Our approach yields accuracy closest to the exact solution while requiring orders of
magnitude less computation time (left plot). Our cost is similar to the other approximations tested (right plot).
Plot titles indicate the y-labels, and error is measured by root-mean-square-error (RMSE).
Figure 6: Three AlexNet conv1 kernels (left squares) and their corresponding four S PH C ONV-P RE kernels at
? ? {9?, 18?, 36?, 72?} (left to right).
First we discuss the three variants of our method. O PT S PH C ONV performs the best in all layers
and ?, validating our main idea of spherical convolution. It performs particularly well in the lower
layers, because the Rf is larger in higher layers and the distortion becomes more significant. Overall,
S PH C ONV-P RE performs the second best, but as to be expected, the gap with O PT C ONV becomes
larger in higher layers because of error propagation. S PH C ONV outperforms S PH C ONV-P RE in
conv5_3 at the cost of larger error in lower layers (as seen here for conv3_3). It also has larger error
at ?=18? for two possible reasons. First, the learning curve indicates that the network learns more
slowly near the pole, possibly because the Rf is larger and the pixels degenerate. Second, we optimize
the joint L2 loss, which may trade the error near the pole with that at the center.
Comparing to the baselines, we see that ours achieves lowest errors. D IRECT performs the worst
among all methods, underscoring that convolutions on the flattened sphere?though fast?are inadequate. I NTERP performs better than D IRECT, and the error decreases in higher layers. This is because
the Rf is larger in the higher layers, so the S-pixel shift in Ie causes relatively smaller changes in
the Rf and therefore the network output. P ERSPECTIVE performs similarly in different layers and
outperforms I NTERP in lower layers. The error of P ERSPECTIVE is particularly large at ?=54?, which
is close to the boundary of the perspective image and has larger perspective distortion.
Fig. 5b shows the accuracy vs. cost tradeoff. We measure computational cost by the number of
Multiply-Accumulate (MAC) operations. The leftmost plot shows cost on a log scale. Here we
see that E XACT?whose outputs we wish to replicate?is about 400 times slower than S PH C ONV,
and S PH C ONV approaches E XACT?s detector accuracy much better than all baselines. The second
plot shows that S PH C ONV is about 34% faster than I NTERP (while performing better in all metrics).
P ERSPECTIVE is the fastest among all methods and is 60% faster than S PH C ONV, followed by
D IRECT which is 23% faster than S PH C ONV. However, both baselines are noticeably inferior in
accuracy compared to S PH C ONV.
To visualize what our approach has learned, we learn the first layer of the AlexNet [22] model
provided by the Caffe package [20] and examine the resulting kernels. Fig. 6 shows the original
kernel Kp and the corresponding kernels Ke at different polar angles ?. Ke is usually the re-scaled
version of Kp , but the weights are often amplified because multiple pixels in Kp fall to the same
pixel in Ke like the second example. We also observe situations where the high frequency signal in
the kernel is reduced, like the third example, possibly because the kernel is smaller. Note that we
learn the first convolution layer for visualization purposes only, since l = 1 (only) has an analytic
solution (cf. Sec 3.3). See Supp. for the complete set of kernels.
4.2
Object detection and proposal accuracy
Having established our approach provides accurate and efficient Ne convolutions, we now examine
how important that accuracy is to object detection on 360? inputs. Fig. 7a shows the result of the
Faster-RCNN detector network on PASCAL in 360? format. O PT S PH C ONV performs almost as well
as E XACT. The performance degrades in S PH C ONV-P RE because of error accumulation, but it still
8
2
Output RMSE
1.5
0.6
1
0.4
0.5
0.2
18? 36? 54? 72? 90?
0
18? 36? 54? 72? 90?
Direct
Interp
Perspective
Exact
OptSphConv
SphConv-Pre
SphConv
Scale = 0.5R
0.3
IoU
Accuracy
0.8
Scale = 1.0R
0.3
0.2
0.2
0.1
0.1
0
18? 36? 54? 72? 90?
0
18? 36? 54? 72? 90?
(a) Detector network performance.
(b) Proposal network accuracy (IoU).
Figure 7: Faster-RCNN object detection accuracy on a 360? version of PASCAL across polar angles ?, for both
the (a) detector network and (b) proposal network. R refers to the Rf of Np . Best viewed in color.
Figure 8: Object detection examples on 360? PASCAL test images. Images show the top 40% of equirectangular
projection; black regions are undefined pixels. Text gives predicted label, multi-class probability, and IoU, resp.
Our method successfully detects objects undergoing severe distortion, some of which are barely recognizable
even for a human viewer.
significantly outperforms D IRECT and is better than I NTERP and P ERSPECTIVE in most regions.
Although joint training (S PH C ONV) improves the output error near the equator, the error is larger near
the pole which degrades the detector performance. Note that the Rf of the detector network spans
multiple rows, so the error is the weighted sum of the error at different rows. The result, together
with Fig. 5a, suggest that S PH C ONV reduces the conv5_3 error in parts of the Rf but increases it at
the other parts. The detector network needs accurate conv5_3 features throughout the Rf in order to
generate good predictions.
D IRECT again performs the worst. In particular, the performance drops significantly at ?=18?,
showing that it is sensitive to the distortion. In contrast, I NTERP performs better near the pole
because the samples are denser on the unit sphere. In fact, I NTERP should converge to E XACT at the
pole. P ERSPECTIVE outperforms I NTERP near the equator but is worse in other regions. Note that
??{18?, 36?} falls on the top face, and ?=54? is near the border of the face. The result suggests that
P ERSPECTIVE is still sensitive to the polar angle, and it performs the best when the object is near the
center of the faces where the perspective distortion is small.
Fig. 7b shows the performance of the object proposal network for two scales (see Supp. for more).
Interestingly, the result is different from the detector network. O PT S PH C ONV still performs almost
the same as E XACT, and S PH C ONV-P RE performs better than baselines. However, D IRECT now
outperforms other baselines, suggesting that the proposal network is not as sensitive as the detector
network to the distortion introduced by equirectangular projection. The performance of the methods
is similar when the object is larger (right plot), even though the output error is significantly different.
The only exception is P ERSPECTIVE, which performs poorly for ??{54?, 72?, 90?} regardless of the
object scale. It again suggests that objectness is sensitive to the perspective image being sampled.
Fig. 8 shows examples of objects successfully detected by our approach in spite of severe distortions.
See Supp. for more examples.
5
Conclusion
We propose to learn spherical convolutions for 360? images. Our solution entails a new form of
distillation across camera projection models. Compared to current practices for feature extraction on
360? images/video, spherical convolution benefits efficiency by avoiding performing multiple perspective projections, and it benefits accuracy by adapting kernels to the distortions in equirectangular
projection. Results on two datasets demonstrate how it successfully transfers state-of-the-art vision
models from the realm of limited FOV 2D imagery into the realm of omnidirectional data.
Future work will explore S PH C ONV in the context of other dense prediction problems like segmentation, as well as the impact of different projection models within our basic framework.
9
References
[1]
[2]
[3]
[4]
[5]
[6]
[7]
[8]
[9]
[10]
[11]
[12]
[13]
[14]
[15]
[16]
[17]
[18]
[19]
[20]
[21]
[22]
[23]
[24]
[25]
[26]
[27]
[28]
[29]
[30]
[31]
[32]
[33]
https://facebook360.fb.com/editing-360-photos-injecting-metadata/.
https://code.facebook.com/posts/1638767863078802/under-the-hood-building-360-video/.
J. Ba and R. Caruana. Do deep nets really need to be deep? In NIPS, 2014.
A. Barre, A. Flocon, and R. Hansen. Curvilinear perspective, 1987.
C. Bucilu?a, R. Caruana, and A. Niculescu-Mizil. Model compression. In ACM SIGKDD, 2006.
T. Cohen, M. Geiger, and M. Welling. Convolutional networks for spherical signals. arXiv preprint
arXiv:1709.04893, 2017.
J. Dai, H. Qi, Y. Xiong, Y. Li, G. Zhang, H. Hu, and Y. Wei. Deformable convolutional networks. arXiv
preprint arXiv:1703.06211, 2017.
J. Deng, W. Dong, R. Socher, L. Li, and L. Fei-Fei. Imagenet: a large-scale hierarchical image database.
In CVPR, 2009.
M. Everingham, S. M. A. Eslami, L. Van Gool, C. K. I. Williams, J. Winn, and A. Zisserman. The pascal
visual object classes challenge: A retrospective. International Journal of Computer Vision, 111(1):98?136,
Jan. 2015.
C. Feichtenhofer, A. Pinz, and A. Zisserman. Convolutional two-stream network fusion for video action
recognition. In CVPR, 2016.
R. Girshick, J. Donahue, T. Darrell, and J. Malik. Rich feature hierarchies for accurate object detection
and semantic segmentation. In CVPR, 2014.
S. Gupta, J. Hoffman, and J. Malik. Cross modal distillation for supervision transfer. In CVPR, 2016.
K. He, G. Gkioxari, P. Doll?r, and R. Girshick. Mask r-cnn. In ICCV, 2017.
K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. In CVPR, 2016.
G. Hinton, O. Vinyals, and J. Dean. Distilling the knowledge in a neural network. arXiv preprint
arXiv:1503.02531, 2015.
H.-N. Hu, Y.-C. Lin, M.-Y. Liu, H.-T. Cheng, Y.-J. Chang, and M. Sun. Deep 360 pilot: Learning a deep
agent for piloting through 360? sports video. In CVPR, 2017.
M. Jaderberg, K. Simonyan, A. Zisserman, et al. Spatial transformer networks. In Advances in Neural
Information Processing Systems, pages 2017?2025, 2015.
S. Jain, B. Xiong, and K. Grauman. Fusionseg: Learning to combine motion and appearance for fully
automatic segmentation of generic objects in video. In CVPR, 2017.
Y. Jeon and J. Kim. Active convolution: Learning the shape of convolution for image classification. arXiv
preprint arXiv:1703.09076, 2017.
Y. Jia, E. Shelhamer, J. Donahue, S. Karayev, J. Long, R. Girshick, S. Guadarrama, and T. Darrell. Caffe:
Convolutional architecture for fast feature embedding. In ACM MM, 2014.
R. Khasanova and P. Frossard. Graph-based classification of omnidirectional images. arXiv preprint
arXiv:1707.08301, 2017.
A. Krizhevsky, I. Sutskever, and G. Hinton. Imagenet classification with deep convolutional neural
networks. In NIPS, 2012.
W.-S. Lai, Y. Huang, N. Joshi, C. Buehler, M.-H. Yang, and S. B. Kang. Semantic-driven generation of
hyperlapse from 360? video. IEEE Transactions on Visualization and Computer Graphics, PP(99):1?1,
2017.
Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-based learning applied to document recognition.
In Proc. of the IEEE, 1998.
J. Long, E. Shelhamer, and T. Darrell. Fully convolutional networks for semantic segmentation. In CVPR,
2015.
E. Parisotto, J. Ba, and R. Salakhutdinov. Actor-mimic: Deep multitask and transfer reinforcement learning.
In ICLR, 2016.
S. Ren, K. He, R. Girshick, and J. Sun. Faster r-cnn: Towards real-time object detection with region
proposal networks. In NIPS, 2015.
A. Romero, N. Ballas, S. E. Kahou, A. Chassang, C. Gatta, and Y. Bengio. Fitnets: Hints for thin deep
nets. In ICLR, 2015.
K. Simonyan and A. Zisserman. Two-stream convolutional networks for action recognition in videos. In
NIPS, 2014.
K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition. In
ICLR, 2015.
Y.-C. Su and K. Grauman. Making 360? video watchable in 2d: Learning videography for click free
viewing. In CVPR, 2017.
Y.-C. Su, D. Jayaraman, and K. Grauman. Pano2vid: Automatic cinematography for watching 360? videos.
In ACCV, 2016.
D. Tran, L. Bourdev, R. Fergus, L. Torresani, and M. Paluri. Learning spatiotemporal features with 3d
convolutional networks. In ICCV, 2015.
10
[34] Y.-X. Wang and M. Hebert. Learning to learn: Model regression networks for easy small sample learning.
In ECCV, 2016.
[35] J. Xiao, K. A. Ehinger, A. Oliva, and A. Torralba. Recognizing scene viewpoint using panoramic place
representation. In CVPR, 2012.
[36] F. Yu and V. Koltun. Multi-scale context aggregation by dilated convolutions. arXiv preprint
arXiv:1511.07122, 2015.
[37] L. Zelnik-Manor, G. Peters, and P. Perona. Squaring the circle in panoramas. In ICCV, 2005.
[38] Y. Zhang, S. Song, P. Tan, and J. Xiao. Panocontext: A whole-room 3d context model for panoramic scene
understanding. In ECCV, 2014.
[39] B. Zhou, A. Lapedriza, J. Xiao, A. Torralba, and A. Oliva. Learning deep features for scene recognition
using places database. In NIPS, 2014.
11
| 6656 |@word multitask:1 cnn:16 version:2 rising:1 compression:2 replicate:4 everingham:1 hu:2 zelnik:1 azimuthal:1 rgb:2 photographer:1 liu:1 offering:1 ours:1 interestingly:1 document:1 outperforms:5 existing:13 current:2 com:3 comparing:1 guadarrama:1 yet:1 must:1 subsequent:1 romero:1 shape:8 analytic:2 remove:2 plot:6 drop:1 v:5 accordingly:2 plane:13 core:1 provides:4 location:2 wxy:4 zhang:3 five:1 height:4 along:4 direct:2 koltun:1 fps:1 consists:2 overhead:1 combine:3 recognizable:1 introduce:3 jayaraman:1 sacrifice:1 mask:1 indeed:1 expected:1 roughly:3 frossard:1 nor:1 examine:2 multi:4 market:1 behavior:3 paluri:1 salakhutdinov:1 voc:3 spherical:29 detects:1 increasing:1 becomes:2 lapedriza:1 project:6 provided:1 underlying:1 unrelated:1 bounded:1 medium:2 alexnet:4 lowest:1 what:1 mountain:2 developer:1 transformation:9 every:8 expands:2 exactly:2 grauman:4 scaled:1 uk:4 control:1 unit:4 stick:1 appear:1 arguably:1 positive:1 before:2 local:1 thinner:1 limit:1 bilinear:1 eslami:1 interpolation:2 mercator:1 black:1 fitnets:1 initialization:1 studied:2 fov:7 suggests:2 fastest:1 limited:3 range:1 unique:1 camera:13 responsible:1 testing:6 hood:1 practice:4 backprojecting:2 union:1 differs:2 lecun:1 procedure:3 jan:1 tera:1 significantly:4 adapting:1 projection:51 pre:16 regular:1 refers:2 spite:1 suggest:1 cannot:4 onto:3 targeting:1 unlabeled:3 operator:5 close:5 context:4 applying:6 transformer:2 py:1 optimize:3 accumulating:1 map:5 accumulation:1 center:6 gkioxari:1 dean:1 go:1 attention:1 starting:1 independently:2 regardless:1 rectangular:1 ke:9 resolution:5 simplicity:1 williams:1 immediately:1 adjusts:1 enabled:1 gigabyte:1 embedding:1 handle:2 autonomous:1 coordinate:3 resp:1 target:20 today:1 heavily:1 massive:2 exact:10 pt:7 hierarchy:1 us:2 tan:1 trend:1 recognition:10 expensive:1 particularly:2 kahou:1 labeled:4 database:2 bottom:4 preprint:6 wang:1 capture:3 worst:2 region:7 sun:3 trade:1 decrease:1 topmost:1 complexity:1 pinz:1 trained:12 raise:1 depend:2 creation:1 incur:1 efficiency:3 basis:1 easily:2 joint:3 convolves:1 surrounding:2 jain:1 fast:6 describe:1 kp:11 detected:1 outside:1 exhaustive:1 caffe:2 whose:1 heuristic:1 widely:3 larger:11 denser:2 distortion:18 relax:1 otherwise:2 compressed:1 cvpr:10 ability:1 simonyan:3 transform:2 jointly:1 final:3 ip:10 karayev:1 net:2 propose:5 tran:1 irect:7 product:1 adaptation:2 rapidly:1 degenerate:1 achieve:1 deformable:2 adapts:1 amplified:1 poorly:1 kh:12 validate:1 curvilinear:1 sutskever:1 double:1 reprojection:1 darrell:3 produce:7 perfect:1 converges:1 object:28 resnet:1 help:2 develop:1 depending:1 derive:1 wider:1 measured:1 ij:3 keywords:1 bourdev:1 conv5:2 eq:5 longitude:2 predicted:2 indicate:1 distilling:1 iou:7 annotated:2 cnns:8 filter:5 centered:1 human:1 viewing:6 enable:1 virtual:1 noticeably:1 require:4 really:2 kristen:1 kel:9 insert:1 viewer:1 mm:1 around:1 ground:1 roi:1 great:1 mapping:1 predict:1 visualize:1 automate:1 achieves:1 torralba:2 purpose:1 polar:8 proc:1 injecting:1 label:2 iw:1 hansen:1 title:1 sensitive:5 bridge:1 bucilu:1 successfully:4 tool:1 weighted:2 hoffman:1 xact:7 aim:2 ke1:1 rather:4 manor:1 avoid:1 shelf:1 zhou:1 varying:1 mobile:1 crawled:1 buehler:1 inherits:1 panoramic:3 indicates:1 contrast:3 sigkdd:1 baseline:10 kim:1 dependent:1 feichtenhofer:1 niculescu:1 inaccurate:2 entire:3 transferring:1 typically:1 initially:1 perona:1 squaring:1 expand:1 reproduce:4 pixel:25 overall:1 classification:4 among:2 pascal:11 art:3 spatial:2 cube:3 field:10 stns:2 saving:1 extraction:5 beach:1 flawed:1 sampling:4 distilled:1 manually:1 kw:9 yu:2 having:1 thin:1 foreground:2 discrepancy:1 mimic:2 np:33 others:2 report:1 future:1 few:1 hint:1 torresani:1 kp1:1 recognize:1 interpolate:1 jeon:1 detection:11 interest:2 possibility:1 multiply:1 evaluation:1 adjust:6 severe:2 introduces:3 analyzed:1 extreme:1 undefined:1 accurate:8 closer:1 necessary:1 experience:1 glimpse:1 xy:3 indexed:1 conv4_1:1 initialized:1 re:7 circle:1 deformation:1 girshick:4 instance:1 column:1 ar:2 cover:4 caruana:2 ordinary:1 cost:17 unwrap:1 pole:8 applicability:1 mac:2 uniform:1 predictor:1 krizhevsky:1 recognizing:1 inadequate:1 graphic:2 too:2 watchable:1 spatiotemporal:1 st:1 explores:1 international:1 ie:27 systematic:1 off:3 dong:1 invertible:1 together:3 quickly:2 imagery:8 again:2 containing:1 huang:1 slowly:2 possibly:3 worse:1 watching:1 expert:1 li:2 supp:7 account:4 potential:1 distribute:1 de:2 suggesting:1 sec:5 dilated:3 coefficient:2 caused:1 depends:2 stream:2 performed:1 view:5 root:2 doing:1 portion:1 recover:1 aggregation:1 capability:1 annotation:2 jia:1 rmse:6 square:9 accuracy:23 convolutional:12 ensemble:1 yield:3 climbing:2 yes:2 raw:2 artist:1 produced:1 accurately:1 none:1 ren:2 researcher:2 chassang:1 detector:17 reach:1 suffers:1 sharing:1 manual:1 facebook:2 distort:1 definition:2 aligns:1 evaluates:1 frequency:1 pp:1 wearable:1 sampled:2 dataset:3 onv:34 adjusting:1 pilot:1 knowledge:3 color:1 improves:1 realm:2 segmentation:5 carefully:1 originally:1 higher:4 supervised:1 planar:8 awkward:1 response:1 wei:1 editing:1 zisserman:5 evaluated:1 though:6 box:10 strongly:1 furthermore:6 until:2 receives:1 horizontal:1 expressive:1 su:3 propagation:1 interp:2 grows:1 usa:1 effect:2 building:1 requiring:2 true:2 contain:1 normalized:1 symmetric:1 omnidirectional:3 semantic:3 during:1 width:2 inferior:1 noted:1 soccer:1 leftmost:1 stress:1 complete:2 demonstrate:2 performs:13 dedicated:1 interface:1 motion:1 fusionseg:1 image:76 wise:9 meaning:2 novel:2 common:4 network2:1 physical:1 cohen:1 ballas:1 million:1 he:8 accumulate:1 significant:5 composition:1 refer:1 distillation:7 tuning:4 automatic:2 grid:3 similarly:2 robot:1 entail:1 supervision:5 longer:1 actor:1 etc:2 closest:1 recent:2 perspective:32 hemisphere:1 driven:1 seen:3 dai:1 care:1 deng:1 converge:1 redundant:1 signal:2 ii:4 relates:1 multiple:7 full:2 reduces:2 faster:13 adapt:3 characterized:2 offer:2 sphere:14 long:4 match:2 cross:1 lin:1 post:1 lai:1 paired:1 impact:3 prediction:2 variant:5 basic:1 qi:1 regression:1 vision:8 metric:4 oliva:2 arxiv:12 kernel:55 adopting:1 represent:2 normalization:1 equator:3 proposal:13 addition:1 whereas:2 want:1 spacing:1 separately:1 fine:6 winn:1 grow:2 source:1 unlike:3 pass:1 backprojected:3 pooling:13 validating:1 joshi:1 near:12 leverage:1 yang:1 intermediate:4 split:1 enough:2 bengio:2 rendering:1 easy:1 independence:1 relu:2 equidistant:1 architecture:4 restrict:5 click:1 idea:2 haffner:1 vgg:5 translates:1 intensive:1 texas:1 shift:3 tradeoff:1 bottleneck:1 whether:1 motivated:2 effort:1 retrospective:1 song:1 peter:1 cause:2 repeatedly:2 action:4 deep:11 dramatically:1 useful:1 tune:2 chuan:1 ten:1 ph:33 nel:2 category:1 reduced:1 generate:4 http:3 northern:1 restricts:1 disjoint:1 popularity:1 per:1 broadly:1 taller:2 key:4 four:2 nevertheless:1 neither:1 graph:1 padded:1 year:2 sum:2 modal:1 angle:9 inverse:3 powerful:3 package:1 distorted:1 place:2 almost:2 throughout:1 geiger:1 decision:1 conv4_3:1 capturing:1 accelerates:1 entirely:1 layer:41 bound:5 npl:4 display:1 followed:1 cheng:1 hiking:1 constraint:1 fei:2 scene:5 flat:7 generates:1 underscoring:1 speed:1 simulate:1 span:2 performing:4 optical:2 format:2 relatively:1 influential:1 across:8 smaller:3 increasingly:1 making:3 projecting:1 iccv:3 computationally:1 vendor:1 visualization:2 remains:1 discus:1 eventually:1 fed:1 serf:2 photo:1 adopted:1 available:1 operation:5 doll:1 apply:3 observe:1 hierarchical:1 spectral:2 generic:1 xiong:2 alternative:1 batch:1 slower:1 original:4 hampered:1 top:5 compress:1 denotes:1 ensure:2 cf:1 exploit:2 restrictive:1 build:1 backprojection:1 objective:2 parade:1 question:1 malik:2 strategy:13 receptive:6 degrades:3 traditional:2 unclear:1 gradient:1 iclr:3 link:1 mapped:2 considers:1 trivial:1 reason:1 barely:1 consumer:2 besides:1 code:1 providing:2 minimizing:4 cij:5 conv5_3:8 ify:1 ba:2 implementation:1 c3d:1 policy:2 perform:2 shallower:1 upper:5 convolution:44 datasets:6 accv:1 situation:1 hinton:2 precise:2 frame:6 arbitrary:2 community:1 iwe:1 inferred:1 introduced:3 namely:1 imagenet:2 security:1 learned:4 tremendous:1 established:1 kang:1 nip:6 address:1 usually:3 latitude:1 challenge:2 rf:19 gaining:2 including:1 video:19 max:10 gool:1 power:1 predicting:1 residual:2 mizil:1 improve:1 github:1 technology:1 ne:22 picture:1 naive:1 coupled:1 metadata:1 text:1 prior:2 geometric:6 l2:4 tangent:9 epoch:1 literature:1 understanding:1 lacking:1 fully:4 par:1 loss:6 parisotto:1 generation:1 limitation:4 versus:1 digital:1 rcnn:9 shelhamer:2 degree:2 agent:1 conv1:1 displaying:1 viewpoint:2 principle:1 xiao:3 amortization:1 austin:1 row:8 course:1 eccv:2 last:1 free:2 hebert:1 wide:2 conv3:1 taking:1 face:4 fall:2 benefit:3 van:1 curve:2 depth:1 boundary:1 world:1 evaluating:1 avoids:1 computes:1 fb:1 rich:1 commonly:1 collection:1 projected:6 reinforcement:1 far:1 welling:1 transaction:1 approximate:2 ignore:1 cutting:1 jaderberg:1 global:1 active:2 overfitting:2 fergus:1 search:1 decade:1 reality:3 learn:16 transfer:6 nature:1 ca:1 channel:2 expansion:1 bottou:1 domain:7 main:3 dense:2 linearly:2 rh:1 bounding:10 border:1 whole:1 repeated:1 augmented:2 site:1 fig:17 ehinger:1 slow:1 vr:2 position:1 originated:1 wish:1 tied:1 third:1 extractor:2 learns:6 donahue:2 specific:1 showing:1 undergoing:1 gupta:1 fusion:1 essential:1 exists:1 socher:1 flattened:1 magnitude:3 illustrates:1 gap:1 generalizing:1 led:1 intersection:1 explore:2 appearance:1 visual:3 gatta:1 prevents:1 horizontally:1 vinyals:1 sport:1 chang:1 applies:2 corresponds:2 truth:1 immersive:1 acm:2 goal:6 narrower:1 targeted:1 viewed:1 towards:1 room:1 shared:1 replace:2 content:8 change:1 youtube:2 objectness:1 typical:1 except:1 uniformly:2 determined:1 panorama:1 invariance:1 ne1:1 exception:1 select:1 support:1 untie:1 evaluate:7 tested:1 avoiding:1 |
6,253 | 6,657 | MarrNet: 3D Shape Reconstruction via 2.5D Sketches
Jiajun Wu*
MIT CSAIL
Yifan Wang*
ShanghaiTech University
Tianfan Xue
MIT CSAIL
William T. Freeman
MIT CSAIL, Google Research
Xingyuan Sun
Shanghai Jiao Tong University
Joshua B. Tenenbaum
MIT CSAIL
Abstract
3D object reconstruction from a single image is a highly under-determined problem, requiring strong prior knowledge of plausible 3D shapes. This introduces
challenges for learning-based approaches, as 3D object annotations are scarce in
real images. Previous work chose to train on synthetic data with ground truth 3D
information, but suffered from domain adaptation when tested on real data.
In this work, we propose MarrNet, an end-to-end trainable model that sequentially
estimates 2.5D sketches and 3D object shape. Our disentangled, two-step formulation has three advantages. First, compared to full 3D shape, 2.5D sketches are much
easier to be recovered from a 2D image; models that recover 2.5D sketches are also
more likely to transfer from synthetic to real data. Second, for 3D reconstruction
from 2.5D sketches, systems can learn purely from synthetic data. This is because
we can easily render realistic 2.5D sketches without modeling object appearance
variations in real images, including lighting, texture, etc. This further relieves the
domain adaptation problem. Third, we derive differentiable projective functions
from 3D shape to 2.5D sketches; the framework is therefore end-to-end trainable on
real images, requiring no human annotations. Our model achieves state-of-the-art
performance on 3D shape reconstruction.
1
Introduction
Humans quickly recognize 3D shapes from a single image. Figure 1a shows a number of images
of chairs; despite their drastic difference in object texture, material, environment lighting, and
background, humans easily recognize they have very similar 3D shapes. What is the most essential
information that makes this happen?
Researchers in human perception argued that our 3D perception could rely on recovering 2.5D
sketches [Marr, 1982], which include intrinsic images [Barrow and Tenenbaum, 1978, Tappen et al.,
2003] like depth and surface normal maps (Figure 1b). Intrinsic images disentangle object appearance
variations in texture, albedo, lighting, etc., with its shape, which retains all information from the
observed image for 3D reconstruction. Humans further combine 2.5D sketches and a shape prior
learned from past experience to reconstruct a full 3D shape (Figure 1c). In the field of computer
vision, there have also been abundant works exploiting the idea for reconstruction 3D shapes of
faces [Kemelmacher-Shlizerman and Basri, 2011], objects [Tappen et al., 2003], and scenes [Hoiem
et al., 2005, Saxena et al., 2009].
Recently, researchers attempted to tackle the problem of single-image 3D reconstruction with deep
learning. These approaches usually regress a 3D shape from a single RGB image directly [Tulsiani
et al., 2017, Choy et al., 2016, Wu et al., 2016b]. In contrast, we propose a two-step while end-to-end
trainable pipeline, sequentially recovering 2.5D sketches (depth and normal maps) and a 3D shape.
? indicates equal contributions.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
(b) 2.5D Sketches
(c) 3D Shape
(a) Images
Figure 1: Objects in real images (a) are subject to appearance variations regarding color, texture,
lighting, material, background, etc. Despite this, their 2.5D sketches like surface normal and depth
maps remain constant (b). The 2.5D sketches can be seen as an abstraction of the image, retaining all
information about the 3D shape of the object inside. We combine the sketches with learned shape
priors to reconstruct the full 3D shape (c).
We use an encoder-decoder structure for each component of the framework, and also enforce the
reprojection consistency between the estimated sketch and the 3D shape. We name it MarrNet, for its
close resemblance to David Marr?s theory of perception [Marr, 1982].
Our approach offers several unique advantages. First, the use of 2.5D sketches releases the burden on
domain transfer. As single-image 3D reconstruction is a highly under-constrained problem, strong
prior knowledge of object shapes is needed. This poses challenges to learning-based methods, as
accurate 3D object annotations in real images are rare. Most previous methods turned to training
purely on synthetic data [Tulsiani et al., 2017, Choy et al., 2016, Girdhar et al., 2016]. However, these
approaches often suffer from the domain adaption issue due to imperfect rendering. Learning 2.5D
sketches from images, in comparison, is much easier and more robust to transfer from synthetic to
real images, as shown in Section 4.
Further, as our second step recovers 3D shape from 2.5D sketches ? an abstraction of the raw input
image, it can be trained purely relying on synthetic data. Though rendering diverse realistic images is
challenging, it is straightforward to obtain almost perfect object surface normals and depths from a
graphics engine. This further relieves the domain adaptation issue.
We also enforce differentiable constraints between 2.5D sketches and 3D shape, making our system
end-to-end trainable, even on real images without any annotations. Given a set of unlabeled images,
our algorithm, pre-trained on synthetic data, can infer the 2.5D sketches of objects in the image,
and use it to refine its estimation of objects? 3D shape. This self-supervised feature enhances its
performance on images from different domains.
We evaluate our framework on both synthetic images of objects from ShapeNet [Chang et al., 2015],
and real images from the PASCAL 3D+ dataset [Xiang et al., 2014]. We demonstrate that our
framework performs well on 3D shape reconstruction, both qualitatively and quantitatively.
Our contributions are three-fold: inspired by visual cognition theory, we propose a two-step, disentangled formulation for single-image 3D reconstruction via 2.5D sketches; we develop a novel,
end-to-end trainable model with a differentiable projection layer that ensures consistency between
3D shape and mid-level representations; we demonstrate its effectiveness on 2.5D sketch transfer and
3D shape reconstruction on both synthetic and real data.
2
Related Work
2.5D Sketch Recovery Estimating 2.5D sketches has been a long-standing problem in computer
vision. In the past, researchers have explored recovering 2.5D shape from shading, texture, or color
images [Horn and Brooks, 1989, Zhang et al., 1999, Tappen et al., 2003, Barron and Malik, 2015,
Weiss, 2001, Bell et al., 2014]. With the development of depth sensors [Izadi et al., 2011] and larger
scale RGB-D datasets [Silberman et al., 2012, Song et al., 2017, McCormac et al., 2017], there
have also been papers on estimating depth [Chen et al., 2016, Eigen and Fergus, 2015], surface
normals [Bansal and Russell, 2016, Wang et al., 2015], and other intrinsic images [Shi et al., 2017,
2
2.5D Sketches
(c) Reprojection Consistency
normal
??
depth
3D Shape
2D Image
(a) 2.5D Sketch Estimation
silhouette
(b) 3D Shape Estimation
Normal Ball
Figure 2: Our model (MarrNet) has three major components: (a) 2.5D sketch estimation, (b) 3D
shape estimation, and (c) a loss function for reprojection consistency. MarrNet first recovers object
normal, depth, and silhouette images from an RGB image. It then regresses the 3D shape from the
2.5D sketches. In both steps, it uses an encoding-decoding network. It finally employs a reprojection
consistency loss to ensure the estimated 3D shape aligns with the 2.5D sketches. The entire framework
can be trained end-to-end.
Janner et al., 2017] with deep networks. Our method employs 2.5D estimation as a component, but
targets reconstructing full 3D shape of an object.
Single-Image 3D Reconstruction The problem of recovering object shape from a single image
is challenging, as it requires both powerful recognition systems and prior shape knowledge. With
the development of large-scale shape repository like ShapeNet [Chang et al., 2015], researchers
developed models encoding shape prior for this task [Girdhar et al., 2016, Choy et al., 2016, Tulsiani
et al., 2017, Wu et al., 2016b, Kar et al., 2015, Kanazawa et al., 2016, Soltani et al., 2017], with
extension to scenes [Song et al., 2017]. These methods typically regress a voxelized 3D shape
directly from an input image, and rely on synthetic data or 2D masks for training. In comparison, our
formulation tackles domain difference better, as it can be end-to-end fine-tuned on images without
any annotations.
2D-3D Consistency It is intuitive and practically helpful to constrain the reconstructed 3D shape
to be consistent with 2D observations. Researchers have explored this idea for decades [Lowe, 1987].
This idea is also widely used in 3D shape completion from depths or silhouettes [Firman et al.,
2016, Rock et al., 2015, Dai et al., 2017]. Recently, a few papers discussed enforcing differentiable
2D-3D constraints between shape and silhouettes, enabling joint training of deep networks for 3D
reconstruction [Wu et al., 2016a, Yan et al., 2016, Rezende et al., 2016, Tulsiani et al., 2017]. In our
paper, we exploit this idea to develop differentiable constraints on the consistency between various
2.5D sketches and 3D shape.
3
Approach
To recover the 3D structure from a single view RGB image, our MarrNet contains three parts: first, a
2.5D sketch estimator, which predicts the depth, surface normal, and silhouette images of the object
(Figure 2a); second, a 3D shape estimator, which infers 3D object shape using a voxel representation
(Figure 2b); third, a reprojection consistency function, enforcing the alignment between the estimated
3D structure and inferred 2.5D sketches (Figure 2c).
3.1
2.5D Sketch Estimation
The first component of our network (Figure 2a) takes a 2D RGB image as input, and predicts its 2.5D
sketch: surface normal, depth, and silhouette. The goal of the 2.5D sketch estimation step is to distill
intrinsic object properties from input images, while discarding properties that are non-essential for
the task of 3D reconstruction, such as object texture and lighting.
We use an encoder-decoder network architecture for 2.5D sketch estimation. Our encoder is a
ResNet-18 [He et al., 2015], encoding a 256?256 RGB image into 512 feature maps of size 8?8.
3
?
?=4
?
?
? =?
Voxels that should be 1 for
reprojected depth consistency
Voxels that should be 0 for
reprojected depth consistency
?=
Voxels that should be 1 for reprojected
surface normal consistency
Figure 3: Reprojection consistency between 2.5D sketches and 3D shape. Left and middle: the
criteria for depths and silhouettes; right: the criterion for surface normals. See Section 3.3 for details.
The decoder contains four sets of 5?5 fully convolutional and ReLU layers, followed by four sets
of 1?1 convolutional and ReLU layers. It outputs the corresponding depth, surface normal, and
silhouette images, also at the resolution of 256?256.
3.2
3D Shape Estimation
The second part of our framework (Figure 2b) infers 3D object shape from estimated 2.5D sketches.
Here, the network focuses on learning the shape prior that explains input well. As it takes only surface
normal and depth images as input, it can be trained on synthetic data, without suffering from the
domain adaption problem: it is straightforward to render nearly perfect 2.5D sketches, but much
harder to render realistic images.
The network architecture is inspired by the TL network [Girdhar et al., 2016], and the 3D-VAEGAN [Wu et al., 2016b], again with an encoding-decoding style. It takes a normal image and a depth
image as input (both masked by the estimated silhouette), maps them to a 200-dim vector via five sets
of convolutional, ReLU, and pooling layers, followed by two fully connected layers. The detailed
encoder structure can be found in Girdhar et al. [2016]. The vector then goes through a decoder,
which consists of five fully convolutional and ReLU layers to output a 128?128?128 voxel-based
reconstruction of the input. The detailed encoder structure can be found in Wu et al. [2016b].
3.3
Reprojection Consistency
There have been works attempting to enforce the consistency between estimated 3D shape and 2D
representations in a neural network [Yan et al., 2016, Rezende et al., 2016, Wu et al., 2016a, Tulsiani
et al., 2017]. Here, we explore novel ways to include a reprojection consistency loss between the
predicted 3D shape and the estimated 2.5D sketch, consisting of a depth reprojection loss and a
surface normal reprojection loss.
We use vx,y,z to represent the value at position (x, y, z) in a 3D voxel grid, assuming that vx,y,z ?
[0, 1], ?x, y, z. We use dx,y to denote the estimated depth at position (x, y), and nx,y = (na , nb , nc )
to denote the estimated surface normal. We assume orthographic projection in this work.
Depths The projected depth loss tries to guarantee that the voxel with depth vx,y,dx,y should be 1,
and all voxels in front of it should be 0. This ensures the estimated 3D shape matches the estimated
depth values.
As illustrated in Figure 3a, we define projected depth loss as follows:
? 2
z < dx,y
?vx,y,z ,
Ldepth (x, y, z) = (1 ? vx,y,z )2 , z = dx,y .
?
0,
z > dx,y
(1)
The gradients are
?
2v
,
?Ldepth (x, y, z) ? x,y,z
= 2(vx,y,z ? 1),
?
?vx,y,z
0,
4
z < dx,y
z = dx,y .
z > dx,y
(2)
When dx,y = ?, our depth criterion reduces to a special case ? the silhouette criterion. As shown
in Figure 3b, for a line that has no intersection with the shape, all voxels in it should be 0.
Surface Normals As vectors nx = (0, ?nc , nb ) and ny = (?nc , 0, na ) are orthogonal to the normal vector nx,y = (na , nb , nc ), we can normalize them to obtain two vectors, n0x = (0, ?1, nb /nc )
and n0y = (?1, 0, na /nc ), both on the estimated surface plane at (x, y, z). The projected surface
normal loss tries to guarantee that the voxels at (x, y, z) ? n0x and (x, y, z) ? n+y should be 1 to
match the estimated surface normals. These constraints only apply when the target voxels are inside
the estimated silhouette.
As shown in Figure 3c, let z = dx,y , the projected surface normal loss is defined as
Lnormal (x, y, z) =
2
+ 1 ? vx,y+1,z? nb +
nc
nc
2
2
1 ? vx?1,y,z+ nna + 1 ? vx+1,y,z? nna .
1 ? vx,y?1,z+ nb
c
2
c
(3)
Then the gradients along the x direction are
?Lnormal (x, y, z)
= 2 vx?1,y,z+ nna ? 1
c
?vx?1,y,z+ nna
and
c
?Lnormal (x, y, z)
= 2 vx+1,y,z? nna ? 1 .
c
?vx+1,y,z? nna
c
(4)
The gradients along the y direction are similar.
3.4
Training Paradigm
We employ a two-step training paradigm. We first train the 2.5D sketch estimation and the 3D shape
estimation components separately on synthetic images; we then fine-tune the network on real images.
For pre-training, we use synthetic images of ShapeNet objects. The 2.5D sketch estimator is trained
using the ground truth surface normal, depth, and silhouette images with a L2 loss. The 3D interpreter
is trained using ground truth voxels and a cross-entropy loss. Please see Section 4.1 for details on
data preparation.
The reprojection consistency loss is used to fine-tune the 3D estimation component of our model on
real images, using the predicted normal, depth, and silhouette. We observe that a straightforward
implementation leads to shapes that explain 2.5D sketches well, but with unrealistic appearance. This
is because the 3D estimation module overfits the images without preserving the learned 3D shape
prior. See Figure 5 for examples, and Section 4.2 for more details.
We therefore choose to fix the decoder of the 3D estimator and only fine-tune the encoder. During
testing, our method can be self-supervised, i.e., we can fine-tune even on a single image without any
annotations. In practice, we fine-tune our model separately on each image for 40 iterations. For each
test image, fine-tuning takes up to 10 seconds on a modern GPU; without fine-tuning, testing time
is around 100 milliseconds. We use SGD for optimization with a batch size of 4, a learning rate of
0.001, and a momentum of 0.9. We implemented our framework in Torch7 [Collobert et al., 2011].
4
Evaluation
In this section, we present both qualitative and quantitative results on single-image 3D reconstruction
using variants of our framework. We evaluate our entire framework on both synthetic and real-life
images on three datasets.
4.1
3D Reconstruction on ShapeNet
Data We start with experiments on synthesized images of ShapeNet chairs [Chang et al., 2015]. We
put objects in front of random backgrounds from the SUN database [Xiao et al., 2010], and render the
corresponding RGB, depth, surface normal, and silhouette images. We use a physics-based renderer,
Mitsuba [Jakob, 2010], to obtain more realistic images. For each of the 6,778 ShapeNet chairs, we
render 20 images of random viewpoints.
5
Images
Estimated
normals
Estimated
depths
Direct predictions
MarrNet
Ground truth
Figure 4: Results on rendered images of ShapeNet objects [Chang et al., 2015]. From left to right:
input, estimated normal map, estimated depth map, our prediction, a baseline algorithm that predicts
3D shape directly from RGB input without modeling 2.5D sketch, and ground truth. Both normal and
depth maps are masked by predicted silhouettes. Our method is able to recover shapes with smoother
surfaces and finer details.
Methods We follow the training paradigm described in Section 3.4, but without the final fine-tuning
stage, as ground truth 3D shapes are available on this synthetic dataset. Specifically, the 2.5D sketch
estimator is trained using ground truth depth, normal and silhouette images and a L2 reconstruction
loss. The 3D shape estimation module takes in the masked ground truth depth and normal images as
input, and predicts 3D voxels of size 128?128?128 with a binary cross entropy loss.
We compare MarrNet with a baseline that predicts 3D shape directly from an RGB image, without
modeling 2.5D sketches. The baseline employs the same architecture as our 3D shape estimator
(Section 3.2). We show qualitative results in Figure 4. Our estimated surface normal and depth images
abstract out non-essential information like textures and lighting in the RGB image, while preserving
intrinsic information about object shape. Compared with the direct prediction baseline, our model
outputs objects with more details and smoother surfaces. For quantitative evaluation, previous works
usually compute the Intersection-over-Union (IoU) [Tulsiani et al., 2017, Choy et al., 2016]. Our full
model achieves a higher IoU (0.57) than the direct prediction baseline (0.52).
4.2
3D Reconstruction on Pascal 3D+
Data PASCAL 3D+ dataset [Xiang et al., 2014] provides (rough) 3D models for objects in real-life
images. Here, we use the same test set of PASCAL 3D+ with earlier works [Tulsiani et al., 2017].
Methods We follow the paradigm described in Section 3.4: we first train each module separately
on the ShapeNet dataset, and then fine-tune them on the PASCAL 3D+ dataset. Unlike previous
works [Tulsiani et al., 2017], our model requires no silhouettes as input during fine-tuning; it instead
estimates silhouette jointly.
As an ablation study, we compare three variants of our model: first, the model trained using ShapeNet
data only, without fine-tuning; second, the fine-tuned model whose decoder is not fixed during
6
Images
Estimated
normals and
silhouettes
Estimated
depths and
silhouettes
Estimated 3D Two views of refined 3D shape Two views of refined 3D shape
not fine-tuned fine-tuned, not fixing decoder
fine-tuned, fixing decoder
Figure 5: We present an ablation study, where we compare variants of our models. From left to
right: input, estimated normal, estimated depth, 3D prediction before fine-tuning, two views of the
3D prediction after fine-tuning without fixing decoder, and two views of the 3D prediction after
fine-tuning with the decoder fixed. When the decoder is not fixed, the model explains the 2.5D sketch
well, but fails to preserve the learned shape prior. Fine-tuning with a fixed decoder resolves the issue.
fine-tuning; and third, the full model whose decoder is fixed during fine-tuning. We also compare
with the state-of-the-art method (DRC) [Tulsiani et al., 2017], and the provided ground truth shapes.
Results The results of our ablation study are shown in Figure 5. The model trained on synthetic
data provides a reasonable shape estimate. If we fine-tune the model on Pascal 3D+ without fixing
the decoder, the output voxels explain the 2.5D sketch data well, but fail to preserve the learned shape
prior, leading to impossible shapes from certain views. Our final model, fine-tuned with the decoder
fixed, keeps the shape prior and provides more details of the shape.
We show more results in Figure 6, where we compare with the state-of-the-art (DRC) [Tulsiani et al.,
2017], and the provided ground truth shapes. Quantitatively, our algorithm achieves a higher IoU
over these methods (MarrNet 0.39 vs. DRC 0.34). However, we find the IoU metric sub-optimal
for three reasons. First, measuring 3D shape similarity is a challenging yet unsolved problem, and
IoU prefers models that predict mean shapes consistently, with no emphasize on details. Second, as
object shape can only be reconstructed up to scale from a single image, it requires searching over all
possible scales during the computation of IoU, making it less efficient. Third, as discussed in Tulsiani
et al. [2017], PASCAL 3D+ has only rough 3D annotations (10 CAD chair models for all images).
Computing IoU with these shapes would thus not be the most informative evaluation metric.
We instead conduct human studies, where we show users the input image and two reconstructions,
and ask them to choose the one that looks closer to the shape in the image. We show each test image
to 10 human subjects. As shown in Table 1, our reconstruction is preferred 74% of the time to DRC,
and 42% of the time to ground truth, showing a clear advantage.
We present some failure cases in Figure 7. Our algorithm does not perform well on recovering
complex, thin structure, and sometimes fails when the estimated mask is very inaccurate. Also, while
DRC may benefit from multi-view supervision, we have only evaluated MarrNet given a single view
of the shape, though adapting our formulation to multi-view data should be straightforward.
7
Images Ground truth
DRC
MarrNet
Images
Ground truth DRC
MarrNet
Figure 6: 3D reconstructions of chairs on the Pascal 3D+ [Xiang et al., 2014] dataset. From left to
right: input, the ground truth shape from the dataset, 3D estimation by DRC [Tulsiani et al., 2017],
and two views of MarrNet predictions. Our model recovers more accurate 3D shapes.
DRC
MarrNet
Ground truth
DRC
MarrNet
GT
50
74
83
26
50
58
17
42
50
Table 1: Human preferences on chairs in PASCAL 3D+ [Xiang et al., 2014]. We compare
MarrNet with the state-of-the-art (DRC) [Tulsiani et al., 2017], and the ground truth provided
by the dataset. Each number shows the percentage of humans prefer the left method to the top
one. MarrNet is preferred 74% of the time to
DRC, and 42% of the time to ground truth.
4.3
Images
Estimated Estimated
normals
depths
MarrNet
Figure 7: Failure cases on Pascal 3D+. Our algorithm does not perform well on recovering complex, thin structure, and sometimes fails when
the estimated mask is very inaccurate.
3D Reconstruction on IKEA
Data The IKEA dataset [Lim et al., 2013] contains images of IKEA furniture, along with accurate
3D shape and pose annotations. These images are challenging, as objects are often heavily occluded
or cropped. We also evaluate our model on the IKEA dataset.
Results We show qualitative results in Figure 8, where we compare with estimations by 3D-VAEGAN [Wu et al., 2016b] and the ground truth. As shown in the figure, our model can deal with mild
occlusions in real life scenarios. We also conduct human studies on the IKEA dataset. Results show
that 61% of the subjects prefer our reconstructions to those of 3D-VAE-GAN.
4.4
Extensions
We also apply our framework on cars and airplanes. We use the same setup as that for chairs. As
shown in Figure 9, shape details like the horizontal stabilizer and rear-view mirrors are recovered
8
Images
Ground truth
3DVAE-GAN
MarrNet
Images
Ground truth
3DVAE-GAN
MarrNet
Figure 8: 3D reconstruction of chairs on the IKEA [Lim et al., 2013] dataset. From left to right:
input, ground truth, 3D estimation by 3D-VAE-GAN [Wu et al., 2016b], and two views of MarrNet
predictions. Our model recovers more details compared to 3D-VAE-GAN.
Images
Ground truth
DRC
MarrNet
Images
Ground truth
DRC
MarrNet
Figure 9: 3D reconstructions of airplanes and cars from PASCAL 3D+. From left to right: input, the
ground truth shape from the dataset, 3D estimation by DRC [Tulsiani et al., 2017], and two views of
MarrNet predictions.
Images
MarrNet
Ground truth
Images
MarrNet
Ground truth
Figure 10: 3D reconstruction of objects from multiple categories on the PASCAL 3D+ [Xiang et al.,
2014] dataset. MarrNet also recovers 3D shape well when it is trained on multiple categories.
by our model. We further train MarrNet jointly on all three object categories, and show results in
Figure 10. Our model successfully recovers shapes of different categories.
5
Conclusion
We proposed MarrNet, a novel model that explicitly models 2.5D sketches for single-image 3D
shape reconstruction. The use of 2.5D sketches enhanced the model?s performance, and made it
easily adaptive to images across domains or even categories. We also developed differentiable loss
functions for the consistency between 3D shape and 2.5D sketches, so that MarrNet can be end-to-end
fine-tuned on real images without annotations. Experiments demonstrated that our model performs
well, and is preferred by human annotators over competitors.
9
Acknowledgements
We thank Shubham Tulsiani for sharing the DRC results, and Chengkai Zhang for the help on shape
visualization. This work is supported by NSF #1212849 and #1447476, ONR MURI N00014-161-2007, the Center for Brain, Minds and Machines (NSF #1231216), Toyota Research Institute,
Samsung, and Shell.
References
Aayush Bansal and Bryan Russell. Marr revisited: 2d-3d alignment via surface normal prediction. In CVPR,
2016. 2
Jonathan T Barron and Jitendra Malik. Shape, illumination, and reflectance from shading. IEEE TPAMI, 37(8):
1670?1687, 2015. 2
Harry G Barrow and Jay M Tenenbaum. Recovering intrinsic scene characteristics from images. Computer
vision systems, 1978. 1
Sean Bell, Kavita Bala, and Noah Snavely. Intrinsic images in the wild. ACM TOG, 33(4):159, 2014. 2
Angel X Chang, Thomas Funkhouser, Leonidas Guibas, Pat Hanrahan, Qixing Huang, Zimo Li, Silvio
Savarese, Manolis Savva, Shuran Song, Hao Su, et al. Shapenet: An information-rich 3d model repository. arXiv:1512.03012, 2015. 2, 3, 5, 6
Weifeng Chen, Zhao Fu, Dawei Yang, and Jia Deng. Single-image depth perception in the wild. In NIPS, 2016.
2
Christopher B Choy, Danfei Xu, JunYoung Gwak, Kevin Chen, and Silvio Savarese. 3d-r2n2: A unified approach
for single and multi-view 3d object reconstruction. In ECCV, 2016. 1, 2, 3, 6
Ronan Collobert, Koray Kavukcuoglu, and Cl?ment Farabet. Torch7: A matlab-like environment for machine
learning. In BigLearn, NIPS Workshop, 2011. 5
Angela Dai, Charles Ruizhongtai Qi, and Matthias Nie?ner. Shape completion using 3d-encoder-predictor cnns
and shape synthesis. In CVPR, 2017. 3
David Eigen and Rob Fergus. Predicting depth, surface normals and semantic labels with a common multi-scale
convolutional architecture. In ICCV, 2015. 2
Michael Firman, Oisin Mac Aodha, Simon Julier, and Gabriel J Brostow. Structured prediction of unobserved
voxels from a single depth image. In CVPR, 2016. 3
Rohit Girdhar, David F Fouhey, Mikel Rodriguez, and Abhinav Gupta. Learning a predictable and generative
vector representation for objects. In ECCV, 2016. 2, 3, 4
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In
CVPR, 2015. 3
Derek Hoiem, Alexei A Efros, and Martial Hebert. Automatic photo pop-up. ACM TOG, 24(3):577?584, 2005.
1
Berthold KP Horn and Michael J Brooks. Shape from shading. MIT press, 1989. 2
Shahram Izadi, David Kim, Otmar Hilliges, David Molyneaux, Richard A. Newcombe, Pushmeet Kohli, Jamie
Shotton, Steve Hodges, Dustin Freeman, Andrew J. Davison, and Andrew W. Fitzgibbon. Kinectfusion:
real-time 3d reconstruction and interaction using a moving depth camera. In UIST, 2011. 2
Wenzel Jakob. Mitsuba renderer, 2010. http://www.mitsuba-renderer.org. 5
Michael Janner, Jiajun Wu, Tejas D Kulkarni, Ilker Yildirim, and Joshua B Tenenbaum. Self-supervised intrinsic
image decomposition. In NIPS, 2017. 3
Angjoo Kanazawa, David W Jacobs, and Manmohan Chandraker. Warpnet: Weakly supervised matching for
single-view reconstruction. In CVPR, 2016. 3
Abhishek Kar, Shubham Tulsiani, Joao Carreira, and Jitendra Malik. Category-specific object reconstruction
from a single image. In CVPR, 2015. 3
10
Ira Kemelmacher-Shlizerman and Ronen Basri. 3d face reconstruction from a single image using a single
reference face shape. IEEE TPAMI, 33(2):394?405, 2011. 1
Joseph J. Lim, Hamed Pirsiavash, and Antonio Torralba. Parsing ikea objects: Fine pose estimation. In ICCV,
2013. 8, 9
David G Lowe. Three-dimensional object recognition from single two-dimensional images. Artif. Intell., 31(3):
355?395, 1987. 3
David Marr. Vision: A computational investigation into the human representation and processing of visual
information. W. H. Freeman and Company, 1982. 1, 2
John McCormac, Ankur Handa, Stefan Leutenegger, and Andrew J Davison. Scenenet rgb-d: Can 5m synthetic
images beat generic imagenet pre-training on indoor segmentation? In ICCV, 2017. 2
Danilo Jimenez Rezende, SM Ali Eslami, Shakir Mohamed, Peter Battaglia, Max Jaderberg, and Nicolas Heess.
Unsupervised learning of 3d structure from images. In NIPS, 2016. 3, 4
Jason Rock, Tanmay Gupta, Justin Thorsen, JunYoung Gwak, Daeyun Shin, and Derek Hoiem. Completing 3d
object shape from one depth image. In CVPR, 2015. 3
Ashutosh Saxena, Min Sun, and Andrew Y Ng. Make3d: Learning 3d scene structure from a single still image.
IEEE TPAMI, 31(5):824?840, 2009. 1
Jian Shi, Yue Dong, Hao Su, and Stella X Yu. Learning non-lambertian object intrinsics across shapenet
categories. In CVPR, 2017. 2
Nathan Silberman, Derek Hoiem, Pushmeet Kohli, and Rob Fergus. Indoor segmentation and support inference
from rgbd images. In ECCV, 2012. 2
Amir Arsalan Soltani, Haibin Huang, Jiajun Wu, Tejas D Kulkarni, and Joshua B Tenenbaum. Synthesizing 3d
shapes via modeling multi-view depth maps and silhouettes with deep generative networks. In CVPR, 2017. 3
Shuran Song, Fisher Yu, Andy Zeng, Angel X Chang, Manolis Savva, and Thomas Funkhouser. Semantic scene
completion from a single depth image. In CVPR, 2017. 2, 3
Marshall F Tappen, William T Freeman, and Edward H Adelson. Recovering intrinsic images from a single
image. In NIPS, 2003. 1, 2
Shubham Tulsiani, Tinghui Zhou, Alexei A Efros, and Jitendra Malik. Multi-view supervision for single-view
reconstruction via differentiable ray consistency. In CVPR, 2017. 1, 2, 3, 4, 6, 7, 8, 9
Xiaolong Wang, David Fouhey, and Abhinav Gupta. Designing deep networks for surface normal estimation. In
CVPR, 2015. 2
Yair Weiss. Deriving intrinsic images from image sequences. In ICCV, 2001. 2
Jiajun Wu, Tianfan Xue, Joseph J Lim, Yuandong Tian, Joshua B Tenenbaum, Antonio Torralba, and William T
Freeman. Single image 3d interpreter network. In ECCV, 2016a. 3, 4
Jiajun Wu, Chengkai Zhang, Tianfan Xue, William T Freeman, and Joshua B Tenenbaum. Learning a Probabilistic Latent Space of Object Shapes via 3D Generative-Adversarial Modeling. In NIPS, 2016b. 1, 3, 4, 8,
9
Yu Xiang, Roozbeh Mottaghi, and Silvio Savarese. Beyond pascal: A benchmark for 3d object detection in the
wild. In WACV, 2014. 2, 6, 8, 9
Jianxiong Xiao, James Hays, Krista A Ehinger, Aude Oliva, and Antonio Torralba. Sun database: Large-scale
scene recognition from abbey to zoo. In CVPR, 2010. 5
Xinchen Yan, Jimei Yang, Ersin Yumer, Yijie Guo, and Honglak Lee. Perspective transformer nets: Learning
single-view 3d object reconstruction without 3d supervision. In NIPS, 2016. 3, 4
Ruo Zhang, Ping-Sing Tsai, James Edwin Cryer, and Mubarak Shah. Shape-from-shading: a survey. IEEE
TPAMI, 21(8):690?706, 1999. 2
11
| 6657 |@word mild:1 kohli:2 repository:2 middle:1 choy:5 rgb:11 decomposition:1 jacob:1 sgd:1 harder:1 shading:4 contains:3 hoiem:4 jimenez:1 tuned:7 past:2 recovered:2 cad:1 yet:1 dx:10 gpu:1 parsing:1 john:1 realistic:4 happen:1 informative:1 ronan:1 shape:97 ashutosh:1 v:1 generative:3 amir:1 plane:1 davison:2 provides:3 revisited:1 preference:1 org:1 tianfan:3 zhang:5 five:2 shubham:3 along:3 direct:3 brostow:1 qualitative:3 consists:1 combine:2 wild:3 ray:1 inside:2 uist:1 angel:2 mask:3 multi:6 brain:1 freeman:6 relying:1 inspired:2 manolis:2 company:1 resolve:1 provided:3 estimating:2 joao:1 what:1 developed:2 interpreter:2 unified:1 unobserved:1 guarantee:2 quantitative:2 saxena:2 jimei:1 tackle:2 before:1 ner:1 despite:2 encoding:4 eslami:1 chose:1 ankur:1 challenging:4 projective:1 tian:1 unique:1 horn:2 camera:1 testing:2 practice:1 orthographic:1 union:1 fitzgibbon:1 shin:1 yan:3 bell:2 adapting:1 projection:2 matching:1 pre:3 close:1 unlabeled:1 nb:6 put:1 impossible:1 transformer:1 www:1 map:9 demonstrated:1 shi:2 center:1 straightforward:4 go:1 reprojected:3 survey:1 resolution:1 recovery:1 zimo:1 estimator:6 marr:5 deriving:1 disentangled:2 searching:1 variation:3 target:2 enhanced:1 heavily:1 user:1 xinchen:1 us:1 designing:1 drc:16 tappen:4 recognition:4 predicts:5 database:2 muri:1 observed:1 module:3 wang:3 ensures:2 connected:1 sun:5 russell:2 environment:2 predictable:1 nie:1 occluded:1 trained:10 weakly:1 ali:1 purely:3 tog:2 edwin:1 easily:3 joint:1 samsung:1 various:1 train:4 jiao:1 kp:1 kevin:1 refined:2 whose:2 larger:1 plausible:1 widely:1 cvpr:13 tested:1 reconstruct:2 encoder:7 jointly:2 final:2 shakir:1 advantage:3 differentiable:7 tpami:4 matthias:1 rock:2 sequence:1 reconstruction:35 propose:3 ment:1 jamie:1 interaction:1 adaptation:3 net:1 turned:1 ablation:3 intuitive:1 normalize:1 exploiting:1 reprojection:11 perfect:2 object:43 resnet:1 derive:1 develop:2 completion:3 fixing:4 pose:3 help:1 andrew:4 tulsiani:17 make3d:1 edward:1 strong:2 recovering:8 janner:2 predicted:3 implemented:1 iou:7 direction:2 cnns:1 human:12 vx:15 material:2 explains:2 argued:1 oisin:1 fix:1 investigation:1 extension:2 kinectfusion:1 practically:1 around:1 ground:26 normal:37 guibas:1 cognition:1 predict:1 major:1 achieves:3 efros:2 torralba:3 abbey:1 albedo:1 battaglia:1 estimation:21 label:1 successfully:1 stefan:1 mit:5 rough:2 sensor:1 biglearn:1 zhou:1 vae:3 ira:1 release:1 rezende:3 focus:1 marrnet:30 consistently:1 indicates:1 contrast:1 adversarial:1 shapenet:11 baseline:5 kim:1 helpful:1 dim:1 inference:1 abstraction:2 rear:1 inaccurate:2 entire:2 typically:1 issue:3 pascal:13 retaining:1 development:2 art:4 constrained:1 special:1 field:1 equal:1 shahram:1 beach:1 koray:1 ng:1 look:1 unsupervised:1 nearly:1 thin:2 yu:3 adelson:1 quantitatively:2 fouhey:2 employ:4 few:1 modern:1 richard:1 preserve:2 recognize:2 intell:1 relief:2 consisting:1 occlusion:1 william:4 kavita:1 detection:1 highly:2 alexei:2 evaluation:3 alignment:2 introduces:1 xiaolong:1 accurate:3 andy:1 fu:1 closer:1 experience:1 shuran:2 orthogonal:1 conduct:2 savarese:3 abundant:1 kemelmacher:2 modeling:5 earlier:1 marshall:1 retains:1 measuring:1 mac:1 distill:1 rare:1 masked:3 predictor:1 graphic:1 front:2 xue:3 synthetic:17 st:1 csail:4 standing:1 probabilistic:1 physic:1 dong:1 decoding:2 lee:1 michael:3 synthesis:1 quickly:1 na:4 again:1 hodges:1 choose:2 huang:2 wenzel:1 zhao:1 style:1 leading:1 preparation:1 li:1 harry:1 jitendra:3 explicitly:1 leonidas:1 collobert:2 view:19 lowe:2 try:2 jason:1 overfits:1 start:1 recover:3 annotation:9 simon:1 jia:1 contribution:2 convolutional:5 characteristic:1 ronen:1 raw:1 kavukcuoglu:1 yildirim:1 ren:1 zoo:1 lighting:6 researcher:5 finer:1 explain:2 hamed:1 ping:1 sharing:1 aligns:1 farabet:1 failure:2 competitor:1 derek:3 mohamed:1 regress:3 james:2 recovers:6 unsolved:1 dataset:14 ask:1 knowledge:3 color:2 infers:2 lim:4 car:2 segmentation:2 sean:1 yumer:1 steve:1 higher:2 supervised:4 follow:2 chengkai:2 danilo:1 wei:2 roozbeh:1 formulation:4 evaluated:1 though:2 stage:1 sketch:51 horizontal:1 christopher:1 su:2 zeng:1 google:1 rodriguez:1 resemblance:1 aude:1 artif:1 usa:1 name:1 requiring:2 funkhouser:2 illustrated:1 deal:1 semantic:2 during:5 self:3 please:1 criterion:4 bansal:2 demonstrate:2 performs:2 dawei:1 ruizhongtai:1 image:110 handa:1 novel:3 recently:2 charles:1 common:1 shanghai:1 discussed:2 shlizerman:2 he:2 julier:1 synthesized:1 yuandong:1 honglak:1 tuning:11 automatic:1 consistency:18 grid:1 gwak:2 moving:1 similarity:1 surface:25 supervision:3 etc:3 renderer:3 gt:1 disentangle:1 perspective:1 scenario:1 certain:1 n00014:1 hay:1 kar:2 binary:1 onr:1 life:3 hanrahan:1 joshua:5 mottaghi:1 leutenegger:1 seen:1 preserving:2 dai:2 tinghui:1 deng:1 xiangyu:1 paradigm:4 smoother:2 full:6 multiple:2 infer:1 reduces:1 match:2 danfei:1 offer:1 long:2 cross:2 qi:1 prediction:12 variant:3 oliva:1 vision:4 metric:2 arxiv:1 iteration:1 represent:1 sometimes:2 background:3 cropped:1 fine:26 separately:3 girdhar:5 jian:2 suffered:1 unlike:1 yue:1 subject:3 pooling:1 effectiveness:1 yang:2 shotton:1 r2n2:1 rendering:2 relu:4 architecture:4 imperfect:1 idea:4 regarding:1 stabilizer:1 airplane:2 intrinsics:1 torch7:2 song:4 render:5 suffer:1 peter:1 shaoqing:1 prefers:1 matlab:1 deep:6 gabriel:1 antonio:3 heess:1 detailed:2 clear:1 tune:7 mid:1 tenenbaum:7 soltani:2 category:7 http:1 percentage:1 nsf:2 millisecond:1 estimated:28 jiajun:5 bryan:1 diverse:1 xingyuan:1 four:2 powerful:1 almost:1 reasonable:1 ikea:7 wu:13 prefer:2 ersin:1 layer:6 completing:1 followed:2 furniture:1 bala:1 fold:1 refine:1 noah:1 constraint:4 constrain:1 scene:6 nathan:1 chair:8 min:1 attempting:1 rendered:1 structured:1 ball:1 remain:1 across:2 reconstructing:1 joseph:2 rob:2 making:2 iccv:4 pipeline:1 izadi:2 visualization:1 fail:1 mccormac:2 needed:1 mind:1 drastic:1 end:16 photo:1 available:1 apply:2 observe:1 lambertian:1 barron:2 enforce:3 generic:1 batch:1 yair:1 shah:1 eigen:2 thomas:2 top:1 angela:1 include:2 ensure:1 gan:5 exploit:1 reflectance:1 silberman:2 malik:4 manmohan:1 snavely:1 enhances:1 gradient:3 thank:1 decoder:15 warpnet:1 nx:3 reason:1 enforcing:2 mikel:1 assuming:1 nc:8 setup:1 voxelized:1 hao:2 synthesizing:1 implementation:1 perform:2 observation:1 datasets:2 sm:1 benchmark:1 enabling:1 sing:1 barrow:2 pat:1 beat:1 jakob:2 inferred:1 david:9 imagenet:1 engine:1 learned:5 pop:1 nip:8 brook:2 able:1 justin:1 beyond:1 usually:2 perception:4 indoor:2 challenge:2 including:1 pirsiavash:1 max:1 unrealistic:1 rely:2 predicting:1 scarce:1 residual:1 abhinav:2 martial:1 stella:1 ilker:1 prior:11 voxels:11 l2:2 acknowledgement:1 rohit:1 xiang:6 loss:15 fully:3 wacv:1 annotator:1 consistent:1 xiao:2 viewpoint:1 eccv:4 supported:1 hebert:1 institute:1 face:3 benefit:1 depth:44 berthold:1 rich:1 qualitatively:1 made:1 projected:4 adaptive:1 voxel:4 pushmeet:2 reconstructed:2 emphasize:1 preferred:3 basri:2 silhouette:21 keep:1 jaderberg:1 chandraker:1 sequentially:2 fergus:3 yifan:1 abhishek:1 latent:1 decade:1 table:2 learn:1 transfer:4 robust:1 ca:1 nicolas:1 complex:2 cl:1 domain:9 cryer:1 suffering:1 rgbd:1 xu:1 junyoung:2 tl:1 ehinger:1 ny:1 tong:1 fails:3 position:2 momentum:1 sub:1 third:4 toyota:1 jay:1 dustin:1 mubarak:1 discarding:1 specific:1 showing:1 krista:1 yijie:1 explored:2 gupta:3 essential:3 intrinsic:10 burden:1 kanazawa:2 workshop:1 mirror:1 texture:7 illumination:1 chen:3 easier:2 entropy:2 intersection:2 savva:2 likely:1 appearance:4 explore:1 visual:2 kaiming:1 chang:6 truth:26 adaption:2 acm:2 shell:1 tejas:2 goal:1 fisher:1 carreira:1 determined:1 specifically:1 silvio:3 attempted:1 qixing:1 support:1 guo:1 jonathan:1 jianxiong:1 kulkarni:2 tsai:1 evaluate:3 trainable:5 |
6,254 | 6,658 | Multimodal Learning and Reasoning for Visual
Question Answering
Ilija Ilievski
Integrative Sciences and Engineering
National University of Singapore
[email protected]
Jiashi Feng
Electrical and Computer Engineering
National University of Singapore
[email protected]
Abstract
Reasoning about entities and their relationships from multimodal data is a key goal
of Artificial General Intelligence. The visual question answering (VQA) problem
is an excellent way to test such reasoning capabilities of an AI model and its
multimodal representation learning. However, the current VQA models are oversimplified deep neural networks, comprised of a long short-term memory (LSTM)
unit for question comprehension and a convolutional neural network (CNN) for
learning single image representation. We argue that the single visual representation
contains a limited and general information about the image contents and thus limits
the model reasoning capabilities. In this work we introduce a modular neural network model that learns a multimodal and multifaceted representation of the image
and the question. The proposed model learns to use the multimodal representation
to reason about the image entities and achieves a new state-of-the-art performance
on both VQA benchmark datasets, VQA v1.0 and v2.0, by a wide margin.
1
Introduction
One of the hallmarks of human intelligence is the ability to reason about entities and their relationships
using data from different modalities [40]. Humans can easily reason about the entities in a complex
scene by building multifaceted mental representations of the scene contents. Thus, any plausible
attempt for Artificial General Intelligence must also be able to reason about entities and their
relationships.
Deep learning models have demonstrated excellent performance on many computer vision tasks such
as image classification, object recognition, and scene classification [37, 23]. However, the models
are limited to a single task over single data modality and thus are still far from complete scene
understanding and reasoning.
The visual question answering (VQA), a task to provide answers to natural language questions about
the contents of an image, has been proposed to fill this gap [3, 22]. Solving the VQA task requires
understanding of the image contents, the question words, and their relationships. The human-posed
questions are arbitrary and thus besides being a challenging machine comprehension problem, their
answering also involves many computer vision tasks such as scene classification, object detection
and classification, and face analysis. Thus, VQA represents a problem comprised of multiple subproblems over multimodal data and as such can serve as a proxy test for the general reasoning
capabilities of an AI model [30].
However, the current state-of-the-art VQA approaches employ oversimplified deep neural network
models, comprised of a long short-term memory (LSTM) unit [9] for question comprehension and
a convolutional neural network (CNN) [24] for learning a single representation of the image. The
single visual representation encodes a limited and general information about the image contents and
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
thus hampers the model?s reasoning about the image entities and their relationships. The analysis of
the model?s behavior is also made difficult because of the single visual representation, as one does
not have evidence of the visual information used by the model to produce the answer. The problem is
partially alleviated by incorporating a visual attention module, as most VQA models do [6, 18, 45],
but still, the VQA models remain black box in most cases. There have also been approaches that
employ modular networks for VQA [2], however they achieved limited success. Most VQA models
join the image and text representations via element-wise multiplication [18, 45], exception is the
model proposed by Fukui et al. [6], which learns the joint representation via compact bilinear pooling.
Goyal et al. [7] recently showed that the current state-of-the-art VQA methods actually learn dataset
biases instead of reasoning about the image objects or learning multifaceted image representations.
In this work we introduce ReasonNet, a model that learns to reason about entities in an image by
learning a multifaceted representation of the image contents. ReasonNet employs a novel multimodal
representation learning and fusion module that enables the model to develop a complete image and
question understanding. ReasonNet then learns to utilize the multiple image representations, each
encoding different visual aspects of the image, by explicitly incorporating a neural reasoning module.
In contrast to current VQA models, the ReasonNet model is fully interpretable and provides evidence
of its reasoning capabilities. The proposed model outperforms the state-of-the-art by a significant
margin on the two largest benchmark VQA datasets, VQA v1.0 [3] and VQA v2.0 [7]. In summary,
the contributions of our work are as follows:
? We develop a novel multimodal representation learning and fusion method, crucial for
obtaining the complete image understanding necessary for multimodal reasoning.
? We introduce a new VQA reasoning model that learns multifaceted image representations to
reason about the image entities.
? We perform an extensive evaluation and achieve new state-of-the-art performance on the
two VQA benchmark datasets.
2
Related work
Multimodal reasoning models Recently, several works introduced modular neural networks for
reasoning evaluated on natural language question answering and visual question answering [2, 1, 11].
The modular networks use conventional natural language parser [21] to obtain a network layout for
composing the network architecture [11]. Later work also incorporated a dynamic network layout
prediction by learning to rank the parser proposed modules [1]. The neural modules are then jointly
trained to execute the task on which they are applied.
In contrast to our model, the existing modular networks learn to compose a network architecture using
a hand-designed or parser proposed layout structure. These models were shown [2] to be unable to
capture the complex nature of the natural language and perform poorly on the complex VQA v1.0 [3].
Concurrently to this work, Hu et al. [10] and Johnson et al. [16] proposed a similar modular network
model that learns the network layout structure and thus do not require a parser. However, the models
have been applied with success only to the synthetic dataset CLEVR [15] because for training the
layout prediction module they require the ground-truth programs that generated the questions.
Visual question answering The visual question answering (VQA) task has received great interest [5, 45, 50, 38] since the release of the first large-scale VQA v1.0 dataset by Antol et al. [3].
Typically, a VQA model is comprised of two modules for learning the question and the image representations, and a third module for fusing the representations into a single multimodal representation.
The multimodal representation is then fed to multiple fully-connected layers and a softmax layer at
the end outputs the probabilities of each answer (e.g. see [12, 6, 44]).
The question representation is learned by mapping each question word to a vector via a lookup
table matrix, which is often initialized with word2vec [31] or skip-thought [20] vectors. The word
vectors are then sequentially fed to a Long Short-Term Memory (LSTM) unit [9], and the final hidden
LSTM state is considered as the question representation. The image representation is obtained from a
pretrained convolutional neural network (CNN) [24], such as ResNet [8] or VGG [39], and the output
from the penultimate layer is regarded as the image representation. Some, increase the information
available in the image representation by using feature maps from a convolutional layer [45, 29, 6].
2
3
ReasonNet
We develop a novel modular deep neural network model that learns to reason about entities and their
relationships in a complex scene using multifaceted scene representations. The model, named as ReasonNet, takes as inputs an image I and a natural language text L. Then, ReasonNet passes the image
and the text through its different modules that encode multiple image aspects into multiple vector
representations. At the same time, ReasonNet uses a language encoder module to encode the text into
a vector representation. Finally, ReasonNet?s reasoning unit fuses the different representations into a
single reasoning vector. In the following we give details of the network modules and its reasoning
unit, while in Section 4 we ensemble and describe a ReasonNet applied to the VQA task (Figure 1).
3.1
Multimodal representation learning
ReasonNet incorporates two types of representation learning modules: visual classification modules
and visual attention modules. A visual classification module outputs a vector that contains the class
probabilities of the specific image aspect assigned to that module. While a visual attention module
outputs a vector that contains visual features focused on the module?s specific image aspect.
We denote the classification modules as ? and the visual attention modules as ?. We use bold lower
case letters to represent vectors and bold upper case letters to represent tensors. Subscript indices
denote countable variables while superscript indices label the variables. For example, P ?k denotes the
k-th matrix P of the module ?. In the following we describe each module type.
Visual classification modules Each visual classification module maps a different aspect of the
image contents to a module specific class representation space. Consequently, ReasonNet needs to
transfer the output of each module to a common representation space. For example, if one module
classifies the image scene as ?train station?, while another module classifies a detected object as
?train?, ReasonNet needs to be aware that the outputs are semantically related. For this reason,
ReasonNet transfers each classification module?s output to a common representation space using a
lookup table matrix.
Formally, a classification module ? outputs a matrix P ? of n one-hot vectors with lookup table
indices of the n highest probability class labels. The matrix P ? is then mapped to a vector c? in a
common representation space with:
c? = vec(P ? W LT ),
where W
LT
(1)
is the lookup table matrix with learned parameters.
Visual attention modules ReasonNet passes the image I through a residual neural network [8] to
obtain a visual feature tensor V ? RF ?W ?H representing the global image contents. Then, each
visual attention module ? focuses the visual feature tensor to the specific image aspect assigned to
the said module using an attention probability distribution ?? ? RW ?H .
The module-specific visual representation vector v ? ? RF is then obtained with:
?
v =
W X
H
X
??
i,j V i,j .
(2)
i=1 j=1
Encoder units A problem with this approach is that the common representation vectors c? from
the classification modules, being distributed representations of the modules? class labels, are highdimensional vectors. On the other hand, the visual feature vectors v ? of the attention modules are
also high-dimensional vectors, but sparse and with different dimensionality and scale.
As a solution, ReasonNet appends to each classification and to each attention modules an encoder
unit Enc that encodes a module?s output vector x (equal to c? if classification module and to v ? if
attention module) to a condensed vector r in a common low-dimensional representation space. The
encoder units are implemented as two fully-connected layers followed by a non-linear activation
function, and max pooling over the magnitude while preserving the sign:
r = Enc(x)
Enc(x) := sgn(f (x)) ? max(|f (x)|),
(3)
E
E
E
E
f (x) := ? W 2 ?(W 1 x + b1 ) + b2 ,
3
E
where W E
k and bk are the parameters and biases of the k-th layer in an encoder E, and ?(?) :=
tanh(?). The max pooling is performed over the row dimension as f (x) outputs a matrix of rowstacked encoded vectors from one mini-batch.
Text encoding ReasonNet treats the words in L as class labels and correspondingly uses the lookup
table from Eq. (1) and an encoder unit to map the text to a vector r l in the same low-dimensional
representation space of the other modules.
3.2
Multimodal reasoning
The reasoning capabilities of the ReasonNet model come from its ability to learn the interaction
between each module?s representation r k and the question representation r l . The ReasonNet model
parameterizes the r k ? r l interaction with a bilinear model [42]. Bilinear models have excellent representational expressiveness by allowing the vector representations to adapt each other multiplicatively.
The bilinear model is defined as:
s
s l
sk = r >
(4)
k W k r + bk ,
where k = 1, . . . , K and K is the number of representation learning modules, provides a rich vector
representation sk of the k-th module?s output ? language interaction.
Note that while other multimodal representation learning works, e.g. [6], have criticized the use of
bilinear models for representation fusion because of their high dimensionality as the tensor W sk is
cubic in the dimensions of r k , r l , and sk . However, ReasonNet mitigates this issue by employing
encoder units to reduce the dimension of the representation vectors and thus also reduce the dimension
of the bilinear model. We evaluate different fusion methods with ablation studies in Section 5.3.
ReasonNet builds a complete image and language representation by concatenating each interaction
fK
f
vector sk into a vector g = k=1 sk , where denotes concatenation of vectors. The vector concatenation is crucial for disentangling the contributions of each module in the model?s task. By partially
ignoring some of the inputs of the vector g, ReasonNet can learn to ?softly? utilize the different
modules only when their outputs help in predicting the correct answer for the given question. In
contrast to the recent works on module networks, e.g. [10, 16], ReasonNet can choose to partially use
a module if it is helpful for the task instead of completely removing a module as other module network
models do. For example, for the visual question answering (VQA) task, the soft module usage is
particularly useful when answering a question which implicitly requires a module, e.g. answering the
question ?Is it raining?? implicitly requires a scene classification module.
The concatenation of the modules? representations also enables the interpretability of the reasoning
behavior of our model. Specifically, by observing which elements of the vector g are most active we
can infer which modules ReasonNet used in the reasoning process and thus explain the reasons for its
behavior. We visualize the reasoning process on the VQA task in Section 5.4.
Finally, the multimodal representation vector g can be used as an input to an answer classification
network, when applied to the VQA task, or as an input to an LSTM unit when applied to the
image captioning task. In the next section, we use the challenging VQA task as a proxy test for
ReasonNet reasoning capabilities.
4
ReasonNet for VQA
The visual question answering problem is an excellent way to test the reasoning capabilities of
ReasonNet and its use of multifaceted image representations. For example, answering the question
?Does the woman look happy?? requires a face detection, gender and emotion classification, while
answering the question ?How many mice are on the desk?? requires object detection and classification.
Thus, for the VQA task, ReasonNet incorporates the following set of modules: 1. question-specific
visual attention module, 2. object-specific visual attention module, 3. face-specific visual attention
module, 4. object classification module, 5. scene classification module, 6. face analysis classification
module. In the following we give a formal definition of the VQA problem and details of the model?s
modules when applied on the VQA task. The network architecture is visualized in Figure 1.
Namely, the VQA problem can be solved by modeling the likelihood probability distribution pvqa
which for each answer a in the answer set ? outputs the probability of being the correct answer, given
a question Q about an image I:
a
? = arg max pvqa (a|Q, I; ?),
(5)
a??
4
Figure 1: Network architecture diagram of the ReasonNet model applied on the VQA task. Round
rectangles represent attention modules, squared rectangles represent classification modules, small
trapezoids represent encoder units (Eq. (3)), thin rectangles represent the learned multimodal representation vectors, N represents the bilinear interaction model (Eq. (4)), and the big trapezoid is a
multi-layer perceptron network that classifies the reasoning vector g to an answer a (Eq. (7))
where ? are the model parameters, a
? is the predicted answer, and ? is the set of possible answers.
4.1
ReasonNet VQA modules
First, ReasonNet obtains the question representation vector r l and a global visual feature tensor V as
described in Section 3.1. ReasonNet then learns a question-specific image representation by using
the question representation vector r l to learn an attention probability distribution ?v ? RW ?H over
the global visual feature tensor V ? RF ?W ?H :
h
i
?v = softmax W ? ? (W v r l + bv ) ? 1 ? ?(W V V + bV ) + b? ,
where ?(?) := tanh(?), W ? , W v , W V and the corresponding biases are learned parameters, 1 ?
1W ?H is used to tile the question vector representation to match the V tensor dimensionality, and ?
denotes element-wise matrix multiplication. The question-specific visual feature vector v v is then
obtained with Eq. (2) and the representation vector r v with Eq. (3).
Naturally, many of the VQA questions are about image objects so ReasonNet incorporates a fully
convolutional network (FCN) [28] for object detection. Given an image, the FCN will output a
set of object bounding boxes and their corresponding confidence scores. Each bounding box is
represented as a four-element vector d> = [x, y, w, h], where (x, y) is the coordinate of the top-left
box corner and w, h is the size of the box. Using a confidence score threshold ReasonNet obtains
a set B containing high confidence bounding boxes. ReasonNet then uses the set B to compute an
attention probability distribution ?o that focuses the visual feature tensor V on the image objects.
To ground the image pixels to visual feature maps, all images are resized to the same pre-fixed
dimension before feeding them to the object detection module. Thus each feature vector v ? RF from
a corresponding element in the tensor slice V slice ? RW ?H from the feature tensor V ? RF ?W ?H
maps to a fixed sized image region.
Formally, for each dk ? B, ReasonNet calculates a single object attention ? k . The |B| object-specific
attentions are then used to calculate the overall object attention distribution ?o :
?o = softmax(?
? o ),
k
?i,j
o
k
?
? i,j
= max (?i,j
),
k=1,...,|B|
?
dkx ? i ? (dkx + dkw ),
?1
dky ? j ? (dky + dkh ),
=
?
0.1 otherwise.
5
(6)
As before, the object-specific visual feature vector v o is then obtained with Eq. (2) and the representation vector r o with Eq. (3).
ReasonNet further uses the object bounding boxes to obtain a vector of object class labels. Namely,
for each bounding box in B, ReasonNet crops out the image part corresponding to the box coordinates
and then uses a residual network to classify the cropped-out image part and obtain a class label. The
n class labels of the boxes with highest class probability are represented as n one-hot vectors of
lookup table indices. The matrix P c , obtained by stacking the n vectors, is then mapped to a dense
low-dimensional vector r c with Eq. (1) and Eq. (3).
Next, ReasonNet uses a scene classification network as many of the questions explicitly or implicitly
necessitate the knowledge of the image setting. The scene classification network is implemented as a
residual network trained on the scene classification task. As before, the top n predicted class labels
are represented as a matrix of n one-hot vectors P s from which the module?s representation vector
r s is obtained (Eq. (1) and Eq. (3)).
Since the VQA datasets [7, 3] contain human-posed questions, many of the questions are about
people. Thus, ReasonNet also incorporates a face detector module, and a face analysis classification
module.
The face detector module is a fully convolutional network that outputs a set of face bounding boxes
and confidence scores. As with the object detector, ReasonNet uses a threshold to filter out bounding
boxes with low confidence scores and obtain a set of face detections F. Then, from F, using Eq. (6),
ReasonNet obtains an attention probability distribution ? f that focuses the visual feature tensor V
on people?s faces. The face-specific visual feature vector v f is then obtained with Eq. (2) and the
representation vector r f with Eq. (3).
The face bounding boxes from F are also used to crop out the image regions that contain a face and
using a convolutional neural network to obtain three class labels for each detected face representing
the age group, the gender, and the emotion. As with the other classification modules, ReasonNet represents the three class labels as a matrix of one-hot vectors P a and uses Equations (1) and (3) to
obtain the face analysis representation vector r a .
ReasonNet obtains a complete multimodal understanding by learning the interaction of the learned
representations H = {r v , r o , r c , r s , r f , r a } with the question representation r q :
n
s q
s
g=
(r >
h W h r + bh ),
r h ?H
f
where W sh is a learned bilinear tensor for a representation r h , and denotes concatenation of vectors.
Finally, ReasonNet forwards the vector g, containing the question representation and multifaceted
image representations, to a two-layer perceptron network which outputs the probabilities pvqa (Eq. (5)):
h
i
pvqa (a|Q, I; ?) = softmax ? W g2 ?(W g1 g + bg1 ) + bg2 ,
(7)
where ? represents all the model parameters.
5
5.1
Experiments
Datasets
We evaluate our model on the two benchmark VQA datasets, VQA v1.0 [3] and VQA v2.0 [7]. The
VQA v1.0 dataset is the first large-scale VQA dataset. The dataset contains three human-posed
questions and answers for each one of the 204,721 images found in the MS-COCO [27] dataset.
We also evaluate our model on the second version of this dataset, the VQA v2.0. The new version
includes about twice as many question-answer pairs and addresses the dataset bias issues [7] of the
VQA v1.0 dataset. We report results according to the evaluation metric provided by the authors of
the VQA datasets, where an answer is counted as correct if at least three annotators gave that answer:
P10
j=1 1(a = aj )
Acc(a) = min(
, 1).
3
For fair evaluation, we use the publicly available VQA evaluation servers to compute the overall and
per question type results.
6
Table 1: Results of the ablation study on the VQA v2.0 validation.
Method
All
Y/N
Num Other Q-type changed
VQA
55.13 69.07 34.29 48.01
VQA+Sc
56.80 70.62 35.14 49.99 +2.74% Which
VQA+Sc+oDec
58.46 71.05 36.16 52.86 +5.73% What color is the
VQA+Sc+oDec+oCls
59.82 72.88 37.38 54.47 +3.68% How
VQA+Sc+oDec+oCls+fDec
60.35 74.21 37.46 53.79 +12.63% Is the man
VQA+Sc+oDec+oCls+fDec+fAna 60.60 73.78 36.98 54.81 +0.88% Is he
ReasonNet-HadamardProduct
58.37 71.05 35.99 52.72
ReasonNet-MCB [6]
58.78 71.04 36.96 53.35
ReasonNet
60.60 73.78 36.98 54.81
5.2
Implementation details
Given a question Q about an image I, our model works as follows. Following [18, 6] the images
are scaled and center-cropped to a dimensionality of 3 ? 448 ? 448, then are fed through a ResNet152 [8] pretrained on ImageNet [36]. We utilize the output of the last convolutional layer as the image
representation V ? R2048?14?14 .
The question words are converted to lowercase and all punctuation characters are removed. We
further remove some uninformative words such as ?a, ?an?, ?the?, etc. We then trim the questions to
contain at most ten question words by removing the words after the first ten. The lookup table matrix
uses 300-dimensional vectors, initialized with word2vec [31] vectors.
The module parameters used to produce the module?s outputs are pretrained on each specific task and
are kept fixed when applying ReasonNet to the VQA problem. In the following we give details of
each module.
The object detection module is implemented and pretrained as in [32, 33]. The object classification
and scene classification modules are implemented as ResNet-152, the only difference is the object
classification module is pretrained on MS-COCO while the scene classification module is pretrained
on Places365 [48, 49]. The object classification module outputs 80 different class labels [27], while
the scene classification module outputs 365 class labels [48].
We implement and pretrain the face detection module following Zhang et al. [46, 47], while the age
and gender classification is performed as [34, 35] and the emotion recognition following [25, 26]. The
output of the age classification network is an integer from zero to hundred, so we group the integers
into four named groups, 0 ? 12 as ?child?, 13 ? 30 as ?young?, 31 ? 65 as ?adult? and +65 as ?old?.
This enables as to map the integer outputs to a class labels. Similarly, the output from the gender
classification module is 0 for ?woman? and 1 for ?man?. Finally, the emotion recognition module
classifies a detected face to the following seven emotions ?Angry?, ?Disgust?, ?Fear?, ?Happy?,
?Neutral?, ?Sad?, and ?Surprise?.
The encoder units encodes the module outputs to 500-dimensional vectors, with a hidden layer of
1,500 dimensions. Each bilinear interaction model outputs a 500-dimensional interaction vector, i.e.
500 ? 500 ? 500. The classification network classifies the reasoning vector g using one hidden
layer of 2,500 dimensions to one of 4,096 most common answers in the training set.
We jointly optimize the parameters of the encoder units, the bilinear models, and the answer classification network using Adam [19] with a learning rate of 0.0007, without learning rate decay. We
apply a gradient clipping threshold of 5 and use dropout[41] (with p(keep) = 0.5) layers before and
batch normalization[13] after each fully-conected layer as a regularization.
5.3
Ablation study
To assess the contribution of each ReasonNet module we perform an ablation study where we train a
model that only uses one module and then subsequently add the rest of the proposed VQA modules.
A VQA model with only question-specific attention module is denoted as ?VQA?, the addition of the
scene classification module is denoted as ?Sc?, the object detection module as ?oDec?, the object
classification module as ?oCls?, the face detection module as ?fDec?, and the face analysis as ?fAna?.
7
To evaluate the bilinear model as representation fusion mechanism, we compare ReasonNet to models
where we only swap the bilinear interaction learning (Eq. (4)) with (1) Hadamard product (denoted
as ReasonNet-HadamardProduct) and with (2) multimodal compact bilinear pooling [6] (denoted
as ReasonNet-MCB). The bilinear interaction model maps the two vectors to an interaction vector,
by learning a projection matrix W that projects the vectorized outer product of the two vectors
to an interaction vector. When using Hadamard product the interaction vector is just an elementwise multiplication of the two vectors. On the other hand, the MCB uses Count Sketch projection
function [4] to project the two vectors to a lower dimension and then applies convolution of the two
count sketch vectors to produce an interaction vector. As opposed to a bilinear model, the MCB does
not learn a projection matrix. We train the models on the VQA v2.0 train set and evaluate them on
the validation set. The results are shown in Table 1.
From Table 1 we can observe that each module addition improves the overall score. The results
show that the object detection module is responsible for the highest increase in accuracy, specifically
for question of type ?Other?. The addition of the object classification module further improves the
accuracy on the ?Other? question types, but the addition of the face detection module reduces the
accuracy of ?Other? question types and increases the accuracy on the ?Yes/No? questions. Possible
reasons for this is that the two attention modules (object detection and face detection) bring too much
noise in the image representation. The increase in accuracy for the ?Yes/No? questions is likely
because most ?Yes/No? questions are about people. Finally, the addition of the face analysis module
brings the highest accuracy by returning the accuracy of the ?Other? question types, possibly due to
the face class labels help in understanding the face attention.
The results in Table 1 clearly show the representational expressiveness of the bilinear models as
representation fusion. The bilinear model improves the accuracy for all question types, while there is
a small difference between the Hadamard product and compact bilinear pooling, as discussed in [18].
Figure 2: Qualitative results: we visualize the concatenation vector g from Eq. (4.1) to investigate
the module utilization given an image and two questions about the same image. The question-image
pairs are from the VQA v2.0 test set.
5.4
Qualitative analysis and failure cases
To investigate the contribution of each of the network modules, we visualize the concatenation vector
g from Eq. (4.1) in Figure 2. We show two images from the VQA v2.0 test set and the corresponding
g vector values for two questions. Higher values are displayed with lighter shade of gray.
From Figure 2 we can observe that for the question ?Does the man look happy?? the network strongly
use the representation from the face analysis module and partially use the question-only representation
and the question-specific attention representation. We can observe the same trend in the next two
questions. It is interesting to observe that for complex questions such as ?What color is the building
the bikes are in front of?? most of the network modules are used which means the network does
actually need multifaceted image representation to answer the questions.
The first example in Figure 2 also serves as a failure case. Namely, for the question ?Does the man
look happy?? the network correctly learns to use the face analysis module when the word ?happy? is
8
Table 2: Results on the VQA v1.0 and v2.0 test-standard datasets for single models and without
data augmentation. NMN is the only other modular neural. Results are taken from the official VQA
evaluation servers.
VQA v1.0 test
VQA v2.0 test
Method
All Y/N Num Other
All
Y/N
Num Other
VQA-LangOnly 48.9 78.1 34.9
27.0 44.26 67.01 31.55 27.37
D-LSTM-nI [14] 58.2 80.6 36.5
43.7 54.22 73.46 35.18 41.83
NMN [2]
58.7
DMN+ [43]
60.4 80.4 36.8
48.3
MRN [17]
61.8 82.4 38.2
49.4
HieCoAtt [29]
62.1 79.9 38.2
51.9
MCB [6] 1
64.7 82.5 37.6
55.6 62.27 78.82 38.28 53.36
MLB [18]
65.1 84.0 38.0
54.8
ReasonNet2
67.9 84.0 38.7
60.4 64.61 78.86 41.98 57.39
present in the question. However, the face analysis module incorrectly classifies the face as ?Angry?
misleading the network to give a wrong answer. Such error propagation from individual network
modules is the main limitation of the proposed model. Future work can possibly overcome this
limitation by backpropagating the error through the network modules. On the other hand, there is a
constant improvement by the research community for each individual computer vision sub-task, that
the limitation might be alleviated by simply incorporating the latest state-of-the-art network module.
5.5
Comparison with the state-of-the-art
Compared with the previous state-of-the-art on the VQA v1.0 dataset, the ReasonNet model achieves
2.8% higher accuracy (Table 2). The improvement in accuracy predominately comes from the
ability of the ReasonNet model to answer complex questions as evident from the 5.6% increase in
accuracy (denoted as ?Other? in Table 2). This validates the ability of the ReasonNet to learn complex
question-image relationships and to perform reasoning over the learned multimodal representations.
We observe the same improvement in accuracy of 2.34% on the more challenging VQA v2.0 dataset.
As on the VQA v1.0, the main improvement comes in answering the questions of type ?Other? where
there is a 4.03% difference. The improvement in the ?Other? questions likely comes from learning
complex interactions of all modules outputs. There is also an improvement of 3.7% in the counting
questions denoted as ?Num?, which serves as evidence of the contribution of the object detection and
object classification modules.
The new state-of-the-art on these datasets indicates the superiority of ReasonNet and the need
for reasoning when answering complicated questions whose answering requires reasoning and
understanding the relationship of multiple image objects.
6
Conclusion
We have presented a novel model for multimodal representation learning and reasoning. Our proposed
reasoning model learns to reason over a learned multifaceted image representation conditioned on a
text data. We validated the proposed reasoning neural network on the challenging VQA task and the
model achieved a new state-of-the-art performance by a wide margin. The proposed reasoning model
is general and probably applicable to other tasks involving multimodal representations, such as image
captioning. We leave this promising direction for future work.
Acknowledgments
The work of Jiashi Feng was partially supported by National University of Singapore startup grant
R-263-000-C08-133, Ministry of Education of Singapore AcRF Tier One grant R-263-000-C21-112
and NUS IDS grant R-263-000-C67-646.
1
Fukui et al. in [6] only report the test-dev results for VQA v1.0. The VQA v2.0 results are obtained from an
implementation of their model.
2
Due to a bug in the answer generating script the reviewed draft reported slightly lower VQA v2.0 results.
9
References
[1] Jacob Andreas, Marcus Rohrbach, Trevor Darrell, and Dan Klein. Learning to compose neural networks
for question answering. In HLT-NAACL, 2016.
[2] Jacob Andreas, Marcus Rohrbach, Trevor Darrell, and Dan Klein. Neural module networks. In Proceedings
of the IEEE Conference on Computer Vision and Pattern Recognition CVPR, 2016.
[3] Stanislaw Antol, Aishwarya Agrawal, Jiasen Lu, Margaret Mitchell, Dhruv Batra, C. Lawrence Zitnick,
and Devi Parikh. VQA: Visual question answering. In The IEEE International Conference on Computer
Vision (ICCV), December 2015.
[4] Moses Charikar, Kevin Chen, and Martin Farach-Colton. Finding frequent items in data streams. Automata,
languages and programming, pages 784?784, 2002.
[5] Kan Chen, Jiang Wang, Liang-Chieh Chen, Haoyuan Gao, Wei Xu, and Ram Nevatia. ABC-CNN: An attention based convolutional neural network for visual question answering. arXiv preprint arXiv:1511.05960,
2015.
[6] Akira Fukui, Dong Huk Park, Daylen Yang, Anna Rohrbach, Trevor Darrell, and Marcus Rohrbach.
Multimodal compact bilinear pooling for visual question answering and visual grounding. In Conference
on Empirical Methods in Natural Language Processing (EMNLP), 2016.
[7] Yash Goyal, Tejas Khot, Douglas Summers-Stay, Dhruv Batra, and Devi Parikh. Making the V in VQA
matter: Elevating the role of image understanding in Visual Question Answering. In Conference on
Computer Vision and Pattern Recognition (CVPR), 2017.
[8] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition.
In Proceedings of the IEEE conference on computer vision and pattern recognition, CVPR, pages 770?778,
2016.
[9] Sepp Hochreiter and J?rgen Schmidhuber. Long short-term memory. Neural computation, 9(8):1735?1780,
1997.
[10] Ronghang Hu, Jacob Andreas, Marcus Rohrbach, Trevor Darrell, and Kate Saenko. Learning to reason:
End-to-end module networks for visual question answering. In Proceedings of the IEEE International
Conference on Computer Vision (ICCV), 2017.
[11] Ronghang Hu, Marcus Rohrbach, Jacob Andreas, Trevor Darrell, and Kate Saenko. Modeling relationships
in referential expressions with compositional modular networks. In Proceedings of the IEEE Conference
on Computer Vision and Pattern Recognition, 2017.
[12] Ilija Ilievski, Shuicheng Yan, and Jiashi Feng. A focused dynamic attention model for visual question
answering. arXiv preprint arXiv:1604.01485, 2016.
[13] Sergey Ioffe and Christian Szegedy. Batch normalization: Accelerating deep network training by reducing
internal covariate shift. In International Conference on Machine Learning, pages 448?456, 2015.
[14] Dhruv Batra Jiasen Lu, Xiao Lin and Devi Parikh. Deeper lstm and normalized cnn visual question
answering model. https://github.com/VT-vision-lab/VQA_LSTM_CNN, 2015.
[15] Justin Johnson, Bharath Hariharan, Laurens van der Maaten, Li Fei-Fei, C Lawrence Zitnick, and Ross
Girshick. CLEVR: A diagnostic dataset for compositional language and elementary visual reasoning. In
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition CVPR, 2017.
[16] Justin Johnson, Bharath Hariharan, Laurens van der Maaten, Judy Hoffman, Li Fei-Fei, C Lawrence Zitnick,
and Ross Girshick. Inferring and executing programs for visual reasoning. In The IEEE International
Conference on Computer Vision (ICCV), 2017.
[17] Jin-Hwa Kim, Sang-Woo Lee, Donghyun Kwak, Min-Oh Heo, Jeonghee Kim, Jung-Woo Ha, and ByoungTak Zhang. Multimodal residual learning for visual qa. In Advances in Neural Information Processing
Systems, pages 361?369, 2016.
[18] Jin-Hwa Kim, Kyoung-Woon On, Jeonghee Kim, Jung-Woo Ha, and Byoung-Tak Zhang. Hadamard
product for low-rank bilinear pooling. In International Conference on Learning Representations, 2017.
[19] Diederik Kingma and Jimmy Ba. ADAM: A method for stochastic optimization. In International
Conference on Learning Representations, 2015.
10
[20] Ryan Kiros, Yukun Zhu, Ruslan R Salakhutdinov, Richard Zemel, Raquel Urtasun, Antonio Torralba,
and Sanja Fidler. Skip-thought vectors. In Advances in neural information processing systems, pages
3294?3302, 2015.
[21] Dan Klein and Christopher D Manning. Accurate unlexicalized parsing. In Proceedings of the 41st
Annual Meeting on Association for Computational Linguistics-Volume 1, pages 423?430. Association for
Computational Linguistics, 2003.
[22] Ranjay Krishna, Yuke Zhu, Oliver Groth, Justin Johnson, Kenji Hata, Joshua Kravitz, Stephanie Chen,
Yannis Kalantidis, Li-Jia Li, David A Shamma, et al. Visual Genome: Connecting language and vision
using crowdsourced dense image annotations. International Journal of Computer Vision, 2017.
[23] Yann LeCun, Yoshua Bengio, and Geoffrey Hinton. Deep learning. Nature, 521(7553):436?444, 2015.
[24] Yann LeCun, Bernhard Boser, John S Denker, Donnie Henderson, Richard E Howard, Wayne Hubbard, and
Lawrence D Jackel. Backpropagation applied to handwritten zip code recognition. Neural computation,
1(4):541?551, 1989.
[25] Gil Levi and Tal Hassner. Emotion recognition in the wild via convolutional neural networks and mapped
binary patterns. In Proc. ACM International Conference on Multimodal Interaction (ICMI), November
2015.
[26] Gil Levi and Tal Hassner. Emotion recognition in the wild via convolutional neural networks and mapped
binary patterns. http://www.openu.ac.il/home/hassner/projects/cnn_emotions/, 2015.
[27] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Doll?r,
and C Lawrence Zitnick. Microsoft COCO: Common objects in context. In European Conference on
Computer Vision, pages 740?755. Springer, 2014.
[28] Jonathan Long, Evan Shelhamer, and Trevor Darrell. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages
3431?3440, 2015.
[29] Jiasen Lu, Jianwei Yang, Dhruv Batra, and Devi Parikh. Hierarchical question-image co-attention for
visual question answering. In Advances in Neural Information Processing Systems 29, pages 289?297,
2016.
[30] Mateusz Malinowski and Mario Fritz. A multi-world approach to question answering about real-world
scenes based on uncertain input. In Advances in Neural Information Processing Systems, pages 1682?1690,
2014.
[31] Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. Distributed representations
of words and phrases and their compositionality. In Advances in neural information processing systems,
pages 3111?3119, 2013.
[32] Pedro O. Pinheiro, Tsung-Yi Lin, Ronan Collobert, and Piotr Doll?r. Learning to refine object segments.
In ECCV, 2016.
[33] Pedro O. Pinheiro, Tsung-Yi Lin, Ronan Collobert, and Piotr Doll?r. Learning to refine object segments.
https://github.com/facebookresearch/deepmask, 2016.
[34] Rasmus Rothe, Radu Timofte, and Luc Van Gool. Deep expectation of real and apparent age from a single
image without facial landmarks. International Journal of Computer Vision (IJCV), July 2016.
[35] Rasmus Rothe, Radu Timofte, and Luc Van Gool. Deep expectation of real and apparent age from a single image without facial landmarks. https://data.vision.ee.ethz.ch/cvl/rrothe/imdb-wiki/,
2016.
[36] Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang,
Andrej Karpathy, Aditya Khosla, Michael Bernstein, et al. Imagenet large scale visual recognition challenge.
International Journal of Computer Vision, 115(3):211?252, 2015.
[37] J?rgen Schmidhuber. Deep learning in neural networks: An overview. Neural networks, 61:85?117, 2015.
[38] Kevin J Shih, Saurabh Singh, and Derek Hoiem. Where to look: Focus regions for visual question
answering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016.
[39] Karen Simonyan and Andrew Zisserman. Very deep convolutional networks for large-scale image recognition. In International Conference on Learning Representations, 2015.
11
[40] Elizabeth S Spelke, Karen Breinlinger, Janet Macomber, and Kristen Jacobson. Origins of knowledge.
Psychological review, 99(4):605, 1992.
[41] Nitish Srivastava, Geoffrey E Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov. Dropout:
a simple way to prevent neural networks from overfitting. Journal of machine learning research, 15(1):1929?
1958, 2014.
[42] Joshua B Tenenbaum and William T Freeman. Separating style and content. Advances in neural information
processing systems, pages 662?668, 1997.
[43] Caiming Xiong, Stephen Merity, and Richard Socher. Dynamic memory networks for visual and textual
question answering. In International Conference on Machine Learning, pages 2397?2406, 2016.
[44] Huijuan Xu and Kate Saenko. Ask, attend and answer: Exploring question-guided spatial attention for
visual question answering. In European Conference on Computer Vision, pages 451?466, 2016.
[45] Zichao Yang, Xiaodong He, Jianfeng Gao, Li Deng, and Alex Smola. Stacked attention networks for image
question answering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,
pages 21?29, 2016.
[46] K. Zhang, Z. Zhang, Z. Li, and Y. Qiao. Joint face detection and alignment using multitask cascaded
convolutional networks. IEEE Signal Processing Letters, 23(10):1499?1503, Oct 2016.
[47] K. Zhang, Z. Zhang, Z. Li, and Y. Qiao. Joint face detection and alignment using multitask cascaded convolutional networks. https://github.com/kpzhang93/MTCNN_face_detection_alignment, 2016.
[48] Bolei Zhou, Agata Lapedriza, Aditya Khosla, Aude Oliva, and Antonio Torralba. Places: A 10 million
image database for scene recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence,
2017.
[49] Bolei Zhou, Agata Lapedriza, Aditya Khosla, Aude Oliva, and Antonio Torralba. Places: A 10 million
image database for scene recognition. https://github.com/CSAILVision/places365, 2017.
[50] Yuke Zhu, Oliver Groth, Michael Bernstein, and Li Fei-Fei. Visual7w: Grounded question answering
in images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages
4995?5004, 2016.
12
| 6658 |@word multitask:2 cnn:5 version:2 hu:3 integrative:1 shuicheng:1 jacob:4 contains:4 score:5 hoiem:1 outperforms:1 existing:1 current:4 com:4 haoyuan:1 activation:1 diederik:1 must:1 parsing:1 john:1 ronan:2 enables:3 christian:1 remove:1 designed:1 interpretable:1 intelligence:4 item:1 kyoung:1 short:4 num:4 mental:1 provides:2 draft:1 zhang:8 donnie:1 qualitative:2 ijcv:1 compose:2 dan:3 wild:2 introduce:3 merity:1 behavior:3 ilievski:3 kiros:1 multi:2 salakhutdinov:2 oversimplified:2 breinlinger:1 freeman:1 dkx:2 provided:1 classifies:6 project:3 lapedriza:2 bike:1 what:2 vqa_lstm_cnn:1 finding:1 returning:1 scaled:1 wrong:1 utilization:1 wayne:1 unit:14 grant:3 superiority:1 ramanan:1 before:4 engineering:2 attend:1 treat:1 limit:1 bilinear:20 encoding:2 id:1 jiang:1 subscript:1 clevr:2 black:1 might:1 twice:1 challenging:4 co:1 limited:4 shamma:1 c21:1 acknowledgment:1 responsible:1 lecun:2 goyal:2 implement:1 backpropagation:1 maire:1 dmn:1 evan:1 empirical:1 yan:1 thought:2 alleviated:2 projection:3 word:10 confidence:5 pre:1 andrej:1 bh:1 janet:1 context:1 applying:1 optimize:1 conventional:1 map:7 demonstrated:1 center:1 www:1 dean:1 layout:5 attention:29 latest:1 sepp:1 automaton:1 focused:2 jimmy:1 tomas:1 regarded:1 fill:1 oh:1 coordinate:2 bg2:1 parser:4 lighter:1 programming:1 us:11 facebookresearch:1 origin:1 element:5 trend:1 recognition:20 particularly:1 database:2 role:1 module:113 preprint:2 electrical:1 capture:1 solved:1 calculate:1 wang:1 region:3 connected:2 sun:1 highest:4 removed:1 dynamic:3 trained:2 singh:1 solving:1 deva:1 segment:2 serve:1 imdb:1 completely:1 swap:1 multimodal:25 easily:1 joint:3 represented:3 yash:1 train:5 stacked:1 describe:2 artificial:2 detected:3 sc:6 startup:1 kevin:2 zemel:1 jianfeng:1 whose:1 modular:9 posed:3 plausible:1 encoded:1 cvpr:4 kai:1 otherwise:1 apparent:2 encoder:10 ability:4 simonyan:1 g1:1 jointly:2 validates:1 final:1 superscript:1 agrawal:1 interaction:16 product:5 frequent:1 enc:3 ablation:4 hadamard:4 poorly:1 achieve:1 representational:2 margaret:1 bug:1 sutskever:2 darrell:6 produce:3 generating:1 captioning:2 adam:2 leave:1 object:34 resnet:2 help:2 develop:3 executing:1 ac:1 andrew:1 received:1 eq:19 implemented:4 predicted:2 involves:1 skip:2 come:4 kenji:1 direction:1 laurens:2 guided:1 correct:3 filter:1 subsequently:1 stochastic:1 human:5 sgn:1 education:1 require:2 hassner:3 feeding:1 kristen:1 elementary:1 comprehension:3 ryan:1 exploring:1 considered:1 ground:2 dhruv:4 elevating:1 great:1 lawrence:5 mapping:1 visualize:3 rgen:2 achieves:2 torralba:3 ruslan:2 proc:1 applicable:1 condensed:1 label:14 tanh:2 yuke:2 jackel:1 ross:2 hubbard:1 largest:1 hoffman:1 concurrently:1 clearly:1 zhou:2 agata:2 resized:1 encode:2 release:1 focus:4 validated:1 improvement:6 kwak:1 rank:2 likelihood:1 indicates:1 pretrain:1 contrast:3 huijuan:1 kim:4 helpful:1 lowercase:1 softly:1 typically:1 hidden:3 perona:1 tak:1 pixel:1 issue:2 classification:42 arg:1 overall:3 denoted:6 art:10 softmax:4 spatial:1 equal:1 aware:1 emotion:7 resnet152:1 beach:1 khot:1 piotr:3 saurabh:1 represents:4 park:1 look:4 thin:1 fcn:2 future:2 report:2 yoshua:1 bg1:1 employ:3 richard:3 kalantidis:1 national:3 hamper:1 individual:2 microsoft:1 william:1 attempt:1 detection:17 interest:1 investigate:2 evaluation:5 henderson:1 alignment:2 punctuation:1 sh:1 jacobson:1 word2vec:2 antol:2 accurate:1 oliver:2 jeonghee:2 necessary:1 dkw:1 predominately:1 facial:2 old:1 initialized:2 girshick:2 hiecoatt:1 criticized:1 uncertain:1 classify:1 soft:1 modeling:2 fdec:3 dev:1 psychological:1 heo:1 clipping:1 stacking:1 fusing:1 phrase:1 neutral:1 hundred:1 comprised:4 krizhevsky:1 jiashi:3 johnson:4 too:1 front:1 reported:1 answer:22 synthetic:1 st:2 fritz:1 lstm:7 international:12 stay:1 lee:1 dong:1 michael:3 connecting:1 mouse:1 ilya:2 sanjeev:1 squared:1 augmentation:1 containing:2 choose:1 opposed:1 woman:2 tile:1 possibly:2 emnlp:1 necessitate:1 corner:1 huang:1 style:1 nevatia:1 sang:1 li:8 szegedy:1 converted:1 lookup:7 bold:2 b2:1 includes:1 matter:1 kate:3 explicitly:2 collobert:2 stream:1 later:1 performed:2 script:1 lab:1 tsung:3 observing:1 mario:1 crowdsourced:1 capability:7 complicated:1 annotation:1 jia:2 contribution:5 ass:1 hwa:2 publicly:1 greg:1 convolutional:15 accuracy:12 ni:1 hariharan:2 ensemble:1 serge:1 farach:1 yes:3 handwritten:1 lu:3 ren:1 russakovsky:1 acc:1 bharath:2 explain:1 detector:3 trevor:6 hlt:1 definition:1 failure:2 derek:1 james:1 naturally:1 dataset:13 ask:1 appends:1 mitchell:1 knowledge:2 color:2 dimensionality:4 improves:3 segmentation:1 visual7w:1 sean:1 actually:2 higher:2 zisserman:1 wei:1 evaluated:1 box:13 execute:1 strongly:1 just:1 smola:1 hand:4 sketch:2 christopher:1 su:1 propagation:1 acrf:1 brings:1 aj:1 gray:1 aude:2 multifaceted:10 xiaodong:1 usa:1 building:2 usage:1 contain:3 naacl:1 grounding:1 normalized:1 regularization:1 assigned:2 fidler:1 nmn:2 semantic:1 round:1 icmi:1 backpropagating:1 m:2 evident:1 complete:5 bring:1 reasoning:35 zhiheng:1 image:76 hallmark:1 wise:2 novel:4 recently:2 parikh:4 common:7 overview:1 volume:1 million:2 discussed:1 he:3 association:2 elementwise:1 fukui:3 significant:1 vec:1 ai:2 fk:1 similarly:1 language:12 mlb:1 sanja:1 etc:1 add:1 showed:1 recent:1 aishwarya:1 coco:3 schmidhuber:2 hay:1 server:2 binary:2 success:2 vt:1 meeting:1 der:2 joshua:2 yi:3 preserving:1 p10:1 ministry:1 akira:1 krishna:1 zip:1 deng:2 xiangyu:1 mrn:1 corrado:1 july:1 stephen:1 signal:1 multiple:6 infer:1 reduces:1 match:1 adapt:1 hata:1 long:6 lin:4 bolei:2 calculates:1 prediction:2 involving:1 crop:2 oliva:2 vision:23 metric:1 expectation:2 arxiv:4 represent:6 normalization:2 sergey:1 grounded:1 achieved:2 hochreiter:1 cropped:2 uninformative:1 addition:5 krause:1 diagram:1 jian:1 modality:2 crucial:2 pinheiro:2 rest:1 pass:2 probably:1 pooling:7 december:1 incorporates:4 integer:3 ee:1 counting:1 yang:3 bernstein:2 bengio:1 gave:1 architecture:4 reduce:2 andreas:4 parameterizes:1 vgg:1 shift:1 expression:1 accelerating:1 karen:2 shaoqing:1 compositional:2 deep:11 antonio:3 useful:1 malinowski:1 vqa:74 karpathy:1 referential:1 desk:1 ten:2 tenenbaum:1 visualized:1 rw:3 http:6 wiki:1 singapore:4 moses:1 sign:1 diagnostic:1 gil:2 per:1 correctly:1 klein:3 group:3 key:1 four:2 levi:2 threshold:3 shih:1 spelke:1 macomber:1 prevent:1 douglas:1 utilize:3 rectangle:3 v1:12 kept:1 fuse:1 ram:1 pietro:1 letter:3 raquel:1 named:2 disgust:1 place:2 places365:2 yann:2 timofte:2 sad:1 home:1 maaten:2 dropout:2 layer:13 followed:1 angry:2 summer:1 refine:2 annual:1 bv:2 fei:6 alex:2 scene:20 encodes:3 tal:2 aspect:6 nitish:1 min:2 mikolov:1 martin:1 charikar:1 radu:2 according:1 manning:1 remain:1 slightly:1 byoung:1 character:1 elizabeth:1 stephanie:1 making:1 iccv:3 taken:1 tier:1 jiasen:3 equation:1 count:2 mechanism:1 fed:3 end:3 serf:2 available:2 doll:3 apply:1 observe:5 denker:1 v2:13 hierarchical:1 caiming:1 xiong:1 batch:3 denotes:4 top:2 linguistics:2 donghyun:1 build:1 chieh:1 feng:3 tensor:12 byoungtak:1 question:90 said:1 gradient:1 unable:1 mapped:4 separating:1 entity:9 penultimate:1 concatenation:6 outer:1 landmark:2 seven:1 argue:1 urtasun:1 reason:12 stanislaw:1 marcus:5 unlexicalized:1 besides:1 code:1 index:4 relationship:9 mini:1 multiplicatively:1 happy:5 trapezoid:2 rasmus:2 liang:1 difficult:1 disentangling:1 zichao:1 subproblems:1 hao:1 ronghang:2 ba:1 implementation:2 countable:1 satheesh:1 perform:4 allowing:1 upper:1 convolution:1 datasets:9 benchmark:4 howard:1 jin:2 november:1 displayed:1 incorrectly:1 c08:1 hinton:2 incorporated:1 station:1 arbitrary:1 community:1 expressiveness:2 compositionality:1 introduced:1 bk:2 namely:3 pair:2 david:1 extensive:1 imagenet:2 qiao:2 learned:8 boser:1 textual:1 kingma:1 nu:3 nip:1 qa:1 address:1 able:1 adult:1 justin:3 pattern:12 ranjay:1 mateusz:1 challenge:1 program:2 rf:5 max:5 memory:5 interpretability:1 gool:2 hot:4 natural:6 predicting:1 cascaded:2 residual:5 zhu:3 representing:2 github:4 misleading:1 dkh:1 woo:3 text:7 review:1 sg:1 understanding:8 multiplication:3 fully:7 interesting:1 limitation:3 geoffrey:2 age:5 annotator:1 validation:2 shelhamer:1 rothe:2 vectorized:1 proxy:2 kravitz:1 xiao:1 eccv:1 row:1 summary:1 changed:1 supported:1 last:1 jung:2 bias:4 formal:1 deeper:1 perceptron:2 wide:2 face:30 correspondingly:1 sparse:1 distributed:2 slice:2 overcome:1 dimension:8 raining:1 van:4 world:2 rich:1 genome:1 forward:1 made:1 author:1 openu:1 counted:1 far:1 employing:1 transaction:1 compact:4 obtains:4 implicitly:3 trim:1 bernhard:1 keep:1 global:3 sequentially:1 active:1 colton:1 ioffe:1 b1:1 overfitting:1 belongie:1 sk:6 khosla:3 reviewed:1 ilija:3 promising:1 table:14 learn:7 nature:2 transfer:2 ca:1 composing:1 obtaining:1 ignoring:1 woon:1 huk:1 excellent:4 complex:8 european:2 zitnick:4 official:1 anna:1 dense:2 main:2 big:1 bounding:8 noise:1 fair:1 child:1 xu:2 join:1 cubic:1 judy:1 sub:1 inferring:1 concatenating:1 answering:32 third:1 learns:11 young:1 yannis:1 removing:2 c67:1 shade:1 specific:16 covariate:1 mitigates:1 dk:1 decay:1 evidence:3 fusion:6 incorporating:3 socher:1 magnitude:1 conditioned:1 margin:3 gap:1 chen:5 surprise:1 lt:2 simply:1 likely:2 rohrbach:6 devi:4 visual:56 gao:2 aditya:3 kaiming:1 partially:5 g2:1 fear:1 pretrained:6 applies:1 springer:1 gender:4 pedro:2 truth:1 kan:1 ch:1 abc:1 groth:2 acm:1 tejas:1 yukun:1 ma:1 goal:1 sized:1 oct:1 consequently:1 jianwei:1 jeff:1 luc:2 man:4 content:9 specifically:2 reducing:1 semantically:1 deepmask:1 olga:1 batra:4 saenko:3 exception:1 formally:2 highdimensional:1 il:1 internal:1 people:3 jonathan:2 ethz:1 evaluate:5 mcb:5 srivastava:1 |
6,255 | 6,659 | Adversarial Surrogate Losses for Ordinal Regression
Rizal Fathony
Mohammad Bashiri
Brian D. Ziebart
Department of Computer Science
University of Illinois at Chicago
Chicago, IL 60607
{rfatho2, mbashi4, bziebart}@uic.edu
Abstract
Ordinal regression seeks class label predictions when the penalty incurred for
mistakes increases according to an ordering over the labels. The absolute error
is a canonical example. Many existing methods for this task reduce to binary
classification problems and employ surrogate losses, such as the hinge loss. We
instead derive uniquely defined surrogate ordinal regression loss functions by
seeking the predictor that is robust to the worst-case approximations of training
data labels, subject to matching certain provided training data statistics. We
demonstrate the advantages of our approach over other surrogate losses based on
hinge loss approximations using UCI ordinal prediction tasks.
1
Introduction
For many classification tasks, the discrete class labels being predicted have an inherent order (e.g.,
poor, fair, good, very good, and excellent labels). Confusing two classes that are distant from one
another (e.g., poor instead of excellent) is more detrimental than confusing two classes that are nearby.
The absolute error, |?
y ? y| between label prediction (?
y ? Y) and actual label (y ? Y) is a canonical
ordinal regression loss function. The ordinal regression task seeks class label predictions for new
datapoints that minimize losses of this kind.
Many prevalent methods reduce the ordinal regression task to subtasks solved using existing supervised learning techniques. Some view the task from the regression perspective and learn both a linear
regression function and a set of thresholds that define class boundaries [1?5]. Other methods take a
classification perspective and use tools from cost-sensitive classification [6?8]. However, since the
absolute error of a predictor on training data is typically a non-convex (and non-continuous) function
of the predictor?s parameters for each of these formulations, surrogate losses that approximate the
absolute error must be optimized instead. Under both perspectives, surrogate losses for ordinal regression are constructed by transforming the surrogate losses for binary zero-one loss problems?such as
the hinge loss, the logistic loss, and the exponential loss?to take into account the different penalties
of the ordinal regression problem. Empirical evaluations have compared the appropriateness of
different surrogate losses, but these still leave the possibility of undiscovered surrogates that align
better with the ordinal regression loss.
To address these limitations, we seek the most robust [9] ordinal regression predictions by focusing
on the following adversarial formulation of the ordinal regression task: what predictor best minimizes
absolute error in the worst case given partial knowledge of the conditional label distribution? We
answer this question by considering the Nash equilibrium for a game defined by combining the loss
function with Lagrangian potential functions [10]. We derive a surrogate loss function for empirical
risk minimization that realizes this same adversarial predictor. We show that different types of
available knowledge about the conditional label distribution lead to thresholded regression-based
predictions or classification-based predictions. In both cases, the surrogate loss is novel compared to
existing surrogate losses. We also show that our surrogate losses enjoy Fisher consistency, a desirable
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
theoretical property guaranteeing that minimizing the surrogate loss produces Bayes optimal decisions
for the original loss in the limit. We develop two different approaches for optimizing the loss: a
stochastic optimization of the primal objective and a quadratic program formulation of the dual
objective. The second approach enables us to efficiently employ the kernel trick to provide a richer
feature representation without an overly burdensome time complexity. We demonstrate the benefits
of our adversarial formulation over previous ordinal regression methods based on hinge loss for a
range of prediction tasks using UCI datasets.
2
Background and Related Work
2.1
Ordinal Regression Problems
Ordinal regression is a discrete label prediction problem characterized by Table 1: Ordinal rean ordered penalty for making mistakes: loss(?
y1 , y) < loss(?
y2 , y) if y < gression loss matrix.
?
?
y?1 < y?2 or y > y?1 > y?2 . Though many loss functions possess this property,
0 1 2 3
the absolute error |?
y ? y| is the most widely studied. We similarly restrict
?1 0 1 2?
our consideration to this loss function in this paper. The full loss matrix
?2 1 0 1?
L for absolute error with four labels is shown in Table 1. The expected
3 2 1 0
loss incurred using a probabilistic predictor P? (?
y |x) evaluated on true data
P
? y |x)Ly?,y . The supervised
distribution P (x, y) is: EX,Y ?P ;Y? |X?P? [LY? ,Y ] =
x,y,?
y P (x, y)P (?
y |x) in a way
learning objective for this problem setting is to construct a probabilistic predictor P? (?
that minimizes this expected loss using training samples distributed according to the empirical
distribution P? (x, y), which are drawn from the unknown true data generating distribution, P (x, y).
A na?ve ordinal regression approach relaxes the task to a continuous prediction problem, minimizes
the least absolute deviation [11], and then rounds predictions to nearest integral label [12]. More
sophisticated methods range from using a cumulative link model [13] that assumes the cumulative
conditional probability P (Y ? j|x) follows a link function, to Bayesian non-parametric approaches
[14] and many others [15?22]. We narrow our focus over this broad range of methods found in the
related work to those that can be viewed as empirical risk minimization methods with piece-wise
convex surrogates, which are more closely related to our approach.
2.2
Threshold Methods for Ordinal Regression
Threshold methods are one popular family of techniques that treat the ordinal response variable,
f? , w ? x, as a continuous real-valued variable and introduce |Y| ? 1 thresholds ?1 , ?2 , ..., ?|Y|?1
that partition the real line into |Y| segments: ?0 = ?? < ?1 < ?2 < ... < ?|Y|?1 < ?|Y| = ?
[4]. Each segment corresponds to a label with y?i assigned label j if ?j?1 < f? ? ?j . There are two
different approaches for constructing surrogate losses based on the threshold methods to optimize the
choice of w and ?1 , . . . , ?|Y|?1 : one is based on penalizing all thresholds involved when a mistake is
made and one is based on only penalizing the most immediate thresholds.
All thresholds methods penalize every erroneous threshold using a surrogate loss, ?, for sets of binary
Py?1
P|Y|
classification problems: lossAT (f?, y) = k=1 ?(?(?k ? f?)) + k=y ?(?k ? f?). Shashua and Levin
[1] studied the hinge loss under the name of support vector machines with a sum-of margin strategy,
while Chu and Keerthi [2] proposed a similar approach under the name of support vector ordinal
regression with implicit constraints (SVORIM). Lin and Li [3] proposed ordinal regression boosting,
an all thresholds method using the exponential loss as a surrogate. Finally, Rennie and Srebro [4]
proposed a unifying approach for all threshold methods under a variety of surrogate losses.
Rather than penalizing all erroneous thresholds when an error is made, immediate thresholds methods
only penalize the threshold of the true label and the threshold immediately beneath the true label:
lossIT (f?, y) = ?(?(?y?1 ? f?)) + ?(?y ? f?).1 Similar to the all thresholds methods, immediate threshold methods have also been studied in the literature under different names. For hinge loss surrogates,
Shashua and Levin [1] called the model support vector with fixed-margin strategy while Chu and
Keerthi [2] use the term support vector ordinal regression with explicit constraints (SVOREX). For
1
For the boundary labels, the method defines ?(?(?0 ? f?)) = ?(?y+1 ? f?) = 0.
2
the exponential loss, Lin and Li [3] introduced ordinal regression boosting with left-right margins.
Rennie and Srebro [4] also proposed a unifying framework for immediate threshold methods.
2.3
Reduction Framework from Ordinal Regression to Binary Classification
Li and Lin [5] proposed a reduction framework to convert ordinal regression problems to binary
classification problems by extending training examples. For each training sample (x, y), the reduction
framework creates |Y| ? 1 extended samples (x(j) , y (j) ) and assigns weight wy,j to each extended
sample. The binary label associated with the extended sample is equivalent to the answer of the
question: ?is the rank of x greater than j?? The reduction framework allows a choice for how extended
samples x(j) are constructed from original samples x and how to perform binary classification. If
the threshold method is used to construct the extended sample and SVM is used as the binary
classification algorithm, the classifier can be obtained by solving a family of quadratic optimization
problems that includes SVORIM and SVOREX as special instances.
2.4
Cost-sensitive Classification Methods for Ordinal Regression
Rather than using thresholding or the reduction framework, ordinal regression can also be cast as a
special case of cost-sensitive multiclass classification. Two of the most popular classification-based
ordinal regression techniques are extensions of one-versus-one (OVO) and one-versus-all (OVA) costsensitive classification [6, 7]. Both algorithms leverage a transformation that converts a cost-sensitive
classification problem to a set of weighted binary classification problems. Rather than reducing
to binary classification, Tu and Lin [8] reduce cost-sensitive classification to one-sided regression
(OSR), which can be viewed as an extension of the one-versus-all (OVA) technique.
2.5
Adversarial Prediction
Foundational results establish a duality between adversarial logarithmic loss minimization and
constrained maximization of the entropy [23]. This takes the form of a zero-sum game between
a predictor seeking to minimize expected logarithmic loss and an adversary seeking to maximize
this same loss. Additionally, the adversary is constrained to choose a distribution that matches
certain sample statistics. Ultimately, through the duality to maximum entropy, this is equivalent
to maximum likelihood estimation of probability distributions that are members of the exponential
family [23]. Gr?nwald and Dawid [9] emphasize this formulation as a justification for the principle of
maximum entropy [24] and generalize the adversarial formulation to other loss functions. Extensions
to multivariate performance measures [25] and non-IID settings [26] have demonstrated the versatility
of this perspective.
Recent analysis [27, 28] has shown that for the special case of zero-one loss classification, this
adversarial formulation is equivalent to empirical risk minimization with a surrogate loss function:
X
AL0-1
max
(?j,yi (xi ) + |S| ? 1)/|S|,
(1)
f (xi , yi ) =
S?{1,...,|Y|},S6=?
j?S
where ?j,yi (xi ) is the potential difference ?j,yi (xi ) = fj (xi ) ? fyi (xi ). This surrogate loss function
provides a key theoretical advantage compared to the Crammer-Singer hinge loss surrogate for
multiclass classification [29]: it guarantees Fisher consistency [27] while Crammer-Singer?despite
its popularity in many applications, such as Structured SVM [30, 31]?does not [32, 33]. We extend
this type of analysis to the ordinal regression setting with the absolute error as the loss function in
this paper, producing novel surrogate loss functions that provide better predictions than other convex,
piece-wise linear surrogates.
3
3.1
Adversarial Ordinal Regression
Formulation as a zero-sum game
We seek the ordinal regression predictor that is the most robust to uncertainty given partial knowledge
of the evaluating distribution?s characteristics. This takes the form of a zero-sum game between a
predictor player choosing a predicted label distribution P? (?
y |x) that minimizes loss and an adversarial
3
player choosing an evaluation distribution P? (?
y |x) that maximizes loss while closely matching the
feature-based statistics of the training data:
min
i
h
?
max EX?P ;Y? |X?P? ;Y? |X?P? Y? ? Y? such that: EX?P ;Y? |X?P? [?(X, Y? )] = ?.
y |x)
P? (?
y |x) P? (?
(2)
The vector of feature moments, ?? = EX,Y ?P? [?(X, Y )], is measured from sample training data
distributed according to the empirical distribution P? (x, y).
An ordinal regression problem can be viewed as a cost-sensitive loss with the entries of the cost
matrix defined by the absolute loss between the row and column labels (an example of the cost
matrix for the case of a problem with four labels is shown in Table 1). Following the construction of
adversarial prediction games for cost-sensitive classification [10], the optimization of Eq. (2) reduces
to minimizing the equilibrium game values of a new set of zero-sum games characterized by matrix
L0xi ,w :
f1 ? fyi
}|
{
f1 ? fyi + 1
Xz
?
? Txi L0xi ,w p
? xi ; L0xi ,w=?
min
max min p
..
?
w
? xi p
p
? xi
.
i
{z
}
|
f1 ? fyi + |Y| ? 1
zero-sum game
?
convex optimization of w
???
???
..
.
???
?
f|Y| ? fyi + |Y| ? 1
f|Y| ? fyi + |Y| ? 2?
? , (3)
..
?
.
f|Y| ? fyi
where: w represents a vector of Lagrangian model parameters; fj = w ? ?(xi , j) is a Lagrangian
? xi is a vector representation of the conditional label distribution, P? (Y? = j|xi ), i.e.,
potential; p
?
? xi = [P (Y? = 1|xi ) P? (Y? = 2|xi ) . . .]T ; and p
? xi is similarly defined. The matrix L0xi ,w =
p
(|?
y ? y?| + fy? ? fyi ) is a zero-sum game matrix for each example. This optimization problem (Eq. (3))
is convex in w and the inner zero-sum game can be solved using a linear program [10]. To address
finite sample estimation errors, the difference between expected and sample feature can be bounded
? ? , leading to Lagrangian parameter regularization in
in Eq. (2), ||EX?P ;Y? |X?P? [?(X, Y? )] ? ?||
Eq. (3) [34].
3.2
Feature representations
We consider two feature representations corresponding to different training data summaries:
?
?
?
?th (x, y) = ?
?
?
yx
I(y ? 1)
I(y ? 2)
..
.
I(y ? |Y| ? 1)
?
?
I(y = 1)x
? I(y = 2)x ?
? I(y = 3)x ?
?.
?mc (x, y) = ?
?
?
..
?
?
.
I(y = |Y|)x
?
?
?
? ; and
?
?
(4)
The first, which we call the thresholded regression representation,
has size m + |Y| ? 1, where m is the dimension of our input space. It
induces a single shared vector of feature weights and a set of thresholds. If we denote the weight vector associated with the yx term as
w and the terms associated with each sum of class indicator functions as ?1 , ?2 , . . ., ?|Y|?1 , then thresholds for switching between
class j and j + 1 (ignoring other classes) occur when w ? xi = ?j .
The second feature representation, ?mc , which we call the multiclass representation, has size m|Y| and can be equivalently interpreted as inducing a set of class-specific feature weights, fj = wj ?xi .
This feature representation is useful when ordered labels cannot be
thresholded according to any single direction in the input space, as Figure 1: Example where mulshown in the example dataset of Figure 1.
tiple weight vectors are useful.
4
3.3
Adversarial Loss from the Nash Equilibrium
We now present the main technical contribution of our paper: a surrogate loss function that, when
minimized, produces a solution to the adversarial ordinal regression problem of Eq. (3).2
Theorem 1. An adversarial ordinal regression predictor is obtained by choosing parameters w that
minimize the empirical risk of the surrogate loss function:
fj + fl + j ? l
fj + j
fl ? l
? fyi = max
+ max
? fyi ,
j
l
2
2
2
j,l?{1,...,|Y|}
ALord
w (xi , yi ) =
max
(5)
where fj = w ? ?(xi , j) for all j ? {1, . . . , |Y|}.
f +f +j?l
Proof sketch. Let j ? , l? be the solution of argmaxj,l?{1,...,|Y|} j 2l
, we show that the Nash
equilibrium value of a game matrix that contains only row j ? and l? and column j ? and l? from
f ? ,+fl? +j ? ?l?
matrix L0xi ,w is exactly j
. We then show that adding other rows and columns in L0xi ,w
2
to the game matrix does not change the game value. Given the resulting closed form solution of the
game (instead of a minimax), we can recast the adversarial framework for ordinal regression as an
empirical risk minimization with the proposed loss.
We note that the ALord
w surrogate is the maximization over pairs of different potential functions
associated with each class (including pairs of identical class labels) added to the distance between the
pair. For both of our feature representations, we make use of the fact that maximization over each
element of the pair can be independently realized, as shown on the right-hand side of Eq. (5).
Thresholded regression surrogate loss
In the thresholded regression feature representation, the parameter contains a single shared vector of
feature weights w and |Y| ? 1 terms ?k associated with thresholds. Following Eq. (5), the adversarial
ordinal regression surrogate loss for this feature representation can be written as:
ord-th
AL
(xi , yi ) = max
j(w ? xi + 1) +
j
P
k?j
?k
2
+ max
l
l(w ? xi ? 1) +
2
P
k?l
?k
? yi w ? x i ?
X
?k .
k?yi
(6)
This loss has a straight-forward interpretation in terms of the thresholded regression
perspective, as shown in Figure 2: it is
based on averaging the thresholded label
predictions for potentials w ? xi + 1 and
w ? xi ? 1. This penalization of the pair of
thresholds differs from the thresholded surrogate losses of related work, which either
penalize all violated thresholds or penalize
only the thresholds adjacent to the actual
Figure 2: Surrogate loss calculation for datapoint xi
class label.
(projected to w ? xi ) with a label prediction of 4 for preUsing a binary search procedure over
dictive purposes, the surrogate loss is instead obtained
?1 , . . . , ?|Y|?1 , the largest lower bounding
using potentials for the classes based on w ?xi +1 (label
threshold for each of these potentials can
5) and w ? xi ? 1 (label 2) averaged together.
be obtained in O(log |Y|) time.
Multiclass ordinal surrogate loss
In the multiclass feature representation, we have a set of specific feature weights wj for each label
and the adversarial multiclass ordinal surrogate loss can be written as:
ALord-mc (xi , yi ) =
max
j,l?{1,...,|Y|}
wj ? xi + wl ? xi + j ? l
? wyi ? xi .
2
2
(7)
The detailed proof of this theorem and others are contained in the supplementary materials. Proof sketches
are presented in the main paper.
5
(a)
(b)
(c)
Figure 3: Loss function contour plots of ALord over the space of potential differences ?j , fj ? fyi
for the prediction task with three classes when the true label is yi = 1 (a), yi = 2 (b), and yi = 3 (c).
We can also view this as the maximization over |Y|(|Y| + 1)/2 linear hyperplanes. For an ordinal
regression problem with three classes, the loss has six facets with different shapes for each true label
value, as shown in Figure 3. In contrast with ALord-th , the class label potentials for ALord-mc may
differ from one another in more-or-less arbitrary ways. Thus, searching for the maximal j and l class
labels requires O(|Y|) time.
3.4
Consistency Properties
The behavior of a prediction method in ideal learning settings?i.e., trained on the true evaluation
distribution and given an arbitrarily rich feature representation, or, equivalently, considering the space
of all measurable functions?provides a useful theoretical validation. Fisher consistency requires that
the prediction model yields the Bayes optimal decision boundary [32, 33, 35] in this setting. Given
the true label conditional probability Pj (x) , P (Y = j|x), a surrogate loss function ? is said to
be Fisher consistent with respect to the loss ` if the minimizer f ? of the surrogate loss achieves the
Bayes optimal risk, i.e.,:
f ? = argmin EY |X?P [?f (X, Y )|X = x]
(8)
f
? EY |X?P [`f ? (X, Y )|X = x] = min EY |X?P [`f (X, Y )|X = x] .
f
Ramaswamy and Agarwal [36] provide a necessary and sufficient condition for a surrogate loss to be
Fisher consistent with respect to general multiclass losses, which includes ordinal regression losses.
A recent analysis by Pedregosa et al. [35] shows that the all thresholds and the immediate thresholds
methods are Fisher consistent provided that the base binary surrogates losses they use are convex
with a negative derivative at zero.
For our proposed approach, the condition for Fisher consistency above is equivalent to:
f ? = argmin
f
X
y
X
fj + fl + j ? l
Py max
? fy ? argmax fj? (x) ? argmin
Py |j ? y| .
j,l
2
j
j
y
(9)
Since adding a constant to all fj does not change the value of both ALord
f and argmaxj fj (x), we
employ the constraint maxj fj (x) = 0, to remove redundant solutions for the consistency analysis.
We establish an important property of the minimizer for ALord
f in the following theorem.
ord
?
Theorem 2. The minimizer vector f of EY |X?P ALf (X, Y )|X = x satisfies the loss reflective
property, i.e., it complements the absolute error by starting with a negative integer value, then
increasing by one until reaching zero, and then incrementally decreases again.
Proof sketch. We show that for any f 0 that does not satisfy the loss reflective property, we can
construct f 1 using several steps that satisfy the loss reflective property and has the expected loss value
less than the expected loss of f 0 .
6
Example vectors f ? that satisfy Theorem 2 are [0, ?1, ?2]T , [?1, 0, ?1]T and [?2, ?1, 0]T for
three-class problems, and [?3, ?2, ?1, 0, ?1] for five-class problems. Using the key property of the
minimizer above, we establish the consistency of our loss functions in the following theorem.
Theorem 3. The adversarial ordinal regression surrogate loss ALord from Eq. (5) is Fisher consistent.
Proof sketch. We only consider |Y| possible values of f that satisfy the loss reflective property. For
the f that corresponds to class j, the value of the expected loss is equal to the Bayes loss if we predict
j as the label. Therefore minimizing over f that satisfy the loss reflective property is equivalent to
finding the Bayes optimal response.
3.5
Optimization
3.5.1
Primal Optimization
To optimize the regularized adversarial ordinal regression loss from the primal, we employ stochastic
average gradient (SAG) methods [37, 38], which have been shown to converge faster than standard
stochastic gradient optimization. The idea of SAG is to use the gradient of each example from the last
iteration where it was selected to take a step. However, the na?ve implementation of SAG requires
storing the gradient of each sample, which may be expensive in terms of the memory requirements.
Fortunately, for our loss ALord
w , we can drastically reduce this memory requirement by just storing a
f +f +j?l
pair of number, (j ? , l? ) = argmaxj,l?{1,...,|Y|} j 2l
, rather than storing the gradient for each
sample. Appendix C explains the details of this technique.
3.5.2
Dual Optimization
Dual optimization is often preferred when optimizing piecewise linear losses, such as the hinge loss,
since it enables one to easily perform the kernel trick and obtain a non-linear decision boundary
without heavily sacrificing computational efficiency. Optimizing the regularized adversarial ordinal
regression loss in the dual can be performed by solving the following quadratic optimization:
max
?,?
X
1 X
(?i,j + ?i,j ) (?k,l + ?k,l ) (?(xi , j) ? ?(xi , yi )) ? (?(xk , l) ? ?(xl , yk ))
2
i,j,k,l
X
X
? 0;
?i,j = C2 ;
?i,j = C2 ; i, k ? {1, . . . , n}; j, l ? {1, . . . , |Y|}.
(10)
j(?i,j ? ?i,j ) ?
i,j
subject to: ?i,j ? 0; ?i,j
j
j
Note that our dual formulation only depends on the dot product of the features. Therefore, we can
also easily apply the kernel trick to our algorithm. Appendix D describes the derivation from the
primal optimization to the dual optimization above.
4
4.1
Experiments
Table 2: Dataset properties.
Experiment Setup
Dataset
We conduct our experiments on a benchmark
dataset for ordinal regression [14], evaluate the
performance using mean absolute error (MAE),
and perform statistical tests on the results of different hinge loss surrogate methods. The benchmark contains datasets taken from the UCI Machine Learning repository [39] ranging from relatively small to relatively large in size. The characteristics of the datasets, including the number
of classes, the training set size, the testing set
size, and the number of features, are described
in Table 2.
#class
diabetes
pyrimidines
triazines
wisconsin
machinecpu
autompg
boston
stocks
abalone
bank
computer
calhousing
5
5
5
5
10
10
5
5
10
10
10
10
#train
#test #features
30
51
130
135
146
274
354
665
2923
5734
5734
14447
13
23
56
59
63
118
152
285
1254
2458
2458
6193
2
27
60
32
6
7
13
9
10
8
21
8
In the experiment, we consider different methods using the original feature space and a kernelized
feature space using the Gaussian radial basis function kernel. The methods that we compare include
two variations of our approach, the threshold based (ALord-th ), and the multiclass-based (ALord-mc ).
7
Table 3: The average of the mean absolute error (MAE) for each model. Bold numbers in each case
indicate that the result is the best or not significantly worse than the best (paired t-test with ? = 0.05).
Threshold-based models
Dataset
ord-th
AL
th
RED
Multiclass-based models
AT
IT
ord-mc
REDmc
CSOSR
CSOVO
AL
CSOVA
diabetes
pyrimidines
triazines
wisconsin
machinecpu
autompg
boston
stocks
abalone
bank
computer
calhousing
0.696
0.654
0.607
1.077
0.449
0.551
0.316
0.324
0.551
0.461
0.640
1.190
0.715
0.678
0.683
1.067
0.456
0.550
0.304
0.317
0.547
0.460
0.635
1.183
0.731
0.615
0.649
1.097
0.458
0.550
0.306
0.315
0.546
0.461
0.633
1.182
0.827
0.626
0.654
1.175
0.467
0.617
0.298
0.324
0.571
0.461
0.683
1.225
0.629
0.509
0.670
1.136
0.518
0.599
0.311
0.168
0.521
0.445
0.625
1.164
0.700
0.565
0.673
1.141
0.515
0.602
0.311
0.175
0.520
0.446
0.624
1.144
0.715
0.520
0.677
1.208
0.646
0.741
0.353
0.204
0.545
0.732
0.889
1.237
0.738
0.576
0.738
1.275
0.602
0.598
0.294
0.147
0.558
0.448
0.649
1.202
0.762
0.526
0.732
1.338
0.702
0.731
0.363
0.213
0.556
0.989
1.055
1.601
average
# bold
0.626
5
0.633
5
0.629
4
0.661
2
0.613
5
0.618
5
0.706
2
0.652
2
0.797
1
The baselines we use for the threshold-based models include a SVM-based reduction framework
algorithm (REDth ) [5], an all threshold method with hinge loss (AT) [1, 2], and an immediate threshold
method with hinge loss (IT) [1, 2]. For the multiclass-based models, we compare our method with an
SVM-based reduction algorithm using multiclass features (REDmc ) [5], with cost-sensitive one-sided
support vector regression (CSOSR) [8], with cost-sensitive one-versus-one SVM (CSOVO) [7], and
with cost-sensitive one-versus-all SVM (CSOVA) [6]. For our Gaussian kernel experiment, we
compare our threshold based model (ALord-th ) with SVORIM and SVOREX [2].
In our experiments, we first make 20 random splits of each dataset into training and testing sets. We
performed two stages of five-fold cross validation on the first split training set for tuning each model?s
regularization constant ?. In the first stage, the possible values for ? are 2?i , i = {1, 3, 5, 7, 9, 11, 13}.
i
Using the best ? in the first stage, we set the possible values for ? in the second stage as 2 2 ?0 , i =
{?3, ?2, ?1, 0, 1, 2, 3}, where ?0 is the best parameter obtained in the first stage. Using the selected
parameter from the second stage, we train each model on the 20 training sets and evaluate the MAE
performance on the corresponding testing set. We then perform a statistical test to find whether the
performance of a model is different with statistical significance from other models. We perform the
Gaussian kernel experiment similarly with model parameter C equals to 2i , i = {0, 3, 6, 9, 12} and
kernel parameter ? equals to 2i , i = {?12, ?9, ?6, ?3, 0} in the first stage. In the second stage, we
set C equals to 2i C0 , i = {?2, ?1, 0, 1, 2} and ? equals to 2i ?0 , i = {?2, ?1, 0, 1, 2}, where C0
and ?0 are the best parameters obtained in the first stage.
4.2
Results
We report the mean absolute error (MAE) averaged over the dataset splits as shown in Table 3 and
Table 4. We highlight the result that is either the best or not worse than the best with statistical
significance (under paired t-test with ? = 0.05) in boldface font. We also provide the summary for
each model in terms of the averaged MAE over all datasets and the number of datasets for which
each model marked with boldface font in the bottom of the table.
As we can see from Table 3, in the experiment with the original feature space, threshold-based
models perform well on relatively small datasets, whereas multiclass-based models perform well on
relatively large datasets. A possible explanation for this result is that multiclass-based models have
more flexibility in creating decision boundaries, hence they perform better if the training data size is
sufficient. However, since multiclass-based models have many more parameters than threshold-based
models (m|Y| parameters rather than m + |Y| ? 1 parameters), multiclass methods may need more
data, and hence, may not perform well on relatively small datasets.
In the threshold-based models comparison, ALord-th , REDth , and AT perform competitively on
relatively small datasets like triazines, wisconsin, machinecpu, and autompg. ALord-th has a
8
slight advantage over REDth on the overall accuracy, and a slight advantage over AT on the number
of ?indistinguishably best? performance on all datasets. We can also see that AT is superior to IT in
the experiments under the original feature space.
Among the multiclass-based models, ALord-mc Table 4: The average of MAE for models with
and REDmc perform competitively on datasets Gaussian kernel.
like abalone, bank, and computer, with a
Dataset
ALord-th SVORIM SVOREX
slight advantage of ALord-mc model on the overall accuracy. In general, the cost-sensitive moddiabetes
0.696
0.665
0688
pyrimidines
0.478
0.539
0.550
els perform poorly compared with ALord-mc and
triazines
0.609
0.612
0.604
REDmc . A notable exception is the CSOVO
wisconsin
1.090
1.113
1.049
model which perform very well on the stocks
machinecpu
0.452
0.652
0.628
and the boston datasets.
autompg
0.529
0.589
0.593
In the Gaussian kernel experiment, we can see
boston
0.278
0.324
0.316
stocks
0.103
0.099
0.100
from Table 4 that the kernelized version of
ALord-th performs significantly better than the
average
0.531
0.574
0.566
threshold-based models SVORIM and SVOREX
# bold
7
3
4
in terms of both the overall accuracy and the
number of ?indistinguishably best? performance
on all datasets. We also note that immediate-threshold-based model (SVOREX) performs better than
all-threshold-based model (SVORIM) in our experiment using Gaussian kernel. We can conclude
that our proposed adversarial losses for ordinal regression perform competitively compared to the
state-of-the-art ordinal regression models using both original feature spaces and kernel feature spaces
with a significant performance improvement in the Gaussian kernel experiments.
5
Conclusion and Future Work
In this paper, we have proposed a novel surrogate loss for ordinal regression, a classification problem
where the discrete class labels have an inherent order and penalty for making mistakes based on that
order. We focused on the absolute loss, which is the most widely used ordinal regression loss. In
contrast with existing methods, which typically reduce ordinal regression to binary classification
problems and then employ surrogates for the binary zero-one loss, we derive a unique surrogate
ordinal regression loss by seeking the predictor that is robust to a worst case constrained approximation of the training data. We derived two versions of the loss based on two different feature
representation approaches: thresholded regression and multiclass representations. We demonstrated
the benefit of our approach on a benchmark of datasets for ordinal regression tasks. Our approach
performs competitively compared to the state-of-the-art surrogate losses based on hinge loss. We
also demonstrated cases when the multiclass feature representations works better than thresholded
regression representation, and vice-versa, in our experiments.
Our future work will investigate less prevalent ordinal regression losses, such as the discrete quadratic
loss and arbitrary losses that have v-shaped penalties. Furthermore, we plan to investigate the
characteristics required of discrete ordinal losses for their optimization to have a compact analytical
solution. In terms of applications, one possible direction of future work is to combine our approach
with deep neural network models to perform end-to-end representation learning for ordinal regression
applications like age estimation and rating prediction. In that setting, our proposed loss can be used
in the last layer of a deep neural network to serve as the gradient source for the backpropagation
algorithm.
Acknowledgments
This research was supported as part of the Future of Life Institute (futureoflife.org) FLI-RFP-AI1
program, grant#2016-158710 and by NSF grant RI-#1526379.
9
References
[1] Amnon Shashua and Anat Levin. Ranking with large margin principle: Two approaches. In
Advances in Neural Information Processing Systems 15, pages 961?968. MIT Press, 2003.
[2] Wei Chu and S Sathiya Keerthi. New approaches to support vector ordinal regression. In
Proceedings of the 22nd international conference on Machine learning, pages 145?152. ACM,
2005.
[3] Hsuan-Tien Lin and Ling Li. Large-margin thresholded ensembles for ordinal regression:
Theory and practice. In International Conference on Algorithmic Learning Theory, pages
319?333. Springer, 2006.
[4] Jason D. M. Rennie and Nathan Srebro. Loss functions for preference levels: Regression with
discrete ordered labels. In Proceedings of the IJCAI Multidisciplinary Workshop on Advances
in Preference Handling, pages 180?186, 2005.
[5] Ling Li and Hsuan-Tien Lin. Ordinal regression by extended binary classification. Advances in
neural information processing systems, 19:865, 2007.
[6] Hsuan-Tien Lin. From ordinal ranking to binary classification. PhD thesis, California Institute
of Technology, 2008.
[7] Hsuan-Tien Lin. Reduction from cost-sensitive multiclass classification to one-versus-one
binary classification. In Proceedings of the Sixth Asian Conference on Machine Learning, pages
371?386, 2014.
[8] Han-Hsing Tu and Hsuan-Tien Lin. One-sided support vector regression for multiclass costsensitive classification. In Proceedings of the 27th International Conference on Machine
Learning (ICML-10), pages 1095?1102, 2010.
[9] Peter D. Gr?nwald and A. Phillip Dawid. Game theory, maximum entropy, minimum discrepancy, and robust Bayesian decision theory. Annals of Statistics, 32:1367?1433, 2004.
[10] Kaiser Asif, Wei Xing, Sima Behpour, and Brian D. Ziebart. Adversarial cost-sensitive classification. In Proceedings of the Conference on Uncertainty in Artificial Intelligence, 2015.
[11] Subhash C Narula and John F Wellington. The minimum sum of absolute errors regression:
A state of the art survey. International Statistical Review/Revue Internationale de Statistique,
pages 317?326, 1982.
[12] Koby Crammer and Yoram Singer. Pranking with ranking. In Advances in Neural Information
Processing Systems 14, 2001.
[13] Peter McCullagh. Regression models for ordinal data. Journal of the royal statistical society.
Series B (Methodological), pages 109?142, 1980.
[14] Wei Chu and Zoubin Ghahramani. Gaussian processes for ordinal regression. Journal of
Machine Learning Research, 6(Jul):1019?1041, 2005.
[15] Krzysztof Dembczy?nski, Wojciech Kot?owski, and Roman S?owi?nski. Ordinal classification
with decision rules. In International Workshop on Mining Complex Data, pages 169?181.
Springer, 2007.
[16] Mark J Mathieson. Ordinal models for neural networks. Neural networks in financial engineering, pages 523?536, 1996.
[17] Shipeng Yu, Kai Yu, Volker Tresp, and Hans-Peter Kriegel. Collaborative ordinal regression.
In Proceedings of the 23rd international conference on Machine learning, pages 1089?1096.
ACM, 2006.
[18] Jianlin Cheng, Zheng Wang, and Gianluca Pollastri. A neural network approach to ordinal
regression. In Neural Networks, 2008. IJCNN 2008.(IEEE World Congress on Computational
Intelligence). IEEE International Joint Conference on, pages 1279?1284. IEEE, 2008.
[19] Wan-Yu Deng, Qing-Hua Zheng, Shiguo Lian, Lin Chen, and Xin Wang. Ordinal extreme
learning machine. Neurocomputing, 74(1):447?456, 2010.
10
[20] Bing-Yu Sun, Jiuyong Li, Desheng Dash Wu, Xiao-Ming Zhang, and Wen-Bo Li. Kernel
discriminant learning for ordinal regression. IEEE Transactions on Knowledge and Data
Engineering, 22(6):906?910, 2010.
[21] Jaime S Cardoso and Joaquim F Costa. Learning to classify ordinal data: The data replication
method. Journal of Machine Learning Research, 8(Jul):1393?1429, 2007.
[22] Yang Liu, Yan Liu, and Keith CC Chan. Ordinal regression via manifold learning. In Proceedings of the Twenty-Fifth AAAI Conference on Artificial Intelligence, pages 398?403. AAAI
Press, 2011.
[23] Flemming Tops?e. Information theoretical optimization techniques. Kybernetika, 15(1):8?27,
1979.
[24] Edwin T Jaynes. Information theory and statistical mechanics. Physical review, 106(4):620?630,
1957.
[25] Hong Wang, Wei Xing, Kaiser Asif, and Brian Ziebart. Adversarial prediction games for
multivariate losses. In Advances in Neural Information Processing Systems, pages 2710?2718,
2015.
[26] Anqi Liu and Brian Ziebart. Robust classification under sample selection bias. In Advances in
Neural Information Processing Systems, pages 37?45, 2014.
[27] Rizal Fathony, Anqi Liu, Kaiser Asif, and Brian Ziebart. Adversarial multiclass classification:
A risk minimization perspective. In Advances in Neural Information Processing Systems 29,
pages 559?567. 2016.
[28] Farzan Farnia and David Tse. A minimax approach to supervised learning. In Advances in
Neural Information Processing Systems, pages 4233?4241. 2016.
[29] Koby Crammer and Yoram Singer. On the algorithmic implementation of multiclass kernelbased vector machines. The Journal of Machine Learning Research, 2:265?292, 2002.
[30] Ioannis Tsochantaridis, Thorsten Joachims, Thomas Hofmann, and Yasemin Altun. Large
margin methods for structured and interdependent output variables. In JMLR, pages 1453?1484,
2005.
[31] Thorsten Joachims. A support vector method for multivariate performance measures. In
Proceedings of the International Conference on Machine Learning, pages 377?384, 2005.
[32] Ambuj Tewari and Peter L Bartlett. On the consistency of multiclass classification methods.
The Journal of Machine Learning Research, 8:1007?1025, 2007.
[33] Yufeng Liu. Fisher consistency of multicategory support vector machines. In International
Conference on Artificial Intelligence and Statistics, pages 291?298, 2007.
[34] Miroslav Dud?k and Robert E Schapire. Maximum entropy distribution estimation with generalized regularization. In International Conference on Computational Learning Theory, pages
123?138. Springer, 2006.
[35] Fabian Pedregosa, Francis Bach, and Alexandre Gramfort. On the consistency of ordinal
regression methods. Journal of Machine Learning Research, 18(55):1?35, 2017.
[36] Harish G Ramaswamy and Shivani Agarwal. Classification calibration dimension for general
multiclass losses. In Advances in Neural Information Processing Systems, pages 2078?2086,
2012.
[37] Mark Schmidt, Nicolas Le Roux, and Francis Bach. Minimizing finite sums with the stochastic
average gradient. Mathematical Programming, pages 1?30, 2013.
[38] Mark Schmidt, Reza Babanezhad, Aaron Defazio, Ann Clifton, and Anoop Sarkar. Non-uniform
stochastic average gradient method for training conditional random fields. 2015.
[39] M. Lichman. UCI machine learning repository, 2013. URL http://archive.ics.uci.edu/
ml.
11
| 6659 |@word repository:2 version:2 nd:1 c0:2 triazine:4 seek:4 moment:1 reduction:8 liu:5 contains:3 series:1 lichman:1 undiscovered:1 existing:4 jaynes:1 anqi:2 chu:4 must:1 written:2 joaquim:1 john:1 indistinguishably:2 chicago:2 distant:1 partition:1 shape:1 enables:2 hofmann:1 remove:1 plot:1 intelligence:4 selected:2 xk:1 provides:2 boosting:2 preference:2 hyperplanes:1 org:1 zhang:1 five:2 mathematical:1 constructed:2 c2:2 replication:1 combine:1 introduce:1 expected:7 behavior:1 xz:1 owski:1 mechanic:1 ming:1 actual:2 considering:2 increasing:1 provided:2 bounded:1 fathony:2 maximizes:1 what:1 kind:1 interpreted:1 minimizes:4 argmin:3 kybernetika:1 finding:1 transformation:1 guarantee:1 every:1 sag:3 exactly:1 classifier:1 grant:2 ly:2 enjoy:1 producing:1 bziebart:1 engineering:2 treat:1 congress:1 mistake:4 limit:1 switching:1 despite:1 studied:3 range:3 averaged:3 unique:1 acknowledgment:1 testing:3 practice:1 revue:1 differs:1 backpropagation:1 procedure:1 foundational:1 empirical:8 yan:1 significantly:2 matching:2 statistique:1 radial:1 zoubin:1 altun:1 cannot:1 selection:1 tsochantaridis:1 risk:7 py:3 optimize:2 equivalent:5 measurable:1 lagrangian:4 demonstrated:3 jaime:1 starting:1 independently:1 convex:6 focused:1 survey:1 hsuan:5 roux:1 immediately:1 assigns:1 rule:1 datapoints:1 s6:1 financial:1 searching:1 variation:1 justification:1 annals:1 construction:1 heavily:1 programming:1 diabetes:2 trick:3 bashiri:1 dawid:2 fyi:11 element:1 expensive:1 bottom:1 solved:2 wang:3 worst:3 futureoflife:1 wj:3 sun:1 ordering:1 decrease:1 yk:1 transforming:1 nash:3 complexity:1 ziebart:5 ultimately:1 trained:1 solving:2 segment:2 serve:1 creates:1 efficiency:1 basis:1 edwin:1 easily:2 joint:1 stock:4 derivation:1 train:2 artificial:3 yufeng:1 choosing:3 hsing:1 richer:1 widely:2 valued:1 supplementary:1 rennie:3 kai:1 statistic:5 uic:1 advantage:5 analytical:1 maximal:1 product:1 tu:2 uci:5 combining:1 beneath:1 calhousing:2 flexibility:1 poorly:1 inducing:1 ijcai:1 requirement:2 extending:1 produce:2 generating:1 guaranteeing:1 leave:1 derive:3 develop:1 measured:1 nearest:1 keith:1 eq:8 predicted:2 indicate:1 appropriateness:1 direction:2 differ:1 closely:2 stochastic:5 material:1 explains:1 f1:3 brian:5 extension:3 ic:1 babanezhad:1 equilibrium:4 algorithmic:2 predict:1 alf:1 achieves:1 purpose:1 estimation:4 realizes:1 label:39 sensitive:13 largest:1 wl:1 vice:1 tool:1 weighted:1 minimization:6 mit:1 gaussian:8 rather:5 reaching:1 volker:1 derived:1 focus:1 joachim:2 improvement:1 methodological:1 prevalent:2 rank:1 likelihood:1 contrast:2 adversarial:24 baseline:1 burdensome:1 pyrimidine:3 el:1 typically:2 kernelized:2 overall:3 classification:34 dual:6 among:1 plan:1 constrained:3 special:3 art:3 gramfort:1 equal:5 construct:3 field:1 shaped:1 beach:1 identical:1 represents:1 broad:1 koby:2 icml:1 yu:4 discrepancy:1 future:4 report:1 others:2 minimized:1 piecewise:1 inherent:2 wen:1 employ:5 roman:1 jianlin:1 ve:2 neurocomputing:1 asian:1 qing:1 maxj:1 argmax:1 keerthi:3 versatility:1 possibility:1 investigate:2 mining:1 zheng:2 evaluation:3 ai1:1 extreme:1 primal:4 integral:1 partial:2 necessary:1 conduct:1 sacrificing:1 theoretical:4 miroslav:1 instance:1 column:3 classify:1 tse:1 facet:1 maximization:4 cost:15 deviation:1 entry:1 predictor:12 uniform:1 levin:3 dembczy:1 gr:2 answer:2 nski:2 st:1 international:10 probabilistic:2 pranking:1 together:1 na:2 again:1 thesis:1 aaai:2 choose:1 wan:1 worse:2 creating:1 derivative:1 leading:1 wojciech:1 li:7 account:1 potential:9 de:1 bold:3 ioannis:1 includes:2 dictive:1 satisfy:5 notable:1 ranking:3 depends:1 piece:2 performed:2 view:2 ramaswamy:2 closed:1 jason:1 francis:2 shashua:3 red:1 bayes:5 xing:2 jul:2 contribution:1 minimize:3 il:1 gression:1 accuracy:3 collaborative:1 characteristic:3 efficiently:1 ensemble:1 yield:1 generalize:1 bayesian:2 iid:1 mc:9 cc:1 straight:1 datapoint:1 sixth:1 pollastri:1 involved:1 associated:5 proof:5 costa:1 dataset:8 popular:2 knowledge:4 sophisticated:1 focusing:1 alexandre:1 supervised:3 response:2 wei:4 formulation:9 evaluated:1 though:1 furthermore:1 just:1 implicit:1 stage:9 until:1 sketch:4 hand:1 incrementally:1 defines:1 logistic:1 costsensitive:2 multidisciplinary:1 usa:1 name:3 phillip:1 y2:1 true:8 regularization:3 assigned:1 hence:2 dud:1 sima:1 round:1 adjacent:1 game:16 uniquely:1 abalone:3 hong:1 generalized:1 mohammad:1 demonstrate:2 txi:1 performs:3 fj:12 ranging:1 wise:2 consideration:1 novel:3 superior:1 physical:1 reza:1 extend:1 interpretation:1 slight:3 mae:6 significant:1 versa:1 tuning:1 rd:1 consistency:10 similarly:3 illinois:1 dot:1 calibration:1 han:2 align:1 base:1 multivariate:3 recent:2 chan:1 perspective:6 optimizing:3 certain:2 asif:3 binary:17 arbitrarily:1 life:1 yi:13 tien:5 yasemin:1 minimum:2 greater:1 fortunately:1 ey:4 deng:1 converge:1 maximize:1 redundant:1 wellington:1 nwald:2 full:1 desirable:1 reduces:1 technical:1 match:1 characterized:2 calculation:1 faster:1 long:1 lin:10 cross:1 bach:2 l0xi:6 paired:2 prediction:21 regression:72 iteration:1 kernel:13 agarwal:2 penalize:4 background:1 whereas:1 source:1 posse:1 tiple:1 archive:1 subject:2 member:1 call:2 reflective:5 integer:1 leverage:1 ideal:1 yang:1 split:3 relaxes:1 variety:1 behpour:1 flemming:1 restrict:1 reduce:5 inner:1 idea:1 multiclass:24 whether:1 six:1 amnon:1 bartlett:1 defazio:1 url:1 penalty:5 peter:4 deep:2 useful:3 tewari:1 detailed:1 cardoso:1 induces:1 mathieson:1 shivani:1 schapire:1 http:1 canonical:2 nsf:1 overly:1 popularity:1 discrete:6 key:2 four:2 threshold:40 drawn:1 penalizing:3 pj:1 internationale:1 thresholded:11 krzysztof:1 sum:11 convert:2 uncertainty:2 family:3 wu:1 decision:6 confusing:2 appendix:2 fl:4 layer:1 dash:1 cheng:1 fold:1 quadratic:4 occur:1 ijcnn:1 constraint:3 ri:1 nearby:1 nathan:1 osr:1 min:4 relatively:6 department:1 structured:2 according:4 poor:2 describes:1 making:2 handling:1 thorsten:2 sided:3 taken:1 bing:1 argmaxj:3 singer:4 ordinal:69 end:2 available:1 competitively:4 apply:1 rizal:2 schmidt:2 original:6 thomas:1 assumes:1 top:1 include:2 harish:1 hinge:12 unifying:2 yx:2 yoram:2 multicategory:1 ghahramani:1 establish:3 society:1 seeking:4 objective:3 question:2 added:1 realized:1 font:2 parametric:1 strategy:2 kaiser:3 surrogate:45 said:1 gradient:8 detrimental:1 distance:1 link:2 manifold:1 fy:2 discriminant:1 rfp:1 boldface:2 minimizing:4 equivalently:2 setup:1 robert:1 negative:2 wyi:1 implementation:2 unknown:1 perform:15 twenty:1 ord:4 datasets:14 benchmark:3 finite:2 fabian:1 immediate:7 extended:6 y1:1 arbitrary:2 subtasks:1 rating:1 introduced:1 complement:1 cast:1 pair:6 required:1 david:1 optimized:1 sarkar:1 california:1 narrow:1 nip:1 address:2 adversary:2 kriegel:1 wy:1 kot:1 program:3 ambuj:1 recast:1 max:11 including:2 memory:2 explanation:1 royal:1 regularized:2 indicator:1 minimax:2 technology:1 al0:1 tresp:1 review:2 literature:1 interdependent:1 wisconsin:4 loss:112 highlight:1 limitation:1 srebro:3 versus:6 age:1 penalization:1 validation:2 incurred:2 sufficient:2 consistent:4 ovo:1 autompg:4 thresholding:1 principle:2 bank:3 xiao:1 storing:3 row:3 gianluca:1 summary:2 supported:1 last:2 ovum:2 side:1 drastically:1 bias:1 institute:2 absolute:16 fifth:1 farzan:1 benefit:2 distributed:2 boundary:5 dimension:2 evaluating:1 cumulative:2 contour:1 rich:1 world:1 forward:1 made:2 projected:1 transaction:1 approximate:1 emphasize:1 compact:1 preferred:1 ml:1 conclude:1 sathiya:1 xi:35 continuous:3 search:1 table:12 additionally:1 learn:1 robust:6 ca:1 nicolas:1 ignoring:1 excellent:2 complex:1 shipeng:1 constructing:1 fli:1 significance:2 main:2 owi:1 bounding:1 ling:2 fair:1 explicit:1 exponential:4 xl:1 jmlr:1 anat:1 theorem:7 erroneous:2 specific:2 svm:6 workshop:2 adding:2 phd:1 margin:6 chen:1 farnia:1 boston:4 entropy:5 logarithmic:2 ordered:3 contained:1 bo:1 hua:1 springer:3 clifton:1 corresponds:2 minimizer:4 satisfies:1 acm:2 conditional:6 viewed:3 marked:1 ann:1 shared:2 fisher:9 subhash:1 change:2 mccullagh:1 reducing:1 averaging:1 called:1 duality:2 xin:1 player:2 pedregosa:2 exception:1 aaron:1 support:9 mark:3 kernelbased:1 crammer:4 anoop:1 violated:1 evaluate:2 lian:1 ex:5 |
6,256 | 666 | Transient Signal Detection with Neural Networks:
The Search for the Desired Signal
Jose C. Principe and Abir Zahalka
Computational NeuroEngineering Laboratory
Department of Electrical Engineering
University of Florida, CSE 447
Gainesville, FL 32611
[email protected]
Abstract
Matched filtering has been one of the most powerful techniques
employed for transient detection. Here we will show that a dynamic
neural network outperforms the conventional approach. When the
artificial neural network (ANN) is trained with supervised learning
schemes there is a need to supply the desired signal for all time,
although we are only interested in detecting the transient. In this
paper we also show the effects on the detection agreement of
different strategies to construct the desired signal. The extension of
the Bayes decision rule (011 desired signal), optimal in static
classification, performs worse than desired signals constructed by
random noise or prediction during the background.
1
INTRODUCTION
Detection of poorly defined waveshapes in a nonstationary high noise background is
an important and difficult problem in signal processing. The matched filter is the
optimal linear filter assuming stationary noise [Thomas, 1969]. The application area
that we are going to discuss is epileptic spike detection in the electrocorticogram
(ECoG), where the matched filtering produce poor results [Barlow and Dubinsky,
1976], [Pola and Romagnoli, 1979] due to the variability of the spike shape and the
ever changing background (the brain electric activity). Recently artificial neural
networks have been applied to spike detection [Eberhart et ai, 1989], [Gabor and
688
Transient Signal Detection with Neural Networks: The Search for the Desired Signal
Seydal, 1992], but static neural network architectures were chosen. Here a static
multilayer perceptron (MLP) will be augmented with a short term memory
mechanism to detect spikes. In the way we utilized the dynamic neural net, the ANN
can be thought of as an extension of the matched filter to nonlinear models, which
we will refer to as a neural template matcher. In our implementation, the ANN looks
directly at the input signal with a time window larger than the longest spike for the
sampling frequency utilized. The input layer of the dynamic network is a delay line,
and the data is clocked one sample at a time through the network. The desired signal
is "I" following the occurrence of a spike. With this strategy we teach the ANN to
produce an output of" 1" when a waveform similar to the spike is present within the
time window. A spike will be recognized when the ANN output is above a given
threshold.
Unlike the matched filter, the ANN does not require a single, explicit waveform for
the template (due to the spike shape variability some form of averaging is needed to
create the "average" spike shape which is normally a poor compromise). Rather, the
ANN will learn the important features of the transient class by training on many
sample spikes, refining its approximation with each presentation. Moreover, the
ANN using the sigmoid nonlinearity will have to necessarily represent the
background activity, since the discriminant function in pattern space is established
from information pertaining to all input classes. Therefore, the nonstationary nature
of the background can be accommodated during network training and we can expect
that the performance of the neural template matcher will be improved with respect to
the matched filter.
We will not address here the normalization of the ECoG, nor the issues associated
with the learning criterion [Zahalka, 1992]. The purpose of this paper is to delve on
the design of the desired signal, and quantify the effect on performance. What should
be the shape of the desired sinal for transient detection, when on-line supervised
learning is employed? In our approach we decided to construct a desired signal that
exists for all time. We shall point out that the existence of a desired signal for every
sample will simplify supervised learning, since in principle the conventional
backpropagation algorithm [Rumelhart et aI, 1986] can be utilized instead of the
more time consuming backpropagation through time [Werbos, 1990] or real-time
recurrent learning [Williams and Zipzer, 1989]. The simplified learning algorithm
may very well be one of the factors which will make ANNs learn on-line as adaptive
linear filters do. The main decision regarding the desired signal for spike detection is
to decide the value of the desired signal during the background (which for spikes
represent 99% of the time), since we decided already that it should be "1" following
the spike. Similar problems have been found in speech recognition, when patterns
that exist in time need to be learned [Unnikrishnan et aI, 1991] [Watrous et aI, 1990].
Here we will experimentally compare three desired signals (Figure 1):
Desired signal 1- Extrapolating the Bayes rule for static patterns [Makhoul], we will
create a target of zero during the background, and a value of 1 following the spike ,
with a duration equal to the amount of time the spike is in the input window.
Desired signal 11- During the background, a random, uniformly distributed (between
-0.5 and 0.5), zero mean target signal. Same target as above following the spike .
689
690
Principe and Zahalka
Desired signal 111- During the background the network will be trained as a one step
predictor. Same targe,t as above following the spike.
+~--+--------
o--~I-foI----'desired
output
-lr----------------
(I). Hardlimited 01+ 1 signal
desired output .
\,
tnput
+~---------+~.-~~----
+
+0.
o
-l~-------------------
-1~--------~-------
(II). Random noise signal
(III). Prediction signal
Figure 1. Desired signals considered.
2
NEURAL NETWORK ARCmTECTURE AND TRAINING
One ECoG channel was sampled at 500 Hz (12 bit AID converter) and was preprocessed for normalization between -1 and 0.4, and DC removal before being
presented to the ANN [Zahalka, 1992]. An epileptic spike has a duration between 20
and 70 msec.
The dynamic network used for this application is a time delay neural network
(TDNN) consisting of an input layer with 45 taps, 12 hidden processing elements
(PEs) and one linear output PE. The input window corresponds to 90 msec, so even
the longest spike is fully present in the window during at least 10 samples.
The training set consists of 60 hand picked 2 sec. segments containing one spike
each, embedded in 6,860 points of background activity. In principle the data could be
streamed on-line to the ANN, provided that we could create also on-line the desired
signal. But at this point of the research we preferred to control the segment length
and choose well defined spikes to study training issues. The test data set consists of
another set (belonging to the same individual) of 49 spikes embedded in 6,970
samples. A spike was defined when the ANN output was above 0.9.
The ANN was trained with the backpropagation algorithm [Rumelhart et al 1986].
The weights were updated after every sample (real-time mode). A momentum term
(a.=0.9) was used, and 0.1 was added to the sigmoid derivative to speedup learning
[Fahlman, 1988]. The training was stopped when the test set performance decreased.
The network typically learned in less than 50 presentations of the training set. All the
Transient Signal Detection with Neural Networks: The Search for the Desired Signal
results presented next use the same trainingltest sets, the same learning and stop
criterion and the same network topology.
3
RESULTS
Desired Signal I.
We begin with the most commonly used desired signal for static classification, the
hardlimited 01+ 1 signal of Figure 1(I), but now extended in time (0 during the
background, 1 after the occurrence of the spike). This desired signal has been shown
to be optimal in the sense that it is a least square approximation of the Bayes decision
rule [Makhoul, 1991 J. However it performs poorly for transient detection (73% of
correct detections and 8% of false positives). We suspect that the problem lies in the
different waveshapes present in the background. When an explicit 0 value is given as
the desired signal for the background, the network has difficulties extracting common
features to all the waves and biases the decision rule. Samples of the network input
and the corresponding output are shown in Figure 2(1). Notice that the ANN output
gets close to "1" during high amplitude background, and fails to be above 0.9 during
small amplitude spikes. In Table 1, the test set performance is better than the training
set due to the fact that the spikes chosen for training happen to include more difficult
cases.
Table 1. Performance Results.
Desired
Training Set
Training Set
Test Set
Signal
detections
false positives
detections
Test Set
false positives
0<--> +1
38/60 = 63%
2/38 = 5%
36/49 = 73%
3/36 = 8%
noise
54/60 = 90%
0154 = 0%
46/49 = 94%
1/47 = 2%
prediction
50/60 = 83%
1150 = 2%
45/49 = 92%
2/45 = 4%
Detections mean number of events in agreement between the human expert
and the ANN (normalized by the number of spikes in the set).
False positives mean the number of events picked by the ANN, but not considered
spikes by the human expert (nonnalized by the number of ANN detections)
The "random noise" desired signal is shown in Figure 1(II). This signal consists
simply of uniformly distributed random values bounded between -0.5 and +0.5 for
the nonspike regions and again a value of "+ 1" after the spikes. This approach is
based on the fact that the random number generator will have a zero mean
distribution. Therefore, during training over the nonspike regions, the errors
backpropagated through the network will normally average out to zero, yielding in
practice a "don't care" target for the background. This effectively should give the
least bias to the decision rule . The net result is that consistent training is only
performed over the spike patterns, allowing the network to better learn these patterns
and provide improved performance. The results of this configuration are very
691
692
Principe and Zahalka
promising, as can be seen in Table 1 and in Figure 2(11). The "noise" signal performs
better than all of the other desired signal configurations examined (94% correct
detections and 2% false detections).
Ok
~~~~V-\f"1t
II. 1\
D."
1\
.'
. Input to the network
~jPL~ if;. ZL:~, (
--- 1 0 0 - 0 2
-0.2
a.t~ ~ ~o
Output of network
(I) 0/+1 signal
D?"b
II
~}\fJrrwrifO
A
0 .7 5
0 ."
1\
""./\
_ r
11 "l!1
~~~;~y ~VdWV'V 7~00
o ? "..
A
. mput to the l~~twork
~~.
~
Input for the network.
~.
~~k~jUl~
?
100
200-0
300
.00
:~:~f' h II
o
500
Uutput ot' network
(II) noise desired signal
""
~~~~~ vV~v1M\[0V\J wrtJ' \.ft
Output for the network
(III) prediction signal
Figure 2. Input and output of neural network
The last signal configuration is the prediction paradigm. The network is performing
one-step prediction during the nonspike portions of the signal and a saturation value
of "+ 1" is added to the desired signal after the spike, as in Figure 1(111). The rationale
behind such a configuration is that the desired signal is readily available and may
require fewer network resources than a target of "0", decreasing the bias in the
decision rule for the spike class. The results are given in Table 1, where we see a
marked improvement over the hardlimited desired signal (92% correct detection and
4% false positives).
4
COMPARISON WITH MATCHED FILTER
The template for the matched filter was formed by averaging five spikes from the
training set. While averaging reduces morphological "crispness," it enhances the
robustness of spike shape representation, since the spike shape is variable. The 45
point template is correlated continuously with the time signal from the test set, hence
Transient Signal Detection with Neural Networks: The Search for the Desired Signal
no correction to the nonwhite nature of the ECoG is being done. Performance in both
the matched filter and the ANN approach will be dependent upon the threshold levels
chosen for detection, which determines the receiver operating characteristic (ROC)
for the detector, but requires large data sets to be meaningful. Table 2 shows a partial
result, wht;~;,! bvth detectors are compared in terms of false positives for 90%
detection rate and detection rate for 2 false positives. The results presented for the
ANN are for the random noise desired signal. In both cases, we see the superiority of
the ANN approach.
Table 2. Comparison ANNlMatched filter
90% detection rate
NN
ofalse +'s
5
MF
21 false +'s
2 false positives
NN
MF
92%
13%
CONCLUSIONS
The ultimate conclusion of this experimental work is that neural networks can be
used to implement transient detectors, outperforming the conventional matched filter
for the detection of spike transients embedded in the ECoG. However, the difficulty
in the ANN approach comes in how to setup the training, and the desired signal.
Although the training was not discussed here, we would like to point several issues
that are important. When a cost function that is sensitive to the a priori probability
of the classes is used (as the error signal in backpropagation) the proper balance of
background versus spike data sizes is relevant for adequate training. Our results
show that at a low spike concentration of 4% (ratio of spike samples over background
samples in the training set), the network never learns the waveform and always
outputs the value chosen to represent the background. For our application, these
problems disappear at an 18% concentration level.
Another important issue is the selection of the class exemplars for training. Transient
detection is a one class classification problem (A versus the universe), rather than a
two class classification problem. It is not possible to cover appropriately the
unconstrained background with examples, and even when a large number of
background waves is utilized (in our case 80% of the training samples belonged to
the background) the training produced bad results. We found out that only the waves
similar in shape to the spikes are important to bound the spike class in feature space.
As a solution, we included in the training set the false positives, i.e. waves that the
system detects as spikes but the human expert classifies as background, to improve
the detection agreement with the human [Zahalka, 1992].
The issues regarding the choice of the desired signal are also important. We found
experimentally that the best desired signal for our case is the one that uses random
noise during the background. We can not, at this point, explain this result
693
694
Principe and Zahalka
theoretically. This desired signal is also the one that uses fewer network resources (i.
e. number of hidden units) without degrading the performance [Zahalka, 1992]. This
result came as a surprise since the 0/1 signal has been shown to be optimal for static
patterns in the sense that makes the classifier approximate the Bayes decision rule.
The explanation may be simply a matter of local minima in the performance surface
of the network trained with the 0/1 desired signal, but it may also reflect deeper
causes. With the OIl desired signal, the network weights are updated equally with the
information of the background waveforms and with the information regarding the
spikes. The training paradigm should emphasize the spikes, since the spikes are the
waves we are interested in. Moreover, since the background class is unconstrained,
many more degrees of freedom are necessary to represent well the background. When
not enough hidden units are available, the network biases the discriminant function,
and the performance is poor. In the prediction paradigm the background is selected
as the desired signal and the required number of hidden nodes to predict the next
sample is much smaller (actually only three hidden nodes are sufficient to keep the
reported performance [Zahalka, 1992]). The network resources naturally self
organize as two template matching nodes and one prediction node. However, we have
found out that the signal has to be properly normalized in the sense that the target for
the spike ("I") must be outside the range of the input signal voltages (that is the
reason we normalize the input signal between -1 and 0.4). From a classification point
of view we "do not care" what the net output is provided it is far from the target of
"1". The random noise target with zero mean achieves this goal very easily, because
the error gradient during the background averages out to zero. Therefore the weights
reflect primarily the information containing in the shape of the spike. We found out
that only two hidden nodes are sufficient to keep the reported performance. [Zahalka,
1992]. It seems that the theory of time varying signal classification with neural
networks is not a straight forward extension of the static classification case, and
requires further theoretical analysis.
An implication of this work regards the role of supervised learning in biological
neural networks. This work shows that during non-interesting events, there is no need
to provide a target to the neural assembly (noise, which is so readily available in
biological systems, suffices). Re-enforcement stimulus are only needed during or
after relevant events, which is compatible with the information processing models of
the olfactory bulb [Freeman and DiPrisco, 1989]. Therefore, looking at learning and
adaptation in biological systems as a signal detection instead of a classification
problem seems promising.
Acknowledgments
This work has been partially supported by NSF grants ECS-9208789 and DDM8914084.
References
Barlow J. S. and Dubinsky J., (1976) "Some Computer Approaches to Continuous
Transient Signal Detection with Neural Networks: The Search for the Desired Signal
695
Automatic Clinical EEG Monitoring," in Quantitative Analytic Studies in Epilepsy, Raven Press,
New York, 309-327.
Eberhart R., Dobbins R., Weber W., (1989) "Casenet: a neural network tool for EEG waveform
classifiaction", Proc IEEE Symp. Compo Based Medical Systems, Minneapolis, 60-68, 1989.
Fahlman S.,(1988) "Faster learning variations on backpropagation: an empirical study", Proc.
1988 Connectionist Summer School, Morgan Kaufmann, 38-51.
Freeman W., DiPrisco V., (1986) "EEG spatial pattern differences with discriminated odors
manifest chaotic and limit cycle attractors in olfactory bulb of rabbits", in Brain Theory, Ed. Palm
and Aertsen, Springer, 97-120.
Gabor A., Seydal M., (1992) "Automated interictal EEG spike detection using artifical neural
networks", Electroenc. Clin. Neurophysiol., (83),271-280.
Makhoul J., (1991) "Pattern recognition properties of neural networks" ,Proc. 1991 IEEE
Workshop Neural Net. in Sig. Proc., 173-187, Princeton.
Pola P. and Romagnoly 0., (1979) "Automatic analysis of interictal epileptic activity related to its
morphological aspects", Electroenceph. Clin. Neurophysiol., #46, 227-231.
Rumelhart,D.E., Hinton,G.E. and Williams,R.J. (1986) "Learning internal representations by
error propagation. in Parallel Distributed Processing (Rumelhart, McClelland, eds.), ch. 8,
Cambridge, MA.
Thomas J., (1969) "An Introduction to statistical Communication Theory", Wiley.
Watrous R., Ladendorf B., Kuhn G., (1990) "Complete gradient optimization of a recurrent
network applied to b,d,g discrimination", 1. Acoust. Soc. Am. 87 (3), 1301-1309.
Werbos, P.J. (1990) "Backpropagation through time: what it does and how to do it", Proc. IEEE,
vol 78, nolO, 1550-1560.
Williams,R.J. and Zipser, D. (1989) "A learning algorithm for continually running fully recurrent
neural networks. in Neural Computation, vol. 1 (2).
Unikrishnan K., Hopfield J., Tank D., (1991) "Connected-Digit Speaker-dependent speech
recognition using a neural network with time delayed connections", IEEE Trans. Sig Proc., vol
39, #3, 698-713.
Zalahka A., (1992) "Signal detection with neural networks: an application to the recognition of
epileptic spikes", Master Thesis, University of FLorida.
| 666 |@word seems:2 gainesville:1 configuration:4 outperforms:1 must:1 readily:2 happen:1 shape:8 analytic:1 extrapolating:1 discrimination:1 stationary:1 fewer:2 selected:1 short:1 lr:1 compo:1 detecting:1 node:5 cse:1 five:1 ladendorf:1 constructed:1 supply:1 consists:3 symp:1 olfactory:2 theoretically:1 nor:1 brain:2 freeman:2 detects:1 decreasing:1 window:5 provided:2 begin:1 matched:10 moreover:2 bounded:1 classifies:1 what:3 watrous:2 degrading:1 acoust:1 quantitative:1 every:2 classifier:1 control:1 normally:2 grant:1 zl:1 superiority:1 unit:2 organize:1 continually:1 before:1 positive:9 engineering:1 local:1 limit:1 examined:1 delve:1 range:1 minneapolis:1 decided:2 acknowledgment:1 practice:1 implement:1 backpropagation:6 chaotic:1 digit:1 area:1 empirical:1 gabor:2 thought:1 matching:1 get:1 close:1 selection:1 conventional:3 williams:3 duration:2 rabbit:1 rule:7 variation:1 updated:2 target:9 dobbin:1 us:2 sig:2 agreement:3 element:1 rumelhart:4 recognition:4 utilized:4 werbos:2 ft:1 role:1 electrical:1 region:2 cycle:1 connected:1 morphological:2 dynamic:4 trained:4 segment:2 compromise:1 upon:1 neurophysiol:2 easily:1 hopfield:1 pertaining:1 artificial:2 outside:1 larger:1 net:4 adaptation:1 relevant:2 poorly:2 wht:1 normalize:1 eberhart:2 produce:2 recurrent:3 exemplar:1 school:1 soc:1 come:1 quantify:1 kuhn:1 waveform:5 correct:3 filter:11 human:4 transient:13 require:2 v1m:1 suffices:1 biological:3 neuroengineering:1 extension:3 ecog:5 zipzer:1 correction:1 considered:2 predict:1 achieves:1 purpose:1 proc:6 sensitive:1 create:3 tool:1 always:1 rather:2 varying:1 voltage:1 unnikrishnan:1 refining:1 longest:2 improvement:1 properly:1 detect:1 sense:3 am:1 dependent:2 nn:2 typically:1 hidden:6 going:1 interested:2 tank:1 issue:5 classification:8 priori:1 spatial:1 equal:1 construct:2 never:1 sampling:1 look:1 connectionist:1 stimulus:1 simplify:1 primarily:1 individual:1 delayed:1 consisting:1 attractor:1 freedom:1 detection:31 mlp:1 yielding:1 behind:1 implication:1 partial:1 necessary:1 accommodated:1 desired:43 re:1 theoretical:1 stopped:1 cover:1 cost:1 predictor:1 delay:2 reported:2 continuously:1 again:1 reflect:2 thesis:1 containing:2 choose:1 worse:1 expert:3 derivative:1 sec:1 matter:1 performed:1 view:1 picked:2 portion:1 wave:5 bayes:4 parallel:1 jul:1 square:1 formed:1 kaufmann:1 characteristic:1 foi:1 produced:1 monitoring:1 straight:1 anns:1 detector:3 explain:1 ed:2 frequency:1 naturally:1 associated:1 static:7 sampled:1 stop:1 crispness:1 manifest:1 amplitude:2 actually:1 ok:1 supervised:4 improved:2 synapse:1 done:1 hand:1 nonlinear:1 propagation:1 mode:1 oil:1 effect:2 normalized:2 barlow:2 hence:1 laboratory:1 during:16 self:1 speaker:1 clocked:1 criterion:2 abir:1 complete:1 performs:3 weber:1 recently:1 sigmoid:2 common:1 discriminated:1 discussed:1 epilepsy:1 refer:1 cambridge:1 ai:4 automatic:2 unconstrained:2 nonlinearity:1 operating:1 surface:1 outperforming:1 came:1 seen:1 minimum:1 morgan:1 care:2 employed:2 recognized:1 paradigm:3 signal:67 ii:6 reduces:1 faster:1 clinical:1 equally:1 prediction:8 multilayer:1 represent:4 normalization:2 background:28 twork:1 decreased:1 appropriately:1 ot:1 unlike:1 hz:1 ufl:1 suspect:1 electrocorticogram:1 nonstationary:2 ee:1 extracting:1 zipser:1 iii:2 enough:1 automated:1 architecture:1 converter:1 topology:1 regarding:3 epileptic:4 ultimate:1 speech:2 york:1 cause:1 adequate:1 amount:1 backpropagated:1 mcclelland:1 exist:1 nsf:1 notice:1 medical:1 shall:1 vol:3 threshold:2 changing:1 preprocessed:1 jose:1 powerful:1 master:1 decide:1 decision:7 bit:1 fl:1 layer:2 bound:1 summer:1 interictal:2 activity:4 aspect:1 performing:1 speedup:1 department:1 palm:1 poor:3 belonging:1 makhoul:3 smaller:1 resource:3 discus:1 mechanism:1 needed:2 enforcement:1 available:3 occurrence:2 robustness:1 odor:1 florida:2 existence:1 thomas:2 running:1 include:1 assembly:1 clin:2 disappear:1 streamed:1 already:1 added:2 spike:50 strategy:2 concentration:2 aertsen:1 enhances:1 gradient:2 discriminant:2 reason:1 assuming:1 length:1 ratio:1 balance:1 difficult:2 setup:1 teach:1 implementation:1 design:1 proper:1 allowing:1 extended:1 variability:2 ever:1 nonnalized:1 dc:1 looking:1 hinton:1 communication:1 required:1 connection:1 tap:1 learned:2 established:1 trans:1 address:1 pattern:8 belonged:1 saturation:1 memory:1 explanation:1 event:4 difficulty:2 scheme:1 improve:1 tdnn:1 removal:1 embedded:3 fully:2 expect:1 rationale:1 interesting:1 filtering:2 versus:2 generator:1 degree:1 bulb:2 sufficient:2 consistent:1 principle:2 compatible:1 supported:1 fahlman:2 last:1 bias:4 vv:1 perceptron:1 deeper:1 template:6 distributed:3 regard:1 electroenceph:1 forward:1 commonly:1 adaptive:1 simplified:1 far:1 ec:1 arcmtecture:1 approximate:1 emphasize:1 preferred:1 keep:2 receiver:1 consuming:1 don:1 search:5 continuous:1 table:6 promising:2 learn:3 nature:2 channel:1 eeg:4 necessarily:1 electric:1 main:1 universe:1 noise:12 augmented:1 roc:1 aid:1 wiley:1 fails:1 momentum:1 explicit:2 msec:2 lie:1 pe:2 learns:1 bad:1 jpl:1 exists:1 raven:1 workshop:1 false:11 effectively:1 mf:2 surprise:1 simply:2 partially:1 springer:1 ch:1 corresponds:1 determines:1 ma:1 marked:1 presentation:2 goal:1 ann:20 pola:2 experimentally:2 included:1 uniformly:2 averaging:3 experimental:1 matcher:2 meaningful:1 principe:5 internal:1 mput:1 artifical:1 princeton:1 correlated:1 |
6,257 | 6,660 | Hypothesis Transfer Learning via
Transformation Functions
Simon S. Du
Carnegie Mellon University
[email protected]
Jayanth Koushik
Carnegie Mellon University
[email protected]
Barnab?s P?czos
Carnegie Mellon University
[email protected]
Aarti Singh
Carnegie Mellon University
[email protected]
Abstract
We consider the Hypothesis Transfer Learning (HTL) problem where one incorporates a hypothesis trained on the source domain into the learning procedure
of the target domain. Existing theoretical analysis either only studies speci?c
algorithms or only presents upper bounds on the generalization error but not on the
excess risk. In this paper, we propose a uni?ed algorithm-dependent framework
for HTL through a novel notion of transformation function, which characterizes
the relation between the source and the target domains. We conduct a general risk
analysis of this framework and in particular, we show for the ?rst time, if two
domains are related, HTL enjoys faster convergence rates of excess risks for Kernel
Smoothing and Kernel Ridge Regression than those of the classical non-transfer
learning settings. Experiments on real world data demonstrate the effectiveness of
our framework.
1
Introduction
In a classical transfer learning setting, we have a large amount of data from a source domain and
a relatively small amount of data from a target domain. These two domains are related but not
necessarily identical, and the usual assumption is that the hypothesis learned from the source domain
is useful in the learning task of the target domain.
In this paper, we focus on the regression problem where the functions we want to estimate of the source
and the target domains are different but related. Figure 1a shows a 1D toy example of this setting,
where the source function is f so (x) = sin(4?x) and the target function is f ta (x) = sin(4?x) + 4?x.
Many real world problems can be formulated as transfer learning problems. For example, in the task
of predicting the reaction time of an individual from his/her fMRI images, we have about 30 subjects
but each subject has only about 100 data points. To learn the mapping from neural images to the
reaction time of a speci?c subject, we can treat all but this subject as the source domain, and this
subject as the target domain. In Section 6, we show how our proposed method helps us learn this
mapping more accurately.
This paradigm, hypothesis transfer learning (HTL) has been explored empirically with success in
many applications [Fei-Fei et al., 2006, Yang et al., 2007, Orabona et al., 2009, Tommasi et al.,
2010, Kuzborskij et al., 2013, Wang and Schneider, 2014]. Kuzborskij and Orabona [2013, 2016]
pioneered the theoretical analysis of HTL for linear regression and recently Wang and Schneider
[2015] analyzed Kernel Ridge Regression. However, most existing works only provide generalization
bounds, i.e. the difference between the true risk and the training error or the leave-one-out error.
These analyses are not complete because minimizing the generalization error does not necessarily
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
12
10
6
1
0.5
2
Offset KS
Only Target KS
fta
Scale KS
Only Target KS
fta
1
0
Y
Y
8
3
fso
Source data
ta
f
Target Data
Y
14
4
?1
0
2
?2
0
?2
0
0.2
0.4
0.6
0.8
1
?0.5
0
0.1
0.2
X
X
(a) A toy example of transfer
learning. We have many more
samples from the source domain
than the target domain.
0.3
0.4
(b) Transfer learning with Offset
Transformation.
?3
0
0.1
0.2
X
0.3
0.4
(c) Transfer learning with Scale
Transformation.
Figure 1: Experimental results on synthetic data.
reduce the true risk. Further, these works often rely on a particular form of transformation from the
source domain to the target domain. For example, Wang and Schneider [2015] studied the offset
transformation that instead of estimating the target domain function directly, they learn the residual
between the target domain function and the source domain function. It is natural to ask what if we
use other transfer functions and how it affects the risk on the target domain.
In this paper, we propose a general framework of HTL. Instead of analyzing a speci?c form of
transfer, we treat it as an input of our learning algorithm. We call this input transformation function
since intuitively, it captures the relevance between these two domains.1 This framework uni?es many
previous works Wang and Schneider [2014], Kuzborskij and Orabona [2013], Wang et al. [2016] and
naturally induces a class of new learning procedures.
Theoretically, we develop excess risk analysis for this framework. The performance depends on the
stability [Bousquet and Elisseeff, 2002] of the algorithm used as a subroutine that if the algorithm
is stable then the estimation error in the source domain will not affect the estimation in the target
domain much. To our knowledge, this connection was ?rst established by Kuzborskij et al. [2013] in
the linear regression setting but here we generalize it to a broader context. In particular, we provide
explicit risk bounds for two widely used nonlinear estimators, Kernel Smoothing (KS) estimators
and Kernel Ridge Regression (KRR) as subroutines. To the best of our knowledge, these are the ?rst
results showing when two domains are related, transfer learning techniques have faster statistical
convergence rate of excess risk than that of non-transfer learning of kernel based methods. Further,
we accompany this framework with a theoretical analysis showing a small amount of data for crossvalidation enables us (1) avoid using HTL when it is not useful and (2) choose the best transformation
function as input from a large pool.
The rest of the paper is organized as follows. In Section 2 we introduce HTL and provide necessary
backgrounds for KS and KRR. We formalize our transformation function based framework in
Section 3. Our main theoretical results are in Section 4 and speci?cally in Section 4.1 and Section 4.2
we provide explicit risk bounds for KS and KRR, respectively. In Section 5 we analyze crossvalidation in HTL setting and in Section 6 we conduct experiments on real world data data. We
conclude with a brief discussion of avenues for future work.
2
Preliminaries
2.1
Problem Setup
In this paper, we assume both X ? Rd and Y ? R lie in compact subsets: ||X||2 ? ?X , |Y | ? ?Y
n
for some ?X , ?Y ? R+ . Throughout the paper, we use T = {(Xi , Yi )}i=1 to denote a set of
so
so
ta
ta
samples. Let (X , Y ) be the sample from the source domain, and (X , Y ) the sample from the
target domain. In our setting, there are nso samples drawn i.i.d from the source distribution: T so =
nta
nso
{(Xiso , Yiso )}i=1
, and nta samples drawn i.i.d from the target distribution: T ta = {(Xita , Yita )}i=1
.
In addition, we also use nval samples drawn i.i.d from the target domain for cross-validation. We
model the joint relation between X and Y by: Y so = f so (X so ) + ?so and Y ta = f ta (X ta ) + ?ta
where f so and f ta are regression functions and we assume the noise E [?so ] = E [?ta ] = 0, i.i.d,
1
We formally de?ne the transformation functions in Section 3.
2
and bounded. We use A : T ? f? to denote an algorithm that takes a set of samples
?? and produce
?2 ?
an estimator. Given an estimator f?, we de?ne the integrated L2 risk as R(f?) = E f?(X) ? Y
where the expectation is taken over the distribution of (X, Y ). Similarly, the empirical L2 risk on
?2
?
? f?) = 1 ?n
a set of sample T is de?ned as R(
Yi ? f? (Xi ) . In HTL setting, we use f?so an
n
i=1
estimator from the source domain to facilitate the learning procedure for f ta .
2.2
Kernel Smoothing
We say a function f is in the (?, ?) H?lder class [Wasserman, 2006], if for any x, x? ? Rd , f satis?es
?
|f (x) ? f (x? )| ? ? ||x ? x? ||2 , for some ? ? (0, 1). The kernel smoothing
? method uses a positive
kernel K on [0, 1], highest at 0, decreasing on [0, 1], 0 outside [0, 1], and Rd u2 K(u) < ?. Using
?n
n
T = {(Xi , Yi )}i=1 , the kernel smoothing estimator is de?ned as follows: f?(x) = i=1 wi (x)Yi ,
i ||/h)
where wi (x) = ?nK(||x?X
K(||x?Xj ||/h) ? [0, 1].
j=1
2.3
Kernel Ridge Regression
Another popular non-linear estimator is the kernel ridge regression (KRR) which uses the theory
of reproducing kernel Hilbert space (RKHS) for regression [Vovk, 2013]. Any symmetric positive
semide?nite kernel function K : Rd ? Rd ? R de?nes a RKHS H. For each x ? Rd , the function
z ? K(z, x) is contained in the Hilbert space H; moreover, the Hilbert space is endowed with
an inner product ??, ??H such that K(?, x) acts as the kernel of the evaluation functional, meaning
?f, K(x, ?)?H = f (x) for f ? H. In this paper we assume K is bounded: supx?Rd K (x, ?
x) = k <
?. Given the inner product, the H norm of a function g ? H is de?ned as ||g||H ? ?g, g?H
??
?1/2
and similarly the L2 norm, ||g||2 ? Rd g(x)2 dPX
for a given
? PX . Also, the kernel induces
an integral operator TK : L2 (PX ) ? L2 (PX ): TK [f ] (x) = Rd K (x? , x) f (x? ) dPx (x? ) with
countably many non-zero eigenvalues: {?i }i?1 . For a given function f , the approximation error is
?
?
2
2
de?ned as: Af (?) ? inf h?H ||h ? f ||L2 (PX ) + ? ||h||H for ? ? 0. Finally the estimated function
?1
evaluated at point x can be written as f? (x) = K(X, x) (K(X, X) + n?I) Y where X ? Rn?d
n?1
are the inputs of training samples and Y ? R
are the training labels Vovk [2013].
2.4
Related work
Before we present our framework, it is helpful to give a brief overview of existing literature on
theoretical analysis of transfer learning. Many previous works focused on the settings when only
unlabeled data from the target domain are available [Huang et al., 2006, Sugiyama et al., 2008, Yu
and Szepesv?ri, 2012]. In particular, a line of research has been established based on distribution
discrepancy, a loss induced metric for the source and target distributions [Mansour et al., 2009,
Ben-David et al., 2007, Blitzer et al., 2008, Cortes and Mohri, 2011, Mohri and Medina, 2012]. For
example, recently Cortes and Mohri [2014] gave generalization bounds for kernel based methods
under convex loss in terms of discrepancy.
In many real world applications such as yield prediction from pictures [Nuske et al., 2014], or
prediction of response time from fMRI [Verstynen, 2014], some labeled data from the target domain
is also available. Cortes et al. [2015] used these data to improve their discrepancy minimization
algorithm. Zhang et al. [2013] focused on modeling target shift (P (Y ) changes), conditional shift
(P (X|Y ) changes), and a combination of both. Recently, Wang and Schneider [2014] proposed a
kernel mean embedding method to match the conditional probability in the kernel space and later
derived generalization bound for this problem Wang and Schneider [2015]. Kuzborskij and Orabona
[2013, 2016], Kuzborskij et al. [2016] gave excess risk bounds for target domain estimator in the form
of a linear combination of estimators from multiple source domains and an additional linear function.
Ben-David and Urner [2013] showed a similar bound of the same setting with different quantities
capturing the relatedness. Wang et al. [2016] showed that if the features of source and target domain
are [0, 1]d , using orthonormal basis function estimator, transfer learning achieves better excess risk
3
guarantee if f ta ? f so can be approximated by the basis functions easier than f ta . Their work can be
viewed as a special case of our framework using the transformation function G(a, b) = a + b.
3
Transformation Functions
In this section, we ?rst de?ne our class of models and give a meta-algorithm to learn the target
regression function. Our models are based on the idea that transfer learning is helpful when one
transforms the target domain regression problem into a simpler regression problem using source
domain knowledge. Consider the following example.
?
?
?
2.1?
and f ta (x) = f so (x) + x. f so
Example: Offset Transfer. Let f so (x) = x (1 ? x) sin x+0.05
is the so called Doppler function. It requires a large number of samples to estimate well because of its
lack of smoothness Wasserman [2006]. For the same reason, f ta is also dif?cult to estimate directly.
However, if we have enough data from the source domain, we can have a fairly good estimate of f so .
Further, notice that the offset function w(x) = f ta (x) ? f so (x) = x, is just a linear function. Thus,
instead of directly using T ta to estimate f ta , we can use the target domain samples to ?nd an estimate
of w(x), denoted by w(x),
?
and our estimator for the target domain is just: f?ta (x) = f?so (x) + w(x).
?
Figure 1b shows this technique gives improved ?tting for f ta .
The previous example exploits the fact that function w(x) = f ta (x) ? f so (x) is a simpler function
than f ta . Now we generalize this idea further. Formally, we de?ne the transformation function as
G(a, b) : R2 ? R where we assume that given a ? R, G(a, ?) is invertible. Here a will be the
regression function of the source domain evaluated at some point and the output of G will be the
regression function of?the target domain
evaluated at the same point. Let G?1
a (?) denote the inverse
?
of G(a, ?) such that G a, G?1
(c)
=
c.
For
example
if
G(a,
b)
=
a
+
b
and
G?1
a
a (c) = c ? a. For a
?1
so
ta
given G and a pair (f , f ), they together induce a function wG (x) = Gf so (x) (f ta (x)). In the offset
transfer example, wG (x) = x. By this de?nition, for any x, we have G (f so (x) , wG (x)) = f ta (x) .
We call wG the auxiliary function of the transformation function G. In the HTL setting, G is a userde?ned transformation that represents users? prior knowledge on the relation between the source and
target domains. Now we list some other examples:
Example: Scale-Transfer. Consider G(a, b) = ab. This transformation function is useful when
f so and f ta satisfy a smooth scale transfer. For example, if f ta = cf so , for some constant c, then
wG (x) = c because f ta (x) = G (f so (x) , wG (x)) = f so (x) wG (x) = f so (x) c. See Figure 1c.
Example: Non-Transfer. Consider G(a, b) = b. Notice that f ta (x) = wG (x) and so f so is
irrelevant. Thus this model is equivalent to traditional regression on the target domain since data from
the source domain does not help.
3.1
A Meta Algorithm
Given the transformation G and data, we provide a general procedure to estimate f ta . The spirit of
the algorithm is turning learning a complex function f ta into an easier function wG . First we use an
algorithm Aso that takes T so to obtain f?so . Since we have suf?cient data from the source domain,
. Second, we? construct a new???
data set using
f?so should be close to the true regression function f so??
nta
wG
ta
so
ta
ta
?
the nta data points from the target domain: T
=
X , HG f (X ) , Y
where
i
HG : R2 ? R and satis?es
?
?
?
?
??
E HG f so Xita , Yita = G?1
f so
(Xita )
?
i
i
i=1
?
??
?
?
f ta Xita = wG Xita
where and the expectation is taken over ?ta . ?Thus, ?we can use these newly constructed data to
?G = AWG T WG . Finally, we plug trained f?so and w
?G into
learn wG with algorithm AWG : w
?G (X)). Pseudocode is
transformation G to obtain an estimation for f ta : f?ta (X) = G(f?so (X) , w
shown in Algorithm 1.
Unbiased Estimator HG (f so (X ta ) , Y ta ): In Algorithm 1, we require an unbiased estimator for
so
wG (X ta ). Note that if G (a, b) is linear b or ?ta = 0, we can simply set
? HG (f (X) , Y? ) =
G?1
f so (X) (Y ).
For other scenarios, G?1
f so
(Y ta ) is biased: E G?1
(Yita )
f so (Xita )
(Xita ) i
ta
G?1
(x)) and we need to design estimator using the structure of G.
f so (x) (f
4
?=
Algorithm 1 Transformation Function based Transfer Learning
so
Inputs: Source domain data: T so = {(Xiso , Yiso )}ni=1
, target domain data: T ta =
ta
ta nta
{(Xi , Yi )}i=1 , transformation function: G, algorithm to train f so : Aso , algorithm to train
wG : AwG and HG an unbiased estimator for estimating wG .
Outputs: Regression function for the target domain: f?ta .
1: Train the source domain regression function f?so = Aso (T so ).
nta
2: Construct new data using f?so and T ta : T wG = {(Xita , Wi )}i=1
, where Wi =
?
?
so
ta
ta
?
HG f (X ) , Y
.
i
i
3: Train the auxiliary function: w
?G = AWG (T wG ).
?
?
4: Output the estimated regression for the target domain: f?ta (X) = G f?so (X), w
?G (X) .
Remark 1: Many transformation functions are equivalent to a transformation function G? (a, b)
2
where G? (a, b) is linear in b. For example, for G (a, b) = ab2 , i.e., f ta (x) = f so (x) wG
(x),
?
?
2
ta
so
?
consider G (a, b) = ab where b in G stands for b in G, i.e., f (x) = f (x) wG (x). Therefore
?
2
?
wG
= wG
and we only need to estimate wG
well instead of estimating wG . More generally, if
G (a, b) can be factorized as G (a, b) = g1 (a) g2 (b), i.e., f ta (x) = g1 (f so (x)) g2 (wG (x)), we
only need to estimate g2 (wG (x)) and the convergence rate depends on the structure of g2 (wG (x)).
Remark 2: When G is not linear in b and ?ta ?= 0, observe that in Algorithm 1, we treat Yita s as
noisy covariates to estimate wG (Xi )s. This problem is called error-in-variable or measurement error
and has been widely studied in statistics literature. For details, we refer the reader to the seminal
work by Carroll et al. [2006]. There is no universal estimator for the measurement error problem. In
Section B, we provide a common technique, regression calibration to deal with measurement error
problem.
4
Excess Risk Analyses
In this section, we present theoretical analyses for the proposed class of models and estimators. First,
we need to impose some conditions on G. The ?rst assures that if the estimations of f so and wG are
close to the source regression and auxiliary function, then our estimator for f ta is close to the true
target regression function. The second assures that we are estimating a regular function.
Assumption 1 G (a, b) is L-Lipschitz: |G(a, b) ? G(a? , b? )| ? L ||(a, b) ? (a? , b? )||2 and is invertible with respect to b given a, i.e. if G (x, y) = z then G?1
x (z) = y.
Assumption 2 Given G, the induced auxiliary function wG is bounded: for x : ||x||2 ? ?X ,
wG (x) ? B for some B > 0.
Offset Transfer and Non-Transfer satisfy these conditions with L = 1 and B = ?Y . Scale Transfer
satis?es these assumptions when f so is lower bounded from away 0. Lastly, we assume our unbiased
estimator is also regular.
Assumption 3 For x : ||x||2 ? ?X and y : |y| ? ?Y , HG (x, y) ? B for some B > 0 and HG is
Lipschitz continuous in the ?rst argument:|HG (x, y) ? HG (x? , y)| ? L |x ? x? | for some L > 0.
We begin with a general result which only requires the stability of AWG :
of samples that have same features but different labels: T =
Theorem 1 Suppose for any
??two sets ??
nta
nta
ta
ta ?
?
{(X , Wi )}
and T =
X , Wi
, the algorithm Aw for training wG satis?es:
i
i=1
i
G
i=1
??
? ???
??
??
??AwG (T ) ? AwG T? ??
?
?
5
nta
?
i=1
?
?
? ??
?i ?? ,
ci Xita ?Wi ? W
(1)
where ci only depends on Xita . Then for any x,
??
?2
?2
?
?
?
?
? ?ta
2
ta
?G (x) ? wG (x)| +
?f (x) ? f (x)? =O ?f?so (x) ? f so (x)? + |w
? ?
?n
ta
? 2
?
? ta ? ?? so ? ta ?
?
?
?
ci Xi ?f? Xi ? f so Xita ? ?
i=1
?
nta ?
where w
?G = AwG {(Xita , HG (f so (Xita ) , Yita ))}i=1
, the estimated auxiliary function trained
based on true source domain regression function.
Theorem 1 shows how the estimation error in the source domain function propagates to our estimation
of ?the target domain function.
Notice that if we happen to know f so , then the error is bounded by
?
2
O |w
?G (x) ? wG (x)| , the estimation error of wG . However, since we are using estimated f so to
construct training samples for wG , the error might accumulate as nta increases. Though the third
term in Theorem 1 might increase with nta , it also depends on the estimation error of f so which is
relatively small because of the large amount of source domain data.
The stability condition (1) we used is related to the uniform stability introduced by Bousquet and
Elisseeff Bousquet and Elisseeff [2002] where they consider how much will the output change if one
of the training instance is removed or replaced by another whereas ours depends on two different
training data sets. The connection between transfer learning and stability has been discovered
by Kuzborskij and Orabona [2013], Liu et al. [2016] and Zhang [2015] in different settings, but they
only showed bounds for generalization, not for excess risk.
4.1
Kernel Smoothing
We ?rst analyze kernel smoothing method.
Theorem 2 Suppose the support of X ta is a subset of the support of X so and the probability density
of PX so and PX ta are uniformly bounded away from below on their supports. Further assume f so is
(?so , ?so ) H?lder and wG is (?wG , ?wG ) H?lder . If we use kernel smoothing estimation for f so and
?1/(2?wG +d)
?1/(2?so +d)
wG with bandwidth hso ? nso
and hwG ? nta
, with probability at least 1 ? ?
the risk satis?es:
?
?
? ?
?2?w
? ? ??
G
?2?so
? ta ?
1
2?w +d
2?
+d
ta
G
so
log
+ nta
E R f?
?R f
= O nso
?
where the expectation is taken over T so and T ta .
Theorem 2 suggests that the risk depends on two sources, one from estimation of f so and one from
estimation of wG . For the ?rst term, since in the typical transfer learning scenarios nso >> nta , it is
relatively small in the setting we focus on. The second terms shows the power of transfer learning on
transforming a possibly complex target regression function into a simpler auxiliary
function. It ?
is
?
?2?
/
2?
+d
ta
ta
(
)
f
f
well known that learning f ta only using target domain has risk of the order ? nta
.
Thus, if the auxiliary function is smoother than the target regression function, i.e. ?wG > ?f ta , we
obtain better statistical rate.
4.2
Kernel Ridge Regression
Next, we give an upper bound for the excess risk using KRR:
Theorem 3 Suppose PX so = PX ta and the eigenvalues of the integral operator TK satisfy ?i ?
ai?1/p for
i ? 1 a ? 16?4Y , p ? (0, 1) and there exists a constant C ? 1 such that for f ? H,
so
p
1?p
||f ||? ? C ||f ||H ? ||f ||L2 (PX ) . Furthur assume that Af (?) ? c??so and AwG (?) ? c??wG .
?1/(?so +p)
If we use KRR for estimating f so and wG with regularization parameters ?so ? nso
6
and
?1/(?wG +p)
, then with probability at least 1 ? ? the excess risk satis?es:
?
??
? ??
??w
2
? ? ??
G
??so
? ta ?
1
?w +p
?w +p
?
+p
ta
G
G
so
log (nta ) ? nso
+ nta
log
?R f
E R f?
=O
nta
?
?wG ? nta
where the expectation is taken over T so and T ta .
Similar to Theorem 2, Theorem 3 suggests that the estimation error comes from two sources. For
estimating the auxiliary function wG , the statistical rate depends on properties of the kernel induced
RKHS, and how far the auxiliary function is from this space. For the ease of presentation, we assume
so
ta
PX so = PX ta , so the approximation errors
Af and A? f are de?ned on the same domain. The
?
error of estimating f so is ampli?ed by O ??2
wG log (nta ) , which is worse than that of nonparametric
kernel smoothing. We believe this ??2
is
nearly
tight because
and Elisseeff have shown
wG
? Bousquet
?
the uniform algorithmic stability parameter for KRR is O ??2
Bousquet
and Elisseeff [2002].
wG
Steinwart et al. Steinwart?et al. [2009]
showed
that
for
non-transfer
learning,
the
optimal statistical
?
??ta
?ta +p
rate for excess risk is ? nta
, so if ?wg ? ?ta and nso is suf?ciently large then we achieve
improved convergence rate through transfer learning.
Remark: Theorem 2 and 3 are not directly comparable because our assumptions on the function
spaces of these two theorems are different. In general, H?lder space is only a Banach space but not a
Hilbert space. We refer readers to Theorem 1 in Zhou [2008] for details.
5
Finding the Best Transformation Function
In the previous section we showed for a speci?c transformation function G, if auxiliary function
is smoother than the target regression function then we have smaller excess risk. In practice,
we would like to try out a class of transformation functions G , which is possibly uncountable.
We can construct a subset of G ? G, which is ?nite and satis?es that each G in G there is a
G in G that is close to G. Here we give an example. Consider the transformation functions
that have the form: G = {G(a, b) = ?a + b where |?| ? L? , |a| ? La } . We can quantize this set
of transformation functions by considering a subset of G: G = {G(a, b) = k?a + b} where ? =
L?
2K , k = ?K, ? ? ? , 0, ? ? ? , K and |a| ? La . Here ? is the quantization unit.
The next theorem shows we only need to search the transformation function G in G whose correta
sponding estimator f?G
has the lowest empirical risk on the validation dataset.
Theorem 4 Let G be a class of transformation functions and G be its ||?||? norm ?-cover.
Suppose wG satis?es the same assumption in Theorem 1 and for any two G1 , G2 ?? G,
?
ta
||wG1 ? wG2 ||? ? L ||G1 ? G2 ||? for some constant L. Denote G? = argminG?G R f?G
?
?
? ?
ta
? ?? ? ??
?
R(f?G
?)
? f?ta . If we choose ? = O ?
and G = argminG?G R
and nval = ? log ?G ? /? , the
nta
G
ci
i=1
? ?
? ? ?
?
??
??
ta
ta
ta
?
with probability at least 1 ? ?, E R fG? ? R (f ) = O E R f?G
? R (f ta ) where the
?
expectation is taken over T so and T ta .
Remark 1: This theorem implies that if no-transfer function (G (a, b) = b) is in G then we will
end up choosing a transformation function that has the same order of excess risk as using no-transfer
learning algorithm, thus avoiding negative transfer.
Remark 2: Note number of validation set is only logarithmically depending on the size of set of
transformation functions. Therefore, we only need to use a very small amount of data from the target
domain to do cross-validation.
6
Experiments
In this section we use robotics and neural imaging data to demonstrate the effectiveness of the
proposed framework. We conduct experiments on real-world data sets with the following procedures.
7
Only Target KS
Only Target KRR
Only Source KRR
Combined KS
Combined KRR
CDM
Offset KS
Offset KRR
Scale KS
Scale KRR
nta = 10
nta = 20
nta = 40
nta = 80
nta = 160
nta = 320
0.086 ? 0.022
0.080 ? 0.017
0.098 ? 0.017
0.092 ? 0.011
0.087 ? 0.025
0.105 ? 0.023
0.080 ? 0.026
0.146 ? 0.112
0.078 ? 0.022
0.102 ? 0.033
0.076 ? 0.010
0.078 ? 0.022
0.098 ? 0.017
0.084 ? 0.008
0.077 ? 0.015
0.074 ? 0.020
0.066 ? 0.023
0.066 ? 0.017
0.065 ? 0.013
0.095 ? 0.100
0.066 ? 0.008
0.063 ? 0.013
0.098 ? 0.017
0.077 ? 0.009
0.062 ? 0.009
0.064 ? 0.008
0.052 ? 0.006
0.053 ? 0.007
0.056 ? 0.009
0.057 ? 0.014
0.064 ? 0.007
0.050 ? 0.007
0.098 ? 0.017
0.075 ? 0.006
0.061 ? 0.005
0.060 ? 0.007
0.054 ? 0.006
0.048 ? 0.006
0.056 ? 0.005
0.052 ? 0.010
0.065 ? 0.006
0.048 ? 0.006
0.098 ? 0.017
0.074 ? 0.006
0.047 ? 0.003
0.053 ? 0.009
0.050 ? 0.003
0.043 ? 0.004
0.054 ? 0.008
0.044 ? 0.004
0.063 ? 0.005
0.040 ? 0.005
0.098 ? 0.017
0.067 ? 0.006
0.041 ? 0.004
0.056 ? 0.004
0.052 ? 0.004
0.041 ? 0.003
0.055 ? 0.004
0.042 ? 0.002
Table 1: 1 standard deviation intervals for the mean squared errors of various algorithms when
transferring from kin-8fm to kin-8nh. The values in bold are the smallest errors for each nta . Only
Source KS has much worse performance than other algorithms so we do not show its result here.
Directly training on the target data T ta (Only Target KS, Only Target KRR).
Only training on the source data T so (Only Source KS, Only Source KRR).
Training on the combined source and target data (Combined KS, Combined KRR).
The CDM algorithm proposed by Wang and Schneider [2014] with KRR (CDM).
The algorithm described in this paper with G(a, b) = (a + ?)b where ? is a hyper-parameter
(Scale KS, Scale KRR).
? The algorithm described in this paper with G(a, b) = ?a + b where ? is a hyper-parameter
(Offset KS, Offset KRR). ?
?
?
?
?
?
For the ?rst experiment, we vary the size of the target domain to study the effect of nta relative
to nso . We use two datasets from the ?kin? family in Delve [Rasmussen et al., 1996]. The two
datasets we use are ?kin-8fm? and ?kin-8nh?, both with 8 dimensional inputs. kin-8fm has fairly linear
output, and low noise. kin-8nh on the other hand has non-linear output, and high noise. We consider
the task of transfer learning from kin-8fm to kin-8nh. In this experiment, We set nso to 320, and
vary nta in {10, 20, 40, 80, 160, 320}. Hyper-parameters were picked using grid search with 10-fold
cross-validation on the target data (or source domain data when not using the target domain data).
Table 1 shows the mean squared errors on the target data. To better understand the results, we show a
box plot of the mean squared errors for nta = 40 onwards in Figure 2(a). The results for nta = 10
and nta = 20 have high variance, so we do not show them in the plot. We also omit the results of
Only Source KRR because of its poor performance. We note that our proposed algorithm outperforms
other methods across nearly all values of nta especially when nta is small. Only when there are as
many points in the target as in the source, does simply training on the target give the best performance.
This is to be expected since the primary purpose in doing transfer learning is to alleviate the problem
of lack of data in the target domain. Though quite comparable, the performance of the scale methods
was worse than the offset methods in this experiment. In general, we would use cross-validation to
choose between the two.
We now consider another real-world dataset where the covariates are fMRI images taken while
subjects perform a Stroop task [Stroop, 1935]. We use the dataset collected by Verstynen [2014]
which contains fMRI data of 28 subjects. A total of 120 trials were presented to each participant and
fMRI data was collected throughout the trials, and went through a standard post-processing scheme.
The result of this is a feature vector corresponding to each trial that describes the activity of brain
regions (voxels), and the goal is to use this to predict the response time.
To frame the problem in the transfer learning setting, we consider as source the data of all but one
subject. The goal is to predict on the remaining subject. We performed ?ve repetitions for each
algorithm by drawing nso = 300 data points randomly from the 3000 points in the source domain.
We used nta = 80 points from the target domain for training and cross-validation; evaluation was
done on the 35 remaining points in the target domain. Figure 2 (b) shows a box plot of the coeffecient
of determination values (R-squared) for the best performing algorithms. R-squared is de?ned as
1 ? SSres /SStot where SSres is the sum of squared residuals, and SStot is the total sum of squares.
Note that R-squared can be negative when predicting on unseen samples ? which were not used to ?t
the model ? as in our case. When positive, it indicates the proportion of explained variance in the
dependent variable (higher the better). From the plot, it is clear that Offset KRR and Only Target
KRR have the best performances on average and Offset KRR has smaller variance.
8
????
??????????????????
???
?????????
????
????
????
???
???
???
????
??
??
???
????????????????????????????????????????????
??????????????????????????????????????????????????
??????????????
???????????????
???????????
????????????
???
?????????
???
??????????????????????????????
??????????
????????
?????????
Figure 2: Box plots of experimental results on real datasets. Each box extends from the ?rst to third
quartile, and the horizontal lines in the middle are medians. For the robotics data, we report mean
squared error (lower the better) and for the fMRI data, we report R-squared (the higher the better).
For the ease of presentation, we only show results of algorithms with good performances.
Only Target KS
Only Target KRR
Only Source KS
Only Source KRR
Combined KS
Combined KRR
CDM
Offset KS
Offset KRR
Scale KS
Scale KRR
Mean
Median
-0.0096
0.1041
-0.4932
-0.8763
-0.7540
-0.5868
-3.1183
0.1190
0.1080
0.0017
0.0897
0.0444
0.1186
-0.5366
-0.9363
-0.2023
-0.0691
-3.4510
0.1081
0.1221
-0.0321
0.1107
Standard Deviation
0.1041
0.2361
0.4555
0.6265
1.5109
1.3223
2.6473
0.0612
0.0682
0.0632
0.1104
Table 2: Mean, median, and standard deviation for the coef?cient of determination (R-squared) of
various algorithms on the fMRI dataset.
Table 2 shows the full table of results for the fMRI task. Using only the source data produces large
negative R-squared, and while Only Target KRR does produce a positive mean R-squared, it comes
with a high variance. On the other hand, both Offset methods have low variance, showing consistent
performance. For this particular case, the Scale methods do not perform as well as the Offset methods,
and as has been noted earlier, in general we would use cross validation to select an appropriate
transfer function.
7
Conclusion and Future Works
In this paper, we proposed a general transfer learning framework for the HTL regression problem
when there is some data available from the target domain. Theoretical analysis shows it is possible to
achieve better statistical rate using transfer learning than standard supervised learning.
Now we list two future directions and how our results could be further improved. First, in many real
world applications, there is also a large amount of unlabeled data from the target domain available.
Combining our proposed framework with previous works for this scenario [Cortes and Mohri, 2014,
Huang et al., 2006] is a promising direction to pursue. Second, we only present upper bounds in
this paper. It is an interesting direction to obtain lower bounds for HTL and other transfer learning
scenarios.
8
Acknowledgements
S.S.D. and B.P. were supported by NSF grant IIS1563887 and ARPA-E Terra program. A.S. was
supported by AFRL grant FA8750-17-2-0212.
9
References
Shai Ben-David and Ruth Urner. Domain adaptation as learning with auxiliary information. In New
Directions in Transfer and Multi-Task-Workshop@ NIPS, 2013.
Shai Ben-David, John Blitzer, Koby Crammer, and Fernando Pereira. Analysis of representations for
domain adaptation. Advances in neural information processing systems, 19:137, 2007.
John Blitzer, Koby Crammer, Alex Kulesza, Fernando Pereira, and Jennifer Wortman. Learning
bounds for domain adaptation. In Advances in neural information processing systems, pages
129?136, 2008.
Olivier Bousquet and Andr? Elisseeff. Stability and generalization. Journal of Machine Learning
Research, 2(Mar):499?526, 2002.
Raymond J Carroll, David Ruppert, Leonard A Stefanski, and Ciprian M Crainiceanu. Measurement
error in nonlinear models: a modern perspective. CRC press, 2006.
Corinna Cortes and Mehryar Mohri. Domain adaptation in regression. In Algorithmic Learning
Theory, pages 308?323. Springer, 2011.
Corinna Cortes and Mehryar Mohri. Domain adaptation and sample bias correction theory and
algorithm for regression. Theoretical Computer Science, 519:103?126, 2014.
Corinna Cortes, Mehryar Mohri, and Andr?s Mu?oz Medina. Adaptation algorithm and theory based
on generalized discrepancy. In Proceedings of the 21th ACM SIGKDD International Conference
on Knowledge Discovery and Data Mining, pages 169?178. ACM, 2015.
Cecil C Craig. On the tchebychef inequality of bernstein. The Annals of Mathematical Statistics, 4
(2):94?102, 1933.
Li Fei-Fei, Rob Fergus, and Pietro Perona. One-shot learning of object categories. IEEE transactions
on pattern analysis and machine intelligence, 28(4):594?611, 2006.
Jiayuan Huang, Arthur Gretton, Karsten M Borgwardt, Bernhard Sch?lkopf, and Alex J Smola.
Correcting sample selection bias by unlabeled data. In Advances in neural information processing
systems, pages 601?608, 2006.
Samory Kpotufe and Vikas Garg. Adaptivity to local smoothness and dimension in kernel regression.
In Advances in Neural Information Processing Systems, pages 3075?3083, 2013.
Ilja Kuzborskij and Francesco Orabona. Stability and hypothesis transfer learning. In ICML (3),
pages 942?950, 2013.
Ilja Kuzborskij and Francesco Orabona. Fast rates by transferring from auxiliary hypotheses. Machine
Learning, pages 1?25, 2016.
Ilja Kuzborskij, Francesco Orabona, and Barbara Caputo. From n to n+ 1: Multiclass transfer
incremental learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, pages 3358?3365, 2013.
Ilja Kuzborskij, Francesco Orabona, and Barbara Caputo. Scalable greedy algorithms for transfer
learning. Computer Vision and Image Understanding, 2016.
Tongliang Liu, Dacheng Tao, Mingli Song, and Stephen Maybank. Algorithm-dependent generalization bounds for multi-task learning. IEEE transactions on pattern analysis and machine
intelligence, 2016.
Yishay Mansour, Mehryar Mohri, and Afshin Rostamizadeh. Domain adaptation: Learning bounds
and algorithms. arXiv preprint arXiv:0902.3430, 2009.
Mehryar Mohri and Andres Munoz Medina. New analysis and algorithm for learning with drifting
distributions. In Algorithmic Learning Theory, pages 124?138. Springer, 2012.
10
Stephen Nuske, Kamal Gupta, Srinivasa Narasimhan, and Sanjiv Singh. Modeling and calibrating
visual yield estimates in vineyards. In Field and Service Robotics, pages 343?356. Springer, 2014.
Francesco Orabona, Claudio Castellini, Barbara Caputo, Angelo Emanuele Fiorilla, and Giulio
Sandini. Model adaptation with least-squares svm for adaptive hand prosthetics. In Robotics and
Automation, 2009. ICRA?09. IEEE International Conference on, pages 2897?2903. IEEE, 2009.
Carl Edward Rasmussen, Radford M Neal, Georey Hinton, Drew van Camp, Michael Revow, Zoubin
Ghahramani, Rafal Kustra, and Rob Tibshirani. Delve data for evaluating learning in valid
experiments. URL http://www. cs. toronto. edu/ delve, 1996.
Ingo Steinwart, Don R Hush, and Clint Scovel. Optimal rates for regularized least squares regression.
In COLT, 2009.
J Ridley Stroop. Studies of interference in serial verbal reactions. Journal of experimental psychology,
18(6):643, 1935.
Masashi Sugiyama, Shinichi Nakajima, Hisashi Kashima, Paul V Buenau, and Motoaki Kawanabe.
Direct importance estimation with model selection and its application to covariate shift adaptation.
In Advances in neural information processing systems, pages 1433?1440, 2008.
Tatiana Tommasi, Francesco Orabona, and Barbara Caputo. Safety in numbers: Learning categories
from few examples with multi model knowledge transfer. In Computer Vision and Pattern
Recognition (CVPR), 2010 IEEE Conference on, pages 3081?3088. IEEE, 2010.
Timothy D Verstynen. The organization and dynamics of corticostriatal pathways link the medial
orbitofrontal cortex to future behavioral responses. Journal of neurophysiology, 112(10):2457?
2469, 2014.
Vladimir Vovk. Kernel ridge regression. In Empirical Inference, pages 105?116. Springer, 2013.
Xuezhi Wang and Jeff Schneider. Flexible transfer learning under support and model shift. In
Advances in Neural Information Processing Systems, pages 1898?1906, 2014.
Xuezhi Wang and Jeff Schneider. Generalization bounds for transfer learning under model shift.
2015.
Xuezhi Wang, Junier B Oliva, Jeff Schneider, and Barnab?s P?czos. Nonparametric risk and stability
analysis for multi-task learning problems. In 25th International Joint Conference on Arti?cial
Intelligence (IJCAI), volume 1, page 2, 2016.
Larry Wasserman. All of nonparametric statistics. Springer Science & Business Media, 2006.
Jun Yang, Rong Yan, and Alexander G Hauptmann. Cross-domain video concept detection using
adaptive svms. In Proceedings of the 15th ACM international conference on Multimedia, pages
188?197. ACM, 2007.
Yaoliang Yu and Csaba Szepesv?ri. Analysis of kernel mean matching under covariate shift. arXiv
preprint arXiv:1206.4650, 2012.
Kun Zhang, Krikamol Muandet, and Zhikun Wang. Domain adaptation under target and conditional
shift. In Proceedings of the 30th International Conference on Machine Learning (ICML-13), pages
819?827, 2013.
Yu Zhang. Multi-task learning and algorithmic stability. In AAAI, volume 2, pages 6?2, 2015.
Ding-Xuan Zhou. Derivative reproducing properties for kernel methods in learning theory. Journal
of computational and Applied Mathematics, 220(1):456?463, 2008.
11
| 6660 |@word neurophysiology:1 trial:3 middle:1 norm:3 proportion:1 nd:1 elisseeff:6 arti:1 shot:1 liu:2 contains:1 rkhs:3 ours:1 fa8750:1 outperforms:1 existing:3 reaction:3 scovel:1 written:1 john:2 sanjiv:1 happen:1 enables:1 krikamol:1 plot:5 stroop:3 medial:1 intelligence:3 greedy:1 cult:1 awg:9 toronto:1 simpler:3 zhang:4 mathematical:1 constructed:1 ilja:4 direct:1 pathway:1 behavioral:1 introduce:1 theoretically:1 expected:1 karsten:1 multi:5 brain:1 decreasing:1 considering:1 begin:1 estimating:7 bounded:6 moreover:1 factorized:1 medium:1 lowest:1 what:1 pursue:1 narasimhan:1 finding:1 transformation:31 csaba:1 guarantee:1 cial:1 masashi:1 act:1 giulio:1 unit:1 grant:2 omit:1 positive:4 before:1 service:1 local:1 treat:3 safety:1 analyzing:1 clint:1 might:2 garg:1 k:22 studied:2 suggests:2 dif:1 ease:2 delve:3 ridley:1 practice:1 dpx:2 procedure:5 nite:2 empirical:3 universal:1 yan:1 matching:1 induce:1 regular:2 zoubin:1 unlabeled:3 close:4 operator:2 selection:2 risk:26 context:1 seminal:1 www:1 equivalent:2 convex:1 focused:2 correcting:1 wasserman:3 estimator:20 orthonormal:1 his:1 stability:10 embedding:1 notion:1 argming:2 tting:1 target:65 suppose:4 annals:1 pioneered:1 user:1 olivier:1 carl:1 us:2 nuske:2 hypothesis:7 logarithmically:1 approximated:1 recognition:2 labeled:1 preprint:2 ding:1 wang:13 capture:1 region:1 went:1 highest:1 removed:1 transforming:1 mu:1 covariates:2 dynamic:1 trained:3 singh:2 tight:1 mingli:1 basis:2 joint:2 various:2 train:4 fast:1 hyper:3 outside:1 choosing:1 whose:1 quite:1 widely:2 cvpr:1 say:1 drawing:1 lder:4 wg:53 statistic:3 g1:4 unseen:1 noisy:1 semide:1 eigenvalue:2 propose:2 product:2 adaptation:10 combining:1 achieve:2 oz:1 crossvalidation:2 rst:10 convergence:4 ijcai:1 produce:3 xuan:1 incremental:1 leave:1 ben:4 tk:3 help:2 blitzer:3 develop:1 depending:1 object:1 edward:1 auxiliary:12 c:3 come:2 implies:1 motoaki:1 direction:4 quartile:1 larry:1 crc:1 require:1 barnab:2 generalization:9 preliminary:1 alleviate:1 rong:1 correction:1 aartisingh:1 mapping:2 algorithmic:4 predict:2 achieves:1 vary:2 smallest:1 aarti:1 purpose:1 estimation:13 angelo:1 label:2 krr:28 repetition:1 aso:3 minimization:1 cdm:4 avoid:1 zhou:2 claudio:1 broader:1 derived:1 focus:2 prosthetics:1 indicates:1 sigkdd:1 rostamizadeh:1 camp:1 helpful:2 inference:1 dependent:3 integrated:1 transferring:2 yaoliang:1 her:1 relation:3 perona:1 subroutine:2 tao:1 colt:1 flexible:1 denoted:1 smoothing:9 special:1 fairly:2 field:1 construct:4 beach:1 ciprian:1 identical:1 represents:1 yu:3 koby:2 nearly:2 icml:2 kamal:1 fmri:8 future:4 discrepancy:4 report:2 few:1 modern:1 randomly:1 ve:1 individual:1 replaced:1 ab:2 detection:1 onwards:1 organization:1 satis:8 mining:1 evaluation:2 analyzed:1 hg:12 ab2:1 integral:2 buenau:1 necessary:1 arthur:1 conduct:3 theoretical:8 arpa:1 instance:1 xuezhi:3 modeling:2 earlier:1 cover:1 deviation:3 subset:4 uniform:2 wortman:1 supx:1 aw:1 synthetic:1 combined:7 muandet:1 st:1 density:1 international:5 terra:1 borgwardt:1 ssdu:1 wg2:1 pool:1 michael:1 invertible:2 together:1 squared:12 aaai:1 rafal:1 choose:3 huang:3 possibly:2 worse:3 derivative:1 toy:2 li:1 de:12 bold:1 hisashi:1 automation:1 satisfy:3 depends:7 performed:1 later:1 try:1 picked:1 analyze:2 characterizes:1 doing:1 participant:1 shai:2 simon:1 square:3 ni:1 variance:5 yield:2 generalize:2 lkopf:1 accurately:1 craig:1 andres:1 coef:1 ed:2 urner:2 naturally:1 newly:1 dataset:4 popular:1 ask:1 knowledge:6 organized:1 formalize:1 hilbert:4 afrl:1 ta:92 higher:2 supervised:1 response:3 improved:3 evaluated:3 though:2 box:4 done:1 mar:1 just:2 smola:1 lastly:1 hand:3 steinwart:3 horizontal:1 nonlinear:2 lack:2 believe:1 usa:1 facilitate:1 effect:1 calibrating:1 true:5 unbiased:4 concept:1 regularization:1 symmetric:1 neal:1 deal:1 sin:3 htl:13 noted:1 generalized:1 ridge:7 demonstrate:2 complete:1 image:4 meaning:1 novel:1 recently:3 srinivasa:1 common:1 pseudocode:1 functional:1 empirically:1 overview:1 nh:4 banach:1 volume:2 accumulate:1 mellon:4 measurement:4 refer:2 dacheng:1 munoz:1 ai:1 maybank:1 smoothness:2 rd:9 grid:1 mathematics:1 similarly:2 sugiyama:2 emanuele:1 stable:1 calibration:1 carroll:2 cortex:1 showed:5 perspective:1 inf:1 irrelevant:1 barbara:4 scenario:4 meta:2 inequality:1 success:1 yi:5 nition:1 additional:1 impose:1 schneider:10 speci:5 paradigm:1 fernando:2 stephen:2 smoother:2 multiple:1 full:1 gretton:1 smooth:1 faster:2 match:1 af:3 cross:7 long:1 plug:1 determination:2 post:1 serial:1 prediction:2 scalable:1 regression:34 oliva:1 vineyard:1 vision:3 cmu:4 expectation:5 metric:1 arxiv:4 kernel:29 sponding:1 nakajima:1 robotics:4 background:1 want:1 addition:1 szepesv:2 whereas:1 interval:1 median:3 source:46 sch:1 biased:1 rest:1 zhikun:1 subject:9 accompany:1 induced:3 incorporates:1 spirit:1 effectiveness:2 call:2 ciently:1 yang:2 bernstein:1 enough:1 affect:2 xj:1 gave:2 nta:38 psychology:1 bandwidth:1 fm:4 reduce:1 inner:2 idea:2 avenue:1 multiclass:1 shift:7 tommasi:2 url:1 song:1 yishay:1 remark:5 useful:3 generally:1 clear:1 amount:6 transforms:1 nonparametric:3 induces:2 svms:1 category:2 http:1 nsf:1 andr:2 notice:3 estimated:4 tibshirani:1 carnegie:4 kuzborskij:11 drawn:3 ampli:1 imaging:1 pietro:1 sum:2 inverse:1 extends:1 throughout:2 reader:2 family:1 comparable:2 orbitofrontal:1 capturing:1 bound:16 fold:1 activity:1 fei:4 alex:2 ri:2 bousquet:6 argument:1 nso:11 performing:1 relatively:3 px:11 ned:7 combination:2 poor:1 smaller:2 across:1 describes:1 wi:7 rob:2 intuitively:1 explained:1 bapoczos:1 interference:1 taken:6 assures:2 jennifer:1 know:1 end:1 available:4 endowed:1 stefanski:1 observe:1 kawanabe:1 away:2 appropriate:1 kashima:1 corinna:3 drifting:1 vikas:1 uncountable:1 remaining:2 cf:1 cally:1 tatiana:1 exploit:1 ghahramani:1 especially:1 classical:2 icra:1 corticostriatal:1 quantity:1 primary:1 usual:1 traditional:1 link:1 collected:2 reason:1 afshin:1 ruth:1 minimizing:1 vladimir:1 setup:1 kun:1 negative:3 design:1 perform:2 kpotufe:1 upper:3 francesco:6 datasets:3 ingo:1 hinton:1 verstynen:3 rn:1 mansour:2 reproducing:2 discovered:1 frame:1 shinichi:1 david:5 introduced:1 pair:1 doppler:1 connection:2 learned:1 established:2 hush:1 nip:2 below:1 pattern:4 kulesza:1 program:1 video:1 power:1 natural:1 rely:1 regularized:1 predicting:2 turning:1 business:1 residual:2 scheme:1 improve:1 brief:2 ne:5 picture:1 jun:1 gf:1 raymond:1 prior:1 literature:2 l2:7 voxels:1 acknowledgement:1 discovery:1 understanding:1 relative:1 loss:2 adaptivity:1 suf:2 interesting:1 validation:8 sandini:1 consistent:1 propagates:1 fta:2 mohri:9 supported:2 czos:2 rasmussen:2 enjoys:1 verbal:1 bias:2 understand:1 fg:1 van:1 dimension:1 world:7 stand:1 evaluating:1 valid:1 adaptive:2 far:1 transaction:2 excess:13 compact:1 uni:2 countably:1 relatedness:1 bernhard:1 conclude:1 xi:7 fergus:1 don:1 continuous:1 search:2 table:5 promising:1 learn:5 transfer:47 ca:1 caputo:4 du:1 quantize:1 mehryar:5 necessarily:2 complex:2 domain:74 main:1 noise:3 paul:1 cient:2 samory:1 medina:3 explicit:2 pereira:2 lie:1 third:2 kin:9 theorem:15 covariate:2 showing:3 explored:1 offset:18 cortes:7 r2:2 list:2 gupta:1 svm:1 exists:1 workshop:1 quantization:1 drew:1 ci:4 importance:1 hauptmann:1 nk:1 easier:2 timothy:1 simply:2 visual:1 contained:1 g2:6 u2:1 springer:5 radford:1 acm:4 conditional:3 viewed:1 formulated:1 presentation:2 goal:2 leonard:1 orabona:11 jeff:3 lipschitz:2 revow:1 change:3 ruppert:1 typical:1 uniformly:1 vovk:3 called:2 total:2 multimedia:1 junier:1 experimental:3 e:9 la:2 jiayuan:1 formally:2 select:1 support:4 crammer:2 alexander:1 relevance:1 avoiding:1 |
6,258 | 6,661 | Controllable Invariance
through Adversarial Feature Learning
Qizhe Xie, Zihang Dai, Yulun Du, Eduard Hovy, Graham Neubig
Language Technologies Institute
Carnegie Mellon University
{qizhex, dzihang, yulund, hovy, gneubig}@cs.cmu.edu
Abstract
Learning meaningful representations that maintain the content necessary for a
particular task while filtering away detrimental variations is a problem of great
interest in machine learning. In this paper, we tackle the problem of learning
representations invariant to a specific factor or trait of data. The representation
learning process is formulated as an adversarial minimax game. We analyze the
optimal equilibrium of such a game and find that it amounts to maximizing the
uncertainty of inferring the detrimental factor given the representation while maximizing the certainty of making task-specific predictions. On three benchmark tasks,
namely fair and bias-free classification, language-independent generation, and
lighting-independent image classification, we show that the proposed framework
induces an invariant representation, and leads to better generalization evidenced by
the improved performance.
1
Introduction
How to produce a data representation that maintains meaningful variations of data while eliminating
noisy signals is a consistent theme of machine learning research. In the last few years, the dominant
paradigm for finding such a representation has shifted from manual feature engineering based on
specific domain knowledge to representation learning that is fully data-driven, and often powered by
deep neural networks [Bengio et al., 2013]. Being universal function approximators [Gybenko, 1989],
deep neural networks can easily uncover the complicated variations in data [Zhang et al., 2017],
leading to powerful representations. However, how to systematically incorporate a desired invariance
into the learned representation in a controllable way remains an open problem.
A possible avenue towards the solution is to devise a dedicated neural architecture that by construction
has the desired invariance property. As a typical example, the parameter sharing scheme and pooling
mechanism in modern deep convolutional neural networks (CNN) [LeCun et al., 1998] take advantage
of the spatial structure of image processing problems, allowing them to induce more generic feature
representations than fully connected networks. Since the invariance we care about can vary greatly
across tasks, this approach requires us to design a new architecture each time a new invariance
desideratum shows up, which is time-consuming and inflexible.
When our belief of invariance is specific to some attribute of the input data, an alternative approach is
to build a probabilistic model with a random variable corresponding to the attribute, and explicitly
reason about the invariance. For instance, the variational fair auto-encoder (VFAE) [Louizos et al.,
2016] employs the maximum mean discrepancy (MMD) to eliminate the negative influence of specific
?nuisance variables?, such as removing the lighting conditions of images to predict the person?s
identity. Similarly, under the setting of domain adaptation, standard binary adversarial cost [Ganin
and Lempitsky, 2015, Ganin et al., 2016] and central moment discrepancy (CMD) [Zellinger et al.,
2017] have been utilized to learn features that are domain invariant. However, all these invariance
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
inducing criteria suffer from a similar drawback, which is they are defined to measure the divergence
between a pair of distributions. Consequently, they can only express the invariance belief w.r.t. a
pair of values of the random variable at a time. When the attribute is a multinomial variable that
takes more than two values, combinatorial number of pairs (specifically, O(n2 )) have to be added to
express the belief that the representation should be invariant to the attribute. The problem is even
more dramatic when the attribute represents a structure that has exponentially many possible values
(e.g. the parse tree of a sentence) or when the attribute is simply a continuous variable.
Motivated by the aforementioned drawbacks and difficulties, in this work, we consider the problem of
learning a feature representation with the desired invariance. We aim at creating a unified framework
that is (1) generic enough such that it can be easily plugged into different models, and (2) more
flexible to express an invariance belief in quantities beyond discrete variables with limited value
choices. Specifically, inspired by the recent advancement of adversarial learning [Goodfellow et al.,
2014], we formulate the representation learning as a minimax game among three players: an encoder
which maps the observed data deterministically into a feature space, a discriminator which looks
at the representation and tries to identify a specific type of variation we hope to eliminate from the
feature, and a predictor which makes use of the invariant representation to make predictions as in
typical discriminative models. We provide theoretical analysis of the equilibrium condition of the
minimax game, and give an intuitive interpretation. On three benchmark tasks from different domains,
we show that the proposed approach not only improves upon vanilla discriminative approaches that do
not encourage invariance, but also outperforms existing approaches that enforce invariant features.
2
Related Work
As a specific case of our problem where s takes two values, domain adaption has attracted a large
amount of research interest. Domain adaptation aims to learn domain-invariant representations that
are transferable to other domains. For example, in image classification, adversarial training has
been shown to able to learn an invariant representation across domains [Ganin and Lempitsky, 2015,
Ganin et al., 2016, Bousmalis et al., 2016, Tzeng et al., 2017] and enables classifiers trained on the
source domain to be applicable to the target domain. Moment discrepancy regularizations can also
effectively remove domain specific information [Zellinger et al., 2017, Bousmalis et al., 2016] for
the same purpose. By learning language-invariant representations, classifiers trained on the source
language can be applied to the target language [Chen et al., 2016b, Xu and Yang, 2017].
Works targeting the development of fair, bias-free classifiers also aim to learn representations invariant
to ?nuisance variables? that could induce bias and hence makes the predictions fair, as data-driven
models trained using historical data easily inherit the bias exhibited in the data. Zemel et al. [2013]
proposes to regularize the `1 distance between representation distributions for data with different
nuisance variables to enforce fairness. The Variational Fair Autoencoder [Louizos et al., 2016] targets
the problem with a Variational Autoencoder [Kingma and Welling, 2014, Rezende et al., 2014]
approach with maximum mean discrepancy regularization.
Our work is also related to learning disentangled representations, where the aim is to separate different
influencing factors of the input data into different parts of the representation. Ideally, each part of
the learned representation can be marginally independent to the other. An early work by Tenenbaum
and Freeman [1997] propose a bilinear model to learn a representation with the style and content
disentangled. From information theory perspective, Chen et al. [2016a] augments standard generative
adversarial networks with an inference network, whose objective is to infer part of the latent code
that leads to the generated sample. This way, the information carried by the chosen part of the latent
code can be retained in the generative sample, leading to disentangled representation.
As we have discussed in Section 1, these methods bear the same drawback that the cost used to
regularize the representation is pairwise, which does not scale well as the number of values that the
attribute can take could be large. Louppe et al. [2016] propose an adversarial training framework to
learn representations independent to a categorical or continuous variable. A basic assumption in their
theoretical analysis is that the attribute is irrelevant to the prediction, making it unsuitable to analyze
the fairness classifications.
2
3
Adversarial Invariant Feature Learning
In this section, we formulate our problem and then present the proposed framework of learning
invariant features.
Given paired observations hx, yi, we are interested in the task of predicting the target y based on the
value of x using a discriminative approach, i.e. directly modeling the conditional distribution p(y | x).
As the input x can have highly complicated structure, we employ a dedicated model or algorithm to
extract an expressive representation h from x. In addition, we have access to some intrinsic attribute s
of x as well as a prior belief that the prediction result should be invariant to s. Thus, when we extract
the representation h from x, we want the representation h to preserve variations that are necessary to
predict y while eliminating information of s.
To achieve the aforementioned goal, we employ a deterministic encoder E to obtain the representation
by encoding x and s into h, namely, h = E(x, s). It should be noted here that we are using s as
an additional input. Intuitively, this can inform and guide the encoder to remove information about
undesired variations within the representation. For example, if we want to learn a representation
of image x that is invariant to the lighting condition s, the model can learn to ?brighten? the
representation if it knows the original picture is dark, and vice versa.
Given the obtained representation h, the target y is predicted by a predictor M , which effectively
models the distribution qM (y | h). By construction, instead of modeling p(y | x) directly, the
discriminative model we formulate captures the conditional distribution p(y | x, s) with additional
information coming from s.
Surely, feeding s into the encoder by no means guarantees the induced feature h will be invariant to s.
Thus, in order to enforce the desired invariance and eliminate variations of factor s from h, we set up
an adversarial game by introducing a discriminator D which inspects the representation h and ensure
that it is invariant to s. Concretely, the discriminator D is trained to predict s based on the encoded
representation h, which effectively maximizes the likelihood qD (s | h). Simultaneously, the encoder
fights to minimize the same likelihood of inferring the correct s by the discriminator. Intuitively, the
discriminator and the encoder form an adversarial game where the discriminator tries to detect an
attribute of the data while the encoder learns to conceal it.
Note that under our framework, in theory, s can be any type of data as long as it represents an attribute
of x. For example, s can be a real value scalar/vector, which may take many possible values, or a
complex sub-structure such as the parse tree of a natural language sentence. But in this paper, we
focus mainly on instances where s is a discrete label with multiple choices. We plan to extend our
framework to deal with continuous s and structured s in the future.
Formally, E, M and D jointly play the following minimax game:
min max J(E, M, D)
E,M
D
where
J(E, M, D) =
E
x,s,y?p(x,s,y)
[? log qD (s | h = E(x, s)) ? log qM (y | h = E(x, s))]
(1)
where ? is a hyper-parameter to adjust the strength of the invariant constraint, and p(x, s, y) is the
true underlying distribution that the empirical observations are drawn from.
Note that the problem of domain adaption can be seen as a special case of our problem, where s is a
Bernoulli variable representing the domain and the model only has access to the target y when s =
?source domain? during training.
4
Theoretical Analysis
In this section, we theoretically analyze, given enough capacity and training time, whether such a
minimax game will converge to an equilibrium where variations of y are preserved and variations of
s are removed. The theoretical analysis is done in a non-parametric limit, i.e., we assume a model
with infinite capacity.
3
Since both the discriminator and the predictor only use h which is transformed deterministically from
x and s, we can substitute x with h and define a joint distribution p?(h, s, y) of h, s and y as follows
Z
Z
Z
p?(h, s, y) =
p?(x, s, h, y)dx =
p(x, s, y)pE (h | x, s)dx =
p(x, s, y)?(E(x, s) = h)dx
x
x
x
Here, we have used the fact that the encoder is a deterministic transformation and thus the distribution
pE (h | x, s) is merely a delta function denoted by ?(?). Intuitively, h absorbs the randomness in x
and has an implicit distribution of its own. Also, note that the joint distribution p?(h, s, y) depends on
the transformation defined by the encoder.
Thus, we can equivalently rewrite objective (1) as
J(E, M, D) =
E
h,s,y?p(h,s,y)
?
[? log qD (s | h) ? log qM (y | h)]
(2)
To analyze the equilibrium condition of the new objective (2), we first deduce the optimal discriminator
D and the optimal predictor M for a given encoder E and then prove the global optimality of the
minimax game.
?
(s | h) = p?(s | h) and the
Claim 1. Given a fixed encoder E, the optimal discriminator outputs qD
?
optimal predictor corresponds to qM (y | h) = p?(y | h).
Proof. The proof uses the fact that the objective is functionally convex w.r.t. each distribution, and
by taking the variations we can obtain the stationary point for qD and qM as a function of q?. The
detailed proof is included in the supplementary material A.
?
?
Note that the optimal qD
(s | h) and qM
(y | h) given in Claim 1 are both functions of the encoder
?
?
E. Thus, by plugging qD
and qM
into the original minimax objective (2), it can be simplified as a
minimization problem only w.r.t. the encoder E with the following form:
min J(E) = min
E
E
[? log q?(s | h) ? log q?(y | h)]
E
h,s,y??
q (h,s,y)
= min ??H(?
q (s | h)) + H(?
q (y | h))
(3)
E
where H(?
q (s | h)) is the conditional entropy of the distribution q?(s | h). As we can see, the objective
(3) consists of two conditional entropies with different signs. Optimizing the first term amounts to
maximizing the uncertainty of inferring s based on h, which is essentially filtering out any information
of s from the representation. On the contrary, optimizing the second term leads to increasing the
certainty of predicting y based on h. Implicitly, the objective defines the equilibrium of the minimax
game.
? Firstly, for cases where the attribute s is entirely irrelevant to the prediction task, the two terms
can reach the optimum at the same time, leading to a win-win equilibrium. For example, with
the lighting condition of an image removed, we can still/better classify the identity of the people
in that image. With enough model capacity, the optimal equilibrium solution would be the same
regardless of the value of ?.
? However, there are cases where these two optimization objectives are competing. For example,
in fair classifications, sensitive factors such as gender and age may help the overall prediction
accuracies. Hence learning a fair/invariant representation is harmful to predictions. In this case,
the optimality of these two entropies cannot be achieved simultaneously, and ? defines the relative
strengths of the two objectives in the final equilibrium.
5
5.1
Parametric Instantiation of the Proposed Framework
Models
To show the general applicability of our framework, we experiment on three different tasks including
sentence generation, image classification and fair classifications. Due to the different natures of data
of x and y, here we present the specific model instantiations we use.
4
Sentence Generation We use multi-lingual machine translation as the testbed for sentence generation. Concretely, we have translation pairs between several source languages and a target language. x
is the source sentence to be translated and s is a scalar denoting which source language x belongs to.
y is the translated sentence for the target language.
Recall that s is used as an input of E to obtain a language-invariant representation. To make full
use of s, we employ separate encoders Encs for sentences in each language s. In other words,
h = E(s, x) = Encs (x) where each Encs is a different encoder. The representation of a sentence is
captured by the hidden states of an LSTM encoder [Hochreiter and Schmidhuber, 1997] at each time
step.
We employ a single LSTM predictor for different encoders. As often used in language generation,
the probability qM output by the predictor is parametrized by an autoregressive process, i.e.,
qM (y1:T | h) =
T
Y
qM (yt |y<t , h)
t=1
where we use an LSTM with attention model [Bahdanau et al., 2015] to compute qM (yt |y<t , h).
The discriminator is also parameterized as an LSTM which gives it enough capacity to deal with
input of multiple timesteps. qD (s | h) is instantiated with the multinomial distribution computed by a
softmax layer on the last hidden state of the discriminator LSTM.
Classification For our classification experiments, the input is either a picture or a feature vector.
All of the three players in the minimax game are constructed by feedforward neural networks. We
feed s to the encoder as an embedding vector.
5.2
Optimization
There are two possible approaches to optimize our framework in an adversarial setting. The first one
is similar to the alternating approach used in Generative Adversarial Nets (GANs) [Goodfellow et al.,
2014]. We can alternately train the two adversarial components while freezing the third one. This
approach has more control in balancing the encoder and the discriminator, which effectively avoids
saturation. Another method is to train all three components together with a gradient reversal layer
[Ganin and Lempitsky, 2015]. In particular, the encoder admits gradients from both the discriminator
and the predictor, with the gradient from the discriminator negated to push the encoder in the opposite
direction desired by the discriminator. Chen et al. [2016b] found the second approach easier to
optimize since the discriminator and the encoder are fully in sync being optimized altogether. Hence
we adopt the latter approach. In all of our experiments, we use Adam [Kingma and Ba, 2014] with a
learning rate of 0.001.
6
Experiments
In this section, we perform empirical experiments to evaluate the effectiveness of proposed framework.
We first introduce the tasks and corresponding datasets we consider. Then, we present the quantitative
results showing the superior performance of our proposed framework, and discuss some qualitative
analysis which verifies the learned representations have the desired invariance property.
6.1
Datasets
Our experiments include three tasks in different domains: (1) fair classification, in which predictions
should be unaffected by nuisance factors; (2) language-independent generation which is conducted
on the multi-lingual machine translation problem; (3) lighting-independent image classification.
Fair Classification For fair classification, we use three datasets to predict the savings, credit ratings
and health conditions of individuals with variables such as gender or age specified as ?nuisance
variable? that we would like to not consider in our decisions [Zemel et al., 2013, Louizos et al.,
2016]. The German dataset [Frank et al., 2010] is a small dataset with 1, 000 samples describing
whether a person has a good credit rating. The sensitive nuisance variable to be factored out is gender.
5
The Adult income dataset [Frank et al., 2010] has 45, 222 data points and the objective is to predict
whether a person has savings of over 50, 000 dollars with the sensitive factor being age. The task of
the health dataset1 is to predict whether a person will spend any days in the hospital in the following
year. The sensitive variable is also the age and the dataset contains 147, 473 entries. We follow the
same 5-fold train/validation/test splits and feature preprocessing used in [Zemel et al., 2013, Louizos
et al., 2016].
Both the encoder and the predictor are parameterized by single-layer neural networks. A three-layer
neural network with batch normalization [Ioffe and Szegedy, 2015] is employed for the discriminator.
We use a batch size of 16 and the number of hidden units is set to 64. ? is set to 1 in our experiments.
Multi-lingual Machine Translation For the multi-lingual machine translation task we use French
to English (fr-en) and German to English (de-en) pairs from IWSLT 2015 dataset [Cettolo et al., 2012].
There are 198, 435 pairs of fr-en sentences and 188, 661 pairs of de-en sentences in the training set. In
the test set, there are 4, 632 pairs of fr-en sentences and 7, 054 pairs of de-en sentences. We evaluate
BLEU scores [Papineni et al., 2002] using the standard Moses multi-bleu.perl script. Here, s
indicates the language of the source sentence.
We use the OpenNMT [Klein et al., 2017] in our multi-lingual MT experiments. The encoder is a
two-layer bidirectional LSTM with 256 units for each direction. The discriminator is a one-layer
single-directional LSTM with 256 units. The predictor is a two-layer LSTM with 512 units and
attention mechanism [Bahdanau et al., 2015]. We follow Johnson et al. [2016] and use Byte Pair
Encoding (BPE) subword units [Sennrich et al., 2016] as the cross-lingual input. Every model is run
for 20 epochs. ? is set to 8 and the batch size is set to 64.
Image Classification We use the Extended Yale B dataset [Georghiades et al., 2001] for our image
classification task. It comprises face images of 38 people under 5 different lighting conditions: upper
right, lower right, lower left, upper left, or the front. The variable s to be purged is the lighting
condition. The label y is the identity of the person. We follow Li et al. [2014], Louizos et al. [2016]?s
train/test split and no validation is used: 38 ? 5 = 190 samples are used for training and all other
1, 096 data points are used for testing.
We use a one-layer neural network for the encoder and a one-layer neural network for prediction. ? is
set to 2. The discriminator is a two-layer neural network with batch normalization. The batch size is
set to 16 and the hidden size is set to 100.
6.2
Results
Fair Classification The results on three fairness tasks are shown in Figure 1. We compare our model
with two prior works on learning fair representations: Learning Fair Representations (LFR) [Zemel
et al., 2013] and Variational Fair Autoencoder (VFAE) [Louizos et al., 2016]. Results of VAE and
directly using x as the representation are also shown.
We first study how much information about s is retained in the learned representation h by using a
logistic regression to predict factor s. In the top row, we see that s cannot be recognized from the
representations learned by three models targeting at fair representations. The accuracy of classifying
s is similar to the trivial baseline predicting the majority label shown by the black line.
The performance on predicting label y is shown in the second row. We see that LFR and VFAE
suffer on Adult and German datasets after removing information of s. In comparison, our model?s
performance does not suffer even when making fair predictions. Specifically, on German, our
model?s accuracy is 0.744 compared to 0.727 and 0.723 achieved by VFAE and LFR. On Adult, our
model?s accuracy is 0.844 while VFAE and LFR have accuracies of 0.813 and 0.823 respectively.
On the health dataset, all models? performances are barely better than the majority baseline. The
unsatisfactory performances of all models may be due to the extreme imbalance of the dataset, in
which 85% of the data has the same label.
We also investigate how fair representations would alleviate biases of machine learning models. We
measure the unbiasedness by evaluating models? performances on identifying minority groups. For
instance, suppose the task is to predict savings with the nuisance factor being age, with savings
1
www.heritagehealthprize.com
6
Adult
German
0.9
0.9
0.78
0.78
Majority
0.67
0.65
0.53
0.4
Health
0.9
Majority
0.8
0.65
0.65
0.53
0.53
LFR
VAE
VFAE
Majority
0.58
0.4
0.4
x
0.78
Ours
x
LFR
VAE
VFAE
x
Ours
LFR
VAE
VFAE
Ours
(a) Accuracy on predicting s. The closer the result is to the majority line, the better the model is in eliminating
the effect of nuisance variables.
Adult
Health
German
0.9
Majority
0.75
0.78
0.9
0.9
0.78
0.78
0.65
0.65
0.53
0.53
Majority
0.71
x
LFR
VAE
VFAE
x
Ours
0.65
0.53
0.4
0.4
Majority
0.84
LFR
VAE
VFAE
Ours
0.4
x
LFR
VAE
VFAE
Ours
(b) Accuracy on predicting y. High accuracy in predicting y is desireable.
x
0.9
Health
German
Adult
Ours
x
0.9
Ours
0.775
0.78
0.78
0.65
0.65
0.65
0.525
0.53
0.53
0.4
Overall
Biased categories
0.4
Overall
Biased categories
x
0.9
Ours
0.4
Overall
Biased categories
(c) Overall performance and performance on biased categories. Fair representations lead to high accuracy on
baised categories.
Figure 1: Fair classification results on different representations. x denotes directly using the observation x as the representation. The black lines in the first and the second row show the performance
of predicting the majority label. ?Biased categories? in the third row are explained in the fourth
paragraph of Section 6.2.
above a threshold of $50, 000 being adequate, otherwise being insufficient. If people of advanced
age generally have fewer savings, then a biased model would tend to predict insufficient savings for
those with an advanced age. In contrast, an unbiased model can better factor out age information and
recognize people that do not fit into these stereotypes.
Concretely, for groups pooled by each possible value of y, we seek for the minority s in each of these
groups and define the minority s as the biased category for the group. Then we first calculate the
accuracy on each biased category and report the average performance for all categories. We do not
compute the instance-level average performance since one category may hold the dominant amount
of data among all categories.
As shown in the third row of Figure 1, on German and Adult, we achieve higher accuracy on the
biased categories, even though our overall accuracy is similar to or lower than the baseline which
does not employ fairness constraints. Specifically, on Adult, our performance on the biased categories
is 0.788 while the baseline?s accuracy is 0.748. On German, our accuracy on biased categories is
0.676 while the baseline achieves 0.648. The results show that our model is able to learn a more
unbiased representation.
Multi-lingual Machine Translation The results of systems on multi-lingual machine translation
are shown in Table 1. We compare our model with attention based encoder-decoder trained on
bilingual data [Bahdanau et al., 2015] and multi-lingual data [Johnson et al., 2016]. The encoderdecoder trained on multi-lingual data employs a single encoder for both source languages. Firstly,
both multi-lingual systems outperform the bilingual encoder-decoder even though multi-lingual
systems use similar number of parameters to translate two languages, which shows that learning
7
Model
test (fr-en) test (de-en)
Bilingual Enc-Dec [Bahdanau et al., 2015]
35.2
27.3
Multi-lingual Enc-Dec [Johnson et al., 2016]
35.5
27.7
Our model
36.1
28.1
w.o. discriminator
35.3
27.6
w.o. separate encoders
35.4
27.7
Table 1: Results on multi-lingual machine translation.
Method
Accuracy of classifying s Accuracy of classifying y
Logistic regression
0.96
0.78
NN + MMD [Li et al., 2014]
0.82
VFAE [Louizos et al., 2016]
0.57
0.85
Ours
0.57
0.89
Table 2: Results on Extended Yale B dataset. A better representation has lower accuracy of classifying
factor s and higher accuracy of classifying label y
invariant representation leads to better generalization in this case. The better generalization may be
due to transferring statistical strength between data in two languages.
Comparing two multi-lingual systems, our model outperforms the baseline multi-lingual system on
both languages, where the improvement on French-to-English is 0.6 BLEU score. We also verify
the design decisions in our framework by ablation studies. Firstly, without the discriminator, the
model?s performance is worse than the standard multi-lingual system, which rules out the possibility
that the gain of our model comes from more parameters of separating encoders. Secondly, when we
do not employ separate encoders, the model?s performance deteriorates and it is more difficult to
learn a cross-lingual representation, which means that the encoder needs to have enough capacity to
reach the equilibrium in the minimax game. Intuitively, German and French have different grammars
and vocabulary, so it is hard to obtain a unified semantic representation by performing the same
operations.
Image Classification We report the results in Table 2 with two baselines [Li et al., 2014, Louizos
et al., 2016] that use MMD regularizations to remove lighting conditions. The advantage of factoring
out lighting conditions is shown by the improved accuracy 89% for classifying identities, while the
best baseline achieves an accuracy of 85%.
(a) Using the original image x as the representation
(b) Representation learned by our model
Figure 2: t-SNE visualizations of images in the Extended Yale B. The original pictures are clustered
by the lighting conditions, while the representation learned by our model is clustered by identities of
individuals
8
In terms of removing s, our framework can filter the lighting conditions since the accuracy of
classifying s drops from 0.96 to 0.57, as shown in Table 2. We also visualize the learned representation
by t-SNE [Maaten and Hinton, 2008] in comparison to the visualization of original pictures in Figure
2. We see that, without removing lighting conditions, the images are clustered based on the lighting
conditions. After removing information of lighting conditions, images are clustered according to the
identity of each person.
7
Conclusion
In sum, we propose a generic framework to learn representations invariant to a specified factor or trait.
We cast the representation learning problem as an adversarial game among an encoder, a discriminator,
and a predictor. We theoretically analyze the optimal equilibrium of the minimax game and evaluate
the performance of our framework on three tasks from different domains empirically. We show that
an invariant representation is learned, resulting in better generalization and improvements on the
three tasks.
Acknowledgement
We thank Shi Feng, Di Wang and Zhilin Yang for insightful discussions. This research was supported
in part by DARPA grant FA8750-12-2-0342 funded under the DEFT program.
9
References
Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. Neural machine translation by jointly
learning to align and translate. ICLR, 2015.
Yoshua Bengio, Aaron Courville, and Pascal Vincent. Representation learning: A review and new
perspectives. IEEE transactions on pattern analysis and machine intelligence, 35(8):1798?1828,
2013.
Konstantinos Bousmalis, George Trigeorgis, Nathan Silberman, Dilip Krishnan, and Dumitru Erhan.
Domain separation networks. In NIPS, 2016.
Mauro Cettolo, Christian Girardi, and Marcello Federico. Wit3: Web inventory of transcribed and
translated talks. In Proceedings of the 16th Conference of the European Association for Machine
Translation (EAMT), volume 261, page 268, 2012.
Xi Chen, Yan Duan, Rein Houthooft, John Schulman, Ilya Sutskever, and Pieter Abbeel. Infogan:
Interpretable representation learning by information maximizing generative adversarial nets. In
NIPS, 2016a.
Xilun Chen, Yu Sun, Ben Athiwaratkun, Claire Cardie, and Kilian Weinberger. Adversarial deep
averaging networks for cross-lingual sentiment classification. arXiv preprint arXiv:1606.01614,
2016b.
Andrew Frank, Arthur Asuncion, et al. Uci machine learning repository, 2010. URL http://
archive.ics.uci.edu/ml.
Yaroslav Ganin and Victor Lempitsky. Unsupervised domain adaptation by backpropagation. ICML,
2015.
Yaroslav Ganin, Evgeniya Ustinova, Hana Ajakan, Pascal Germain, Hugo Larochelle, Fran?ois
Laviolette, Mario Marchand, and Victor Lempitsky. Domain-adversarial training of neural networks.
Journal of Machine Learning Research, 17(59):1?35, 2016.
Athinodoros S. Georghiades, Peter N. Belhumeur, and David J. Kriegman. From few to many:
Illumination cone models for face recognition under variable lighting and pose. IEEE transactions
on pattern analysis and machine intelligence, 23(6):643?660, 2001.
Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair,
Aaron Courville, and Yoshua Bengio. Generative adversarial nets. In NIPS, 2014.
G Gybenko. Approximation by superposition of sigmoidal functions. Mathematics of Control,
Signals and Systems, 2(4):303?314, 1989.
Sepp Hochreiter and J?rgen Schmidhuber. Long short-term memory. Neural computation, 1997.
Sergey Ioffe and Christian Szegedy. Batch normalization: Accelerating deep network training by
reducing internal covariate shift. arXiv preprint arXiv:1502.03167, 2015.
Melvin Johnson, Mike Schuster, Quoc V Le, Maxim Krikun, Yonghui Wu, Zhifeng Chen, Nikhil
Thorat, Fernanda Vi?gas, Martin Wattenberg, Greg Corrado, et al. Google?s multilingual neural
machine translation system: Enabling zero-shot translation. arXiv preprint arXiv:1611.04558,
2016.
Diederik Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprint
arXiv:1412.6980, 2014.
Diederik P Kingma and Max Welling. Auto-encoding variational bayes. ICLR, 2014.
G. Klein, Y. Kim, Y. Deng, J. Senellart, and A. M. Rush. OpenNMT: Open-Source Toolkit for Neural
Machine Translation. ArXiv e-prints, 2017.
Yann LeCun, L?on Bottou, Yoshua Bengio, and Patrick Haffner. Gradient-based learning applied to
document recognition. Proceedings of the IEEE, 86(11):2278?2324, 1998.
10
Yujia Li, Kevin Swersky, and Richard Zemel.
arXiv:1412.5244, 2014.
Learning unbiased features.
arXiv preprint
Christos Louizos, Kevin Swersky, Yujia Li, Max Welling, and Richard Zemel. The variational fair
autoencoder. ICLR, 2016.
Gilles Louppe, Michael Kagan, and Kyle Cranmer. Learning to pivot with adversarial networks.
arXiv preprint arXiv:1611.01046, 2016.
Laurens van der Maaten and Geoffrey Hinton. Visualizing data using t-sne. JMLR, 2008.
Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. Bleu: a method for automatic
evaluation of machine translation. In ACL, 2002.
Danilo Jimenez Rezende, Shakir Mohamed, and Daan Wierstra. Stochastic backpropagation and
approximate inference in deep generative models. ICML, 2014.
Rico Sennrich, Barry Haddow, and Alexandra Birch. Neural machine translation of rare words with
subword units. ACL, 2016.
Joshua B Tenenbaum and William T Freeman. Separating style and content. NIPS, 1997.
Eric Tzeng, Judy Hoffman, Kate Saenko, and Trevor Darrell. Adversarial discriminative domain
adaptation. arXiv preprint arXiv:1702.05464, 2017.
Ruochen Xu and Yiming Yang. Cross-lingual distillation for text classification. ACL, 2017.
Werner Zellinger, Thomas Grubinger, Edwin Lughofer, Thomas Natschl?ger, and Susanne SamingerPlatz. Central moment discrepancy (cmd) for domain-invariant representation learning. ICLR,
2017.
Richard S Zemel, Yu Wu, Kevin Swersky, Toniann Pitassi, and Cynthia Dwork. Learning fair
representations. ICML, 2013.
Chiyuan Zhang, Samy Bengio, Moritz Hardt, Benjamin Recht, and Oriol Vinyals. Understanding
deep learning requires rethinking generalization. ICLR, 2017.
11
A
Supplementary Material: Proofs
The proof for Claim 1:
?
(s | h) = p?(s | h). The
Claim. Given a fixed encoder E, the optimal discriminator outputs qD
?
optimal predictor corresponds to qM
(y | h) = p?(y | h).
Proof. We first prove the optimal solution of the discriminator. With a fixed encoder, we have the
following optimization problem
min
qD
s.t.
? J(E, M, D)
X
qD (s | h) = 1, ?h
s
P
P
Then L = J(E, M, D) ? h ?(h)( s qD (s | h) ? 1) is the Lagrangian dual function of the above
optimization problem where ?(h) are the dual variables introduced for equality constraints.
The optimal D satisfies the following equation
0=
??
??
??
?L
? (s
?qD
| h)
?J
? ?(h)
0=? ?
?qD (s | h)
P
?(h, s, y)
yq
?(h) = ? ?
qD (s | h)
?
?(h)qD (s | h) = ??
q (s, h)
Summing w.r.t. s on both sides of the last line of Eqn. (4) and using the fact that
we get
?(h) = ??
q (h)
(4)
?
s qD (s
P
| h) = 1,
(5)
Substituting Eqn. 5 back into Eqn. 4, we can prove the optimal discriminator is
?
qD
(s | h) = q?(s | h)
?
Similarly, taking derivation w.r.t. qM (y | h) and setting it to 0, we can prove qM
(y | h) = q?(y |
h).
12
| 6661 |@word cnn:1 repository:1 eliminating:3 open:2 pieter:1 seek:1 dramatic:1 shot:1 moment:3 contains:1 score:2 jimenez:1 denoting:1 ours:10 document:1 subword:2 fa8750:1 outperforms:2 existing:1 com:1 comparing:1 diederik:2 dx:3 attracted:1 john:1 enables:1 christian:2 remove:3 drop:1 interpretable:1 stationary:1 generative:6 fewer:1 advancement:1 intelligence:2 krikun:1 short:1 firstly:3 sigmoidal:1 zhang:2 melvin:1 wierstra:1 constructed:1 qualitative:1 prove:4 consists:1 sync:1 absorbs:1 paragraph:1 introduce:1 theoretically:2 pairwise:1 multi:17 inspired:1 freeman:2 duan:1 increasing:1 underlying:1 maximizes:1 evgeniya:1 unified:2 finding:1 transformation:2 guarantee:1 certainty:2 quantitative:1 every:1 tackle:1 classifier:3 qm:14 ustinova:1 control:2 unit:6 grant:1 sherjil:1 engineering:1 influencing:1 todd:1 limit:1 bilinear:1 encoding:3 ajakan:1 black:2 acl:3 limited:1 lecun:2 testing:1 cmd:2 backpropagation:2 universal:1 empirical:2 yan:1 word:2 induce:2 get:1 cannot:2 targeting:2 salim:1 influence:1 zihang:1 optimize:2 www:1 map:1 deterministic:2 yt:2 maximizing:4 shi:1 sepp:1 regardless:1 attention:3 jimmy:1 convex:1 lagrangian:1 formulate:3 identifying:1 pouget:1 factored:1 rule:1 regularize:2 disentangled:3 deft:1 embedding:1 variation:10 construction:2 target:8 play:1 suppose:1 hana:1 us:1 samy:1 goodfellow:3 recognition:2 utilized:1 observed:1 louppe:2 mike:1 preprint:7 wang:1 capture:1 calculate:1 connected:1 sun:1 kilian:1 removed:2 benjamin:1 ideally:1 warde:1 kriegman:1 trained:6 rewrite:1 upon:1 eric:1 edwin:1 translated:3 easily:3 joint:2 georghiades:2 darpa:1 talk:1 derivation:1 train:4 instantiated:1 zemel:7 hyper:1 kevin:3 rein:1 whose:1 encoded:1 supplementary:2 spend:1 jean:1 nikhil:1 otherwise:1 federico:1 encoder:31 grammar:1 ward:1 jointly:2 noisy:1 final:1 shakir:1 advantage:2 net:3 propose:3 coming:1 adaptation:4 fr:4 uci:2 enc:2 ablation:1 translate:2 achieve:2 papineni:2 intuitive:1 inducing:1 sutskever:1 optimum:1 jing:1 darrell:1 produce:1 bousmalis:3 adam:2 ben:1 yiming:1 help:1 andrew:1 ganin:7 pose:1 c:1 predicted:1 come:1 larochelle:1 qd:18 stereotype:1 direction:2 ois:1 laurens:1 drawback:3 correct:1 attribute:12 filter:1 stochastic:2 material:2 hx:1 feeding:1 abbeel:1 generalization:5 clustered:4 alleviate:1 secondly:1 hold:1 credit:2 eduard:1 ic:1 great:1 equilibrium:10 predict:9 visualize:1 claim:4 rgen:1 substituting:1 vary:1 early:1 adopt:1 achieves:2 chiyuan:1 purpose:1 applicable:1 combinatorial:1 label:7 superposition:1 sensitive:4 vice:1 hoffman:1 hope:1 minimization:1 aim:4 vae:7 rezende:2 focus:1 improvement:2 unsatisfactory:1 zellinger:3 likelihood:2 mainly:1 indicates:1 greatly:1 bernoulli:1 adversarial:20 contrast:1 baseline:8 detect:1 dollar:1 kim:1 inference:2 dilip:1 factoring:1 nn:1 eliminate:3 transferring:1 fight:1 hidden:4 transformed:1 interested:1 overall:6 classification:20 aforementioned:2 flexible:1 among:3 denoted:1 pascal:2 development:1 proposes:1 spatial:1 plan:1 tzeng:2 special:1 softmax:1 saving:6 beach:1 represents:2 look:1 unsupervised:1 marcello:1 fairness:4 yu:2 icml:3 discrepancy:5 future:1 report:2 yoshua:4 mirza:1 richard:3 few:2 employ:8 modern:1 preserve:1 divergence:1 simultaneously:2 recognize:1 individual:2 dual:2 maintain:1 william:1 interest:2 highly:1 investigate:1 possibility:1 dwork:1 evaluation:1 adjust:1 extreme:1 farley:1 encourage:1 closer:1 necessary:2 arthur:1 tree:2 plugged:1 harmful:1 desired:6 rush:1 theoretical:4 neubig:1 instance:4 classify:1 modeling:2 desireable:1 werner:1 cost:2 introducing:1 applicability:1 entry:1 rare:1 predictor:12 conducted:1 johnson:4 front:1 encoders:5 cho:1 unbiasedness:1 person:6 st:1 lstm:8 trigeorgis:1 recht:1 probabilistic:1 michael:1 together:1 ilya:1 gans:1 central:2 transcribed:1 worse:1 creating:1 leading:3 style:2 li:5 szegedy:2 de:4 yaroslav:2 pooled:1 kate:1 explicitly:1 depends:1 vi:1 script:1 try:2 analyze:5 mario:1 bayes:1 maintains:1 complicated:2 asuncion:1 minimize:1 hovy:2 convolutional:1 accuracy:21 greg:1 identify:1 directional:1 vincent:1 marginally:1 cardie:1 lighting:15 unaffected:1 randomness:1 sennrich:2 inform:1 reach:2 manual:1 sharing:1 trevor:1 mohamed:1 proof:6 di:1 gain:1 dataset:9 hardt:1 birch:1 recall:1 knowledge:1 improves:1 uncover:1 back:1 feed:1 bidirectional:1 higher:2 rico:1 xie:1 day:1 follow:3 danilo:1 improved:2 wei:1 done:1 though:2 implicit:1 eqn:3 parse:2 expressive:1 mehdi:1 freezing:1 web:1 google:1 french:3 defines:2 logistic:2 alexandra:1 usa:1 effect:1 verify:1 true:1 unbiased:3 regularization:3 hence:3 kyunghyun:1 alternating:1 moritz:1 equality:1 semantic:1 undesired:1 deal:2 visualizing:1 game:14 during:1 nuisance:8 transferable:1 noted:1 criterion:1 dedicated:2 image:17 variational:6 kyle:1 superior:1 multinomial:2 mt:1 empirically:1 hugo:1 exponentially:1 volume:1 discussed:1 louizos:9 interpretation:1 extend:1 trait:2 functionally:1 association:1 mellon:1 distillation:1 versa:1 automatic:1 vanilla:1 mathematics:1 similarly:2 language:19 funded:1 toolkit:1 access:2 haddow:1 deduce:1 align:1 patrick:1 dominant:2 pitassi:1 own:1 recent:1 perspective:2 optimizing:2 irrelevant:2 driven:2 belongs:1 schmidhuber:2 cettolo:2 wattenberg:1 binary:1 fernanda:1 approximators:1 yi:1 joshua:1 devise:1 victor:2 seen:1 captured:1 dai:1 care:1 additional:2 george:1 employed:1 belhumeur:1 surely:1 converge:1 paradigm:1 recognized:1 barry:1 signal:2 corrado:1 deng:1 multiple:2 full:1 infer:1 cross:4 long:3 opennmt:2 paired:1 plugging:1 prediction:11 desideratum:1 basic:1 regression:2 essentially:1 cmu:1 arxiv:15 normalization:3 sergey:1 mmd:3 achieved:2 hochreiter:2 dec:2 preserved:1 addition:1 want:2 source:9 biased:11 yonghui:1 exhibited:1 archive:1 lughofer:1 natschl:1 pooling:1 induced:1 tend:1 bahdanau:5 contrary:1 effectiveness:1 encoderdecoder:1 yang:3 feedforward:1 bengio:6 enough:5 split:2 krishnan:1 fit:1 timesteps:1 architecture:2 competing:1 opposite:1 avenue:1 haffner:1 konstantinos:1 shift:1 pivot:1 whether:4 motivated:1 url:1 accelerating:1 sentiment:1 suffer:3 peter:1 adequate:1 deep:7 generally:1 detailed:1 amount:4 dark:1 tenenbaum:2 induces:1 augments:1 category:14 http:1 outperform:1 shifted:1 der:1 moses:1 sign:1 delta:1 deteriorates:1 klein:2 discrete:2 carnegie:1 express:3 group:4 threshold:1 zhilin:1 drawn:1 merely:1 cone:1 year:2 sum:1 run:1 houthooft:1 parameterized:2 uncertainty:2 powerful:1 fourth:1 swersky:3 wu:2 yann:1 separation:1 fran:1 decision:2 maaten:2 graham:1 entirely:1 layer:10 courville:2 yale:3 fold:1 marchand:1 strength:3 constraint:3 grubinger:1 nathan:1 min:5 optimality:2 performing:1 martin:1 structured:1 according:1 lingual:20 across:2 inflexible:1 making:3 quoc:1 intuitively:4 invariant:23 iwslt:1 explained:1 equation:1 visualization:2 remains:1 bing:1 discus:1 german:10 mechanism:2 describing:1 know:1 bpe:1 reversal:1 brighten:1 operation:1 away:1 generic:3 enforce:3 alternative:1 batch:6 weinberger:1 altogether:1 original:5 substitute:1 top:1 denotes:1 ensure:1 include:1 conceal:1 thomas:2 laviolette:1 unsuitable:1 build:1 silberman:1 feng:1 objective:10 added:1 quantity:1 print:1 kagan:1 parametric:2 gradient:4 detrimental:2 win:2 distance:1 separate:4 thank:1 separating:2 capacity:5 parametrized:1 majority:10 decoder:2 mauro:1 rethinking:1 trivial:1 reason:1 bleu:4 barely:1 minority:3 dzmitry:1 ozair:1 code:2 senellart:1 retained:2 insufficient:2 equivalently:1 difficult:1 sne:3 frank:3 negative:1 ba:2 susanne:1 design:2 negated:1 gilles:1 allowing:1 perform:1 upper:2 observation:3 imbalance:1 datasets:4 benchmark:2 enabling:1 daan:1 gas:1 extended:3 hinton:2 y1:1 rating:2 purged:1 evidenced:1 namely:2 pair:10 specified:2 lfr:10 sentence:14 discriminator:25 optimized:1 cast:1 germain:1 learned:9 testbed:1 kingma:4 nip:5 alternately:1 adult:8 beyond:1 able:2 pattern:2 yujia:2 saturation:1 perl:1 program:1 max:3 including:1 memory:1 belief:5 athinodoros:1 david:2 difficulty:1 natural:1 predicting:8 advanced:2 zhu:1 minimax:11 scheme:1 representing:1 technology:1 yq:1 picture:4 carried:1 categorical:1 auto:2 autoencoder:4 extract:2 health:6 byte:1 prior:2 understanding:1 epoch:1 acknowledgement:1 powered:1 review:1 schulman:1 relative:1 text:1 toniann:1 fully:3 bear:1 generation:6 filtering:2 inspects:1 ger:1 geoffrey:1 age:8 validation:2 introduced:1 consistent:1 systematically:1 classifying:7 roukos:1 balancing:1 translation:15 row:5 claire:1 supported:1 last:3 free:2 english:3 bias:5 guide:1 side:1 institute:1 taking:2 face:2 cranmer:1 van:1 vocabulary:1 evaluating:1 avoids:1 dataset1:1 autoregressive:1 concretely:3 preprocessing:1 simplified:1 historical:1 erhan:1 income:1 welling:3 transaction:2 kishore:1 approximate:1 implicitly:1 multilingual:1 ml:1 global:1 instantiation:2 ioffe:2 summing:1 consuming:1 discriminative:5 xi:1 continuous:3 latent:2 table:5 learn:11 nature:1 ca:1 controllable:2 du:1 inventory:1 bottou:1 complex:1 european:1 domain:22 inherit:1 bilingual:3 n2:1 verifies:1 fair:22 xu:3 en:8 judy:1 christos:1 sub:1 inferring:3 theme:1 deterministically:2 comprises:1 pe:2 infogan:1 third:3 jmlr:1 learns:1 zhifeng:1 ian:1 removing:5 dumitru:1 specific:9 covariate:1 showing:1 insightful:1 cynthia:1 abadie:1 admits:1 intrinsic:1 effectively:4 maxim:1 illumination:1 push:1 chen:6 easier:1 entropy:3 simply:1 vinyals:1 scalar:2 gender:3 corresponds:2 satisfies:1 adaption:2 lempitsky:5 conditional:4 identity:6 formulated:1 goal:1 consequently:1 towards:1 content:3 hard:1 included:1 typical:2 specifically:4 infinite:1 reducing:1 averaging:1 hospital:1 invariance:14 iclr:5 player:2 meaningful:2 saenko:1 aaron:2 formally:1 internal:1 people:4 latter:1 oriol:1 incorporate:1 evaluate:3 schuster:1 |
6,259 | 6,662 | Convergence Analysis of Two-layer Neural Networks
with ReLU Activation
Yuanzhi Li
Computer Science Department
Princeton University
[email protected]
Yang Yuan
Computer Science Department
Cornell University
[email protected]
Abstract
In recent years, stochastic gradient descent (SGD) based techniques has become
the standard tools for training neural networks. However, formal theoretical understanding of why SGD can train neural networks in practice is largely missing.
In this paper, we make progress on understanding this mystery by providing a
convergence analysis for SGD on a rich subset of two-layer feedforward networks
with ReLU activations. This subset is characterized by a special structure called
?identity mapping?.?We prove that, if input follows from Gaussian distribution,
with standard O(1/ d) initialization of the weights, SGD converges to the global
minimum in polynomial number of steps. Unlike normal vanilla networks, the
?identity mapping? makes our network asymmetric and thus the global minimum is
unique. To complement our theory, we are also able to show experimentally that
multi-layer networks with this mapping have better performance compared with
normal vanilla networks.
Our convergence theorem differs from traditional non-convex optimization techniques. We show that SGD converges to optimal in ?two phases?: In phase I, the
gradient points to the wrong direction, however, a potential function g gradually
decreases. Then in phase II, SGD enters a nice one point convex region and converges. We also show that the identity mapping is necessary for convergence, as it
moves the initial point to a better place for optimization. Experiment verifies our
claims.
1
Introduction
Deep learning is the mainstream technique for many machine learning tasks, including image
recognition, machine translation, speech recognition, etc. [17]. Despite its success, the theoretical
understanding on how it works remains poor. It is well known that neural networks have great
expressive power [22, 7, 3, 8, 31]. That is, for every function there exists a set of weights on the
neural network such that it approximates the function everywhere. However, it is unclear how to
obtain the desired weights. In practice, the most commonly used method is stochastic gradient
descent based methods (e.g., SGD, Momentum [40], Adagrad [10], Adam [25]), but to the best of
our knowledge, there were no theoretical guarantees that such methods will find good weights.
In this paper, we give the first convergence analysis of SGD for two-layer feedforward network with
ReLU activations. For this basic network, it is known that even in the simplified setting where the
weights are initialized symmetrically and the ground truth forms orthonormal basis, gradient descent
might get stuck at saddle points [41].
Inspired by the structure of residual network (ResNet) [21], we add an extra identity mapping for
the hidden layer (see Figure 1). Surprisingly, we show that simply by adding this mapping, with the
standard initialization scheme and small step size, SGD always converges to the ground truth. In other
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
output
Take sum
ReLU(W> x)
output
Take sum
> x)
L ReLU((I + W)
I + W?
Identity
Link +x
W> x
I+W
Identity mapping
W> x
I
Easy for SGD
Seems hard
input x
O
input x
Figure 1: Vanilla network (left), with identity mapping (right)
Unknown
Figure 2: Illustration for our result.
words, the optimization becomes significantly easier, after adding the identity mapping. See Figure
2, based on our analysis, the region near the identity matrix I contains only one global minimum
without any saddle points or local minima, thus is easy for SGD to optimize. The role of the identity
mapping here, is to move the initial point to this easier region (better initialization).
Other than being feedforward and shallow, our network is different from ResNet in the sense that
our identity mapping skips one layer instead of two. However, as we will show in Section 5.1, the
skip-one-layer identity mapping already brings significant improvement to vanilla networks.
Formally, we consider the following function.
f (x, W) = kReLU((I + W)> x)k1
(1)
where ReLU(v) = max(v, 0) is the ReLU activation function. x ? Rd is the input vector sampled
from a Gaussian distribution, and W ? Rd?d is the weight matrix, where d is the number of input
units. Notice that I adds ei to column i of W, which makes f asymmetric in the sense that by
switching any two columns in W, we get different functions.
Following the standard setting [34, 41], we assume that there exists a two-layer teacher network with
weight W? . We train the student network using `2 loss:
L(W) = Ex [(f (x, W) ? f (x, W? ))2 ]
(2)
We will define a potential function g, and show that if g is small, the gradient points to partially
correct direction and we get closer to W? after every SGD step. However, g could be large and thus
gradient might point to the reverse direction. Fortunately, we also show that if g is large, by doing
SGD, it will keep decreasing until it is small enough while maintaining the weight W in a nice region.
We call the process of decreasing g as Phase I, and the process of approaching W? as Phase II. See
Figure 3 and simulations in Section 5.3.
Our two phases framework is fundamentally different from any type of local convergence, as in Phase
I, the gradient is pointing to the wrong direction to W? , so the path from W to W? is non-convex,
and SGD takes a long detour to arrive W? . This framework could be potentially useful for analyzing
other non-convex problems.
To support our theory, we have done a few other experiments and got interesting observations.
For example, as predicted by our theorem, we found that for multilayer feedforward network with
identity mappings, zero initialization performs as good as random initialization. At the first glance, it
contradicts the common belief ?random initialization is necessary to break symmetry?, but actually
the identity mapping itself serves as the asymmetric component. See Section 5.4.
Another common belief is that neural network has lots of local minima and saddle points [9], so
even if there exists a global minimum, we may not be able to arrive there. As a result, even when
the teacher network is shallow, the student network usually needs to be deeper, otherwise it will
underfit. However, both our theorem and our experiment show that if the shallow teacher network
is in a pretty large region near identity (Figure 2), SGD always converges to the global minimum
by initializing the weights I + W in this region, with equally shallow student network. By contrast,
wrong initialization gets stuck at local minimum and underfit. See Section 5.2.
2
Related Work
Expressivity. Even two-layer network has great expressive power. For example, two-layer network
with sigmoid activations could approximate any continuous function [22, 7, 3]. ReLU is the state-ofthe-art activation function [30, 13], and has great expressive power as well [29, 32, 31, 4, 26].
Learning. Most previous results on learning neural network are negative [39, 28, 38], or positive but
with algorithms other than SGD [23, 43, 37, 14, 15, 16], or with strong assumptions on the model
[1, 2]. [35] proved that with high probability, there exists a continuous decreasing path from random
initial point to the global minimum, but SGD may not follow this path. Recently, Zhong et al. showed
that with initialization point found using tensor decomposition, gradient descent could find the ground
truth for one hidden layer network [44].
Linear network and independent activation. Some previous works simplified the model by ignoring the activation functions and considering deep linear networks [36, 24] or deep linear residual
networks [19], which can only learn linear functions. Some previous results are based on independent
activation assumption that the activations of ReLU and the input are independent [5, 24].
Saddle points. It is observed that saddle point is not a big problem for neural networks [9, 18]. In
general, if the objective is strict-saddle [11], SGD could escape all saddle points.
2
Preliminaries
Denote x as the input vector in Rd . For now, we first consider x sampled from normal distribution
N (0, I). Denote W? = (w1? , ? ? ? , wn? ) ? Rd?d as the weights for the teacher network, W =
(w1 , ? ? ? , wn ) ? Rd?d as the weights for the student network, where wi? , wi ? Rd are column
vectors. f (x, W? ), f (x, W) are defined in (1), representing the teacher and student network.
We want to know whether a randomly initialized W will converge to W? , if we run SGD with l2
loss defined in (2). Alternatively, we can write the loss L(W) as
Ex [(?i ReLU(hei + wi , xi) ? ?i ReLU(hei + wi? , xi))2 ]
Taking derivative with respect to wj , we get
"
!
#
X
X
?
?L(W)j = 2Ex
ReLU(hei + wi , xi) ?
ReLU(hei + wi , xi) x1hej +wj ,xi?0
i
i
where 1e is the indicator function that equals 1 if the event e is true, and 0 otherwise. Here
?L(W) ? Rd?d , and ?L(W)j is its j-th column.
Denote ?i,j as the angle between ei + wi and ej + wj , ?i? ,j as the angle between ei + wi? and ej + wj .
v
Denote v? = kvk
. Denote I + W? and I + W? as the column-normalized version of I + W? and
2
I + W such that every column has unit norm. Since the input is from a normal distribution, one can
compute the expectation inside the gradient as follows.
Pd ? ?
Lemma 2.1 (Eqn (13) from [41]). If x ? N (0, I), then ??L(W)j =
i=1 2 (wi ? wi ) +
?
?
?
?
?
?
2 ? ?i ,j (ei + wi ) ? 2 ? ?i,j (ei + wi ) + kei + wi k2 sin ?i ,j ? kei + wi k2 sin ?i,j ej + wj
Remark. Although the gradient of ReLU is not well defined at the point of zero, if we assume input x
is from the Gaussian distribution, the loss function becomes smooth, and the gradient is well defined
everywhere.
Denote u ? Rd as the all one vector. Denote Diag(W) as the diagonal matrix of matrix W,
Diag(v) as a diagonal matrix whose main diagonal equals to the vector v. Denote Off-Diag(W) ,
W ? Diag(W). Denote [d] as the set {1, ? ? ? , d}. Throughout the paper, we abuse the notation of
inner product between matrices W, W? , ?L(W), such that h?L(W), Wi means the summation of
the entrywise products. kWk2 is the spectral norm of W, and kWkF is the Frobenius norm of W.
We define the potential function g and variables gj , Aj , A below, which will be useful in the proof.
Pd
Definition 2.2. We define the potential function g , i=1 (kei + wi? k2 ? kei + wi k2 ), and variable
P
gj , i6=j (kei + wi? k2 ? kei + wi k2 ).
3
W1
15
W6
10
5
W10
0
W?
?5
200
150
200
150
100
100
50
50
0
Figure 3: Phase I: W1 ? W6 , W may go to
the wrong direction but the potential is shrinking.
Phase II: W6 ? W10 , W gets closer to W? in
every step by one point convexity.
Definition 2.3. Denote Aj ,
wi? )ei
+
>
wi?
? (ei + wi )ei +
0
Figure 4: The function is one point strongly convex as every point?s negative gradient points to
the center, but not convex as any line between
the center and the red region is below surface.
?
? > ? (e + w )e + w > ), A
i
i i
i
i6=j ((ei + wi )ei + wi
>
>
>
?
?
wi ) = (I + W )I + W ? (I + W)I + W .
P
,
Pd
i=1 ((ei
+
In this paper, we consider the standard SGD with mini batch method for training the neural network.
Assume W0 is the initial point, and in step t > 0, we have the following updating rule:
Wt+1 = Wt ? ?t Gt
where the stochastic gradient Gt = ?L(Wt ) + Et with E[Et ] = 0 and kEt kF ? ?. Let G2 ,
6d? + ?, GF , 6d1.5 ? + ?, where ? is the upper bound of kW? k2 and kW0 k2 (defined later). As
we will see in Lemma C.2, they are the upper bound of kGt k2 and kGt kF respectively.
It?s clear that L is not convex, In order to get convergence guarantees, we need a weaker condition
called one point convexity.
Definition 2.4 (One point strongly convexity). A function f (x) is called ?-one point strongly convex
in domain D with respect to point x? , if ?x ? D, h??f (x), x? ? xi > ?kx? ? xk22 .
By definition, if a function f is strongly convex, it is also one point strongly convex in the entire space
with respect to the global minimum. However, the reverse is not necessarily true, e.g., see Figure
4. If a function is one point strongly convex, then in every step a positive fraction of the negative
gradient is pointing to the optimal point. As long as the step size is small enough, we will finally
arrive the optimal point, possibly by a winding path. See Figure 3 for illustration, where starting from
W6 (Phase II), we get closer to W? in every step. Formally, we have the following lemma.
Lemma 2.5. For function f (W), consider the SGD update Wt+1 = Wt ? ?Gt , where E[Gt ] =
?f (Wt ), E[kGt k2F ] ? G2 . Suppose for all t, Wt is always inside the ?-one point strongly convex
region with diameter D, i.e., kWt ? W? kF ? D. Then for any ? > 0 and any T such that
(1+?) log T
T G2
D2 ?2
, we have EkWT ? W? k2F ? (1+?)?2log
.
T ? log T ? (1+?)G
2 , if ? =
?T
T
The proof can be found in Appendix J. Lemma 2.5 uses fixed step size, so it easily fits the standard
practical scheme that shrinks ? by a factor of 10 after every a few epochs. For example, we may
apply Lemma 2.5 every time ? gets changed. Notice that our lemma does not imply that WT will
converge to W? . Instead, it only says WT will be sufficiently close to W? with small step size ?.
3
Main Theorem
Theorem 3.1 (Main Theorem). There exists constants ? > ?0 > 0 such that If x ? N (0, I),
kW0 k2 , kW? k2 ? ?0 , d ? 100, ? ? ? 2 , then SGD for L(W) will find the ground truth W? by
?2
two phases. In Phase I, by setting ? ? G
2 , the potential function will keep decreasing until it is
?
T log T ?
36d
1004 (1+?)G2F
, if we set ? =
2
1
16? steps.
(1+?) log T
,
?T
smaller than 197? 2 , which takes at most
In Phase II, for any ? > 0 and any T such that
we have EkWT ? W? k2F ?
1002 (1+?) log T G2F
9T
.
?
Remarks. Randomly initializing the weights with O(1/ d) is standard ?
in deep learning, see
[27, 12, 20]. It is also well known that if the entries are initialized with O(1/ d), the spectral norm
4
of the random matrix is O(1) [33]. So our result matches with the common practice. Moreover, as we
will show in Section 5.5, networks with small average spectral norm already have good performance.
Thus, our assumption kW? k2 = O(1) is reasonable. Notice that here we assume the spectral
norm
?
?
?
of W to be constant, which means the Frobenius norm kW kF could be as big as O( d).
The assumption that the input follows a Gaussian distribution is not necessarily true in practice
(Although this is a common assumption appeared in the previous papers [5, 41, 42], and also
considered plausible in [6]). We could easily generalize the analysis to rotation invariant distributions,
and potentially more general distributions (see Section 6). Moreover, previous analyses either ignore
the nonlinear activations and thus consider linear model [36, 24, 19], or directly [5, 24] or indirectly
[41]1 assume that the activations are independent. By contrast, in our model the ReLU activations
are highly correlated2 as kWk2 , kW? k2 = ?(1). As pointed out by [6], eliminating the unrealistic
assumptions on activation independence is the central problem of analyzing the loss surface of neural
network, which was not fully addressed by the previous analyses.
To prove the main theorem, we split the process and present the following two theorems, which will
be proved in Appendix C and D.
Theorem 3.2 (Phase I). There exists a constant ? > ?0 > 0 such that If kW0 k2 , kW? k2 ? ?0 ,
?2
2
d ? 100, ? ? G
2 , ? ? ? , then gt will keep decreasing by a factor of 1 ? 0.5?d for every step,
2
until gt1 ? 197? 2 for step t1 ?
1
kWT k2 ? 100
and gT ? 0.1.
1
16? .
After that, Phase II starts. That is, for every T > t1 , we have
Theorem 3.3 (Phase II). There exists a constant ? such that if kWk2 , kW? k2 ? ?, and g ? 0.1,
Pd
then h??L(W), W? ? Wi = j=1 h??L(W)j , wj? ? wj i > 0.03kW? ? Wk2F .
With these two theorems, we get the main theorem immediately.
Proof for Theorem 3.1. By Theorem 3.2, we know the statement for Phase I is true, and we will enter
1
phase II in 16?
steps. After entering Phase II, based on Theorem 3.3, we simply use Lemma 2.5 by
setting ? = 0.03, D =
4
?
d
50 , G
= GF to get the convergence guarantee.
Overview of the Proofs
General Picture. In many convergence analyses for non-convex functions, one would like to show
that L is one point strongly convex, and directly apply Lemma 2.5 to get the convergence result.
However, this is not true for 2-layer neural network, as the gradient may point to the wrong direction,
see Section 5.3.
So when is our L one point convex? Consider the following thought experiment: First, suppose
kWk2 , kW? k2 ? 0, we know kwi k2 , kwi? k2 also go to 0. Thus, ei + wi and ei + wi? are close to ei .
As a result, ?i,j , ?i? ,j ? ?2 , and ?i? ,i ? 0. Based on Lemma 2.1, this gives us a na?ve approximation
Pd
P
of the negative gradient, i.e., ??L(W)j ? ?2 (wj? ? wj ) + ?2 i=1 (wi? ? wi ) + ej + wj i6=j (kei +
wi? k2 ? kei + wi k2 ) .
Pd
While the first two terms ?2 (wj? ?wj ) and ?2 i=1 (wi? ?wi ) have positive inner product with W? ?W,
P
the last term gj = ej + wj i6=j (kei +wi? k2 ? kei +wi k2 ) can point to arbitrary direction. If the last
term is small, it can be covered by the first two terms, and L becomes one point strongly convex. So
Pd
we define a potential function closely related to the last term: g = i=1 (kei + wi? k2 ? kei + wi k2 ).
We show that if g is small enough, L is also one point strongly convex (Theorem 3.3).
?
However, from random initialization, g can be as large as of ?( d), which is too big to be covered.
Fortunately, we show that if g is big, it will gradually decrease simply by doing SGD on L. More
specifically, we introduce a two phases convergence analysis framework:
1
They assume input is Gaussian and the W? is orthonormal, which means the activations are independent in
teacher network.
P
2
Let ?i be the output of i-th ReLU unit, then in our setting, i,j Cov[?i , ?j ] can be as large as ?(d), which
is far from being independent.
5
h
Constant
Part
+
First
Order
+
Higher
Order
, W? ? W
i
?
2
?
2
?
2
? [?
2 ? O(g)]kW ? WkF ?1.3kW ? WkF ?0.085kW ? WkF
Lemma D.1
Lemma B.2
Lemma D.2 + Lemma D.3
Figure 5: Lower bounds of inner product using Taylor expansion
1. In Phase I, the potential function g is decreasing to a small value.
2. In Phase II, g remains small, so L is one point convex and thus W starts to converge to W? .
We believe that this framework could be helpful for other non-convex problems.
Technical difficulty: Phase I. Our key technical challenge is to show that in Phase I, the potential
function actually decreases to O(1) after polynomial number of iterations. However, we cannot show
this by merely looking at g itself. Instead, we introduce an auxiliary variable s = (W? ? W)u,
where u is the all one vector. By doing a careful calculation, we get their joint update rules (Lemma
C.3 and Lemma C.4):
?
?
st+1 ? st ? ??d
d?)
2 st + ?O( ?dgt +
gt+1 ? gt ? ?dgt + ?O(? dkst k2 + d? 2 )
Solving this dynamics, we can show that gt will approach to (and stay around) O(?), thus we enter
Phase II.
Technical difficulty: Phase II. Although the overall approximation in the thought experiment looks
simple, the argument is based on an over simplified assumption that ?i? ,j , ?i,j ? ?2 for i 6= j.
However, when W? has constant spectral norm, even when W is very close to W? , ?i,j ? could be
constantly far away from ?2 , which prevents us from applying this approximation directly. To get a
formal proof, we use the standard Taylor expansion and control the higher order terms. Specifically,
we write ?i? ,j as ?i? ,j = arccoshei + wi? , ej + wj i and expand arccos at point 0, thus,
?i? ,j =
?
? hei + wi? , ej + wj i + O(hei + wi? , ej + wj i3 )
2
However, even when W ? W? , the higher order term O(hei + wi? , ej + wj i3 ) still can be as large
as a constant, which is too big for us. Our trick here is to consider the ?joint Taylor expansion?:
?i? ,j ? ?i,j = hei + wi ? ei + wi? , ej + wj i + O(|hei + wi? , ej + wj i3 ? hei + wi , ej + wj i3 |)
As W approaches W? , |hei + wi? , ej + wj i3 ? hei + wi , ej + wj i3 | also tends to zero, therefore
our approximation has bounded error.
In the thought experiment, we already know that the constant part in the Taylor expansion of ?L(W)
is ?2 ? O(g)-one point convex. We show that after taking inner product with W? ? W, the first
order terms are lower bounded by (roughly) ?1.3kW? ? Wk2F and the higher order terms are lower
bounded by ?0.085kW? ? Wk2F . Adding them together, we can see that L(W) is one point convex
as long as g is small. See Figure 5.
Geometric Lemma. In order to get through the whole analysis, we need tight bounds on a few
common terms that appear everywhere. Instead of using na?ve algebraic techniques, we come up with
a nice geometric proof to get nearly optimal bounds. Due to space limit, we defer it to Appendix E.
5
Experiments
In this section, we present several simulation results to support our theory. Our code can be found in
the supplementary materials.
5.1
Importance of identity mapping
In this experiment, we compare the standard ResNet [21] and single skip model where identity
mapping skips only one layer. See Figure 6 for the single skip model. We also ran the vanilla network,
where the identity mappings are completely removed.
6
Single skip
9.01%
?
Vanilla
12.04%
ReLU
ResNet
6.97%
BatchNorm
Test Err
Convolution
input
Table 1: Test error of three 56-layer networks on
Cifar-10
output
Identity
Figure 6: Illustration of one block in single skip
model in Sec 5.1
10
20.0
Test (ResLink)
Test (Vanilla)
Test (3-Block)
Train (ResLink)
Train (Vanilla)
Train (3-Block)
loss
6
4
17.5
15.0
W (ResLink)
W-W* (ResLink)
W (Vanilla)
W-W* (Vanilla)
12.5
l2 norm
8
10.0
7.5
5.0
2
2.5
0
0
25
50
75
100
epochs
125
150
175
0.0
200
(a) Test Error, Train Error
0
(b)
25
50
75
100
epochs
125
150
175
200
?
kW ? WkF , kWkF
Figure 7: Verifying the global convergence
In this experiment, we choose Cifar-10 as the dataset, and all the networks have 56-layers. Other than
the identity mappings, all other settings are identical and default. We run the experiments for 5 times
and report the average test error. As we can see in Table 1, compared with vanilla network, by simply
using a single skip identity mapping, one can already improve the test error by 3.03%, and is 2.04%
close to the ResNet. So single skip identity mapping brings significant improvement on test accuracy.
5.2
Global minimum convergence
In this experiment, we verify our main theorem that for two-layer teacher network and student network
with identity mappings, as long as kW0 k2 , kW? k2 is small, SGD always converges to the global
minimum W? , thus gives almost 0 training error and test error. We consider three student networks.
The first one (ResLink) is defined using (2), the second one (Vanilla) is the same model without the
identity mapping. The last one (3-Block) is a three block network with each block containing a linear
layer (500 hidden nodes), a batch normalization and a ReLU layer. The teacher network always
shares the same structure as the student network.
The input dimension is 100. We generated a fixed W? for all the trials with kW? k2 ? 0.6, kW? kF ?
5.7. We generated a training set of size 100, 000, and test set of size 10, 000, sampled from a Gaussian
distribution. We use batch size 200, step size 0.001. We run ResLink for 5 times with random
initialization (kWk2 ? 0.6 and kWkF ? 5), and plot the curves by taking the average.
Figure 7(a) shows test error and training error of the three networks. Comparing Vanilla with 3-Block,
we find that 3-Block is more expressive, so its training error is smaller compared with vanilla network;
but it suffers from overfitting and has bigger test error. This is the standard overfitting vs underfitting
tradeoff. Surprisingly, with only one hidden layer, ResLink has both zero test error and training
error. If we look at Figure 7(b), we know the distance between W and W? converges to 0, meaning
ResLink indeed finds the global optimal in all 5 trials. By contrast, for vanilla network, which is
essentially the same network with different initialization, kW ? W? k2 does not converge to zero3 .
This is exactly what our theory predicted.
5.3
Verify the dynamics
In this experiment, we verify our claims on the dynamics. Based on the analysis, we construct a
1500 ? 1500 matrix W s.t. kWk2 ? 0.15, kWkF ? 5 , and set W? = 0. By plugging them into (2),
one can see that even in this simple case that W? = 0, initially the gradient is pointing to the wrong
direction, i.e., not one point convex. We then run SGD on W by using samples x from Gaussian
distribution, with batch size 300, step size 0.0001.
3
To make comparison meaningful, we set W ? I to be the actual weight for Vanilla as its identity mapping
is missing, which is why it has a much bigger initial norm.
7
25
16
P-I P-II
20
Distance to optimal
Inner product
Potential g
Loss
14
12
15
10
10
8
5
6
4
0
2
?5
?10
0
0
20
40
60
80
100
?2
(a) First 100 iterations
0
50000
100000 150000 200000 250000 300000 350000
(b) The entire process
Figure 8: Verifying the dynamics
Figure 8(a) shows the first 100 iterations. We can see that initially the inner product defined in
Definition 2.4 is negative, then after about 15 iterations, it turns positive, which means W is in the
one point strongly convex region. At the same time, the potential g keeps decreasing to a small value,
while the distance to optimal (which also equals to kWkF in this experiment) is not affected. They
precisely match with our description of Phase I in Theorem 3.2.
After that, we enter Phase II and slowly approach to W? , see Figure 8(b). Notice that the potential
g is always very small, the inner product is always positive, and the distance to optimal is slowly
decreasing. Again, they precisely match with our Theorem 3.3.
5.4
Zero initialization works
In this experiment, we used a simple 5-block neural network on MNIST, where every block contains
a 784 ? 784 feedforward layer, an identity mapping, and ?
a ReLU layer. Cross entropy criterion is
used. We compare zero initialization with standard O(1/ d) random initialization. We found that
for zero initialization, we can get 1.28% test error, while for random initialization, we can get 1.27%
test error. Both results were obtained by taking average among 5 runs and use step size 0.1, batch
size 256. If the identity mapping is removed, zero initialization no longer works.
5.5
Spectral norm of W?
We also applied the exact model f defined in (1) to distinguish two classes in MNIST. For any input
image x, We say it?s in class A if f (x, W) < TA,B , and in class B otherwise. Here TA,B is the
optimal threshold for the function f (x, 0) to distinguish A and B. If W = 0, we get 7% training
error for distinguish class 0 and class 1. However, it can be improved to 1% with kWk2 = 0.6.
We tried this experiment for all possible 45 pairs of classes in MNIST, and improve the average
training error from 34% (using W = 0) to 14% (using kWk2 = 0.6). Therefore our model with
kWk2 = ?(1) has reasonable expressive power, and is substantially different from just using the
identity mapping alone.
6
Discussions
The assumption that the input is Gaussian can be relaxed in several ways. For example, when
the distribution is N (0, ?) where k? ? Ik2 is bounded by a small constant, the same result holds
with slightly worse constants. Moreover, since the analysis relies Lemma 2.1, which is proved by
converting the original input space into polar space, it is easy to generalize the calculation to rotation
invariant distributions. Finally, for more general distributions, as long as we could explicitly compute
the expectation, which is in the form of O(W? ? W) plus certain potential function, our analysis
framework may also be applied.
There are many exciting open problems. For example, Our paper is the first one that gives solid
SGD analysis for neural network with nonlinear activations, without unrealistic assumptions like
independent activation assumption. It would be great if one could further extend it to multiple layers,
which would be a major breakthrough of understanding optimization for deep learning. Moreover,
our two phase framework could be applied to other non-convex problems as well.
8
Acknowledgement
The authors want to thank Robert Kleinberg, Kilian Weinberger, Gao Huang, Adam Klivans and
Surbhi Goel for helpful discussions, and the anonymous reviewers for their comments.
References
[1] Alexandr Andoni, Rina Panigrahy, Gregory Valiant, and Li Zhang. Learning polynomials with
neural networks. In ICML, pages 1908?1916, 2014.
[2] Sanjeev Arora, Aditya Bhaskara, Rong Ge, and Tengyu Ma. Provable bounds for learning
some deep representations. In Proceedings of the 31th International Conference on Machine
Learning, ICML 2014, Beijing, China, 21-26 June 2014, pages 584?592, 2014.
[3] Andrew R. Barron. Universal approximation bounds for superpositions of a sigmoidal function.
IEEE Trans. Information Theory, 39(3):930?945, 1993.
[4] Leo Breiman. Hinging hyperplanes for regression, classification, and function approximation.
IEEE Trans. Information Theory, 39(3):999?1013, 1993.
[5] Anna Choromanska, Mikael Henaff, Micha?l Mathieu, G?rard Ben Arous, and Yann LeCun.
The loss surfaces of multilayer networks. In AISTATS, 2015.
[6] Anna Choromanska, Yann LeCun, and G?rard Ben Arous. Open problem: The landscape of
the loss surfaces of multilayer networks. In Proceedings of The 28th Conference on Learning
Theory, COLT 2015, Paris, France, July 3-6, 2015, pages 1756?1760, 2015.
[7] George Cybenko. Approximation by superpositions of a sigmoidal function. MCSS, 5(4):455,
1992.
[8] Amit Daniely, Roy Frostig, and Yoram Singer. Toward deeper understanding of neural networks:
The power of initialization and a dual view on expressivity. In NIPS, pages 2253?2261, 2016.
[9] Yann N Dauphin, Razvan Pascanu, Caglar Gulcehre, Kyunghyun Cho, Surya Ganguli, and
Yoshua Bengio. Identifying and attacking the saddle point problem in high-dimensional nonconvex optimization. In NIPS 2014, pages 2933?2941, 2014.
[10] John C. Duchi, Elad Hazan, and Yoram Singer. Adaptive subgradient methods for online
learning and stochastic optimization. Journal of Machine Learning Research, 12:2121?2159,
2011.
[11] Rong Ge, Furong Huang, Chi Jin, and Yang Yuan. Escaping from saddle points - online
stochastic gradient for tensor decomposition. In COLT 2015, volume 40, pages 797?842, 2015.
[12] Xavier Glorot and Yoshua Bengio. Understanding the difficulty of training deep feedforward
neural networks. In AISTATS, pages 249?256, 2010.
[13] Xavier Glorot, Antoine Bordes, and Yoshua Bengio. Deep sparse rectifier neural networks. In
AISTATS, pages 315?323, 2011.
[14] Surbhi Goel, Varun Kanade, Adam R. Klivans, and Justin Thaler. Reliably learning the relu in
polynomial time. CoRR, abs/1611.10258, 2016.
[15] Surbhi Goel and Adam Klivans. Eigenvalue decay implies polynomial-time learnability for
neural networks. In NIPS 2017, 2017.
[16] Surbhi Goel and Adam Klivans. Learning Depth-Three Neural Networks in Polynomial Time.
ArXiv e-prints, 2017.
[17] Ian Goodfellow, Yoshua Bengio, and Aaron Courville. Deep Learning. MIT Press, 2016.
http://www.deeplearningbook.org.
[18] Ian J. Goodfellow and Oriol Vinyals. Qualitatively characterizing neural network optimization
problems. CoRR, abs/1412.6544, 2014.
9
[19] Moritz Hardt and Tengyu Ma. Identity matters in deep learning. CoRR, abs/1611.04231, 2016.
[20] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Delving deep into rectifiers:
Surpassing human-level performance on imagenet classification. In ICCV, pages 1026?1034,
2015.
[21] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image
recognition. In CVPR, pages 770?778, 2016.
[22] Kurt Hornik, Maxwell B. Stinchcombe, and Halbert White. Multilayer feedforward networks
are universal approximators. Neural Networks, 2(5):359?366, 1989.
[23] Majid Janzamin, Hanie Sedghi, and Anima Anandkumar. Beating the perils of non-convexity:
Guaranteed training of neural networks using tensor methods. arXiv preprint arXiv:1506.08473,
2015.
[24] Kenji Kawaguchi. Deep learning without poor local minima. In NIPS, pages 586?594, 2016.
[25] Diederik P. Kingma and Jimmy Ba. Adam: A method for stochastic optimization. CoRR,
abs/1412.6980, 2014.
[26] J. M. Klusowski and A. R. Barron. Risk Bounds for High-dimensional Ridge Function Combinations Including Neural Networks. ArXiv e-prints, July 2016.
[27] Yann LeCun, Leon Bottou, Genevieve B. Orr, and Klaus Robert M?ller. Efficient BackProp,
pages 9?50. Springer Berlin Heidelberg, Berlin, Heidelberg, 1998.
[28] Roi Livni, Shai Shalev-Shwartz, and Ohad Shamir. On the computational efficiency of training
neural networks. In NIPS, pages 855?863, 2014.
[29] Guido F. Mont?far, Razvan Pascanu, KyungHyun Cho, and Yoshua Bengio. On the number of
linear regions of deep neural networks. In NIPS, pages 2924?2932, 2014.
[30] Vinod Nair and Geoffrey E. Hinton. Rectified linear units improve restricted boltzmann
machines. In ICML, pages 807?814, 2010.
[31] Xingyuan Pan and Vivek Srikumar. Expressiveness of rectifier networks. In ICML, pages
2427?2435, 2016.
[32] Razvan Pascanu, Guido Mont?far, and Yoshua Bengio. On the number of inference regions of
deep feed forward networks with piece-wise linear activations. CoRR, abs/1312.6098, 2013.
[33] M. Rudelson and R. Vershynin. Non-asymptotic theory of random matrices: extreme singular
values. ArXiv e-prints, 2010.
[34] David Saad and Sara A. Solla. Dynamics of on-line gradient descent learning for multilayer
neural networks. Advances in Neural Information Processing Systems, 8:302?308, 1996.
[35] Itay Safran and Ohad Shamir. On the quality of the initial basin in overspecified neural networks.
In ICML, pages 774?782, 2016.
[36] Andrew M. Saxe, James L. McClelland, and Surya Ganguli. Exact solutions to the nonlinear
dynamics of learning in deep linear neural networks. CoRR, abs/1312.6120, 2013.
[37] Hanie Sedghi and Anima Anandkumar. Provable methods for training neural networks with
sparse connectivity. ICLR, 2015.
[38] Ohad Shamir.
Distribution-specific hardness of learning neural networks.
abs/1609.01037, 2016.
CoRR,
[39] Jir? S?ma. Training a single sigmoidal neuron is hard. Neural Computation, 14(11):2709?2728,
2002.
[40] Ilya Sutskever, James Martens, George E. Dahl, and Geoffrey E. Hinton. On the importance of
initialization and momentum in deep learning. In ICML, pages 1139?1147, 2013.
10
[41] Yuandong Tian. Symmetry-breaking convergence analysis of certain two-layered neural networks with relu nonlinearity. In Submitted to ICLR 2017, 2016.
[42] Bo Xie, Yingyu Liang, and Le Song. Diversity leads to generalization in neural networks. In
AISTATS, 2017.
[43] Yuchen Zhang, Jason D. Lee, Martin J. Wainwright, and Michael I. Jordan. Learning halfspaces
and neural networks with random initialization. CoRR, abs/1511.07948, 2015.
[44] Kai Zhong, Zhao Song, Prateek Jain, Peter L. Bartlett, and Inderjit S. Dhillon. Recovery
guarantees for one-hidden-layer neural networks. In ICML 2017, 2017.
11
| 6662 |@word trial:2 version:1 eliminating:1 polynomial:6 seems:1 norm:11 open:2 d2:1 simulation:2 tried:1 decomposition:2 sgd:26 solid:1 arous:2 initial:6 contains:2 kurt:1 err:1 comparing:1 activation:18 diederik:1 john:1 hanie:2 plot:1 update:2 v:1 alone:1 pascanu:3 node:1 hyperplanes:1 sigmoidal:3 org:1 zhang:4 become:1 yuan:2 prove:2 inside:2 yingyu:1 underfitting:1 introduce:2 hardness:1 indeed:1 roughly:1 multi:1 chi:1 inspired:1 decreasing:8 actual:1 considering:1 becomes:3 notation:1 moreover:4 bounded:4 what:1 prateek:1 substantially:1 guarantee:4 every:12 exactly:1 wrong:6 k2:31 control:1 unit:4 appear:1 positive:5 t1:2 local:5 tends:1 limit:1 switching:1 despite:1 analyzing:2 path:4 abuse:1 might:2 plus:1 initialization:20 china:1 sara:1 micha:1 tian:1 unique:1 practical:1 lecun:3 alexandr:1 practice:4 block:10 differs:1 razvan:3 universal:2 significantly:1 got:1 thought:3 word:1 klusowski:1 get:19 cannot:1 close:4 layered:1 risk:1 applying:1 optimize:1 www:1 reviewer:1 missing:2 center:2 marten:1 go:2 starting:1 jimmy:1 convex:24 identifying:1 immediately:1 recovery:1 rule:2 surbhi:4 orthonormal:2 shamir:3 suppose:2 itay:1 exact:2 guido:2 us:1 goodfellow:2 trick:1 roy:1 recognition:3 updating:1 asymmetric:3 srikumar:1 observed:1 role:1 preprint:1 enters:1 initializing:2 verifying:2 region:11 wj:22 rina:1 kilian:1 sun:2 solla:1 decrease:3 removed:2 ran:1 halfspaces:1 pd:7 convexity:4 dynamic:6 solving:1 tight:1 efficiency:1 basis:1 completely:1 easily:2 joint:2 leo:1 train:6 jain:1 krelu:1 klaus:1 shalev:1 whose:1 supplementary:1 plausible:1 elad:1 say:2 cvpr:1 otherwise:3 kai:1 cov:1 itself:2 online:2 eigenvalue:1 product:8 description:1 frobenius:2 sutskever:1 convergence:14 adam:6 converges:7 ben:2 resnet:5 batchnorm:1 andrew:2 progress:1 strong:1 auxiliary:1 c:2 skip:9 predicted:2 implies:1 kenji:1 come:1 direction:8 closely:1 correct:1 kgt:3 stochastic:6 human:1 saxe:1 material:1 backprop:1 generalization:1 preliminary:1 anonymous:1 cybenko:1 safran:1 summation:1 rong:2 hold:1 sufficiently:1 considered:1 ground:4 normal:4 around:1 great:4 roi:1 mapping:26 claim:2 pointing:3 major:1 polar:1 superposition:2 tool:1 gt1:1 mit:1 gaussian:8 always:7 i3:6 ej:14 cornell:2 zhong:2 breiman:1 june:1 improvement:2 contrast:3 sense:2 helpful:2 inference:1 ganguli:2 entire:2 initially:2 hidden:5 expand:1 choromanska:2 france:1 overall:1 among:1 classification:2 colt:2 dual:1 dauphin:1 arccos:1 art:1 yuanzhil:1 special:1 breakthrough:1 equal:3 construct:1 beach:1 identical:1 kw:19 look:2 k2f:3 nearly:1 icml:7 report:1 yoshua:6 fundamentally:1 escape:1 few:3 randomly:2 ve:2 kwt:2 phase:29 ab:8 highly:1 genevieve:1 extreme:1 kvk:1 closer:3 necessary:2 janzamin:1 ohad:3 detour:1 taylor:4 initialized:3 desired:1 yuchen:1 halbert:1 theoretical:3 column:6 subset:2 entry:1 daniely:1 too:2 learnability:1 teacher:8 gregory:1 cho:2 vershynin:1 st:4 international:1 stay:1 lee:1 off:1 michael:1 together:1 ilya:1 na:2 connectivity:1 w1:4 again:1 central:1 sanjeev:1 containing:1 choose:1 possibly:1 slowly:2 huang:2 ket:1 worse:1 wkf:4 derivative:1 zhao:1 li:2 potential:13 diversity:1 orr:1 sec:1 student:8 matter:1 explicitly:1 piece:1 later:1 break:1 lot:1 view:1 jason:1 doing:3 hazan:1 red:1 start:2 shai:1 defer:1 accuracy:1 largely:1 ofthe:1 landscape:1 peril:1 generalize:2 mc:1 ren:2 rectified:1 anima:2 submitted:1 suffers:1 overspecified:1 definition:5 james:2 proof:6 sampled:3 proved:3 dataset:1 hardt:1 knowledge:1 actually:2 furong:1 maxwell:1 higher:4 ta:2 varun:1 follow:1 feed:1 xie:1 improved:1 entrywise:1 rard:2 done:1 shrink:1 strongly:11 just:1 until:3 eqn:1 expressive:5 ei:15 nonlinear:3 dgt:2 glance:1 brings:2 aj:2 quality:1 believe:1 usa:1 normalized:1 true:5 verify:3 xavier:2 kyunghyun:2 entering:1 moritz:1 dhillon:1 white:1 vivek:1 sin:2 criterion:1 ridge:1 performs:1 duchi:1 image:3 meaning:1 wise:1 recently:1 common:5 sigmoid:1 rotation:2 deeplearningbook:1 overview:1 volume:1 extend:1 he:2 approximates:1 yuandong:1 surpassing:1 kwk2:9 significant:2 enter:3 rd:8 vanilla:16 i6:4 pointed:1 frostig:1 nonlinearity:1 longer:1 mainstream:1 gj:3 etc:1 add:2 surface:4 gt:9 recent:1 showed:1 henaff:1 reverse:2 certain:2 nonconvex:1 success:1 approximators:1 minimum:13 fortunately:2 relaxed:1 george:2 goel:4 converting:1 converge:4 attacking:1 xiangyu:2 ller:1 july:2 ii:14 multiple:1 smooth:1 technical:3 match:3 characterized:1 calculation:2 cross:1 long:6 cifar:2 equally:1 bigger:2 plugging:1 basic:1 regression:1 multilayer:5 essentially:1 expectation:2 arxiv:5 iteration:4 normalization:1 want:2 addressed:1 singular:1 jian:2 extra:1 saad:1 unlike:1 strict:1 kwi:2 comment:1 majid:1 kwkf:5 jordan:1 call:1 anandkumar:2 near:2 yang:2 symmetrically:1 feedforward:7 split:1 easy:3 enough:3 wn:2 bengio:6 independence:1 relu:21 fit:1 vinod:1 approaching:1 escaping:1 inner:7 tradeoff:1 whether:1 bartlett:1 song:2 peter:1 algebraic:1 speech:1 shaoqing:2 remark:2 deep:17 useful:2 clear:1 covered:2 mcclelland:1 diameter:1 http:1 notice:4 write:2 xingyuan:1 affected:1 key:1 threshold:1 dahl:1 subgradient:1 merely:1 fraction:1 year:1 sum:2 beijing:1 run:5 mystery:1 everywhere:3 angle:2 place:1 arrive:3 throughout:1 reasonable:2 almost:1 yann:4 appendix:3 layer:23 bound:8 guaranteed:1 distinguish:3 courville:1 precisely:2 kleinberg:1 argument:1 klivans:4 leon:1 tengyu:2 martin:1 department:2 combination:1 poor:2 smaller:2 slightly:1 pan:1 contradicts:1 wi:48 shallow:4 gradually:2 invariant:2 iccv:1 restricted:1 xk22:1 remains:2 hei:12 kw0:4 turn:1 singer:2 know:5 ge:2 serf:1 gulcehre:1 yuanzhi:1 apply:2 away:1 spectral:6 indirectly:1 barron:2 batch:5 weinberger:1 original:1 rudelson:1 maintaining:1 mikael:1 yoram:2 k1:1 amit:1 kawaguchi:1 tensor:3 move:2 objective:1 already:4 print:3 traditional:1 diagonal:3 unclear:1 antoine:1 gradient:19 iclr:2 distance:4 link:1 thank:1 berlin:2 w0:1 toward:1 provable:2 sedghi:2 w6:4 panigrahy:1 code:1 illustration:3 providing:1 mini:1 liang:1 robert:2 potentially:2 wk2f:3 statement:1 negative:5 ba:1 reliably:1 boltzmann:1 unknown:1 upper:2 observation:1 convolution:1 neuron:1 caglar:1 descent:5 jin:1 hinton:2 looking:1 arbitrary:1 expressiveness:1 david:1 complement:1 pair:1 paris:1 imagenet:1 expressivity:2 kingma:1 nip:7 trans:2 able:2 justin:1 usually:1 below:2 beating:1 appeared:1 challenge:1 including:2 max:1 belief:2 stinchcombe:1 power:5 event:1 unrealistic:2 difficulty:3 wainwright:1 indicator:1 residual:3 representing:1 scheme:2 improve:3 thaler:1 imply:1 picture:1 mathieu:1 arora:1 jir:1 gf:2 nice:3 understanding:6 l2:2 epoch:3 kf:5 geometric:2 adagrad:1 acknowledgement:1 asymptotic:1 loss:9 fully:1 interesting:1 geoffrey:2 g2f:2 basin:1 exciting:1 share:1 bordes:1 translation:1 changed:1 surprisingly:2 last:4 mont:2 formal:2 weaker:1 deeper:2 ik2:1 taking:4 characterizing:1 livni:1 sparse:2 curve:1 default:1 dimension:1 depth:1 rich:1 stuck:2 commonly:1 author:1 adaptive:1 simplified:3 qualitatively:1 kei:12 far:4 forward:1 approximate:1 ignore:1 keep:4 global:11 overfitting:2 xi:6 shwartz:1 alternatively:1 surya:2 continuous:2 why:2 pretty:1 table:2 kanade:1 learn:1 delving:1 ca:1 ignoring:1 symmetry:2 hornik:1 heidelberg:2 expansion:4 bottou:1 necessarily:2 domain:1 diag:4 anna:2 aistats:4 main:6 underfit:2 big:5 whole:1 verifies:1 shrinking:1 momentum:2 breaking:1 ian:2 bhaskara:1 theorem:19 rectifier:3 specific:1 decay:1 glorot:2 exists:7 mnist:3 andoni:1 adding:3 valiant:1 importance:2 corr:8 kx:1 easier:2 hinging:1 entropy:1 simply:4 saddle:9 gao:1 prevents:1 aditya:1 vinyals:1 kaiming:2 partially:1 g2:3 bo:1 inderjit:1 springer:1 truth:4 constantly:1 relies:1 ma:3 w10:2 nair:1 identity:29 careful:1 experimentally:1 hard:2 specifically:2 wt:9 lemma:18 called:3 meaningful:1 aaron:1 formally:2 support:2 oriol:1 princeton:2 d1:1 ex:3 |
6,260 | 6,663 | Doubly Accelerated
Stochastic Variance Reduced Dual Averaging Method
for Regularized Empirical Risk Minimization
Tomoya Murata
NTT DATA Mathematical Systems Inc. , Tokyo, Japan
[email protected]
Taiji Suzuki
Department of Mathematical Informatics
Graduate School of Information Science and Technology, The University of Tokyo, Tokyo, Japan
PRESTO, Japan Science and Technology Agency, Japan
Center for Advanced Integrated Intelligence Research, RIKEN, Tokyo, Japan
[email protected]
Abstract
We develop a new accelerated stochastic gradient method for efficiently solving
the convex regularized empirical risk minimization problem in mini-batch settings.
The use of mini-batches has become a golden standard in the machine learning
community, because the mini-batch techniques stabilize the gradient estimate and
can easily make good use of parallel computing. The core of our proposed method
is the incorporation of our new ?double acceleration? technique and variance reduction technique. We theoretically analyze our proposed method and show that
our method much improves the mini-batch efficiencies
of previous accelerated
?
stochastic methods, and essentially only needs size n mini-batches for achieving
the optimal iteration complexities for both non-strongly and strongly convex objectives, where n is the training set size. Further, we show that even in non-mini-batch
settings, our method achieves the best known convergence rate for non-strongly
convex and strongly convex objectives.
1
Introduction
We consider a composite convex optimization problem associated with regularized empirical risk
minimization, which often arises in machine learning. In particular, our goal is to minimize the
sum of finite smooth convex functions and a relatively simple (possibly) non-differentiable convex
function by using first order methods in mini-batch settings. The use of mini-batches is now a
golden standard in the machine learning community, because it is generally more efficient to execute
matrix-vector multiplications over a mini-batch than an equivalent amount of vector-vector ones each
over a single instance; and more importantly, mini-batch techniques can easily make good use of
parallel computing.
Traditional and effective methods for solving the abovementioned problem are the ?proximal gradient?
(PG) method and ?accelerated proximal gradient? (APG) method [10, 3, 20]. These methods are
well known to achieve linear convergence for strongly convex objectives. Particularly, APG achieves
optimal iteration complexities for both non-strongly and strongly convex objectives. However, these
methods need a per iteration cost of O(nd), where n denotes the number of components of the
finite sum, and d is the dimension of the solution space. In typical machine learning tasks, n and d
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
correspond to the number of instances and features respectively, which can be very large. Then, the
per iteration cost of these methods can be considerably high.
A popular alternative is the ?stochastic gradient descent? (SGD) method [19, 5, 17]. As the per
iteration cost of SGD is only O(d) in non-mini-batch settings, SGD is suitable for many machine
learning tasks. However, SGD only achieves sublinear rates and is ultimately slower than PG and
APG.
Recently, a number of stochastic gradient methods have been proposed; they use a variance reduction
technique that utilizes the finite sum structure of the problem (?stochastic averaged gradient? (SAG)
method [15, 16], ?stochastic variance reduced gradient" (SVRG) method [6, 22] and SAGA [4]).
Even though the per iteration costs of these methods are same as that of SGD, they achieve a linear
convergence for strongly convex objectives. Consequently, these methods dramatically improve the
total computational cost of PG. However, in size b mini-batch settings, the rate is essentially b times
worse than in non-mini-batch settings (the extreme situation is b = n which corresponds to PG). This
means that there is little benefit in applying mini-batch scheme to these methods.
More recently, several authors have proposed accelerated stochastic methods for the composite finite
sum problem (?accelerated stochastic dual coordinate ascent? (ASDCA) method [18], Universal
Catalyst (UC) [8], ?accelerated proximal coordinate gradient? (APCG) method [9], ?stochastic
primal-dual coordinate? (SPDC) method [23], and Katyusha [1]). ASDCA (UC), APCG, SPDC and
2
Katyusha essentially achieve the optimal total computational cost1 for strongly convex objectives
?
in non-mini-batch settings. However, in size b mini-batch settings, the rate is essentially b times
worse than that in non-mini-batch settings, and these methods need size O(n) mini-batches for
achieving the optimal iteration complexity3 , which is essentially the same as APG. In addition,
[12, 13] has proposed the ?accelerated mini-batch proximal stochastic variance reduced gradient?
(AccProxSVRG) method and its variant, the ?accelerated efficient mini-batch stochastic variance
reduced gradient? (AMSVRG) method. In non-mini-batch settings, AccProxSVRG only achieves the
same rate as SVRG. However, in mini-batch settings, AccProxSVRG significantly improves the minibatch efficiency of non-accelerated
variance reduction methods, and surprisingly, AccProxSVRG
?
essentially only needs size O( ?) mini-batches for achieving the optimal iteration complexity for
strongly convex objectives, where ? is the condition number of the problem. However, the necessary
size of mini-batches depends on the condition number and gradually increases when the condition
number increases and ultimately matches with O(n) for a large condition number.
Main contribution
We propose a new accelerated stochastic variance reduction method that achieves better convergence
than existing methods do, and it particularly takes advantages of mini-batch settings well; it is called
the ?doubly accelerated stochastic variance reduced dual averaging? (DASVRDA) method. We
describe the main feature of our proposed method below and list the comparisons of our method with
several preceding methods in Table 1.
Our method significantly improves the mini-batch?
efficiencies of state-of-the-art methods. As a
result, our method essentially only needs size O( n) mini-batches4 for achieving the optimal
iteration complexities for both non-strongly and strongly convex objectives.
1
More precisely, the rate of ASDCA (UC) is with extra log-factors, and near but worse than the one of APCG,
SPDC and Katyusha. This means that ASDCA (UC) cannot be optimal.
2
Katyusha also achieves a near optimal total computational cost for non-strongly convex objectives.
3
We refer to ?optimal iteration complexity? as the iteration complexity of deterministic Nesterov?s acceleration method [11].
p
p
4
Actually, when L/? ? n and L/??? n, our method needs size O(n ?/L) and O(n ?/L) mini-batches,
respectively, which are larger than O( n), but smaller than O(n). Achieving optimal iteration complexity for
solving high accuracy and bad conditioned problems is much more important than doing ones with low accuracy
and well-conditioned ones, because the former needs more overall computational cost than the latter.
2
Table 1: Comparisons of our method with SVRG (SVRG++ [2]), ASDCA (UC), APCG, SPDC,
Katyusha and AccProxSVRG. n is the number of components of the finite sum, d is the dimension of
the solution space, b is the mini-batch size, L is the smoothness parameter of the finite sum, ? is the
strong convexity parameter of objectives, and ? is accuracy. ?Necessary mini-batch size? indicates
thep
order of the necessary p
size of mini-batches for achieving the optimal iteration complexities
O( L/?log(1/?)) and O( L/?) for strongly and non-strongly convex objectives, respectively.
We regard one computation of a full gradient as n/b iterations in size b mini-batch settings, for
a fair comparison. ?Unattainable? implies that the algorithm cannot achieve the optimal iteration
e hides extra log-factors.
complexity even if it uses size n mini-batches. O
SVRG (++ )
ASDCA (UC)
APCG
SPDC
Katyusha
AccProxSVRG
DASVRDA
2
?-strongly convex
Total computational cost
Necessary mini-batch size
in size
L/? ? n
otherwise
settings
b mini-batch
1
bL
Unattainable Unattainable
O d n + ? log ?
q
e d n + nbL log 1
O
Unattainable Unattainable
?
q ?
log 1?
O(n)
O(n)
O d n + nbL
q ?
O d n + nbL
log 1?
O(n)
O(n)
?
q
O d n + nbL
log 1?
O(n)
O(n)
?
q
q
p?
L
1
L
n?b L
O
O n L
O d n + n?1
? +b
? log ?
?
p?
? qL
?
1
O n L
O d n + (b + n) ? log ?
O ( n)
Non-strongly convex
Total computational cost
Necessary mini-batch size
in size b mini-batch settings
L/? ? nlog2 (1/?)
otherwise
bL
1
O d nlog ? + ?
Unattainable
Unattainable
?
e d n+? nbL
O
Unattainable
Unattainable
?
No direct analysis
Unattainable
Unattainable
No direct analysis
q
O d nlog 1? + nbL
?
Unattainable
Unattainable
O(n)
O(n)
No direct analysis
? q
O d nlog 1? + (b + n) L?
Unattainable
?
O ( n)
Unattainable
p
e n ?
O
L
Preliminary
In this section, we formally describe the problem to be considered in this paper and the assumptions
for our theory.
pP
2
We use k ? k to denote the Euclidean L2 norm k ? k2 : kxk = kxk2 =
i xi . For natural number n,
[n] denotes set {1, . . . , n}.
In this paper, we consider the following composite convex minimization problem:
def
min {P (x) = F (x) + R(x)},
x?Rd
(1)
Pn
where F (x) = n1 i=1 fi (x). Here each fi : Rd ? R is a Li -smooth convex function and
R : Rd ? R is a relatively simple and (possibly) non-differentiable convex function. Problems
of this form often arise in machine learning and fall under regularized empirical risk minimization
(ERM). In ERM problems, we are given n training examples {(ai , bi )}ni=1 , where each ai ? Rd
is the feature vector of example i, and each bi ? R is the label of example i. Important examples
of ERM in our setting include linear regression and logistic regression with Elastic Net regularizer
R(x) = ?1 k ? k1 + (?2 /2)k ? k22 (?1 , ?2 ? 0).
We make the following assumptions for our analysis:
Assumption 1. There exists a minimizer x? of (1).
Assumption 2. Each fi is convex, and is Li -smooth, i.e.,
k?fi (x) ? ?fi (y)k ? Li kx ? yk (?x, y ? Rd ).
Assumption 3. Regularization function R is convex, and is relatively
simple, which means
that
computing the proximal mapping of R at y, proxR (y) = argmin x?Rd 21 kx ? yk2 + R(x) , takes
O(d) computational cost, for any y ? Rd .
We always consider Assumptions 1, 2 and 3 in this paper.
Assumption 4. There exists ? > 0 such that objective function P is ?-optimally strongly convex,
i.e., P has a minimizer and satisfies
?
kx ? x? k2 ? P (x) ? P (x? ) (?x ? Rd , ?x? ? argminx?Rd f (x)).
2
Note that the requirement of optimally strong convexity is weaker than the one of ordinary strong
convexity (for the definition of ordinary strong convexity, see [11]).
We further consider Assumption 4 when we deal with strongly convex objectives.
3
3
Our Approach: Double Acceleration
In this section, we provide high-level ideas of our main contribution called ?double acceleration.?
First, we consider deterministic PG (Algorithm 1) and (non-mini-batch) SVRG (Algorithm 2). PG
is an extension of the steepest descent to proximal settings. SVRG is a stochastic gradient method
using the variance reduction technique, which utilizes the finite sum structure of the problem, and
it achieves a faster convergence rate than PG does. As SVRG (Algorithm 2) matches with PG
(Algorithm 1) when the number of inner iterations m equals 1, SVRG can be seen as a generalization
of PG. The key element of SVRG is employing a simple but powerful technique called the variance
reduction technique for gradient estimate. The variance reduction of the gradient is realized by
setting gk = ?fik (xk?1 ) ? ?fik (e
x) + ?F (e
x) rather than vanilla stochastic gradient ?fik (xk?1 ).
Generally, stochastic gradient ?fik (xk?1 ) is an unbiased estimator of ?F (xk?1 ), but it may have
high variance. In contrast, gk is also unbiased, and one can show that its variance is ?reduced?; that
is, the variance converges to zero as xk?1 and x
e to x? .
Algorithm 1: PG (e
x0 , ?, S)
for s = 1 to S do
x
es = One Stage PG(e
xs?1 , ?).
end for P
S
return S1 s=1 x
es .
Algorithm 3: One Stage PG (e
x, ?)
x
e+ = prox?R (e
x ? ??F (e
x)).
return x
e+ .
Algorithm 4: One Stage SVRG (e
x, ?, m)
x0 = x
e.
for k = 1 to m do
Pick ik ? [1, n] randomly.
gk = ?fik (xk?1 ) ? ?fik (e
x) + ?F (e
x).
xk = prox?R (xk?1 ? ?gk ).
end for P
n
return n1 k=1 xk .
Algorithm 2: SVRG (e
x0 , ?, m, S)
for s = 1 to S do
x
es = One Stage SVRG(e
xs?1 , ?, m).
end for P
S
return S1 s=1 x
es .
Algorithm 5: APG (e
x0 , ?, S)
x
e?1 = x
e0 , ?e0 = 0.
for s = 1 to S do
es?1 +
?es = s+1 , yes = x
2
?es?1 ?1
(e
xs?1
?es
?x
es?2 ).
x
es = One Stage PG(e
ys , ?).
end for
return xS .
Next, we explain the method of accelerating SVRG and obtaining an even faster convergence rate
based on our new but quite natural idea ?outer acceleration.? First, we would like to remind you
that the procedure of deterministic APG is given as described in Algorithm 5. APG uses the famous
?momentum? scheme and achieves the optimal iteration complexity. Our natural idea is replacing
One Stage PG in Algorithm 5 with One Stage SVRG. With slight modifications, we can show
that this algorithm improves the rates of PG, SVRG and APG, and is optimal. We call this new
algorithm outerly accelerated SVRG. However, this algorithm has poor mini-batch
efficiency, because
?
in size b mini-batch settings, the rate of this algorithm is essentially b times worse than that of
non-mini-batch settings. State-of-the-art methods APCG, SPDC, and Katyusha also suffer from the
same problem in the mini-batch setting.
Now, we illustrate that for improving the mini-batch efficiency, using the ?inner acceleration? technique is beneficial. The author of [12] has proposed AccProxSVRG in mini-batch settings. AccProxSVRG applies the momentum scheme to One Stage SVRG, and we call this technique ?inner?
acceleration. He showed that the inner acceleration could significantly improve the mini-batch
efficiency of vanilla SVRG. This fact indicates that inner acceleration is essential to fully utilize
the mini-batch settings. However, AccProxSVRG is not a truly accelerated method, because in
non-mini-batch settings, the rate of AccProxSVRG is same as that of vanilla SVRG.
In this way, we arrive at our main high-level idea called ?double? acceleration, which involves
applying momentum scheme to both outer and inner algorithms. This enables us not only to lead to
4
the optimal total computational cost in non-mini-batch settings, but also to improving the mini-batch
efficiency of vanilla acceleration methods.
We have considered SVRG and its accelerations so far; however, we actually adopt stochastic variance
reduced dual averaging (SVRDA) rather than SVRG itself, because we can construct lazy update
rules of (innerly) accelerated SVRDA for sparse data. In Section G of supplementary material, we
briefly discuss a SVRG version of our proposed method and provide its convergence analysis.
4
Algorithm Description
In this section, we describe the concrete procedure of the proposed algorithm in detail.
4.1
DASVRDA for non-strongly convex objectives
Algorithm 6: DASVRDAns (e
x0 , ze0 , ?, {Li }ni=1 , m, b, S)
? = 1 Pn Li , Q = {qi } = L?i , ? =
x
e?1 = ze0 , ?e0 = 0, L
i=1
n
nL
for s =
1 to S do
e
?s = 1 ? ?1 s+2
es = x
es?1 +
2 , y
?es?1 ?1
(e
xs?1
?es
1
(1+ ?(m+1)
)L?
b
?x
es?2 ) +
?es?1
(e
zs?1
?es
.
?x
es?1 ).
(e
xs , zes ) = One Stage AccSVRDA(e
ys , x
es?1 , ?, m, b, Q).
end for
return x
eS .
Algorithm 7: One Stage AccSVRDA (e
y, x
e, ?, m, b, Q)
x0 = z0 = ye, g?0 = 0, ?0 = 12 .
for k = 1 to m do
Pick independently i1k , . . . ,ibk ? [1, n] according to Q, set Ik = {i`k }b`=1 .
?k =
k+1
2 ,
yk = 1 ?
1
?k
xk?1 +
1
?k zk?1 .
(?fi (yk ) ? ?fi (e
x)) + ?F (e
x), g?k = 1 ? ?1k g?k?1 + ?1k gk .
o
n
zk = argmin h?
gk , zi + R(z) + 2??k1?k?1 kz ? z0 k2 = prox??k ?k?1 R (z0 ? ??k ?k?1 g?k ) .
z?Rd
xk = 1 ? ?1k xk?1 + ?1k zk .
end for
return (xm , zm ).
gk =
1
b
1
i?Ik nqi
P
We provide details of the doubly accelerated SVRDA (DASVRDA) method for non-strongly convex
objectives in Algorithm 6. Our momentum step is slightly different from that of vanilla deterministic
accelerated methods: we not only add momentum term ((?es?1 ? 1)/?es )(e
xs?1 ? x
es?2 ) to the
e
e
current solution x
es?1 but also add term (?s?1 /?s )(e
zs?1 ? x
es?1 ), where zes?1 is the current more
?aggressively? updated solution rather than x
es?1 ; thus, this term also can be interpreted as momentum5 .
Then, we feed yes to One Stage Accelerated SVRDA (Algorithm 7) as an initial point. Note that
Algorithm 6 can be seen as a direct generalization of APG, because if we set m = 1, One Stage
Accelerated SVRDA is essentially the same as one iteration PG with initial point yes ; then, we can
see that zes = x
es , and Algorithm 6 essentially matches with deterministic APG. Next, we move
to One Stage Accelerated SVRDA (Algorithm 7). Algorithm 7 is essentially a combination of the
?accelerated regularized dual averaging? (AccSDA) method [21] with the variance reduction technique
of SVRG. It updates zk by using the weighted average of all past variance reduced gradients g?k
instead of only using a single variance reduced gradient gk . Note that for constructing variance
reduced gradient gk , we use the full gradient of F at x
es?1 rather than the initial point yes . The
5
This form also arises in Monotone APG [7]. In Algorithm 7, x
e = xm can be rewritten as (2/(m(m +
P
1)) m
e is updated more ?aggressively? than
k=1 kzk , which is a weighted average of zk ; thus, we can say that z
x
e. For the outerly accelerated SVRGP
(that is a combination of Algorithm 6 with vanilla SVRG, see section 3), ze
and x
e correspond to xm and (1/m) m
e is updated more
k=1 xk in [22], respectively. Thus, we can also see that z
?aggressively? than x
e.
5
Adoption of (Innerly) Accelerated SVRDA rather than (Innerly) Accelerated SVRG enables us to
construct lazy updates for sparse data (for more details, see Section E of supplementary material).
4.2
DASVRDA for strongly convex objectives
Algorithm 8: DASVRDAsc (?
x0 , ?, {Li }ni=1 , m, b, S, T )
for t = 1 to T do
x
?t = DASVRDAns (?
xt?1 , x
?t?1 , ?, {Li }ni=1 , m, b, S).
end for
return x
?T .
Algorithm 8 is our proposed method for strongly convex objectives. Instead of directly accelerating
the algorithm using a constant momentum rate, we restart Algorithm 6. Restarting scheme has several
advantages both theoretically and practically. First, the restarting scheme only requires the optimal
strong convexity of the objective instead of the ordinary strong convexity. Whereas, non-restarting
accelerated algorithms essentially require the ordinary strong convexity of the objective. Second,
for restarting algorithms, we can utilize adaptive restart schemes [14]. The adaptive restart schemes
have been originally proposed for deterministic cases. The schemes are heuristic but quite effective
empirically. The most fascinating property of these schemes is that we need not prespecify the
strong convexity parameter ?, and the algorithms adaptively determine the restart timings. [14]
have proposed two heuristic adaptive restart schemes: the function scheme and gradient scheme.
We can easily apply these ideas to our method, because our method is a direct generalization of
the deterministic APG. For the function scheme, we restart Algorithm 6 if P (e
xs ) > P (e
xs?1 ). For
the gradient scheme, we restart the algorithm if (e
ys ? x
es )> (e
ys+1 ? x
es ) > 0. Here yes ? x
es can be
interpreted as a ?one stage? gradient mapping of P at yes . As yes+1 ? x
es is the momentum, this scheme
can be interpreted such that we restart whenever the momentum and negative one Stage gradient
mapping form an obtuse angle (this means that the momentum direction seems to be ?bad?). We
numerically demonstrate the effectiveness of these schemes in Section 6.
Parameter tunings
For DASVRDAns , only learning rate ? needs to be tuned, because we can theoretically obtain the
optimal choice of ?, and we can naturally use m = n/b as a default epoch length (see Section 5). For
DASVRDAsc , both learning rate ? and fixed restart interval S need to be tuned.
5
Convergence Analysis of DASVRDA Method
In this section, we provide the convergence analysis of our algorithms. Unless otherwise specified,
serial computation is assumed. First, we consider the DASVRDAns algorithm.
e0 , ze0 ? Rd , ? ? 3, m ? N, b ? [n]
Theorem 5.1. Suppose that Assumptions 1, 2 and 3 hold. Let x
and S ? N. Then DASVRDAns (e
x0 , ze0 , ?, {Li }ni=1 , m, b, S) satisfies
4
E [P (e
xS ) ? P (x? )] ?
2
1 ? ?1 (S + 2)2
1
1?
?
2
2
(P (e
x0 ) ? P (x? )) +
ke
z0 ? x? k2
?(m + 1)m
!
.
The proof of Theorem 5.1 is found
p in the supplementary material (Section A). We can easily see that
the optimal choice of ? is (3+ 9 + 8b/(m + 1))/2 = O(1+b/m) (see Section B of supplementary
material). We denote this value as ?? . From Theorem 5.1, we obtain the following corollary:
Corollary 5.2. Suppose that Assumptions
e0 ? Rd , ??= ?? p
, m ? n/b and b ?
p 1, 2, and 3 hold. Let x
? x0 ? x? k2 /?),
[n]. If we appropriately choose S = O( (P (e
x0 ) ? P (x? ))/?+(1/m+1/ mb) Lke
ns
n
then a total computational cost of DASVRDA (e
x0 , ?? , {Li }i=1 , m, b, S) for E [P (e
xS ) ? P (x? )] ?
? is
!!
r
r
? x0 ? x? k2
? Lke
P (e
x0 ) ? P (x? )
O d n
+ b+ n
.
?
?
6
Remark. If we adopt a warm start scheme for DASVRDAns , we can further improve the rate to
!!
r
?
Lke
x0 ? x? k 2
P (e
x0 ) ? P (x? )
+ (b + n)
O d nlog
?
?
(see Section C and D of supplementary material).
Next, we analyze the DASVRDAsc algorithm for optimally strongly convex objectives. Combining
Theorem 5.1 with the optimal strong convexity of the objective function immediately yields the
following theorem, which implies that the DASVRDAsc algorithm achieves a linear convergence.
?0 ? Rd , ? = ?? , m ? N, b ? [n]
Theorem 5.3. Suppose that Assumptions 1, 2, 3 and 4 hold. Let x
def
and T ? N. Define ? = 4{(1 ? 1/?? )2 + 4/(?(m + 1)m?)}/{(1 ? 1/?? )2 (S + 2)2 }. If S is
sufficiently large such that ? ? (0, 1), then DASVRDAsc (?
x0 , ?? , {Li }ni=1 , m, b, S, T ) satisfies
E[P (?
xT ) ? P (x? )] ? ?T [P (?
x0 ) ? P (x? )].
From Theorem 5.3, we have the following corollary.
Corollary 5.4. Suppose that Assumptions 1, 2,p
3 and 4 hold. Let x
?0 ? Rd , ? = ?? , m ? n/b,
?
?
b ? [n]. There exists S = O(1 + (b/n + 1/ n) L/?), such that 1/log(1/?) = O(1). Moreover,
if we appropriately choose T = O(log(P (?
x0 ) ? P (x? )/?), then a total computational cost of
DASVRDAsc (?
x0 , ?? , {Li }ni=1 , m, b, S, T ) for E [P (?
xT ) ? P (x? )] ? ? is
? ?
?
O ?d ?n + b + n
?
s ?
?
P
(?
x
)
?
P
(x
)
L
0
?
? log
?.
?
?
?
sc
(?
x0 , ?? ,
Remark. Corollary 5.4 implies that if the mini-batch size b is O( n), DASVRDA
p
n
?
{Li }i=1 , n/b, b, S, T ) still achievesp
the total computational cost of O(d(n + nL/?)log(1/?)),
?
which is much better than O(d(n + nbL/?)log(1/?))
of APCG, SPDC, and Katyusha.
?
sc
Remark. Corollary 5.4 also implies that DASVRDA
only needs size O( n) mini-batches for
p
achieving the optimal iteration complexity O( L/?log(1/?)), when L/? ? n. Inp
contrast, APCG,
SPDC and Katyusha need size O(n) mini-batches and AccProxSVRG does O( L/?) ones for
achievingpthe optimal iteration complexity. Note that even when L/? ? n, our method only needs
size O(n ?/L) mini-batches 6 . This size is smaller than O(n) of APCG, SPDC, and Katyusha, and
the same as that of AccProxSVRG.
6
Numerical Experiments
In this section, we provide numerical experiments to demonstrate the performance of DASVRDA.
We numerically compare our method with several well-known stochastic gradient methods in minibatch settings: SVRG [22] (and SVRG++ [2]), AccProxSVRG [12], Universal Catalyst [8] , APCG
[9], and Katyusha [1]. The details of the implemented algorithms and their parameter tunings
are found in the supplementary material. In the experiments, we focus on the regularized logistic
regression problem for binary classification, with regularizer ?1 k ? k1 + (?2 /2)k ? k22 .
We used three publicly available data sets in the experiments. Their sizes n and dimensions d, and
common min-batch sizes b for all implemented algorithms are listed in Table 2.
Table 2: Summary of the data sets and mini-batch size used in our numerical experiments
Data sets
a9a
rcv1
sido0
n
32, 561
20, 242
12, 678
d
123
47, 236
4, 932
b
180
140
100
For regularization parameters, we used three settings (?1 , ?2 ) = (10?4 , 0), (10?4 , 10?6 ), and
(0, 10?6 ). For the former case, the objective is non-strongly convex, and for the latter two cases,
6
p
p
Note that the required size is O(n ?/L)(? O(n)), which is not O(n L/?) ? O(n).
7
(a) a9a, (?1 , ?2 ) = (10?4 , 0)
(b) a9a, (?1 , ?2 ) = (10?4 , 10?6 )
(c) a9a, (?1 , ?2 ) = (0, 10?6 )
(d) rcv1, (?1 , ?2 ) = (10?4 , 0)
(e) rcv1, (?1 , ?2 ) = (10?4 , 10?6 )
(f) rcv1, (?1 , ?2 ) = (0, 10?6 )
(g) sido0, (?1 , ?2 ) = (10?4 , 0)
(h) sido0, (?1 , ?2 ) = (10?4 , 10?6 )
(i) sido0, (?1 , ?2 ) = (0, 10?6 )
Figure 1: Comparisons on a9a (top), rcv1 (middle) and sido0 (bottom) data sets, for regularization
parameters (?1 , ?2 ) = (10?4 , 0) (left), (?1 , ?2 ) = (10?4 , 10?6 ) (middle) and (?1 , ?2 ) = (0, 10?6 )
(right).
the objectives are strongly convex. Note that for the latter two cases, the strong convexity of the
objectives is ? = 10?6 and is relatively small; thus, it makes acceleration methods beneficial.
Figure 1 shows the comparisons of our method with the different methods described above on several
settings. ?Objective Gap? means P (x) ? P (x? ) for the output solution x. ?Elapsed Time [sec]?
means the elapsed CPU time (sec). ?Restart_DASVRDA? means DASVRDA with heuristic adaptive
restarting (Section 4). We can observe the following from these results:
? Our proposed DASVRDA and Restart DASVRDA significantly outperformed all the other
methods overall.
? DASVRDA with the heuristic adaptive restart scheme efficiently made use of the local
strong convexities of non-strongly convex objectives and significantly outperformed vanilla
DASVRDA. For the other settings, the algorithm was still comparable to vanilla DASVRDA.
? UC+AccProxSVRG7 outperformed vanilla AccProxSVRG but was outperformed by our
methods overall.
7
Although there has been no theoretical guarantee for UC + AccProxSVRG, we thought that it was fair
to include experimental results about that because UC + AccProxSVRG gives better performances than the
vanilla AccProxSVRG. Through some theoretical analysis, we can prove that UC + AccProxSVRG also has the
similar rate and mini-batch efficiency to our proposed method, although these results are not obtained in any
literature. However, our proposed method is superior to this algorithm both theoretically and practically, because
the algorithm has several drawbacks due to the use of UC as follows. First, the algorithm has an additional
logarithmic factor in its convergence rate. This factor is generally not negligible and slows down its practical
performances. Second, the algorithm has more tuning parameters than our method. Third, the stopping criterion
of each sub-problem of UC is hard to be tuned.
8
? APCG sometimes performed unstably and was outperformed by vanilla SVRG. On sido0
data set, for Ridge Setting, APCG significantly outperformed all the other methods.
? Katyusha always outperformed vanilla SVRG, but was significantly outperformed by our
methods.
7
Conclusion
In this paper, we developed a new accelerated stochastic variance reduced gradient method for
regularized empirical risk minimization problems in mini-batch settings: DASVRDA. Wephave shown
?
that DASVRDA achieves
the total computational costs of O(d(nlog(1/?) + (b + n) L/?)) and
? p
O(d(n + (b + n) L/?)log(1/?)) in size b mini-batch settings for non-strongly and optimally
strongly convex objectives, respectively. In
? addition, DASVRDA essentially achieves the optimal
iteration complexities only with size O( n) mini-batches for both settings. In the numerical
experiments, our method significantly outperformed state-of-the-art methods, including Katyusha
and AccProxSVRG.
Acknowledgment
This work was partially supported by MEXT kakenhi (25730013, 25120012, 26280009 and
15H05707), JST-PRESTO and JST-CREST.
References
[1] Z. Allen-Zhu. Katyusha: The First Direct Acceleration of Stochastic Gradient Methods. In 48th
Annual ACM Symposium on the Theory of Computing, pages 19?23, 2017.
[2] Z. Allen-Zhu and Y. Yuan. Improved SVRG for Non-Strongly-Convex or Sum-of-Non-Convex
Objectives. In Proceedings of the 33rd International Conference on Machine Learning, pages
1080?1089, 2016.
[3] A. Beck and M. Teboulle. A fast iterative shrinkage-thresholding algorithm for linear inverse
problems. SIAM Journal on Imaging Sciences, 2(1):183?202, 2009.
[4] A. Defazio, F. Bach, and S. Lacoste-Julien. Saga: A fast incremental gradient method with
support for non-strongly convex composite objectives. In Advances in Neural Information
Processing Systems 27, pages 1646?1654, 2014.
[5] E. Hazan, A. Agarwal, and S. Kale. Logarithmic regret algorithms for online convex optimization. Machine Learning, 69(2-3):169?192, 2007.
[6] R. Johnson and T. Zhang. Accelerating stochastic gradient descent using predictive variance
reduction. In Advances in Neural Information Processing Systems 26, pages 315?323, 2013.
[7] H. Li and Z. Lin. Accelerated proximal gradient methods for nonconvex programming. In
Advances in Neural Information Processing Systems 28, pages 379?387, 2015.
[8] H. Lin, J. Mairal, and Z. Harchaoui. A universal catalyst for first-order optimization. In
Advances in Neural Information Processing Systems 28, pages 3384?3392, 2015.
[9] Q. Lin, Z. Lu, and L. Xiao. An accelerated proximal coordinate gradient method. In Advances
in Neural Information Processing Systems 27, pages 3059?3067, 2014.
[10] Y. Nesterov. Gradient methods for minimizing composite objective function. Mathematical
Programming, 140(1):125?161, 2013.
[11] Y. Nesterov. Introductory lectures on convex optimization: A basic course, volume 87. Springer
Science & Business Media, 2013.
[12] A. Nitanda. Stochastic proximal gradient descent with acceleration techniques. In Advances in
Neural Information Processing Systems 27, pages 1574?1582, 2014.
9
[13] A. Nitanda. Accelerated stochastic gradient descent for minimizing finite sums. In Proceedings
of the 19th International Conference on Artificial Intelligence and Statistics, pages 195?203,
2016.
[14] B. O?Donoghue and E. Candes. Adaptive restart for accelerated gradient schemes. Foundations
of computational mathematics, 15(3):715?732, 2015.
[15] N. L. Roux, M. Schmidt, and F. R. Bach. A stochastic gradient method with an exponential
convergence _rate for finite training sets. In Advances in Neural Information Processing Systems
25, pages 2663?2671, 2012.
[16] M. Schmidt, N. L. Roux, and F. Bach. Minimizing finite sums with the stochastic average
gradient. Mathematical Programming, 162(1):83?112, 2017.
[17] S. Shalev-Shwartz and Y. Singer. Logarithmic regret algorithms for strongly convex repeated
games. Technical report, The Hebrew University, 2007.
[18] S. Shalev-Shwartz and T. Zhang. Stochastic dual coordinate ascent methods for regularized
loss. The Journal of Machine Learning Research, 14(1):567?599, 2013.
[19] Y. Singer and J. C. Duchi. Efficient learning using forward-backward splitting. In Advances in
Neural Information Processing Systems 22, pages 495?503, 2009.
[20] P. Tseng. On accelerated proximal gradient methods for convex-concave optimization. Technical
report, University of Washington, Seattle, 2008.
[21] L. Xiao. Dual averaging method for regularized stochastic learning and online optimization. In
Advances in Neural Information Processing Systems 22, pages 2116?2124, 2009.
[22] L. Xiao and T. Zhang. A proximal stochastic gradient method with progressive variance
reduction. SIAM Journal on Optimization, 24(4):2057?2075, 2014.
[23] Y. Zhang and L. Xiao. Stochastic primal-dual coordinate method for regularized empirical
risk minimization. In Proceedings of the 32nd International Conference on Machine Learning,
pages 353?361, 2015.
10
| 6663 |@word version:1 briefly:1 middle:2 norm:1 seems:1 nd:2 pg:16 unstably:1 pick:2 sgd:5 reduction:10 initial:3 tuned:3 past:1 existing:1 current:2 numerical:4 enables:2 update:3 intelligence:2 xk:13 steepest:1 core:1 zhang:4 mathematical:4 direct:6 become:1 symposium:1 ik:3 yuan:1 prove:1 doubly:3 nlog2:1 introductory:1 x0:21 theoretically:4 little:1 cpu:1 moreover:1 medium:1 argmin:2 interpreted:3 z:5 developed:1 guarantee:1 golden:2 h05707:1 concave:1 sag:1 k2:6 negligible:1 timing:1 local:1 co:1 graduate:1 adoption:1 bi:2 averaged:1 acknowledgment:1 practical:1 regret:2 procedure:2 universal:3 empirical:6 significantly:8 composite:5 thought:1 inp:1 cannot:2 risk:6 applying:2 equivalent:1 deterministic:7 center:1 kale:1 independently:1 convex:40 ke:1 roux:2 splitting:1 immediately:1 fik:6 estimator:1 rule:1 importantly:1 coordinate:6 updated:3 suppose:4 programming:3 us:2 element:1 ze:1 taiji:2 particularly:2 bottom:1 yk:3 agency:1 convexity:11 complexity:13 nesterov:3 ultimately:2 solving:3 predictive:1 efficiency:8 easily:4 regularizer:2 riken:1 fast:2 effective:2 describe:3 artificial:1 sc:2 shalev:2 quite:2 heuristic:4 larger:1 supplementary:6 say:1 otherwise:3 statistic:1 itself:1 lke:3 online:2 advantage:2 differentiable:2 net:1 propose:1 nlog:5 mb:1 zm:1 combining:1 achieve:4 description:1 seattle:1 convergence:12 double:4 requirement:1 incremental:1 converges:1 illustrate:1 develop:1 ac:1 school:1 strong:11 implemented:2 involves:1 implies:4 direction:1 drawback:1 tokyo:5 stochastic:30 jst:2 material:6 sido0:6 require:1 generalization:3 preliminary:1 extension:1 hold:4 practically:2 sufficiently:1 considered:2 mapping:3 achieves:11 adopt:2 outperformed:9 label:1 weighted:2 minimization:7 always:2 rather:5 pn:2 shrinkage:1 corollary:6 focus:1 kakenhi:1 indicates:2 a9a:5 contrast:2 stopping:1 integrated:1 overall:3 dual:9 classification:1 art:3 uc:12 equal:1 construct:2 beach:1 washington:1 progressive:1 report:2 randomly:1 beck:1 argminx:1 n1:2 truly:1 extreme:1 nl:2 primal:2 accproxsvrg:19 necessary:5 obtuse:1 unless:1 euclidean:1 e0:5 theoretical:2 instance:2 teboulle:1 ordinary:4 cost:15 johnson:1 optimally:4 unattainable:15 proximal:11 considerably:1 adaptively:1 st:1 international:3 siam:2 informatics:1 concrete:1 choose:2 possibly:2 worse:4 return:8 li:13 japan:5 prox:3 sec:2 stabilize:1 inc:1 depends:1 performed:1 analyze:2 doing:1 hazan:1 start:1 parallel:2 candes:1 contribution:2 minimize:1 ni:7 accuracy:3 publicly:1 variance:23 efficiently:2 murata:2 correspond:2 yield:1 yes:7 famous:1 lu:1 explain:1 whenever:1 definition:1 pp:1 naturally:1 associated:1 proof:1 popular:1 improves:4 actually:2 feed:1 originally:1 katyusha:14 improved:1 execute:1 though:1 strongly:32 stage:15 replacing:1 minibatch:2 logistic:2 usa:1 k22:2 ye:1 unbiased:2 former:2 regularization:3 aggressively:3 deal:1 game:1 criterion:1 mist:1 ridge:1 demonstrate:2 duchi:1 allen:2 recently:2 fi:7 common:1 superior:1 empirically:1 jp:2 volume:1 slight:1 he:1 numerically:2 refer:1 ai:2 smoothness:1 rd:15 vanilla:12 tuning:3 mathematics:1 yk2:1 add:2 hide:1 showed:1 nonconvex:1 binary:1 seen:2 additional:1 preceding:1 determine:1 full:2 harchaoui:1 ntt:1 technical:2 smooth:3 match:3 faster:2 bach:3 long:1 lin:3 serial:1 y:4 qi:1 variant:1 regression:3 basic:1 essentially:13 iteration:21 asdca:6 sometimes:1 agarwal:1 addition:2 whereas:1 interval:1 appropriately:2 extra:2 ascent:2 effectiveness:1 call:2 near:2 zi:1 inner:6 idea:5 donoghue:1 defazio:1 accelerating:3 suffer:1 remark:3 dramatically:1 generally:3 listed:1 amount:1 reduced:11 per:4 key:1 achieving:7 utilize:2 lacoste:1 backward:1 imaging:1 monotone:1 sum:10 angle:1 inverse:1 powerful:1 you:1 arrive:1 utilizes:2 comparable:1 apg:12 def:2 fascinating:1 annual:1 precisely:1 incorporation:1 i1k:1 min:2 rcv1:5 relatively:4 department:1 according:1 combination:2 poor:1 smaller:2 beneficial:2 slightly:1 modification:1 s1:2 gradually:1 erm:3 discus:1 singer:2 nitanda:2 end:7 presto:2 available:1 rewritten:1 apply:1 observe:1 batch:58 alternative:1 schmidt:2 slower:1 denotes:2 top:1 include:2 k1:3 bl:2 objective:30 move:1 realized:1 traditional:1 abovementioned:1 gradient:40 nbl:7 restart:12 outer:2 tseng:1 length:1 msi:1 remind:1 mini:58 minimizing:3 hebrew:1 ql:1 gk:9 negative:1 slows:1 finite:10 descent:5 situation:1 community:2 required:1 specified:1 elapsed:2 nip:1 below:1 xm:3 including:1 suitable:1 natural:3 warm:1 regularized:10 business:1 advanced:1 zhu:2 scheme:20 improve:3 technology:2 julien:1 epoch:1 literature:1 l2:1 multiplication:1 catalyst:3 fully:1 lecture:1 loss:1 sublinear:1 foundation:1 xiao:4 thresholding:1 course:1 summary:1 surprisingly:1 supported:1 svrg:31 weaker:1 fall:1 sparse:2 benefit:1 regard:1 apcg:12 dimension:3 kzk:1 default:1 kz:1 author:2 suzuki:1 adaptive:6 made:1 forward:1 employing:1 far:1 crest:1 restarting:5 mairal:1 assumed:1 xi:1 thep:1 shwartz:2 iterative:1 table:4 zk:5 ca:1 elastic:1 obtaining:1 improving:2 constructing:1 main:4 arise:1 fair:2 repeated:1 ibk:1 n:1 sub:1 momentum:9 saga:2 exponential:1 kxk2:1 third:1 z0:4 theorem:7 down:1 bad:2 xt:3 list:1 x:11 exists:3 essential:1 conditioned:2 kx:3 spdc:9 gap:1 logarithmic:3 lazy:2 kxk:1 partially:1 applies:1 springer:1 corresponds:1 minimizer:2 satisfies:3 acm:1 goal:1 acceleration:15 consequently:1 hard:1 typical:1 averaging:5 called:4 total:10 e:32 experimental:1 formally:1 mext:1 support:1 latter:3 arises:2 accelerated:31 |
6,261 | 6,664 | Langevin Dynamics with Continuous Tempering for
Training Deep Neural Networks
Nanyang Ye
University of Cambridge
Cambridge, United Kingdom
[email protected]
Zhanxing Zhu
Center for Data Science, Peking University
Beijing Institute of Big Data Research (BIBDR)
Beijing, China
[email protected]
Rafal K.Mantiuk
University of Cambridge
Cambridge, United Kingdom
[email protected]
Abstract
Minimizing non-convex and high-dimensional objective functions is challenging,
especially when training modern deep neural networks. In this paper, a novel approach is proposed which divides the training process into two consecutive phases
to obtain better generalization performance: Bayesian sampling and stochastic
optimization. The first phase is to explore the energy landscape and to capture
the ?fat? modes; and the second one is to fine-tune the parameter learned from the
first phase. In the Bayesian learning phase, we apply continuous tempering and
stochastic approximation into the Langevin dynamics to create an efficient and
effective sampler, in which the temperature is adjusted automatically according to
the designed ?temperature dynamics?. These strategies can overcome the challenge
of early trapping into bad local minima and have achieved remarkable improvements in various types of neural networks as shown in our theoretical analysis and
empirical experiments.
1
Introduction
Minimizing non-convex error functions over continuous and high-dimensional spaces has been a
primary challenge. Specifically, training modern deep neural networks presents severe difficulties,
mainly because of the large number of critical points with respect to the number of dimensions,
including various saddle points and local minima [9, 5]. In addition, the landscapes of the error
functions are theoretically and computationally impossible to characterize rigidly.
Recently, some researchers have attempted to investigate the landscapes of the objective functions
for several types of neural networks. Under some strong assumptions, previous works [21, 4, 12]
showed that there exists multiple, almost equivalent local minima for deep neural networks, using a
wide variety of theoretical analysis and empirical observations. Despite of the nearly equivalent local
minima during training, obtaining good generalization performance is often more challenging with
current stochastic gradient descent (SGD) or some of its variants. It was demonstrated in [22] that
deep network structures are sensitive to initialization and learning rates. And even networks without
nonlinear activation functions may have degenerate or hard to escape saddle points [12].
One important reason of the difficulty to achieve good generalization is, that SGD and some variants
may tend to trap into a certain local minima or flat regions with poor generalization property [25, 1, 13].
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
In other words, most of existing optimization methods do not explore the landscapes of the error
functions efficiently and effectively. To increase the possibility of sufficient exploration of the
parameter space, [25] proposed to train multiple deep networks in parallel and made individual
networks explore by modulating their distance to the ensemble average.
Another kind of approaches attempt to tackle this issue through borrowing the idea of classical
simulated annealing or tempering [15, 6, 10]. The authors of [19] proposed to inject Gaussian
noise with annealed variance (corresponding to the annealed temperature in simulated annealing)
into the standard SGD to make the original optimization dynamics more ?stochastic?. In essence,
this approach is the same as a scalable Bayesian learning method, Stochastic Gradient Langevin
Dynamics (SGLD [24]) with decayed stepsizes. The Santa algorithm [1] incorporated a similar idea
into a more sophisticated stochastic gradient Markov Chain Monte Carlo (SG-MCMC) framework.
However, previous studies show that the efficiency and performance of these methods for training deep
neural networks is very sensitive to the annealing schedule of the temperature in these methods. Slow
annealing will lead to significantly slow optimization process as observed in the literature of simulated
annealing [10], while fast annealing hinders the exploration dramatically, leading to the optimizer
trapped in poor local minima too early. Unfortunately, searching for a suitable annealing schedule
for training deep neural network is hard and time-consuming according to empirical observations in
these works.
To facilitate more efficient and effective exploration for training deep networks, we divide the whole
training process into two phases: Bayesian sampling for exploration and optimization for fine-tuning.
The motivation of implementing a sampling phase is that sampling is theoretically capable of fully
exploring the parameter space and can provide a good initialization for optimization phase. This
strategy is motivated by the sharp minima theory [13] and its validity will be verified by our empirical
experiments.
Crucially, in the sampling phase, we employ the idea of continuous tempering [8, 17] in molecule
dynamics [20], and implement an extended stochastic gradient second-order Langevin dynamics with
smoothly varying temperatures. Importantly, the change of temperature is governed automatically
by a specifically designed dynamics coupled with the original Langevin dynamics. This is different
from the idea of simulated annealing adopted in [19, 1], in which the temperature is only allowed to
decrease according to a manually predefined schedule. Our ?temperature dynamics? is beneficial in
the sense that it increases the capability of exploring the energy landscapes and hopping between
different modes of the sampling distributions. Thus, it may avoid the problem of early trapping into
bad local minima that exists in other algorithms. We name our approach CTLD, abbreviated for
?Continuously Tempered Langevin Dynamics?. With support of extensive empirical evidence, we
demonstrated the efficiency and effectiveness of our proposed algorithm for training various types of
deep neural networks. To the best of our knowledge, this is the first attempt that adopts continuous
tempering into training modern deep networks and produces remarkable improvements over the
state-of-the-art techniques.
2
Preliminaries
The goal of training deep neural network is to minimize the objective function U (?) corresponding
to a non-convex model of interest, where ? 2 Rd are the model parameters. In a Bayesian setting,
the objective U (?) can be treated as the potential energy function, i.e., the negative log posterior,
PN
U (?) =
log 0 (?), where xi represents the i-th observed data point, 0 (?) is
i=1 log (xi |?)
the prior distribution for the model parameters and (xi |?) is the likelihood term for each observation.
In optimization scenario, the counterpart of the complete negative log likelihood is the loss function
and log 0 (?) is typically referred to as a regularization term.
2.1 Stochastic Gradient MCMC
In the scenario of Bayesian learning, obtaining the samples of a high-dimensional distribution is a
necessary procedure for many tasks. Classic dynamics offers such a way to sample the distribution.
The Hamiltonian in classic dynamics is H(?, r) = U (?)+ 12 rT r, the sum of the potential energy U (?)
and kinetic energy 12 rT r, where r 2 Rd is the momentum term Standard (second-order) Langevin dy2
namics1 with constant temperature Tc can be described by following stochastic differential equations
(SDEs),
p
1 dW
d? = rdt, dr = r? U (?)dt
rdt + 2
(1)
where r? U (?) is the gradient of the potential energy w.r.t. the configuration states ?, denotes
the friction coefficient, 1 = kB Tc with Boltzmann constant kB , and dW is the standard Wiener
process. In the context of this work for Markov Chain Monte Carlo (MCMC) and optimization theory,
we always assume = 1 for simplicity.
If we simulate the dynamics in Eqs (1), a well-known
stationary distribution can be achieved [20],
RR
(?, r) = exp ( H(?, r)) /Z, where Z =
exp ( H(?, r)) d?dr is the normalization constant for the probability density. The desired probability distribution Rassociated with the parameters
? can be obtained by marginalizing the joint distribution, (?) =
(?, r)dr / exp ( U (?)).
The MCMC procedures using the analogy of dynamics described by SDEs are often referred to as
dynamics-based MCMC methods.
However, in the ?Big Data? settings with large N , evaluating the full gradient term r? U (?) is
computationally expensive. The usage of stochastic approximation reduces the computational burden
dramatically, where a much smaller subset of the data, {xk1 , . . . , xkm }, is selected randomly to
approximate the full one,
? (?) =
U
m
NX
log (xkj |?)
m j=1
log
0 (?).
(2)
? (?) is an unbiased estimation of the true gradient. Then the
And the resulting stochastic gradient rU
stochastic gradient approximation can be used in the dynamics-based MCMC methods, often referred
to as SG-MCMC, such as [24, 3].
2.2
Simulated Annealing for Global Optimization
Simulated annealing (SA [15, 6, 10]) is a probabilistic technique for approximating the global
optimum of a given function U (?). A Brownian-type of diffusion algorithm was proposed [6] for
continuous optimization by discretizing the following SDE,
p
d? = rU (?)dt + 2 1 (t)dW,
(3)
1
where
(t) = kB T (t) decays as T (t) = c/ log(2 + t) with a sufficiently large constant c, to
ensure theoretical convergence. Unfortunately, this logarithmic annealing schedule is extremely
slow for optimization. In practice, the polynomial schedules are often adopted to accelerate the
optimization processes though without any theoretical guarantee, such as T (t) = c/(a + t)b , where
a > 0, b 2 (0.5, 1), c > 0 are hyperparameters. Recently, [19, 1] incorporated the simulated
annealing with this polynomial cooling schedule into the training of neural networks. A critical issue
behind these methods is that the generalization performance and efficiency of the optimization are
highly sensitive to the cooling schedule. Unfortunately, searching for a suitable annealing schedule
for training deep neural network is hard and time-consuming according to empirical observations in
these works.
These challenges motivate our work. We proposed to divide the whole optimization process into two
phases: Bayesian sampling based on stochastic gradient for parameter space exploration and standard
SGD with momentum for parameters optimization. The key step in the first phase is that we employ
a new tempering scheme to facilitate more effective exploration over the whole energy landscape.
Now, we will elaborate on the proposed approach.
3
Two Phases for Training Neural Networks
As mentioned in Section 1, the objective functions of deep networks contain multiple, nearly equivalent local minima. The key difference between these local minima is whether they are ?flat? or
?sharp?, i.e., lying in ?wide valleys? or ?stiff valleys?. A recent study by [13] showed that sharp
1
Standard Langevin dynamics is different from that used in SGLD [24], which is the first-order Langevin
dynamics, i.e., Brownian dynamics.
3
minima often lead to poorer generalization performance. Flat minimizers of the energy landscape
tend to generalize better due to their robustness to data perturbations, noise in the activations as well
as perturbations of the parameters. However, most of existing optimization methods lack the ability
to efficiently explore the flat minima, often trapping into sharp minima too early.
We consider this issue in a Bayesian way: the flat minima corresponds to ?fat? modes of the induced
probability distribution over ?, (?) / exp ( U (?)). Obviously, these fat modes own much more
probability mass than ?thin? ones since they are nearly as ?tall? as each other. Based on this simple
observation, we propose to implement a Bayesian sampling procedure before the optimization phase.
Bayesian learning is capable of exploring the energy landscape more thoroughly. Due to the large
probability mass, the sampler tends to capture the desired regions near the ?flat? minima. This
provides a good starting region for optimization phase to fine-tune the parameters learning.
When sampling the distribution (?), the multi-modality issue demands the samplers to transit
between isolated modes efficiently. To this end, we incorporate the continuous tempering and
stochastic approximation techniques into the Langevin dynamics to derive an efficient and effective
sampling process for training deep neural networks.
4
CTLD: Continuously Tempered Langevin Dynamics
Faced with high-dimensional and non-convex energy landscapes U (?), such as the error functions
in deep neural networks, the key challenge is how to efficiency and effectively explore the energy
landscapes. Inspired by the idea of continuous tempering [8, 17] in molecule dynamics, we incorporate
the ?temperature dynamics? and stochastic approximation into the Langevin dynamics in a principled
way to allow a more effective exploration of the energy landscape. The temperature in CTLD
evolves automatically governed by the embedded ?temperature dynamics?, which is different from
the predefined annealing schedules used in [19, 1].
The primary dynamics we use for Bayesian sampling is as follows,
q
d? = rdt, dr = r? U (?)dt
rdt + 2 ? 1 (?)dW
p
d? = r? dt, dr? = h(?, r, ?)dt
2 ? dW? ,
? r? dt +
(4)
where ? is the newly augmented variable to control the inverse temperature ?, ? is the corresponding
friction coefficient. Note that ? 1 (?) = kB T (?) = 1/g(?), depending on the augmented variable
? to dynamically adjust the temperature. The function g(?) plays the role as scaling the constant
temperature Tc . The dynamics of ? and ? are coupled through the function h(?, r, ?). Both of the
two functions will be described later.
It can be shown that if we simulate the SDEs described in Eq (4), the following stationary distribution
will be achieved [8],
(?, r, ?, r? ) / exp ( He (?, r, ?, r? )) ,
(5)
with the extended Hamiltonian and the coupling function h(?) as
He (?, r, ?, r? ) = g(?)H(?, r) + (?) + r?2 /2,
h(?, r, ?) =
@? g(?)H(?, r)
@? (?), (6)
where (?) is some confining potential to enforce additional properties of ?, discussed in Section 4.2. The proof of achievement of this stationary distribution (?, r, ?, r? ) is provided in the
Supplementary Material for completeness.
In order to allow the system to overcome the issue of muli-modality efficiently, the temperature
scaling function g(?) can be any convenient form that satisfies: 0 < g(?) ? 1 and being smooth.
This will allow the system to experience different temperature configurations smoothly. A simple
choice would be the following piecewise polynomial function, with z(?) = |?|0 ,
8
<1, if |?| ? ,
g(?) = 1 S 3z 2 (?) 2z 3 (?) , if < |?| < 0
(7)
:
0
1 S, if |?|
,
Figure 1 presents this temperature scaling function with = 0.4, 0 = 1.5 and S = 0.85. In this
case, ? 1 (?) 2 [1 S, 1]. Experiencing high temperature configurations continuously allows the
4
1
g( )
0.8
0.6
0.4
0.2
0
-2
-1
0
1
2
Figure 1: Temperature scaling function g(?).
sampler to explore the parameter space more ?wildly?, significantly alleviating the issue of trapping
into local minima or flat regions. Moreover, it can be easily seen that when g(?) = 1, we can recover
the desired distribution (?) / exp( U (?)).
4.1 Stochastic Approximation for CTLD
With large-scale datasets, we adopt the technique of stochastic approximation to estimate the full
potential term U (?) and its gradient rU (?), as shown in Eq. (2). One way to analyse the impact of
the stochastic approximation is to make use of the central limit theorem,
? (?) = U (?) + N 0, 2 (?) , r? U
? (?) = r? U (?) + N (0, ?(?))
U
(8)
The usage of stochastic approximation results in a noisy potential term and gradient. Simply plugging
in the the noisy estimation into the original dynamics will lead to a dynamics with additional
noise terms. To dissipate the introduced noise, we assume the covariance matrices, 2 (?) and
?(?), are available, and satisfy the positive semi-definiteness, 2 ? 1 (?)I ??(?) < 0 and
2 ? ?@? g(?) 2 (?) 0 with ? as the associated step size of numerical integration for the SDEs.
With ? small enough, this is always true since the introduced stochastic noise scales down faster than
the added noise. Then, we propose CTLD with stochastic approximation,
q
? (?)dt
d? = rdt, dr = r? U
rdt + 2 ? 1 (?)I ??(?)dW
(9)
p
2 (?)dW ,
? r, ?)dt
d? = r? dt, dr? = h(?,
r
dt
+
2
?@
g(?)
? ?
?
?
?
?
?
T
?
?
where the coupling function h(?, r, ?) = @? g(?) U (?) + r r/2
@? (?). Then the following
theorem to show the stationary distribution of the dynamics described in Eq. (9).
Theorem 1. (?, r, ?, r? ) / exp ( He (?, r, ?, r? )) is the stationary distribution of the dynamics
SDEs Eq. (9), when the variance terms 2 (?) and ?(?) are available.
The proof for this theorem is provided in the Supplementary Materials. In practical implementation
of simulating the r and r? of Eq. (9), we have
?
?
2? (t)
(t)
(t)
(t 1)
(t) (t)
(t) 2 ?
(t 1)
?
r = (1 ?
)r
r? U (? )? + N 0,
I (? ) ?(?
)
g(?(t 1) )
(10)
? (t) , r(t) , ?(t) )? (t) + N (0, 2? (t) ? (? (t) )2 ? 2 (?)),
r?(t) = (1 ? (t) ? )r?(t 1) + h(?
?
where ?(?)
and ? 2 (?) are the estimation of the noise variance terms. In Eq. (10), the noise introduced
by the stochastic approximation is compensated by multiplying (? (t) )2 . To avoid the estimation of the
?
variance terms, we often choose ? (t) = ? small enough and , ? large enough to make the ? 2 ?(?)
2 2
and ? ? (?) numerically negligible, and thus ignored in practical use.
4.2 Control of The Augmented Variable
It is expected that the distribution of experienced temperatures of the system should only depend on
the form of the scaling function g(?). This would help us achieve the desired temperature distribution,
thus resulting in a more controllable system. To this end, two strategies are shown in this part.
Firstly, we confine the augmented variable ? to be in the interval [
achieve this is to configure its gradient as a ?force well?:
?
0, if |?| ? 0
@? (?) =
C, otherwise,
5
0
, 0 ]. One simple choice to
(11)
where C is some appropriate constant. Intuitively, when the particle ? ?escapes? from the interval
[ 0 , 0 ], a force induced by @? (?) will ?pull? it back.
Secondly, we restrict the distribution of ? to be uniform over the specified range. Together with the
design of g(?), this restriction can guarantee the required percent of running time for sampling with
the original inverse temperature = 1, and the remaining for high temperatures. For example, in
case of g(?) in Eq.(7), the percent of simulation time for high temperatures is (1
/ 0 )100%.
An adaptive biasing method metadynamics [16] can be used to achieve a flat density across a bounded
range of ?. Metadynamics incorporates a history-dependent potential term to gradually fill the
minima of energy surface corresponding to ??s marginal density, resulting in a uniform distribution
of ?. In essence, metadynamics biases the extended Hamiltonian by an additional potential Vb (?),
Hm (?, r, ?, r? ) = g(?)H(?, r) + (?) + r?2 /2 + Vb (?)
(12)
(0)
Vb (?)
The bias potential term is initialized
= 0, and then updated by iteratively adding Gaussian
kernel terms,
?
?
(t+1)
(t)
Vb
(?) = Vb (?) + w exp
(? ?(t) )2 /(2 2 ) ,
(13)
where ?(t) is the value of the t-th time step during simulation, the magnitude w and variance term
2
are hyperparameters. To update the bias potential over the range [ 0 , ], we can discretize this
(t)
(t)
interval into K equal bins, { 0 , ?1 , . . . , ?K 1 , 0 } and in each time step update ? in each bin.
Thus, the force induced by the bias potential can be approximated by the difference between adjacent
bins divided by the length of each bin. The force h(? (t) , r(t) , ?t ) over the particle ? will be biased
due to the force induced by metadynamics,
? (t) , r(t) , ?(t) )
h(?
? (t) , r(t) , ?(t) )
h(?
(t)
Vb
(t)
(?k? +1 ) Vb
2 0 /K
(?k? )
(14)
where k ? denotes the bin index inside which ?(t) is located. Finally, we summarize CTLD in Alg. 1.
Algorithm 1: Continuously Tempered Langevin Dynamics
Input: m, ?, number of steps for sampling Ls , , ? ; metadynamics parameters: C, w,
(0)
(0)
Initialize ? (0) , r(0) ? N (0, I), ?(0) = 0, r? ? N (0, 1), and Vb (?(0) ) = 0.
for t = 1, 2, . . . do
? (? (t) );
Randomly sample a minibatch of the dataset with size m to obtain U
if t < Ls then
Sample ? ? N (0, I), ?? ? N (0, 1);
q
2?
? (? (t) )? +
? (t) = ? (t 1) + ?r(t 1) , r(t) = (1 ? )r(t 1) r? U
?
g(?(t 1) )
2
and K.
(t 1)
?(t) = ?(t 1) + ?r?
.
Update Vb (?) according to Eq. (13); Find the k ? indexing which bin ?(t) is located in.
(t)
(t)
? (t) , r(t) , ?(t) ) = @? g(?(t) )H(?
? (t) , r(t) ) @? (?(t) ) Vb (?k? +10) Vb (?k? )
h(?
2 /K
(t)
(t 1)
? (t) , r(t) , ?(t) )? + p2? ? ??
r? = (1 ? ? )r?
+ h(?
else
? (? (t) )?
? (t) = ? (t 1) + ?r(t 1) , r(t) = (1 ? )r(t 1) r? U
end if
end for
4.3
Connection with Other Methods
There is a direct relationship between the proposed method and SGD with momentum. In the
optimization phase, CTLD essentially implements SGD with momentum: as shown in SGHMC
[3], the learning rate in the SGD with momentum corresponds to ? 2 in our method, the momentum
coefficient the SGD is equivalent q
to 1 ? . The key difference appears in the Bayesian learning
phase, a dynamical diffusion term g(?2?
(t 1) ) ? is added to the update of the momentum to empower
the sampler/optimizer to explore the parameter space more thoroughly. This directly avoids the issue
of being stuck into poor local minima too early. CTLD introduces stochastic approximation and
temperature dynamics into the Langevin dynamics in a principled way. This distinguishes it from the
deterministic annealing schedules adopted in Santa [1] and SGLD/AnnealSGD [24, 19].
6
4.4
Parameter Settings, Computational Time and Convergence Analysis
Though there exists several hyperparameters in our method, in practice, the only parameters we
need to tune are the learning rate and the momentum (i.e. related to friction coefficients). Across
all the experiments, for other hyperparamters, including those in confining potential function (?)
and metadynamics, we fix them with our empirical formulae relying on the learning rate. See
Supplementary Materials for a thorough analysis on hyperparameter settings. Moreover, through our
sensitivity analysis for these hyperparameters, we find they are quite robust to algorithm performance
within our estimation range, as shown in Section 5.3. Therefore, practical users can just tune the
learning rate and momentum to use CTLD for training neural networks, which is as simple as SGD
with momentum.
Compared with SGD with momentum, our proposal CTLD only introduces an additional 1D augmented variable ?, and its computational cost in almost negligible, as shown in Supplementary Materials.
The convergence analysis of CTLD is also provided in the Supplementary Materials to demonstrate
its stability.
5
Experiments
To evaluate the proposed method, we conduct experiments on stacked denoising autoencoders and
character-level recurrent neural networks. The comparing methods include SGD with momentum,
RMSprop, Adam [14], AnnealSGD [19], Santa [1] and our proposal CTLD. The same parameter
initialization method ?Xavier? [7] is used except for character recurrent neural networks. The
hyperparameter settings for each compared method are implemented by grid search, provided in the
Supplementary Materials.
5.1
Stacked Denoising Autoencoders
Stacked denoising autoencoders (SdA) [23] have been proven to be useful in pre-training neural
networks for improved performance. We focus on the greedy layer-wise training procedure of SdAs.
Dropout layers are appended to each layer with a rate of 0.2 except for the first and last layer. We use
the training set of MNIST data consisting of 60000 training images for this task. The network is fully
connected, 784-500-500-2000-10. The learning curves of mean square errors (MSE) for each method
are shown in Figure. 2(a). The bumps in iteration 1, 2, 3 ? 105 is due to the switching to next layer
during training. Though CTLD in the sampling phase is not as fast as other methods, it can find the
regions of good minima, and fine-tune to the best results in final stage.
7
SGD-M
RMSProp
Adam
AnnealSGD
Santa
CLTD
10 -1
6
5
1/g( )
MSE
10 0
4
3
10
2
-2
1
0
1
2
Iteration
3
4
3
3.1
3.2
Iteration
10 5
3.3
10 5
Figure 2: (Left) Learning curves of SdAs; (Right) The evolution of the noise magnitude ? = 1/g(?) during
the training the final layer.
We also track the evolution of the augmented variable ? and plot the noise magnitude ?(?) = 1/g(?)
during the training the final layer, shown in the right panel of Fig. 2. We can observe that the behavior
of the magnitude of the noise term is dramatically different from the predefined decreasing schedule
used in Santa and AnnealSGD. The temperature dynamics introduced in CTLD adjusts the noise term
adaptively. This helps the system to explore the landscape of the loss function more thoroughly and
find the regions of good local minima with a higher probability.
5.2
Character Recurrent Neural Networks for Language Modeling
We test our method on the task of character prediction
is to
?Pusing LSTM networks. The objective
?
PN
Ti
1
i i
i
log p(xt |x1 , ..., xt 1 ; ?) , where ?
minimize the per-character perplexity, N i=1 exp
t=1
7
is a set of parameters for the model, xnt is the observed data and Ti is the length of i-th sentence. The
hidden units are set as LSTM units. We run the models with different methods on the Wikipedia
Hutter Prize 100MB dataset with a setting of 3-layer LSTM, 64 hidden layers, the same with the
original paper [11]. The training and test perplexity are shown in Fig. 3.
The best training and test perplexities are reached by our method CTLD, which also has the fastest
convergence speed. RMSProp and Adam converge very fast in the early iterations, but they seem to
be trapped in some poor local minima.
RMSProp
Adam
AnnealSGD
Santa
CTLD
Training Perplexity
2.6
2.5
RMSProp
Adam
AnnealSGD
Santa
CTLD
2.7
2.6
Test Perplexity
2.7
2.4
2.3
2.2
2.1
2.5
2.4
2.3
2.2
2.1
2
2
1.9
1.9
0
5
10
15
20
Epoch
0
5
10
15
20
Epoch
Figure 3: (Left) CharRNN on Wiki training set; (Right) CharRNN on Wiki test set. Note that SGD-M doest
not appear in the Figure because the training and test perplexities for SGD-M are much higher.
5.3
Sensitivity Analysis
Since there exist several hyperparameters in CTLD, we analyze the sensitivity of hyperparameter
settings within our estimation range (provided in Supplementary Materials). We implement the
character-level RNN on War and Peace by Leo Tolstoy instead of Wiki dataset, considering the
computational speed. The same model architecture is used as [11]. The learning rate is set as 2?10 4 ,
momentum as 0.7. We train the model for 50 epochs until full convergence. The results are shown in
Fig. 4. According to Fig. 4, the setting of hyperparameter w and is robust within our estimation
Figure 4: (Left) test perplexity versus w and
w = 20,
(S = 0.85); (Right) test perplexity versus of S ( with
= 0.04)
range. This also shows metadynamics performs quite stable for training neural networks. For the
sensitivity of S, with the increase of S, the range of temperature enlarges accordingly. Larger range of
temperature slightly enhances the ability of CTLD to explore the energy landscape, and leads to better
local minima. However, this improvement is quite limited, as shown in Fig. 4, demonstrating the
robustness of the hyperparameter S. Therefore, we can conclude that with the continuous tempering
scheme, our proposed method remains relatively stable under different hyperparameter settings.
Practical users only need tune the learning rate and momentum to use CTLD.
6
Conclusion & Future Directions
We propose CTLD, an effective and efficient approach for training modern deep neural networks.
It involves scalable Bayesian sampling combined with continuous tempering to capture the ?fat?
modes, and thus avoiding the issue of getting trapped into poor local minima too early. Extensive
theoretical and empirical evidence verify the superiority of our proposal over the existing methods.
Future directions includes theoretically analyzing the effects of metadynamics and hyperparameter
settings, and usage of high-order integrators and preconditioners to improve convergence speed.
8
References
[1] C. Chen, D. Carlson, Z. Gan, C. Li, and L. Carin. Bridging the gap between stochastic gradient MCMC
and stochastic optimization. In AISTATS, 2016.
[2] C. Chen, N. Ding, and L. Carin. On the convergence of stochastic gradient MCMC algorithms with
high-order integrators. In NIPS, 2015.
[3] T. Chen, E. B. Fox, and C. Guestrin. Stochastic gradient Hamiltonian Monte Carlo. In Proceedings of the
31st International Conference on Machine Learning, pages 1683?1691, 2014.
[4] A. Choromanska, M. Henaff, M. Mathieu, G.B. Arous, and Y. LeCun. The loss surfaces of multilayer
networks. In AISTATS, 2015.
[5] Y. N. Dauphin, R. Pascanu, C. Gulcehre, K. Cho, S. Ganguli, and Y. Bengio. Identifying and attacking the
saddle point problem in high-dimensional non-convex optimization. In NIPS, 2014.
[6] S. Geman and C. Hwang. Diffusions for global optimization. SIAM Journal on Control and Optimization,
24(5):1031?1043, 1986.
[7] Xavier Glorot and Yoshua Bengio. Understanding the difficulty of training deep feedforward neural
networks. In Aistats, volume 9, pages 249?256, 2010.
[8] G. Gobbo and B.J. Leimkuhler. Extended hamiltonian approach to continuous tempering. Physical Review
E, 91(6):061301, 2015.
[9] I. Goodfellow, Y. Bengio, and A. Courville. Deep Learning. MIT Press, 2016.
[10] L. Ingber. Simulated annealing: Practice versus theory. Mathematical and computer modelling, 18(11):29?
57, 1993.
[11] Andrej Karpathy, Justin Johnson, and Fei-Fei Li. Visualizing and understanding recurrent networks. CoRR,
abs/1506.02078, 2015.
[12] K. Kawaguchi. Deep learning without poor local minima. In Advances in Neural Information Processing
Systems 29, pages 586?594. 2016.
[13] N. S. Keskar, D. Mudigere, J. Nocedal, M. Smelyanskiy, and P. T. Tang. On large-batch training for deep
learning: Generalization gap and sharp minima. arXiv preprint arXiv:1609.04836, 2016.
[14] D. Kingma and J. Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
[15] S. Kirkpatrick, C. D. Gelatt, and M. P. Vecchi.
220(4598):671?680, 1983.
Optimization by simulated annealing.
Science,
[16] A. Laio and M. Parrinello. Escaping free-energy minima. Proceedings of the National Academy of Sciences,
99(20):12562?12566, 2002.
[17] N. Lenner and G. Mathias. Continuous tempering molecular dynamics: A deterministic approach to
simulated tempering. Journal of Chemical Theory and Computation, 12(2):486?498, 2016.
[18] J. C. Mattingly, A. M. Stuart, and M. V. Tretyakov. Convergence of numerical time-averaging and stationary
measures via poisson equations. SIAM Journal on Numerical Analysis, 48(2):552?577, 2010.
[19] A. Neelakantan, L. Vilnis, Q. V. Le, I. Sutskever, L. Kaiser, K. Kurach, and J. Martens. Adding gradient
noise improves learning for very deep networks. arXiv preprint arXiv:1511.06807, 2015.
[20] D. C. Rapaport. The art of molecular dynamics simulation. Cambridge university press, 2004.
[21] A. M. Saxe, J. L. McClelland, and S. Ganguli. Exact solutions to the nonlinear dynamics of learning in
deep linear neural networks. In ICLR, 2014.
[22] I. Sutskever, J. Martens, G. E. Dahl, and G. E. Hinton. On the importance of initialization and momentum
in deep learning. ICML, 2013.
[23] P. Vincent, H. Larochelle, I. Lajoie, Y. Bengio, and P. Manzagol. Stacked denoising autoencoders:
Learning useful representations in a deep network with a local denoising criterion. J. Mach. Learn. Res.,
11:3371?3408, 2010.
[24] M. Welling and Y. W. Teh. Bayesian learning via stochastic gradient langevin dynamics. In ICML, 2011.
[25] S. Zhang, A. Choromanska, and Y. LeCun. Deep learning with elastic averaging SGD. In NIPS, 2015.
9
| 6664 |@word polynomial:3 simulation:3 crucially:1 covariance:1 sgd:15 arous:1 configuration:3 united:2 existing:3 current:1 comparing:1 activation:2 numerical:3 sdes:5 designed:2 plot:1 update:4 stationary:6 greedy:1 selected:1 accordingly:1 trapping:4 hamiltonian:5 prize:1 provides:1 completeness:1 pascanu:1 firstly:1 zhang:1 mathematical:1 direct:1 differential:1 inside:1 theoretically:3 expected:1 behavior:1 multi:1 integrator:2 inspired:1 relying:1 decreasing:1 automatically:3 considering:1 provided:5 moreover:2 bounded:1 panel:1 mass:2 sde:1 kind:1 guarantee:2 thorough:1 ti:2 tackle:1 fat:4 uk:2 control:3 unit:2 appear:1 superiority:1 before:1 positive:1 negligible:2 local:17 tends:1 limit:1 switching:1 despite:1 mach:1 analyzing:1 rigidly:1 initialization:4 china:1 dynamically:1 challenging:2 fastest:1 limited:1 range:8 practical:4 lecun:2 nanyang:1 practice:3 implement:4 procedure:4 empirical:8 rnn:1 significantly:2 parrinello:1 convenient:1 word:1 pre:1 leimkuhler:1 valley:2 andrej:1 context:1 impossible:1 restriction:1 equivalent:4 deterministic:2 demonstrated:2 center:1 compensated:1 annealed:2 marten:2 starting:1 l:2 convex:5 simplicity:1 identifying:1 adjusts:1 importantly:1 fill:1 pull:1 dw:7 classic:2 searching:2 stability:1 updated:1 hyperparamters:1 play:1 experiencing:1 alleviating:1 user:2 exact:1 goodfellow:1 expensive:1 approximated:1 located:2 cooling:2 geman:1 observed:3 role:1 preprint:3 ding:1 capture:3 region:6 connected:1 hinders:1 decrease:1 mentioned:1 principled:2 rmsprop:5 cam:2 dynamic:40 motivate:1 depend:1 efficiency:4 accelerate:1 joint:1 easily:1 various:3 leo:1 train:2 stacked:4 fast:3 effective:6 monte:3 quite:3 supplementary:7 larger:1 otherwise:1 enlarges:1 ability:2 analyse:1 noisy:2 final:3 obviously:1 rr:1 propose:3 sda:1 mb:1 degenerate:1 achieve:4 academy:1 getting:1 achievement:1 sutskever:2 convergence:8 optimum:1 produce:1 adam:6 tall:1 derive:1 depending:1 ac:2 coupling:2 help:2 recurrent:4 sa:1 p2:1 eq:9 implemented:1 strong:1 involves:1 larochelle:1 direction:2 stochastic:29 kb:4 exploration:7 saxe:1 material:7 implementing:1 bin:6 fix:1 generalization:7 preliminary:1 secondly:1 adjusted:1 exploring:3 lying:1 sufficiently:1 confine:1 sgld:3 exp:9 bump:1 optimizer:2 consecutive:1 early:7 adopt:1 estimation:7 sensitive:3 modulating:1 create:1 mit:1 gaussian:2 always:2 avoid:2 pn:2 stepsizes:1 varying:1 focus:1 improvement:3 modelling:1 likelihood:2 mainly:1 sense:1 ganguli:2 dependent:1 minimizers:1 dy2:1 typically:1 borrowing:1 hidden:2 mattingly:1 choromanska:2 issue:8 dauphin:1 art:2 integration:1 initialize:1 marginal:1 equal:1 beach:1 sampling:15 manually:1 represents:1 stuart:1 icml:2 nearly:3 thin:1 carin:2 future:2 yoshua:1 piecewise:1 escape:2 employ:2 distinguishes:1 modern:4 randomly:2 national:1 individual:1 phase:16 consisting:1 attempt:2 ab:1 interest:1 investigate:1 possibility:1 highly:1 severe:1 adjust:1 introduces:2 kirkpatrick:1 behind:1 configure:1 chain:2 predefined:3 poorer:1 capable:2 necessary:1 experience:1 fox:1 conduct:1 pku:1 divide:3 initialized:1 desired:4 re:1 isolated:1 theoretical:5 hutter:1 modeling:1 cost:1 subset:1 uniform:2 johnson:1 too:4 characterize:1 cho:1 adaptively:1 combined:1 thoroughly:3 international:1 lstm:3 st:2 decayed:1 density:3 sensitivity:4 probabilistic:1 siam:2 together:1 continuously:4 central:1 rafal:2 choose:1 ingber:1 dr:7 inject:1 leading:1 li:2 potential:12 includes:1 coefficient:4 satisfy:1 dissipate:1 later:1 analyze:1 reached:1 recover:1 parallel:1 capability:1 minimize:2 appended:1 square:1 wiener:1 variance:5 keskar:1 efficiently:4 ensemble:1 landscape:13 generalize:1 bayesian:14 vincent:1 carlo:3 multiplying:1 researcher:1 history:1 rdt:6 mudigere:1 energy:15 proof:2 associated:1 newly:1 dataset:3 knowledge:1 improves:1 schedule:11 sophisticated:1 back:1 appears:1 higher:2 dt:10 improved:1 though:3 wildly:1 xk1:1 just:1 stage:1 autoencoders:4 until:1 nonlinear:2 lack:1 minibatch:1 mode:6 hwang:1 facilitate:2 effect:1 ye:1 verify:1 validity:1 usa:1 name:1 counterpart:1 usage:3 regularization:1 unbiased:1 true:2 contain:1 iteratively:1 xavier:2 evolution:2 chemical:1 adjacent:1 visualizing:1 during:5 essence:2 criterion:1 complete:1 demonstrate:1 performs:1 temperature:29 percent:2 image:1 wise:1 novel:1 recently:2 xkj:1 wikipedia:1 physical:1 volume:1 discussed:1 he:3 numerically:1 cambridge:5 tuning:1 rd:2 grid:1 particle:2 language:1 stable:2 surface:2 posterior:1 brownian:2 showed:2 recent:1 own:1 stiff:1 henaff:1 perplexity:8 scenario:2 certain:1 discretizing:1 tempered:3 seen:1 minimum:26 additional:4 guestrin:1 sdas:2 converge:1 attacking:1 semi:1 doest:1 multiple:3 full:4 reduces:1 smooth:1 faster:1 offer:1 long:1 divided:1 molecular:2 plugging:1 peking:1 impact:1 prediction:1 variant:2 scalable:2 peace:1 multilayer:1 essentially:1 poisson:1 arxiv:6 iteration:4 normalization:1 kernel:1 achieved:3 proposal:3 addition:1 fine:4 annealing:17 interval:3 else:1 modality:2 biased:1 induced:4 tend:2 incorporates:1 effectiveness:1 seem:1 near:1 empower:1 feedforward:1 bengio:4 enough:3 variety:1 architecture:1 restrict:1 escaping:1 idea:5 cn:1 kurach:1 whether:1 motivated:1 war:1 bridging:1 deep:26 dramatically:3 ignored:1 useful:2 santa:7 tune:6 zhanxing:2 karpathy:1 neelakantan:1 mcclelland:1 wiki:3 exist:1 trapped:3 track:1 per:1 hyperparameter:7 key:4 demonstrating:1 tempering:13 verified:1 dahl:1 diffusion:3 nocedal:1 sum:1 beijing:2 run:1 inverse:2 muli:1 almost:2 scaling:5 vb:11 dropout:1 layer:9 courville:1 fei:2 flat:8 simulate:2 friction:3 extremely:1 speed:3 preconditioners:1 vecchi:1 relatively:1 smelyanskiy:1 according:6 poor:6 beneficial:1 smaller:1 across:2 character:6 slightly:1 evolves:1 intuitively:1 gradually:1 indexing:1 computationally:2 equation:2 remains:1 abbreviated:1 end:4 adopted:3 available:2 gulcehre:1 sghmc:1 apply:1 observe:1 enforce:1 appropriate:1 simulating:1 gelatt:1 batch:1 robustness:2 original:5 denotes:2 running:1 ensure:1 remaining:1 include:1 gan:1 hopping:1 carlson:1 especially:1 kawaguchi:1 approximating:1 classical:1 objective:6 added:2 kaiser:1 strategy:3 primary:2 rt:2 enhances:1 gradient:19 iclr:1 distance:1 simulated:10 nx:1 lajoie:1 transit:1 reason:1 ru:3 length:2 index:1 relationship:1 manzagol:1 minimizing:2 kingdom:2 unfortunately:3 negative:2 ba:1 xnt:1 implementation:1 design:1 boltzmann:1 confining:2 teh:1 discretize:1 observation:5 markov:2 datasets:1 tretyakov:1 descent:1 langevin:15 extended:4 incorporated:2 hinton:1 perturbation:2 sharp:5 introduced:4 required:1 specified:1 extensive:2 connection:1 sentence:1 learned:1 kingma:1 nip:4 justin:1 dynamical:1 biasing:1 challenge:4 summarize:1 including:2 critical:2 suitable:2 difficulty:3 treated:1 force:5 zhu:2 scheme:2 improve:1 mathieu:1 hm:1 coupled:2 pusing:1 faced:1 prior:1 sg:2 literature:1 epoch:3 understanding:2 review:1 marginalizing:1 embedded:1 fully:2 loss:3 analogy:1 proven:1 remarkable:2 versus:3 rapaport:1 sufficient:1 last:1 free:1 bias:4 allow:3 institute:1 wide:2 overcome:2 dimension:1 curve:2 evaluating:1 avoids:1 author:1 made:1 adopts:1 adaptive:1 stuck:1 welling:1 approximate:1 laio:1 global:3 conclude:1 consuming:2 xi:3 continuous:12 search:1 learn:1 molecule:2 ca:1 elastic:1 controllable:1 obtaining:2 robust:2 alg:1 mse:2 cl:1 aistats:3 big:2 noise:13 whole:3 motivation:1 hyperparameters:5 allowed:1 x1:1 augmented:6 fig:5 referred:3 elaborate:1 definiteness:1 slow:3 experienced:1 momentum:15 governed:2 tang:1 theorem:4 down:1 formula:1 bad:2 xt:2 decay:1 evidence:2 glorot:1 exists:3 trap:1 burden:1 mnist:1 adding:2 effectively:2 corr:1 importance:1 magnitude:4 demand:1 chen:3 gap:2 smoothly:2 tc:3 logarithmic:1 simply:1 explore:9 saddle:3 corresponds:2 satisfies:1 kinetic:1 goal:1 hard:3 change:1 xkm:1 specifically:2 except:2 sampler:5 averaging:2 denoising:5 mathias:1 attempted:1 support:1 vilnis:1 incorporate:2 evaluate:1 mcmc:9 avoiding:1 |
6,262 | 6,665 | Efficient Online Linear Optimization
with Approximation Algorithms
Dan Garber
Technion - Israel Institute of Technology
[email protected]
Abstract
We revisit the problem of online linear optimization in case the set of feasible actions is accessible through an approximated linear optimization oracle with a factor
? multiplicative approximation guarantee. This setting is in particular interesting
since it captures natural online extensions of well-studied offline linear optimization problems which are NP-hard, yet admit efficient approximation algorithms.
The goal here is to minimize the ?-regret which is the natural extension of the
standard regret in online learning to this setting. We present new algorithms with
significantly improved oracle complexity for both the full information and bandit
variants of the problem. Mainly, for both variants, we present ?-regret bounds of
O(T 1/3 ), were T is the number of prediction rounds, using only O(log(T )) calls
to the approximation oracle per iteration, on average. These are the first results to
obtain both average oracle complexity of O(log(T )) (or even poly-logarithmic in
T ) and ?-regret bound O(T c ) for a constant c > 0, for both variants.
1
Introduction
In this paper we revisit the problem of Online Linear Optimization (OLO) [14], which is a specialized
case of Online Convex Optimization (OCO) [12] with linear loss functions, in case the feasible set of
actions is accessible through an oracle for approximated linear optimization with a multiplicative
approximation error guarantee. In the standard setting of OLO, a decision maker is repeatedly
required to choose an action, a vector in some fixed feasible set in Rd . After choosing his action,
the decision maker incurs loss (or payoff) given by the inner product between his selected vector
and a vector chosen by an adversary. This game between the decision maker and the adversary then
repeats itself. In the full information variant of the problem, after the decision maker receives his
loss (payoff) on a certain round, he gets to observe the vector chosen by the adversary. In the bandit
version of the problem, the decision maker only observes his loss (payoff) and does not get to observe
the adversary?s vector. The standard goal of the decision maker in OLO is to minimize a quantity
known as regret, which measures the difference between the average loss of the decision maker on
a game of T consecutive rounds (where T is fixed and known in advance), and the average loss
of the best feasible action in hindsight (i.e., chosen with knowledge of all actions of the adversary
throughout the T rounds) (in case of payoffs this difference is reversed). The main concern when
designing algorithms for choosing the actions of the decision maker, is guaranteeing that the regret
goes to zero as the length of the game T increases, as fast as possible (i.e., the rate of the regret in
terms of T ). It should be noted that in this paper we focus on the case in which the adversary is
oblivious (a.k.a. non-adaptive), which means the adversary chooses his entire sequence of actions for
the T rounds beforehand.
While there exist well known algorithms for choosing the decision maker?s actions which guarantee
optimal regret bounds in T , such as the celebrated Follow the Perturbed Leader (FPL) and Online
Gradient Descent (OGD) algorithms [14, 17, 12], efficient implementation of these algorithms hinges
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
on the ability to efficiently solve certain convex optimization problems (e.g., linear minimization
for FPL or Euclidean projection for OGD) over the feasible set (or the convex hull of feasible
points). However, when the feasible set corresponds for instance to the set of all possible solutions
to some NP-Hard optimization problem, no such efficient implementations are known (or even
widely believed to exist), and thus these celebrated regret-minimizing procedures cannot be efficiently
applied. Luckily, many NP-Hard linear optimization problems (i.e., the objective function to either
minimize or maximize is linear) admit efficient approximation algorithms with a multiplicative
approximation guarantee. Some examples include MAX-CUT (factor 0.87856 approximation due to
[9]) , M ETRIC TSP (factor 1.5 approximation due to [6]), M INIMUM W EIGHTED V ERTEX C OVER
(factor 2 approximation [4]), and W EIGHTED S ET C OVER (factor (log n + 1) approximation due to
[7]). It is thus natural to ask wether an efficient factor ? approximation algorithm for an NP-Hard
offline linear optimization problem could be used to construct, in a generic way, an efficient algorithm
for the online version of the problem. Note that in this case, even efficiently computing the best fixed
action in hindsight is not possible, and thus, minimizing regret via an efficient algorithm does not
seem likely (given an approximation algorithm we can however compute in hindsight a decision
that corresponds to at most (at least) ? times the average loss (payoff) of the best fixed decision in
hindsight).
In their paper [13], Kakade, Kalai and Ligett were the first to address this question in a fully generic
way. They showed that using only an ?-approximation oracle for the set of feasible actions, it
is possible, at a high level, to construct an online algorithm which achieves vanishing (expected)
?-regret, which is the difference between the average loss of the decision maker and ? times the
average loss of the best fixed point in hindsight (for loss minimization problems and ?
1; a
corresponding definition exists for payoff maximization problems and ? < 1). Concretely, [13]
showed that one can guarantee O(T 1/2 ) expected ?-regret in the full-information setting, which is
optimal, and O(T 1/3 ) in the bandit setting under the additional assumption of the availability of a
Barycentric Spanner (which we discuss in the sequel).
While the algorithm in [13] achieves an optimal ?-regret bound (in terms of T ) for the full information
setting, in terms of computational complexity, the algorithm requires, in worst case, to perform on
each round O(T ) calls to the approximation oracle, which might be prohibitive and render the
algorithm inefficient, since as discussed, in general, T is assumed to grow to infinity and thus the
dependence of the runtime on T is of primary interest. Similarly, their algorithm for the bandit setting
requires O(T 2/3 ) calls to the approximation oracle per iteration.
The main contribution of our work is in providing new low ?-regret algorithms for the full information
and bandit settings with significantly improved oracle complexities. A detailed comparison with [13]
is given in Table 1. Concretely, for the full-information setting, we show it is possible to achieve
O(T 1/3 ) expected ?-regret using only O(log(T )) calls to the approximation oracle per iteration,
on average, which significantly improves over the O(T ) bound of [13]1 . pWe also show a bound
of O(T 1/2 ) on the expected ?-regret (which is optimal) using only O( T log(T )) calls to the
oracle per iteration, on average, which gives nearly quadratic improvement over [13]. In the bandit
setting we show it is possible to obtain a O(T 1/3 ) bound on the expected ?-regret (same as in [13])
using only O(log(T )) calls to the oracle per iteration, on average, under the same assumption on the
availability of a Barycentric Spanner (BS). It is important to note that while there exist algorithms
1/2
?
?
for OLO with bandit feedback which guarantee O(T
) expected regret [1, 11] (where the O(?)
hides poly-logarithmic factors in T ), these require on each iteration to either solve to arbitrarily small
accuracy a convex optimization problem over the feasible set [1], or sample a point from the feasible
set according to a specified distribution [11], both of which cannot be implemented efficiently in our
setting. On the other-hand, as we formally show in the sequel, at a high level, using a BS (originally
introduced in [2]) simply requires to find a single set of d points from the feasible set which span the
entire space Rd (assuming this is possible, otherwise the set could be mapped to a lower dimensional
space). The process of finding these vectors can be viewed as a preprocessing step and thus can be
carried out offline. Moreover, as discussed in [13], for many NP-Hard problems it is possible to
compute a BS in polynomial time and thus even this preprocessing step is efficient. Importantly, [13]
shows that the approximation oracle by itself is not strong enough to guarantee non-trivial ?-regret in
the bandit setting, and hence this assumption on the availability of a BS seems reasonable. Since the
1
as we show in the appendix, even if we relax the algorithm of [13] to only guarantee O(T
will still require O(T 2/3 ) calls to the oracle per iteration, on average.
2
1/3
) ?-regret, it
Reference
KKL [13]
This paper (Thm. 4.1, 4.2)
This paper (Thm. 4.1)
?
full information
regret oracle complexity
T 1/2
T
T 1/3
log(T )
p
T 1/2
T log(T )
?
bandit information
regret oracle complexity
T 1/3
T 2/3
1/3
T
log(T )
-
Table 1: comparison of expected ? regret bounds and average number of calls to the approximation
oracle per iteration. In all bounds we give only the dependence on the length of the game T and
omit all other dependencies which we treat as constants. In the bandit setting we report the expected
number of calls to the oracle per iteration.
best general regret bound known using a BS is O(T 1/3 ), the ?-regret bound of our bandit algorithm
is the best achievable to date via an efficient algorithm.
Technically, the main challenge in the considered setting is that as discussed, we cannot readily apply
standard tools such as FPL and OGD. At a high level, in [13] it was shown that it is possible to
apply the OGD method by replacing the exact projection step of OGD with an iterative algorithm
which finds an infeasible point, but one that both satisfies the projection property required by OGD
and is dominated by a convex combination of feasible points for every relevant linear loss (payoff)
function. Unfortunately, in worst case, the number of queries to the approximation oracle required by
this so-called projection algorithm per iteration is linear in T . While our online algorithms are also
based on an application of OGD, our approach to computing the so-called projections is drastically
different than [13], and is based on a coupling of two cutting plane methods, one that is based on
the Ellipsoid method, and the other that resembles Gradient Descent. This approach might be of
independent interest and might prove useful to similar problems.
1.1
Additional related work
Kalai and Vempala [14] showed that approximation algorithms which have point-wise approximation
guarantee, such as the celebrated MAX-CUT algorithm of [9], could be used to instantiate their
Follow the Perturbed Leader framework to achieve low ?-regret. However this construction is far
from generic and requires the oracle to satisfy additional non-trivial conditions. This approach was
also used in [3]. In [14] it was also shown that FPL could be instantiated with a FPTAS to achieve low
?-regret, however the approximation factor in the FPTAS needs to be set to roughly (1 + O(T 1/2 )),
which may result in prohibitive running times even if a FPTAS for the underlying problem is available.
Similarly, in [8] it was shown that if the approximation algorithm is based on solving a convex
relaxation of the original, possibly NP-Hard, problem, this additional structure can be used with
the FPL framework to achieve low ?-regret efficiently. To conclude all of the latter works consider
specialized cases in which the approximation oracle satisfies additional non-trivial assumptions
beyond its approximation guarantee, whereas here, similarly to [13], we will be interested in a generic
as possible conversion from the offline problem to the online one, without imposing additional
structure on the offline oracle.
2
2.1
Preliminaries
Online linear optimization with approximation oracles
Let K, F be compact sets of points in Rd+ (non-negative orthant in Rd ) such that maxx2K kxk ?
R, maxf 2F kf k ? F , for some R > 0, F > 0 (throughout this work we let k ? k denote the standard
Euclidean norm), and for all x 2 K, f 2 F it holds that C x ? f 0, for some C > 0.
We assume K is accessible through an approximated linear optimization oracle OK : Rd+ ! K with
?
parameter ? > 0 such that:
OK (c) ? c ? ? minx2K x ? c if ? 1;
d
8c 2 R+ :
OK (c) 2 K and
OK (c) ? c ? maxx2K x ? c if ? < 1.
Here K is the feasible set of actions for the player, and F is the set of all possible loss/payoff vectors2 .
2
we note that both of our assumptions that K ? Rd+ , F ? Rd+ and that the oracle takes inputs from Rd+ are
made for ease of presentation and clarity, and since these naturally hold for many NP-Hard optimization problem
that are relevant to our setting. Nevertheless, these assumptions could be easily generalized as done in [13].
3
Since naturally a factor ? > 1 for the approximation oracle is reasonable only for loss minimization
problems, and a value ? < 1 is reasonable for payoff maximization problems, throughout this
work it will be convenient to use the value of ? to differentiate between minimization problems and
maximization problems.
Given a sequence of linear loss/payoff functions {f1 , ..., fT } 2 F T and a sequence of feasible points
{x1 , ...., xT } 2 KT , we define the ? regret of the sequence {xt }t2[T ] with respect to the sequence
8 PT
P
{ft }t2[T ] as
< T1 t=1 xt ? ft ? ? minx2K T1 Tt=1 x ? ft if ? 1;
? regret({(xt , ft )}t2[T ] ) :=
(1)
PT
PT
:
1
1
? ? maxx2K T t=1 x ? ft T t=1 xt ? ft if ? < 1.
When the sequences {xt }t2[T ] , {ft }t2[T ] are obvious from context we will simply write ? regret
without stating these sequences. Also, when the sequence {xt }t2[T ] is randomized we will use
E[? regret] to denote the expected ?-regret.
2.1.1
Online linear optimization with full information
In OLO with full information, we consider a repeated game of T prediction rounds, for a fixed
T , where on each round t, the decision maker is required to choose a feasible action xt 2 K.
After committing to his choice, a linear loss function ft 2 F is revealed, and the decision maker
incurs loss of xt ? ft . In the payoff version, the decision maker incurs payoff of xt ? ft . The
game then continues to the next round. The overall goal of the decision maker is to guarantee
that ? regret({(xt , ft )}t2[T ] ) = O(T c ) for some c > 0, at least in expectation (in fact using
randomization is mandatory since K need not be convex). Here we assume that the adversary is
oblivious (aka non-adaptive), i.e., the sequence of losses/payoffs f1 , ..., fT is chosen in advance
(before the first round), and does not depend on the actions of the decision maker.
2.1.2
Bandit feedback
The bandit version of the problem is identical to the full information setting with one crucial difference:
on each round t, after making his choice, the decision maker does not observe the vector ft , but only
the value of his loss/payoff, given by xt ? ft .
2.2
Additional notation
For any two sets S, K ? Rd and a scalar 2 R we define the sets S + K := {x + y | x 2 S, y 2 K},
S := { x | x 2 S}. We also denote by CH(K) the convex-hull of all points in a set K. For a
convex and compact set S ? Rd and a point x 2 Rd we define dist(x, S) := minz2S kz xk. We
let B(c, r) denote the Euclidean ball or radius r centered in c.
2.3
Basic algorithmic tools
We now briefly describe two very basic ideas that are essential for constructing our algorithms,
namely the extended approximation oracle and the online gradient descent without feasibility method.
These were already suggested in [13] to obtain their low ?-regret algorithms. We note that in the
appendix we describe in more detail the approach of [13] and discuss its shortcomings in obtaining
oracle-efficient algorithms.
2.3.1
The extended approximation oracle
As discussed, a key difficulty of our setting that prevents us from directly applying well studied
algorithms for OLO, is that essentially all standard algorithms require to exactly solve (or up to
arbitrarily small error) some linear/convex optimization problem over the convexification of the
feasible set CH(K). However, not only that our approximation oracle OK (?) cannot perform exact
minimization, even for ? = 1 it is applicable only with inputs in Rd+ , and hence cannot optimize in
all directions. A natural approach, suggested in [13], to overcome the approximation error of the
oracle OK (?), is to consider optimization with respect to the convex set CH(?K) (i.e. convex hull
of all points in K scaled by a factor of ?) instead of CH(K). Indeed, if we consider for instance
the case ?
1, it is straightforward to see that for any c 2 Rd+ , OK (c) ? c ? ? minx2K x ? c =
4
? minx2CH(K) x ? c = minx2CH(?K) x ? c. Thus, in a certain sense, OK (?) can optimize with respect
to CH(?K) for all directions in Rd+ , although the oracle returns points in the original set K.
The following lemma shows that one can easily extend the oracle OK (?) to optimize with respect to
all directions in Rd .
Lemma 2.1 (Extended approximation oracle). Given c 2 Rd write c = c+ + c where c+ equals
to c on all non-negative coordinates of c and zero everywhere else, and c equals c on all negative
? K : Rd !
coordinates and zero everywhere else. The extended approximation oracle is a mapping O
(K + B(0, (1 + ?)R), K) defined as:
?
(OK (c+ ) ?R?
c , OK (c+ ))
if ? 1;
?K (c) = (v, s) :=
O
(2)
(OK ( c ) R?
c+ , OK ( c )) if ? < 1,
? = v/kvk if kvk > 0 and v
? = 0 otherwise, and it satisfies
where for any vector v 2 Rd we denote v
the following three properties:
1. v ? c ? minx2?K x ? c
2. 8f 2 F: s ? f ? v ? f if ?
3. kvk ? (? + 2)R
1 and s ? f
v ? f if ? < 1
The proof is given in the appendix for completeness.
It is important to note that while the extended oracle provides solutions with values at least as low as
any point in CH(?K), still in general the output point v need not be in either K or CH(?K), which
means that it is not a feasible point to play in our OLO setting, nor does it allow us to optimize
over CH(?K). This is why we also need the oracle to output the feasible point s 2 K which
dominates v for any possible loss/payoff vector in F. While we will use the outputs v to solve a
certain optimization problem involving CH(?K), this dominance relation will be used to convert the
solutions to these optimization problems into feasible plays for our OLO algorithms.
2.3.2
Online gradient descent with and without feasibility
As in [13], our online algorithms will be based on the well known Online Gradient Descent method
(OGD) for online convex optimization, originally due to [17]. For a sequence of loss vectors
{f1 , ..., fT } ? Rd OGD produces a sequence of plays {x1 , ..., xT } ? S, for a convex and compact
set S ? Rd via the following updates: 8t 1 : yt+1
xt ?ft , xt+1
arg minx2S kx yt+1 k2 ,
where x1 is initialized to some arbitrary point in S and ? is some pre-determined step-size. The
obvious difficulty in applying OGD to online linear optimization over S = CH(?K) is the step of
computing xt+1 by projecting yt+1 onto the feasible set S, since as discussed, even with the extended
approximation oracle, one cannot exactly optimize over CH(?K). Instead we will consider a variant
of OGD which may produce infeasible points, i.e., outside of S, but which guarantees low regret
with respect to any point in S. This algorithm, which we refer to as online gradient descent without
feasibility, is given below (Algorithm 1).
Algorithm 1 Online Gradient Descent Without Feasibility
1: input: learning rate ? > 0
2: x1
some point in S
3: for t = 1 . . . T do
d
4:
play xt and
? receive loss/payoff vector ft 2 R
5:
yt+1
6:
find xt+1
7: end for
xt ?ft for losses
xt + ?ft for payoffs
2 Rd such that
8z 2 S :
kz
xt+1 k2 ? kz
yt+1 k2
(3)
Lemma 2.2. [Online gradient descent without feasibility] Fix ? > 0. Suppose Algorithm 1 is applied
for T rounds and let {ft }Tt=1 ? Rd be the sequence of observed loss/payoff vectors, and let {xt }Tt=1
5
be the sequence of points played by the algorithm. Then for any x 2 S it holds that
PT
PT
? PT
1
1
1
2
xk2 + 2T
for losses;
t=1 xt ? ft
t=1 x ? ft ? 2T ? kx1
t=1 kft k
T
T
1
T
PT
t=1
x ? ft
1
T
PT
t=1
x t ? ft ?
1
2T ? kx1
xk2 +
The proof is given in the appendix for completeness.
3
?
2T
PT
t=1
kft k2
for payoffs.
Oracle-efficient Computation of (infeasible) Projections onto CH(?K)
In this section we detail our main technical tool for obtaining oracle-efficient online algorithms,
i.e., our algorithm for computing projections, in the sense of Eq. (3), onto the convex set CH(?K).
Before presenting our projection algorithm, Algorithm 2 and detailing its theoretical guarantees,
we first present the main algorithmic building block in the algorithm, which is described in the
following lemma. Lemma 3.1 shows that for any point x 2 Rd , we can either find a near-by point
p which is a convex combination of points outputted by the extended approximation oracle (and
hence, p is dominated by a convex combination of feasible points in K for any vector in F, as
discussed in Section 2.3.1), or we can find a separating hyperplane that separates x from CH(?K)
with sufficiently large margin. We achieve this by running the well known Ellipsoid method [10, 5]
in a very specialized way. This application of the Ellipsoid method is similar in spirit to those in
[15, 16], which applied this idea to computing correlated equilibrium in games and algorithmic
mechanism design, though the implementation details and the way in which we apply this technique
are quite different.
The proof of the following lemma is given in the appendix.
Lemma 3.1 (Separation-or-Decomposition via the
Fix x 2 Rd , ? 2
? Ellipsoid method).
?
(0, (? + 2)R], and a positive integer N
cd2 ln (?+1)R+kxk
, where c is a positive univer?
sal constant. Consider an attempt to apply the Ellipsoid method for N iterations to the following
feasibility problem:
find w 2 Rd such that:
8z 2 ?K :
(x
z) ? w
?
and
kwk ? 1,
(4)
such that each iteration of the Ellipsoid method applies the following consecutive steps:
?K ( w), where w is the current iterate. If (x v) ? w < ?, use v x as a
1. (v, s)
O
separating hyperplane for the Ellipsoid method and continue to to the next iteration
2. if kwk > 1, use w as a separating hyperplane for the Ellipsoid method and continue to the
next iteration
3. otherwise (kwk ? 1 and (x
vector w.
?), declare Problem (4) feasible and return the
v) ? w
Then, if the Ellipsoid method terminates declaring Problem 4 feasible, the returned vector w is a
feasible solution to Problem (4). Otherwise (the Ellipsoid method completes N iterations without
declaring Problem (4) feasible), let (v1 , s1 ), ..., (vN , sN ) be the outputs of the extended approximation oracle gathered throughout the run of the algorithm, and let (a1 , ..., aN ) be an optimal solution
to the following convex optimization problem:
2
N
N
X
1 X
min
ai v i x
such that 8i 2 {1, ..., N } : ai 0,
ai = 1.
(5)
(a1 ,...,aN ) 2
i=1
i=1
Then the point p =
PN
i=1
ai vi satisfies kx
pk ? 3?.
We are now ready to present our algorithm for computing projections onto CH(?K) (in the sense of
Eq. (3)). Consider now an attempt to project a point y 2 Rd , and note that in particular, y itself is a
valid projection (again, in the sense of Eq. (3)), however, in general, it is not a feasible point nor is
it dominated by a convex combination of feasible points. When attempting to project y 2 Rd , our
algorithm continuously applies the separation-or-decomposition procedure described in Lemma 3.1.
6
In case the procedure returns a decomposition, then by Lemma 3.1, we have a point that is sufficiently
close to y and is dominated for any vector in F by a convex combination (given explicitly) of feasible
points in K. Otherwise, the procedure returns a separating hyperplane which can be used to to ?pull
y closer" to CH(?K) in a way that the resulting point still satisfies the projection inequality given in
Eq. (3), and the process then repeats itself. Since each time we obtain a hyperplane separating our
current iterate from CH(?K), we pull the current iterate sufficiently towards CH(?K), this process
must terminate. Lemma 3.2 gives exact bounds on the performance of the algorithm.
Algorithm 2 (infeasible) Projection onto CH(?K)
1:
2:
3:
4:
5:
6:
7:
8:
9:
10:
11:
input: point y 2 Rd , tolerance ? > 0
?
y
y/ max{1, kyk/(?R)}
for t = 1 . . . do
call the S EPARATION - OR -D ECOMPOSTION procedure (Lemma 3.1) with parameters (?
y, ?)
if the procedure outputs a separating hyperplane w then
?
? ?w
y
y
else
let (a1 , ..., aN ), {(v1 , s1 ), ..., (vN , sN )} be the decomposition returned
? , (a1 , ..., aN ), {(v1 , s1 ), ..., (vN , sN )}
return y
end if
end for
Lemma 3.2. Fix y 2 Rd and ? 2 (0, (? + 2)R]. Algorithm 2 terminates after at most d?2 R2 /?2 e
? 2 Rd , a distribution (a1 , ..., aN ) and a set {(v1 , s1 ), ..., (vN , sN )}
iterations, returning a point y
outputted by the extended approximation oracle, where N is as defined in Lemma 3.1, such that
X
? k ? 3? for p :=
1. 8z 2 CH(?K) : k?
y zk2 ? ky zk2 ,
2. kp y
ai v i .
i2[N ]
? satisfies
Moreover, if the for loop was entered a total number of k times, then the final value of y
dist2 (?
y, CH(?K)) ? min{2?2 R2 , dist2 (y, CH(?K))
(k
1)?2 },
and the overall number of queries to the approximation oracle is O kd2 ln ((? + 1)R/?) .
It is important to note that the worst case iteration bound in Lemma 3.2 does not seem so appealing
for our purposes, since it depends polynomially on 1/?, and in our online algorithms naturally we
will need to take ? = O(T c ) for some c > 0, which seems to contradict our goal of achieving
poly-logarithmic in T oracle complexity, at least on average. However, as Lemma 3.2 shows, the
more iterations Algorithm 2 performs, the closer it brings its final iterate to the set CH(?K). Thus, as
we will show when analyzing the oracle complexity of our online algorithms, while a single call to
Algorithm 2 can be expensive, when calling it sequentially, where each input is a small perturbation
of the output of the previous call, the average number of iterations performed per such call cannot be
too high.
4
Efficient Algorithms for the Full Information and Bandit Settings
We now turn to present our online algorithms for the full-information and bandit settings together
with their regret bounds and oracle-complexity guarantees.
4.1
Algorithm for the full information setting
Our algorithm for the full-information setting, Algorithm 3, is given below.
Theorem 4.1. [Main Theorem] Fix ? > 0, ? 2 (0, (? + 2)R]. Suppose Algorithm 3 is applied for T
rounds and let {ft }Tt=1 ? F be the sequence of observed loss/payoff vectors, and let {st }Tt=1 be the
sequence of points played by the algorithm. Then it holds that
?
?
E ? regret {(st , ft )}t2[T ] ? ?2 R2 T 1 ? 1 + ?F 2 /2 + 3F ?,
and the average number of calls to the approximation oracle of K per iteration is upper bounded by
K(?, ?) := O
1 + ??RF + ? 2 F 2 ?
7
2
d2 ln ((? + 1)R/?) .
Algorithm 3 Online Gradient Descent with Infeasible Projections onto CH(?K)
1: input: learning rate ? > 0, projection error parameter ? > 0
?1
2: s1
some point in K, y
?s1
3: for t = 1 . . . T do
4:
play st and
? receive loss/payoff vector ft 2 F
y?t ?ft if ? 1
y?t + ?ft if ? < 1
? t+1 , a distribution
6:
call Algorithm 2 with inputs (yt+1 , ?) to obtain an approximated projection y
(a1 , ..., aN ) and {(v1 , s1 ), ..., (vN , sN )} ? Rd ? K, for some N 2 N.
7:
sample st+1 2 {s1 , ..., sN } according to distribution (a1 , ..., aN )
8: end for
5:
yt+1
In particular, setting ? = ?RT 2/3 /F , ? = ?RT 1/3 gives E [? regret] = O ?RF T 1/3 ,
1/2
K = O d2 ln ?+1
. Alternatively,? setting ? = ?RT
/F , ? = ?RT 1/2 gives
? T
?
p
E [? regret] = O ?RF T 1/2 , K = O
T d2 ln ?+1
.
? T
The proof is given in the appendix.
4.2
Algorithm for the bandit information setting
Our algorithm for the bandit setting follows from a very well known reduction from the bandit setting
to the full information setting, also applied in the bandit algorithm of [13]. The algorithm simply
simulates the full information algorithm, Algorithm 3, by providing it with estimated loss/payoff
vectors ?f1 , ..., ?fT instead of the true vectors f1 , ..., fT which are not available in the bandit setting.
This reduction is based on the use of a Barycentric Spanner (defined next) for the feasible set K.
As standard, we assume the points in K span the entire space Rd , otherwise we can reformulate the
problem in a lower-dimensional space, in which this assumption holds.
Definition 4.1 (Barycentric Spanner3 ). We say that a set of d vectors {q1 , ..., qd } ? Rd is a
Barycentric Spanner with parameter > 0 for a set S ? Rd , denoted by -BS(S), if it holds that
Pd
1
{q1 , ..., qd } ? S, and the matrix Q := i=1 qi q>
qi k ? .
i is not singular and maxi2[d] kQ
Importantly, as discussed in [13], the assumption on the availability of such a set -BS(K) seems
reasonable, since i) for many sets that correspond to the set of all possible solutions to some wellstudied NP-Hard optimization problem, one can still construct in poly(d) time a barycentric spanner
with = poly(d), ii) -BS(K) needs to be constructed only once and then stored in memory (overall
d vectors in Rd ), and hence its construction can be viewed as a pre-processing step, and iii) as
illustrated in [13], without further assumptions, the approximation oracle by itself is not sufficient to
guarantee nontrivial regret bounds in the bandit setting.
The algorithm and the proof of the following theorem are given in the appendix.
Theorem 4.2. Fix ? > 0, ? 2 (0, (? + 2)R], 2 (0, 1). Suppose Algorithm 5 is applied for T
rounds and let {ft }Tt=1 ? F be the sequence of observed loss/payoff vectors, and let {?st }Tt=1 be the
sequence of points played by the algorithm. Then it holds that
?
?
E ? regret {(?st , ft )}t2[T ] ? ?2 R2 ? 1 T 1 + ?d2 C 2 2 1 /2 + 3?F + C,
and the expected number of calls to the approximation oracle of K per iteration is upper bounded by
E [K(?, ?, )] := O
In particular, setting ? =
O (? dCR + ?RF + C)T
1 + ?? dCR + (?dC )2 /
, ? = ?RT
, E[K] = O d2 ln
?R
dC T
1/3
2/3
3
?
,
?+1
? T
1/3
2
d2 ln ((? + 1)R/?) .
= T
.
1/3
gives E [?
regret] =
this definition is somewhat different than the classical one given in [2], however it is equivalent to a
C-approximate barycentric spanner [2], with an appropriately chosen constant C( ).
8
References
[1] Jacob Abernethy, Elad Hazan, and Alexander Rakhlin. Competing in the dark: An efficient algorithm for
bandit linear optimization. In COLT, pages 263?274, 2008.
[2] Baruch Awerbuch and Robert D Kleinberg. Adaptive routing with end-to-end feedback: Distributed
learning and geometric approaches. In Proceedings of the thirty-sixth annual ACM symposium on Theory
of computing, pages 45?53. ACM, 2004.
[3] Maria-Florina Balcan and Avrim Blum. Approximation algorithms and online mechanisms for item pricing.
In Proceedings of the 7th ACM Conference on Electronic Commerce, pages 29?35. ACM, 2006.
[4] Reuven Bar-Yehuda and Shimon Even. A linear-time approximation algorithm for the weighted vertex
cover problem. Journal of Algorithms, 2(2):198?203, 1981.
[5] S?bastien Bubeck. Convex optimization: Algorithms and complexity. Foundations and Trends R in
Machine Learning, 8(3-4):231?357, 2015.
[6] Nicos Christofides. Worst-case analysis of a new heuristic for the travelling salesman problem. Technical
report, DTIC Document, 1976.
[7] V. Chvatal. A greedy heuristic for the set-covering problem. Mathematics of Operations Research,
4(3):233?235, 1979.
[8] Takahiro Fujita, Kohei Hatano, and Eiji Takimoto. Combinatorial online prediction via metarounding. In
ALT, pages 68?82. Springer, 2013.
[9] Michel X Goemans and David P Williamson. Improved approximation algorithms for maximum cut and
satisfiability problems using semidefinite programming. Journal of the ACM (JACM), 42(6):1115?1145,
1995.
[10] M. Gr?tschel, L. Lov?sz, and A. Schrijver. The ellipsoid method and its consequences in combinatorial
optimization. Combinatorica, 1(2):169?197, 1981.
[11] Elad Hazan, Zohar Shay Karnin, and Raghu Meka. Volumetric spanners: an efficient exploration basis for
learning. In COLT, volume 35, pages 408?422, 2014.
[12] Elad Hazan and Haipeng Luo. Variance-reduced and projection-free stochastic optimization. In Proceedings
of the 33nd International Conference on Machine Learning, ICML 2016, New York City, NY, USA, June
19-24, 2016, pages 1263?1271, 2016.
[13] Sham M. Kakade, Adam Tauman Kalai, and Katrina Ligett. Playing games with approximation algorithms.
SIAM J. Comput., 39(3):1088?1106, 2009.
[14] Adam Kalai and Santosh Vempala. Efficient algorithms for online decision problems. Journal of Computer
and System Sciences, 71(3):291?307, 2005.
[15] Christos H Papadimitriou and Tim Roughgarden. Computing correlated equilibria in multi-player games.
Journal of the ACM (JACM), 55(3):14, 2008.
[16] S Matthew Weinberg. Algorithms for strategic agents. PhD thesis, Massachusetts Institute of Technology,
2014.
[17] Martin Zinkevich. Online convex programming and generalized infinitesimal gradient ascent. In Machine
Learning, Proceedings of the Twentieth International Conference (ICML 2003), August 21-24, 2003,
Washington, DC, USA, pages 928?936, 2003.
9
| 6665 |@word briefly:1 version:4 achievable:1 polynomial:1 seems:3 norm:1 nd:1 d2:6 decomposition:4 jacob:1 q1:2 incurs:3 reduction:2 celebrated:3 etric:1 document:1 current:3 luo:1 yet:1 kft:2 readily:1 must:1 ligett:2 update:1 greedy:1 selected:1 prohibitive:2 instantiate:1 kyk:1 item:1 plane:1 xk:1 vanishing:1 completeness:2 provides:1 constructed:1 symposium:1 prove:1 dan:1 lov:1 indeed:1 expected:10 roughly:1 dist:1 nor:2 multi:1 project:2 moreover:2 underlying:1 notation:1 bounded:2 israel:1 hindsight:5 finding:1 guarantee:15 every:1 runtime:1 exactly:2 returning:1 scaled:1 k2:4 omit:1 t1:2 before:2 positive:2 declare:1 treat:1 consequence:1 analyzing:1 minx2k:3 might:3 studied:2 resembles:1 ease:1 commerce:1 thirty:1 yehuda:1 regret:43 block:1 procedure:6 kohei:1 significantly:3 outputted:2 projection:16 convenient:1 pre:2 get:2 cannot:7 onto:6 close:1 context:1 applying:2 optimize:5 equivalent:1 zinkevich:1 yt:7 go:1 straightforward:1 convex:22 importantly:2 pull:2 his:8 coordinate:2 construction:2 pt:9 play:5 suppose:3 exact:3 programming:2 ogd:11 designing:1 trend:1 approximated:4 expensive:1 continues:1 cut:3 convexification:1 observed:3 ft:34 capture:1 eighted:2 worst:4 observes:1 sal:1 pd:1 complexity:10 inimum:1 depend:1 solving:1 technically:1 basis:1 easily:2 instantiated:1 fast:1 committing:1 describe:2 shortcoming:1 kp:1 query:2 choosing:3 outside:1 abernethy:1 quite:1 garber:1 widely:1 solve:4 elad:3 say:1 relax:1 otherwise:6 katrina:1 heuristic:2 ability:1 itself:5 tsp:1 final:2 online:31 christofides:1 differentiate:1 sequence:17 product:1 relevant:2 loop:1 date:1 entered:1 achieve:5 kx1:2 haipeng:1 ky:1 dist2:2 produce:2 guaranteeing:1 adam:2 tim:1 coupling:1 ac:1 stating:1 eq:4 strong:1 implemented:1 qd:2 direction:3 radius:1 hull:3 luckily:1 centered:1 exploration:1 routing:1 stochastic:1 require:3 f1:5 fix:5 preliminary:1 randomization:1 univer:1 extension:2 hold:7 sufficiently:3 considered:1 equilibrium:2 algorithmic:3 mapping:1 matthew:1 achieves:2 consecutive:2 xk2:2 purpose:1 applicable:1 combinatorial:2 maker:16 city:1 tool:3 weighted:1 minimization:5 kalai:4 pn:1 focus:1 june:1 improvement:1 maria:1 mainly:1 aka:1 sense:4 entire:3 fptas:3 bandit:22 relation:1 interested:1 fujita:1 overall:3 arg:1 colt:2 denoted:1 equal:2 construct:3 once:1 karnin:1 beach:1 santosh:1 washington:1 identical:1 kd2:1 icml:2 oco:1 nearly:1 papadimitriou:1 np:8 report:2 t2:9 oblivious:2 attempt:2 interest:2 wellstudied:1 kvk:3 semidefinite:1 kt:1 beforehand:1 maxi2:1 closer:2 euclidean:3 detailing:1 initialized:1 theoretical:1 instance:2 cover:1 maximization:3 strategic:1 vertex:1 vectors2:1 kq:1 technion:2 gr:1 too:1 stored:1 reuven:1 dependency:1 perturbed:2 chooses:1 st:7 international:2 randomized:1 siam:1 accessible:3 sequel:2 together:1 continuously:1 again:1 thesis:1 choose:2 possibly:1 admit:2 inefficient:1 return:5 michel:1 availability:4 satisfy:1 explicitly:1 vi:1 depends:1 multiplicative:3 wether:1 performed:1 kwk:3 hazan:3 contribution:1 minimize:3 il:1 accuracy:1 variance:1 efficiently:5 gathered:1 correspond:1 volumetric:1 infinitesimal:1 definition:3 sixth:1 obvious:2 naturally:3 proof:5 massachusetts:1 ask:1 knowledge:1 improves:1 eparation:1 satisfiability:1 ok:13 originally:2 follow:2 improved:3 done:1 though:1 hand:1 receives:1 replacing:1 brings:1 pricing:1 usa:3 building:1 true:1 awerbuch:1 hence:4 i2:1 illustrated:1 pwe:1 round:14 game:9 covering:1 noted:1 generalized:2 presenting:1 tt:7 performs:1 balcan:1 wise:1 specialized:3 volume:1 discussed:7 he:1 extend:1 refer:1 imposing:1 ai:5 meka:1 rd:35 mathematics:1 similarly:3 hatano:1 showed:3 hide:1 mandatory:1 certain:4 inequality:1 arbitrarily:2 continue:2 additional:7 somewhat:1 maximize:1 ii:1 full:16 sham:1 technical:2 believed:1 long:1 a1:7 feasibility:6 qi:2 prediction:3 variant:5 basic:2 involving:1 florina:1 essentially:1 expectation:1 iteration:21 receive:2 whereas:1 else:3 grow:1 completes:1 singular:1 crucial:1 appropriately:1 takahiro:1 ascent:1 simulates:1 spirit:1 seem:2 call:16 integer:1 near:1 revealed:1 iii:1 enough:1 chvatal:1 iterate:4 competing:1 inner:1 idea:2 maxf:1 render:1 returned:2 york:1 action:14 repeatedly:1 useful:1 detailed:1 dark:1 eiji:1 reduced:1 exist:3 revisit:2 estimated:1 per:12 write:2 dominance:1 key:1 nevertheless:1 blum:1 achieving:1 clarity:1 takimoto:1 v1:5 baruch:1 relaxation:1 convert:1 run:1 everywhere:2 throughout:4 reasonable:4 electronic:1 vn:5 separation:2 decision:19 appendix:7 bound:15 played:3 quadratic:1 oracle:50 annual:1 nontrivial:1 roughgarden:1 infinity:1 calling:1 dominated:4 kleinberg:1 span:2 min:2 attempting:1 vempala:2 martin:1 according:2 combination:5 ball:1 terminates:2 kakade:2 appealing:1 b:8 making:1 s1:8 projecting:1 ln:7 discus:2 turn:1 mechanism:2 dcr:2 end:6 zk2:2 travelling:1 salesman:1 available:2 operation:1 raghu:1 apply:4 observe:3 generic:4 original:2 running:2 include:1 hinge:1 classical:1 objective:1 question:1 quantity:1 already:1 primary:1 dependence:2 rt:5 gradient:10 reversed:1 separate:1 mapped:1 separating:6 trivial:3 kkl:1 assuming:1 length:2 ellipsoid:11 providing:2 minimizing:2 reformulate:1 unfortunately:1 robert:1 weinberg:1 negative:3 implementation:3 design:1 perform:2 conversion:1 upper:2 descent:9 orthant:1 payoff:23 extended:9 dc:3 barycentric:7 perturbation:1 arbitrary:1 thm:2 august:1 introduced:1 david:1 namely:1 required:4 specified:1 nip:1 address:1 beyond:1 adversary:8 suggested:2 below:2 bar:1 zohar:1 challenge:1 spanner:7 rf:4 max:3 memory:1 natural:4 difficulty:2 technology:2 carried:1 ready:1 fpl:5 sn:6 geometric:1 kf:1 loss:28 fully:1 interesting:1 declaring:2 foundation:1 shay:1 agent:1 sufficient:1 playing:1 repeat:2 free:1 infeasible:5 offline:5 drastically:1 allow:1 institute:2 tauman:1 tolerance:1 distributed:1 feedback:3 overcome:1 valid:1 kz:3 concretely:2 made:1 adaptive:3 preprocessing:2 far:1 polynomially:1 approximate:1 compact:3 contradict:1 cutting:1 sz:1 sequentially:1 assumed:1 conclude:1 leader:2 alternatively:1 iterative:1 why:1 table:2 terminate:1 ca:1 obtaining:2 williamson:1 tschel:1 poly:5 constructing:1 pk:1 main:6 repeated:1 x1:4 ny:1 christos:1 comput:1 shimon:1 theorem:4 xt:23 bastien:1 r2:4 rakhlin:1 alt:1 concern:1 dominates:1 exists:1 essential:1 cd2:1 avrim:1 phd:1 kx:2 margin:1 dtic:1 logarithmic:3 simply:3 likely:1 bubeck:1 jacm:2 twentieth:1 prevents:1 kxk:2 scalar:1 applies:2 springer:1 ch:24 corresponds:2 satisfies:6 acm:6 goal:4 viewed:2 presentation:1 towards:1 feasible:29 hard:8 determined:1 hyperplane:6 lemma:15 called:2 total:1 goemans:1 schrijver:1 player:2 formally:1 combinatorica:1 latter:1 alexander:1 olo:8 correlated:2 |
6,263 | 6,666 | Geometric Descent Method for
Convex Composite Minimization
Shixiang Chen1 , Shiqian Ma2 , and Wei Liu3
1
Department of SEEM, The Chinese University of Hong Kong, Hong Kong
2
Department of Mathematics, UC Davis, USA
3
Tencent AI Lab, China
Abstract
In this paper, we extend the geometric descent method recently proposed by Bubeck,
Lee and Singh [1] to tackle nonsmooth and strongly convex composite problems.
We prove that our proposed algorithm, dubbed geometric
proximal gradient method
?
(GeoPG), converges with a linear rate (1 ? 1/ ?) and thus achieves the optimal
rate among first-order methods, where ? is the condition number of the problem.
Numerical results on linear regression and logistic regression with elastic net
regularization show that GeoPG compares favorably with Nesterov?s accelerated
proximal gradient method, especially when the problem is ill-conditioned.
1
Introduction
Recently, Bubeck, Lee and Singh proposed a geometric descent method (GeoD) for minimizing a
smooth and strongly convex function [1]. They showed that GeoD achieves the same optimal rate
as Nesterov?s accelerated gradient method (AGM) [2, 3]. In this paper, we provide an extension of
GeoD that minimizes a nonsmooth function in the composite form:
min F (x) := f (x) + h(x),
x?Rn
(1.1)
where f is ?-strongly convex and ?-smooth (i.e., ?f is Lipschitz continuous with Lipschitz constant
?), and h is a closed nonsmooth convex function with simple proximal mapping. Commonly seen
examples of h include `1 norm, `2 norm, nuclear norm, and so on.
If h vanishes, then the objective function of (1.1) becomes?smooth and strongly convex. In this case,
it is known that AGM converges with a linear rate (1 ? 1/ ?), which is optimal among all first-order
methods, where ? = ?/? is the condition number of the problem. However, AGM lacks a clear
geometric intuition, making it difficult to interpret. Recently, there has been much work on attempting
to explain AGM or designing new algorithms with the same optimal rate (see, [4, 5, 1, 6, 7]). In
particular, the GeoD method proposed in [1] has a clear geometric intuition that is in the flavor
of the ellipsoid method [8]. The follow-up work [9, 10] attempted to improve the performance of
GeoD by exploiting the gradient information from the past with a ?limited-memory? idea. Moreover,
Drusvyatskiy, Fazel and Roy [10] showed how to extend the suboptimal version of GeoD (with the
convergence rate (1 ? 1/?)) to solve the composite problem (1.1). However, it was not clear how to
extend the optimal version of GeoD to address (1.1), and the authors posed this as an open question.
In this paper, we settle this question by proposing a geometric proximal gradient (GeoPG) algorithm
which can solve the composite problem (1.1). We further show how to incorporate various techniques
to improve the performance of the proposed algorithm.
Notation. We use B(c, r2 ) = x|kx ? ck2 ? r2 to denote the ball with center c and radius r.
We use Line(x, y) to denote the line that connects x and y, i.e., {x + s(y ? x), s ? R}. For fixed
t ? (0, 1/?], we denote x+ := Proxth (x ? t?f (x)), where the proximal mapping Proxh (?) is
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
defined as Proxh (x) := argminz h(z) + 12 kz ? xk2 . The proximal gradient of F is defined as
Gt (x) := (x?x+ )/t. It should be noted that x+ = x?tGt (x). We also denote x++ := x?Gt (x)/?.
Note that both x+ and x++ are related to t, and we omit t whenever there is no ambiguity.
The rest of this paper is organized as follows. In Section 2, we briefly review the GeoD method for
solving smooth and strongly convex problems. In Section 3, we provide our GeoPG algorithm for
solving nonsmooth problem (1.1) and analyze its convergence rate. We address two practical issues of
the proposed method in Section 4, and incorporate two techniques: backtracking and limited memory,
to cope with these issues. In Section 5, we report some numerical results of comparing GeoPG with
Nesterov?s accelerated proximal gradient method in solving linear regression and logistic regression
problems with elastic net regularization. Finally, we conclude the paper in Section 6.
2
Geometric Descent Method for Smooth Problems
The GeoD method [1] solves (1.1) when h ? 0, in which the problem reduces to a smooth and
strongly convex problem min f (x). We denote its optimal solution and optimal value as x? and
f ? , respectively. Throughout this section, we fix t = 1/?, which together with h ? 0 implies that
x+ = x ? ?f (x)/? and x++ = x ? ?f (x)/?. We first briefly describe the basic idea of the
suboptimal GeoD. Since f is ?-strongly convex, the following inequality holds
?
f (x) + h?f (x), y ? xi + ky ? xk2 ? f (y), ?x, y ? Rn .
(2.1)
2
By letting y = x? in (2.1), it is easy to obtain
x? ? B x++ , k?f (x)k2 /?2 ? 2(f (x) ? f ? )/? , ?x ? Rn .
(2.2)
Note that the ?-smoothness of f implies
f (x+ ) ? f (x) ? k?f (x)k2 /(2?), ?x ? Rn .
(2.3)
Combining (2.2) and (2.3) yields x ? B x , (1 ? 1/?)k?f (x)k /? ? 2(f (x ) ? f )/? . As
a result, suppose that initially we have a ball B(x0 , R02 ) that contains x? , then it follows that
+
2
2
?
x? ? B x0 , R02 ? B x++
(2.4)
0 , (1 ? 1/?)k?f (x0 )k /? ? 2(f (x0 ) ? f )/? .
Some simple algebraic calculations show that the squared radius of the minimum enclosing ball
of the right hand side of (2.4) is no larger than R02 (1 ? 1/?), i.e., there exists some x1 ? Rn such
that x? ? B x1 , R02 (1 ? 1/?) . Therefore, the squared radius of the initial ball shrinks by a factor
(1 ? 1/?). Repeating this process yields a linear convergent sequence {xk } with the convergence
rate (1 ? 1/?): kxk ? x? k2 ? (1 ? 1/?)k R02 .
?
The optimal GeoD (with the linear convergence rate (1 ? 1/ ?)) maintains two balls containing
x? in each iteration, whose centers are ck and x++
k+1 , respectively. More specifically, suppose that in
the k-th iteration we have ck and xk , then ck+1 and xk+1 are obtained as follows. First, xk+1 is the
2
minimizer of f on Line(ck , x+
k ). Second, ck+1 (resp. Rk+1 ) is the center (resp. squared radius) of
the ball (given by Lemma 2.1) that contains
2
2
B ck , Rk2 ? k?f (xk+1 )k2 /(?2 ?) ? B x++
k+1 , (1 ? 1/?)k?f (xk+1 )k /? .
Calculating ck+1 and Rk+1 is easy and we refer to Algorithm 1 of [1] for details. By applying
Lemma 2.1 with xA = ck , rA = Rk , rB = k?f (xk+1 )k/?, = 1/? and ? = ?2 (f (x+
) ? f (x? )),
?
? kk 2
2
2
?
2
we obtain Rk+1 = (1 ? 1/ ?)Rk , which further implies kx ? ck k ? (1 ? 1/ ?) R0 , i.e., the
?
optimal GeoD converges with the linear rate (1 ? 1/ ?).
2
2
Lemma 2.1 (see [1, 10]). Fix centers xA , xB ? Rn and squared radii rA
, rB
> 0. Also fix ? (0, 1)
2
2
n
and suppose kxA ? xB k ? rB . There exists a new center c ? R such that for any ? > 0, we have
? 2
2
2
2
B(xA , rA
? rB
? ?) ? B xB , rB
(1 ? ) ? ? ? B c, (1 ? )rA
?? .
?
3
++
2
2
+
?
Geometric Descent Method for Nonsmooth Convex Composite Problems
Drusvyatskiy, Fazel and Roy [10] extended the suboptimal GeoD to solve the composite problem
(1.1). However, it was not clear how to extend the optimal GeoD to solve problem (1.1). We resolve
this problem in this section.
The following lemma is useful to our analysis. Its proof is in the supplementary material.
2
Lemma 3.1. Given point x ? Rn and step size t ? (0, 1/?], denote x+ = x ? tGt (x). The following
inequality holds for any y ? Rn :
t
?
F (y) ? F (x+ ) + hGt (x), y ? xi + kGt (x)k2 + ky ? xk2 .
(3.1)
2
2
3.1
GeoPG Algorithm
In this subsection, we describe our proposed geometric proximal gradient method (GeoPG) for
solving (1.1). Throughout Sections 3.1 and 3.2, t ? (0, 1/?] is a fixed scalar. The key observation for
designing GeoPG is that in the k-th iteration one has to find xk that lies on Line(x+
k?1 , ck?1 ) such
that the following two inequalities hold:
t
1
+
++
2
(3.2)
F (x+
? ck?1 k2 ? 2 kGt (xk )k2 .
k ) ? F (xk?1 ) ? kGt (xk )k , and kxk
2
?
Intuitively, the first inequality in (3.2) requires that there is a function value reduction on x+
k from
x+
,
and
the
second
inequality
requires
that
the
centers
of
the
two
balls
are
far
away
from
each
other
k?1
so that Lemma 2.1 can be applied.
The following lemma gives a sufficient condition for (3.2). Its proof is in the supplementary material.
Lemma 3.2. (3.2) holds if xk satisfies
+
+
hx+
(3.3)
k ? xk , xk?1 ? xk i ? 0, and hxk ? xk , xk ? ck?1 i ? 0.
Therefore, we only need to find xk such that (3.3) holds. To do so, we define the following functions
for given x, c (x 6= c) and t ? (0, ?]:
?t,x,c (z) = hz + ? z, x ? ci, ?z ? Rn , and ??t,x,c (s) = ?t,x,c x + s(c ? x) , ?s ? R.
The functions ?t,x,c (z) and ??t,x,c (s) have the following properties. Its proof can be found in the
supplementary material.
Lemma 3.3. (i) ?t,x,c (z) is Lipschitz continuous. (ii) ??t,x,c (s) strictly monotonically increases.
We are now ready to describe how to find xk such that (3.3) holds. This is summarized in Lemma 3.4.
Lemma 3.4. The following two ways find xk satisfying (3.3).
(i) If ??t,x+
(1) ? 0, then (3.3) holds by setting xk := ck?1 ; if ??t,x+ ,ck?1 (0) ? 0, then
k?1
(3.3) holds by setting xk := x+ ; if ?? +
(1) > 0 and ?? +
(0) < 0, then
k?1 ,ck?1
k?1
there exists s ? [0, 1] such that ??t,x+
t,xk?1 ,ck?1
k?1
t,xk?1 ,ck?1
,ck?1 (s) = 0. As a result, (3.3) holds by setting
+
xk := x+
k?1 + s(ck?1 ? xk?1 ).
(ii) If ??t,x+
?
(0) ? 0, then (3.3) holds by setting xk := x+
(0) < 0,
k?1 ; if ?t,x+
k?1 ,ck?1
?
then there exists s ? 0 such that ?t,x+ ,ck?1 (s) = 0. As a result, (3.3) holds by setting
k?1 ,ck?1
k?1
+
xk := x+
k?1 + s(ck?1 ? xk?1 ).
Proof. Case (i) directly follows from the Mean-Value Theorem. Case (ii) follows from the monotonicity and continuity of ??t,x+ ,ck?1 from Lemma 3.3.
k?1
It is indeed very easy to find xk satisfying the two cases in Lemma 3.4, since we are tackling a
univariate Lipschitz continuous function ??t,x,c (s) . Specifically, for case (i) of Lemma 3.4, we can
use the bisection method to find the zero of ??t,x+ ,ck?1 in the closed interval [0, 1]. In practice, we
k?1
found that the Brent-Dekker method [11, 12] performs much better than the bisection method, so
we use the Brent-Dekker method in our numerical experiments. For case (ii) of Lemma 3.4, we
can use the semi-smooth Newton method to find the zero of ??t,x+ ,ck?1 in the interval [0, +?).
k?1
In our numerical experiments, we implemented the global semi-smooth Newton method [13, 14]
and obtained very encouraging results. These two procedures are described in Algorithms 1 and 2,
respectively. Based on the discussions above, we know that xk generated by these two algorithms
satisfies (3.3) and hence (3.2).
We are now ready to present our GeoPG algorithm for solving (1.1) as in Algorithm 3.
3
Algorithm 1 : The first procedure for finding xk from given x+
k?1 and ck?1 .
+
+
+
1: if h(x+
k?1 ) ? xk?1 , xk?1 ? ck?1 i ? 0 then
2:
set xk := x+
k?1 ;
3: else if hc+
?
ck?1 , x+
k?1
k?1 ? ck?1 i ? 0 then
4:
set xk := ck?1 ;
5: else
use the Brent-Dekker method to find s ? [0, 1] such that ??t,x+
6:
k?1 ,ck?1
xk :=
7: end if
x+
k?1
+ s(ck?1 ?
(s) = 0, and set
x+
k?1 );
Algorithm 2 : The second procedure for finding xk from given x+
k?1 and ck?1 .
+
+
+
1: if h(x+
k?1 ) ? xk?1 , xk?1 ? ck?1 i ? 0 then
+
2:
set xk := xk?1 ;
3: else
use the global semi-smooth Newton method [13, 14] to find the root s ? [0, +?) of
4:
+
??t,x+ ,ck?1 (s), and set xk := x+
k?1 + s(ck?1 ? xk?1 );
k?1
5: end if
3.2
Convergence Analysis of GeoPG
We are now ready to present our main convergence result for GeoPG.
2
(x0 )k
Theorem 3.5. Given initial point x0 and step size t ? (0, 1/?], we set R02 = kGt ?
(1 ? ?t).
2
?
Suppose that sequence {(xk , ck , Rk )} is generated by Algorithm 3, and that x is the optimal
?
?
2
solution of (1.1)
? and F2 is the optimal objective value. For any k ? 0, one has x ? B(ck , Rk ) and
2
Rk+1 ? (1 ? ?t)Rk , and thus
?
?
?
?
(1 ? ?t)k R02 .
kx? ? ck k2 ? (1 ? ?t)k R02 , and F (x+
(3.4)
k+1 ) ? F ?
2
?
Note that when t = 1/?, (3.4) implies the linear convergence rate (1 ? 1/ ?).
Proof. We prove a stronger result by induction that for every k ? 0, one has
?
x? ? B ck , Rk2 ? 2(F (x+
k ) ? F )/? .
(3.5)
Let y = x? in (3.1). We have kx? ? x++ k2 ? (1 ? ?t)kGt (x)2 k/?2 ? 2(F (x+ ) ? F ? )/?, implying
x? ? B x++ , kGt (x)k2 (1 ? ?t)/?2 ? 2(F (x+ ) ? F ? )/? .
(3.6)
Setting x = x0 in (3.6) shows that (3.5) holds for k = 0. We now assume that (3.5) holds for some
k ? 0, and in the following we will prove that (3.5) holds for k + 1. Combining (3.5) and the first
inequality of (3.2) yields
?
x? ? B ck , Rk2 ? tkGt (xk+1 )k2 /? ? 2(F (x+
(3.7)
k+1 ) ? F )/? .
By setting x = xk+1 in (3.6), we obtain
+
2
2
?
x? ? B x++
k+1 , kGt (xk+1 )k (1 ? ?t)/? ? 2(F (xk+1 ) ? F )/? .
x++
k+1 ,
(3.8)
We now apply Lemma 2.1 to (3.7) and (3.8). Specifically, we set xB =
xA = ck , = ?t,
+
2
?
2
2
rA = Rk , rB = kGt (xk+1 )k/?, ? = ? (F (xk ) ? F ), and note that kxA ? xB k ? rB
because of
the second inequality of (3.2). Then Lemma 2.1 indicates that there exists ck+1 such that
?
?
x? ? B ck+1 , (1 ? 1/ ?)Rk2 ? 2(F (x+
(3.9)
k+1 ) ? F )/? ,
?
2
? (1 ? ?t)Rk2 . Note that ck+1 is the center of the minimum
i.e., (3.5) holds for k + 1 with Rk+1
enclosing ball of the intersection of the two balls in (3.7) and (3.8), and can be computed
in the
?
same?way as Algorithm 1 of [1]. From (3.9) we obtain that kx? ? ck+1 k2 ? (1?? ?t)Rk2 ?
? 2
?
?
(1 ? ?t)k+1 R02 . Moreover, (3.7) indicates that F (x+
?t)k R02 .
k+1 ) ? F ? 2 Rk ? 2 (1 ?
4
Algorithm 3 : GeoPG: geometric proximal gradient descent for convex composite minimization.
Require: Parameters ?, ?, initial point x0 and step size t ? (0, 1/?].
2
2
2
1: Set c0 = x++
0 , R0 = kGt (x0 )k (1 ? ?t)/? ;
2: for k = 1, 2, . . . do
3:
Use Algorithm 1 or 2 to find xk ;
2
4:
Set xA := x++
= xk ? Gt (xk )/?, and RA
= kGt (xk )k2 (1 ? ?t)/?2 ;
k
+
2
2
5:
Set xB := ck?1 , and RB = Rk?1 ? 2(F (xk?1 ) ? F (x+
k ))/?;
2
2
Compute B(ck , Rk2 ): the minimum enclosing ball of B(xA , RA
) ? B(xB , RB
), which can be
6:
done using Algorithm 1 in [1];
7: end for
4
4.1
Practical Issues
GeoPG with Backtracking
In practice, the Lipschitz constant ? may be unknown to us. In this subsection, we describe a
backtracking strategy for GeoPG in which ? is not needed. From the ?-smoothness of f , we have
f (x+ ) ? f (x) ? th?f (x), Gt (x)i + tkGt (x)k2 /2.
(4.1)
Note that inequality (3.1) holds because of (4.1), which holds when t ? (0, 1/?]. If ? is unknown,
we can perform backtracking on t such that (4.1) holds, which is a common practice for proximal
gradient method, e.g., [15?17]. Note that the key step in our analysis of GeoPG is to guarantee that
the two inequalities in (3.2) hold. According to Lemma 3.2, the second inequality in (3.2) holds as
long as we use Algorithm 1 or Algorithm 2 to find xk , and it does not need the knowledge of ?.
However, the first inequality in (3.2) requires t ? 1/?, because its proof in Lemma 3.2 needs (3.1).
Thus, we need to perform backtracking on t until (4.1) is satisfied, and use the same t to find xk by
Algorithm 1 or Algorithm 2. Our GeoPG algorithm with backtracking (GeoPG-B) is described in
Algorithm 4.
Algorithm 4 : GeoPG with Backtracking (GeoPG-B)
Require: Parameters ?, ? ? (0, 1), ? ? (0, 1), initial step size t0 > 0 and initial point x0 .
Repeat t0 := ?t0 until (4.1) holds for t = t0 ;
kGt0 (x0 )k2
2
Set c0 = x++
(1 ? ?t0 );
0 , R0 =
?2
for k = 1, 2, . . . do
if no backtracking was performed in the (k ? 1)-th iteration then
Set tk := tk?1 /?;
else
Set tk := tk?1 ;
end if
Compute xk by Algorithm 1 or Algorithm 2 with t = tk ;
tk
2
while f (x+
k ) > f (xk ) ? tk h?f (xk ), Gtk (xk )i + 2 kGtk (xk )k do
Set tk := ?tk (backtracking);
Compute xk by Algorithm 1 or Algorithm 2 with t = tk ;
end while
kG (x )k2
2
Set xA := x++
= xk ? Gtk (xk )/?, RA
= tk?2 k (1 ? ?tk );
k
+
2
2
Set xB := ck?1 , RB
= Rk?1
? ?2 (F (x+
k?1 ) ? F (xk ));
2
2
Compute B(ck , Rk2 ): the minimum enclosing ball of B(xA , RA
) ? B(xB , RB
);
end for
Note that the sequence {tk } generated in Algorithm 4 is uniformly bounded away from 0. This is
because (4.1) always holds when tk ? 1/?. As a result, we know tk ? tmin := mini=0,...,k ti ? ?/?.
It is easy to see that in the k-th iteration of Algorithm 4, x? is contained in two balls:
2
?
x? ? B ck?1 , Rk?1
? tk kGtk (xk )k2 /? ? 2(F (x+
k ) ? F )/?,
+
2
2
?
x? ? B x++
k , kGtk (xk )k (1 ? ?tk )/? ? 2(F (xk ) ? F )/? .
5
Therefore, we have the following convergence result for Algorithm 4, whose proof is similar to that
for Algorithm 3. We thus omit the proof for succinctness.
Theorem 4.1. Suppose that {(xk , ck , Rk , tk )} is generated by Algorithm 4. For any k ? 0, one
Qk
?
?
2
has x? ? B(ck , Rk2 ) and Rk+1
? (1 ? ?tk )Rk2 , and thus kx? ? ck k2 ? i=0 (1 ? ?ti )i R02 ?
?
(1 ? ?tmin )k R02 .
4.2
GeoPG with Limited Memory
The basic idea of GeoD is that in each iteration we maintain two balls B(y1 , r12 ) and B(y2 , r22 ) that
both contain x? , and then compute the minimum enclosing ball of their intersection, which is expected
to be smaller than both B(y1 , r12 ) and B(y2 , r22 ). One very intuitive idea that can possibly improve
the performance of GeoD is to maintain more balls from the past, because their intersection should be
smaller than the intersection of two balls. This idea has been proposed by [9] and [10]. Specifically,
Bubeck and Lee [9] suggested to keep all the balls from past iterations and then compute the minimum
enclosing ball of their intersection. For a given bounded set Q, the center of its minimum enclosing
ball is known as the Chebyshev center, and is defined as the solution to the following problem:
min max ky ? xk2 = min max kyk2 ? 2y > x + Tr(xx> ).
y
y
x?Q
x?Q
(4.2)
2
(4.2) is not easy to solve for a general set Q. However, when Q := ?m
i=1 B(yi , ri ), Beck [18] proved
that the relaxed Chebyshev center (RCC) [19], which is a convex quadratic program, is equivalent to
(4.2) if m < n. Therefore, we can solve (4.2) by solving a convex quadratic program (RCC):
min max kyk2 ?2y > x+Tr(4) = max min kyk2 ?2y > x+Tr(4) = max ?kxk2 +Tr(4),
y
(x,4)??
(x,4)??
y
(x,4)??
2
where ? = {(x, 4) : x ? Q, 4 xx> }. If Q = ?m
i=1 B(ci , ri ), then the dual of (4.3) is
min kC?k2 ?
m
X
i=1
?i kci k2 +
m
X
?i ri2 , s.t.
i=1
m
X
?i = 1,
?i ? 0, i = 1, . . . , m,
(4.3)
(4.4)
i=1
where C = [c1 , . . . , cm ] and ?i , i = 1, 2, . . . , m are the dual variables. Beck [18] proved that the
optimal solutions of (4.2) and (4.4) are linked by x? = C?? if m < n.
Now we can give our limited-memory GeoPG algorithm (L-GeoPG) as in Algorithm 5.
Algorithm 5 : L-GeoPG: Limited-memory GeoPG
Require: Parameters ?, ?, memory size m > 0 and initial point x0 .
2
2
2
2
1: Set c0 = x++
0 , r0 = R0 = kGt (x0 )k (1 ? 1/?)/? , and t = 1/?;
2: for k = 1, 2, . . . do
3:
Use Algorithm 1 or 2 to find xk ;
4:
Compute rk2 = kGt (xk )k2 (1 ? 1/?)/?2 ;
2
5:
Compute B(ck , Rk2 ): an enclosing ball of the intersection of B(ck?1 , Rk?1
) and Qk :=
++ 2
++ 2
k
k
?i=k?m+1 B(xi , ri ) (if k ? m, then set Qk := ?i=1 B(xi , ri )). This is done by setting
ck = C?? , where ?? is the optimal solution of (4.4);
6: end for
Remark 4.2. Backtracking can also be incorporated into L-GeoPG. We denote the resulting algorithm as L-GeoPG-B.
L-GeoPG has the same linear convergence rate as GeoPG, as we show in Theorem 4.3.
Theorem 4.3. Consider the L-GeoPG algorithm. For any k ? 0, one has x? ? B(ck , Rk2 ) and
?
?
2
Rk2 ? (1 ? 1/ ?)Rk?1
, and thus kx? ? ck k2 ? (1 ? 1/ ?)k R02 .
++ 2
2
Proof. Note that Qk := ?ki=k?m+1 B(x++
i , ri ) ? B(xk , rk ). Thus, the minimum enclosing ball
++ 2
2
2
of B(ck?1 , Rk?1 )? B(xk , rk ) is an enclosing ball of B(ck?1 , Rk?1
)?Qk . The proof then follows
from the proof of Theorem 3.5, and we omit it for brevity.
6
5
Numerical Experiments
In this section, we compare our GeoPG algorithm with Nesterov?s accelerated proximal gradient
(APG) method for solving two nonsmooth problems: linear regression and logistic regression, both
with elastic net regularization. Because of the elastic net term, the strong convexity parameter ? is
known. However, we assume that ? is unknown, and implement backtracking for both GeoPG and
APG, i.e., we test GeoPG-B and APG-B (APG with backtracking). We do not target at comparing
with other efficient algorithms for solving these two problems. Our main purpose here is to illustrate
the performance of this new first-order method GeoPG. Further improvement of this algorithm and
comparison with other state-of-the-art methods will be a future research topic.
The initial points were set to zero. To obtain the optimal objective function value F ? , we ran
APG-B and GeoPG-B for a sufficiently long time and the smaller function value returned by the two
algorithms is selected as F ? . APG-B was terminated if (F (xk ) ? F ? )/F ? ? tol, and GeoPG-B was
?
?
?8
terminated if (F (x+
is the accuracy tolerance. The parameters
k ) ? F )/F ? tol, where tol = 10
used in backtracking were set to ? = 0.5 and ? = 0.9. In GeoPG-B, we used Algorithm 2 to find xk ,
because we found that the performance of Algorithm 2 is slightly better than Algorithm 1 in practice.
In the experiments, we ran Algorithm 2 until the absolute value of ?? is smaller than 10?8 . The code
was written in Matlab and run on a standard PC with 3.20 GHz I5 Intel microprocessor and 16GB of
memory. In all figures we reported, the x-axis denotes the CPU time (in seconds) and y-axis denotes
?
?
(F (x+
k ) ? F )/F .
5.1
Linear regression with elastic net regularization
In this subsection, we compare GeoPG-B and APG-B in terms of solving linear regression with
elastic net regularization, a popular problem in machine learning and statistics [20]:
1
?
min
kAx ? bk2 + kxk2 + ?kxk1 ,
(5.1)
x?Rn 2p
2
where A ? Rp?n , b ? Rp , and ?, ? > 0 are the weighting parameters.
We conducted tests on two real datasets downloaded from the LIBSVM repository: a9a, RCV1. The
results are reported in Figure 1. In particular, we tested ? = 10?8 and ? = 10?3 , 10?4 , 10?5 . Note
that since ? is very small, the problems are very likely to be ill-conditioned. We see from Figure 1
that GeoPG-B is faster than APG-B on these real datasets, which indicates that GeoPG-B is preferable
than APG-B. In the supplementary material, we show more numerical results on varying ?, which
further confirm that GeoPG-B is faster than APG-B when the problems are more ill-conditioned.
2
2
10
10
?3
GeoPG?B: ? =10?3
GeoPG?B: ? =10
APG?B: ? =10?3
0
10
APG?B: ? =10?3
0
10
GeoPG?B: ? =10?4
GeoPG?B: ? =10?4
?4
?4
APG?B: ? =10
?2
10
APG?B: ? =10
GeoPG?B: ? =10?5
10
?5
APG?B: ? =10
?4
(F ?F*)/F*
10
APG?B: ? =10?5
?4
10
k
(Fk?F*)/F*
GeoPG?B: ? =10?5
?2
?6
10
?6
10
?8
10
?8
10
?10
10
?12
10
?10
0
10
20
30
40
10
50
0
CPU(s)
(a) Dataset a9a
5
10
15
CPU(s)
20
25
30
(b) Dataset RCV1
Figure 1: GeoPG-B and APG-B for solving (5.1) with ? = 10?8 .
5.2
Logistic regression with elastic net regularization
In this subsection, we compare the performance of GeoPG-B and APG-B in terms of solving the
following logistic regression problem with elastic net regularization:
p
?
1X
min
log 1 + exp(?bi ? a>
kxk2 + ?kxk1 ,
(5.2)
i x) +
x?Rn p
2
i=1
7
where ai ? Rn and bi ? {?1} are the feature vector and class label of the i-th sample, respectively,
and ?, ? > 0 are the weighting parameters.
We tested GeoPG-B and APG-B for solving (5.2) on the three real datasets a9a, RCV1 and Gisette
from LIBSVM, and the results are reported in Figure 2. In particular, we tested ? = 10?8 and
? = 10?3 , 10?4 , 10?5 . Figure 2 shows that with the same ?, GeoPG-B is much faster than APG-B.
More numerical results are provided in the supplementary material, which also indicate that GeoPG-B
is much faster than APG-B, especially when the problems are more ill-conditioned.
2
2
10
4
10
10
GeoPG?B: ? =10?3
GeoPG?B: ? =10?3
APG?B: ? =10?3
0
10
10
GeoPG?B: ? =10?4
10
?4
10
?6
APG?B: ? =10?4
GeoPG?B: ? =10?5
APG?B: ? =10?5
APG?B: ? =10
*
(Fk?F*)/F*
APG?B: ? =10
GeoPG?B: ? =10?4
?5
?4
?2
10
*
?5
0
10
GeoPG?B: ? =10?5
?2
(Fk?F )/F
10
(Fk?F*)/F*
APG?B: ? =10?4
GeoPG?B: ? =10?5
APG?B: ? =10?3
2
10
GeoPG?B: ? =10?4
APG?B: ? =10?4
?2
GeoPG?B: ? =10?3
APG?B: ? =10?3
0
10
?4
10
?6
10
10
?6
10
?8
?8
10
10
?10
10
?8
10
?10
0
10
20
30
40
50
CPU(s)
60
70
80
10
90
?10
0
1
2
3
4
CPU(s)
5
6
7
10
8
0
500
1000
1500
2000
CPU(s)
2500
3000
3500
Figure 2: GeoPG-B and APG-B for solving (5.2) with ? = 10?8 . Left: dataset a9a; Middle: dataset
RCV1; Right: dataset Gisette.
5.3
Numerical results of L-GeoPG-B
In this subsection, we test GeoPG with limited memory described in Algorithm 5 in solving (5.2)
on the Gisette dataset. Since we still need to use the backtracking technique, we actually tested
L-GeoPG-B. The results with different memory sizes m are reported in Figure 3. Note that m = 0
corresponds to the original GeoPG-B without memory. The subproblem (4.4) is solved using the
function ?quadprog? in Matlab. From Figure 3 we see that roughly speaking, L-GeoPG-B performs
better for larger memory sizes, and in most cases, the performance of L-GeoPG-B with m = 100
is the best among the reported results. This indicates that the limited-memory idea indeed helps
improve the performance of GeoPG.
2
2
10
4
10
memorysize=0
memorysize=5
memorysize=20
memorysize=100
0
10
10
memorysize=0
memorysize=5
memorysize=20
memorysize=100
0
10
memorysize=0
memorysize=5
memorysize=20
memorysize=100
2
10
0
?2
?6
?4
?2
10
*
10
(Fk?F )/F
?4
*
10
(Fk?F*)/F*
(Fk?F*)/F*
10
?2
10
10
?4
10
?6
10
10
?6
10
?8
?8
10
10
?10
10
?8
10
?10
0
50
100
150
CPU(s)
200
250
10
?10
0
50
100
150
200
CPU(s)
250
300
350
10
0
50
100
150
200
CPU(s)
250
300
350
400
Figure 3: L-GeoPG-B for solving (5.2) on the dataset Gisette with ? = 10?8 . Left: ? = 10?3 ;
Middle: ? = 10?4 ; Right: ? = 10?5 .
6
Conclusions
In this paper, we proposed a GeoPG algorithm for solving nonsmooth convex composite problems,
which is an extension of the recent method GeoD that can only handle smooth problems. We proved
that GeoPG enjoys the same optimal rate as Nesterov?s accelerated gradient method for solving
strongly convex problems. The backtracking technique was adopted to deal with the case when the
Lipschitz constant is unknown. Limited-memory GeoPG was also developed to improve the practical
performance of GeoPG. Numerical results on linear regression and logistic regression with elastic net
regularization demonstrated the efficiency of GeoPG. It would be interesting to see how to extend
GeoD and GeoPG to tackle non-strongly convex problems, and how to further accelerate the running
time of GeoPG. We leave these questions in future work.
Acknowledgements. Shiqian Ma is supported by a startup funding in UC Davis.
8
References
[1] S. Bubeck, Y.-T. Lee, and M. Singh. A geometric alternative to Nesterov?s accelerated gradient
descent. arXiv preprint arXiv:1506.08187, 2015.
[2] Y. E. Nesterov. A method for unconstrained convex minimization problem with the rate of
convergence O(1/k 2 ). Dokl. Akad. Nauk SSSR, 269:543?547, 1983.
[3] Y. E. Nesterov. Introductory lectures on convex optimization: A basic course. Applied
Optimization. Kluwer Academic Publishers, Boston, MA, 2004. ISBN 1-4020-7553-7.
[4] W. Su, S. Boyd, and E. J. Cand?s. A differential equation for modeling Nesterov?s accelerated
gradient method: Theory and insights. In NIPS, 2014.
[5] H. Attouch, Z. Chbani, J. Peypouquet, and P. Redont. Fast convergence of inertial dynamics
and algorithms with asymptotic vanishing viscosity. Mathematical Programming, 2016.
[6] L. Lessard, B. Recht, and A. Packard. Analysis and design of optimization algorithms via
integral quadratic constraints. SIAM Journal on Optimization, 26(1):57?95, 2016.
[7] A. Wibisono, A. Wilson, and M. I. Jordan. A variational perspective on accelerated methods in
optimization. Proceedings of the National Academy of Sciences, 133:E7351?E7358, 2016.
[8] R. G. Bland, D. Goldfarb, and M. J. Todd. The ellipsoid method: A survey. Operations
Research, 29:1039?1091, 1981.
[9] S. Bubeck and Y.-T. Lee. Black-box optimization with a politician. ICML, 2016.
[10] D. Drusvyatskiy, M. Fazel, and S. Roy. An optimal first order method based on optimal quadratic
averaging. SIAM Journal on Optimization, 2016.
[11] R. P. Brent. An algorithm with guaranteed convergence for finding a zero of a function. In
Algorithms for Minimization without Derivatives. Englewood Cliffs, NJ: Prentice-Hall, 1973.
[12] T. J. Dekker. Finding a zero by means of successive linear interpolation. In Constructive Aspects
of the Fundamental Theorem of Algebra. London: Wiley-Interscience, 1969.
[13] M. Gerdts, S. Horn, and S. Kimmerle. Line search globalization of a semismooth Newton
method for operator equations in Hilbert spaces with applications in optimal control. Journal of
Industrial And Management Optimization, 13(1):47?62, 2017.
[14] E. Hans and T. Raasch. Global convergence of damped semismooth Newton methods for L1
Tikhonov regularization. Inverse Problems, 31(2):025005, 2015.
[15] A. Beck and M. Teboulle. A fast iterative shrinkage-thresholding algorithm for linear inverse
problems. SIAM J. Imaging Sciences, 2(1):183?202, 2009.
[16] K. Scheinberg, D. Goldfarb, and X. Bai. Fast first-order methods for composite convex
optimization with backtracking. Foundations of Computational Mathematics, 14(3):389?417,
2014.
[17] Y. E. Nesterov. Gradient methods for minimizing composite functions. Mathematical Programming, 140(1):125?161, 2013.
[18] A. Beck. On the convexity of a class of quadratic mappings and its application to the problem of
finding the smallest ball enclosing a given intersection of balls. Journal of Global Optimization,
39(1):113?126, 2007.
[19] Y. C. Eldar, A. Beck, and M. Teboulle. A minimax Chebyshev estimator for bounded error
estimation. IEEE Transactions on Signal Processing, 56(4):1388?1397, 2008.
[20] H. Zou and T. Hastie. Regularization and variable selection via the elastic net. Journal of the
Royal Statistical Society, Series B, 67(2):301?320, 2005.
9
| 6666 |@word kong:2 repository:1 briefly:2 version:2 middle:2 norm:3 stronger:1 c0:3 dekker:4 open:1 tr:4 reduction:1 bai:1 initial:7 contains:2 series:1 past:3 comparing:2 tackling:1 written:1 numerical:9 implying:1 selected:1 xk:77 vanishing:1 ck2:1 successive:1 mathematical:2 differential:1 prove:3 introductory:1 interscience:1 x0:13 indeed:2 expected:1 ra:9 cand:1 roughly:1 resolve:1 encouraging:1 cpu:9 redont:1 becomes:1 provided:1 xx:2 moreover:2 notation:1 bounded:3 gisette:4 kg:1 cm:1 minimizes:1 developed:1 proposing:1 finding:5 dubbed:1 nj:1 guarantee:1 every:1 ti:2 tackle:2 preferable:1 k2:22 control:1 omit:3 rcc:2 todd:1 cliff:1 interpolation:1 black:1 china:1 limited:8 bi:2 fazel:3 practical:3 horn:1 practice:4 implement:1 procedure:3 ri2:1 composite:11 boyd:1 selection:1 operator:1 prentice:1 applying:1 equivalent:1 demonstrated:1 center:10 convex:19 survey:1 insight:1 estimator:1 nuclear:1 handle:1 resp:2 target:1 suppose:5 programming:2 quadprog:1 designing:2 roy:3 satisfying:2 kxk1:2 subproblem:1 preprint:1 solved:1 ran:2 intuition:2 vanishes:1 convexity:2 kgtk:3 nesterov:10 dynamic:1 singh:3 solving:17 algebra:1 f2:1 efficiency:1 accelerate:1 various:1 fast:3 describe:4 london:1 startup:1 whose:2 posed:1 solve:6 larger:2 supplementary:5 statistic:1 sequence:3 net:10 isbn:1 combining:2 nauk:1 academy:1 intuitive:1 ky:3 exploiting:1 convergence:13 r02:13 converges:3 leave:1 tk:19 help:1 illustrate:1 strong:1 solves:1 implemented:1 implies:4 indicate:1 radius:5 sssr:1 kgt:12 settle:1 material:5 require:3 hx:1 fix:3 extension:2 strictly:1 hold:22 sufficiently:1 hall:1 exp:1 mapping:3 achieves:2 smallest:1 xk2:4 purpose:1 estimation:1 label:1 minimization:4 always:1 ck:62 shrinkage:1 varying:1 wilson:1 improvement:1 indicates:4 a9a:4 industrial:1 initially:1 kc:1 issue:3 among:3 ill:4 dual:2 eldar:1 art:1 uc:2 beach:1 icml:1 future:2 nonsmooth:7 report:1 national:1 beck:5 connects:1 maintain:2 englewood:1 pc:1 damped:1 xb:9 integral:1 agm:4 politician:1 modeling:1 teboulle:2 hgt:1 conducted:1 reported:5 proximal:11 st:1 recht:1 fundamental:1 siam:3 lee:5 together:1 squared:4 ambiguity:1 satisfied:1 management:1 containing:1 possibly:1 shiqian:2 brent:4 derivative:1 summarized:1 performed:1 root:1 lab:1 closed:2 analyze:1 linked:1 maintains:1 accuracy:1 qk:5 yield:3 bisection:2 explain:1 whenever:1 proof:11 proved:3 dataset:7 popular:1 subsection:5 knowledge:1 organized:1 hilbert:1 inertial:1 actually:1 globalization:1 proxh:2 follow:1 wei:1 done:2 shrink:1 strongly:9 box:1 xa:8 until:3 hand:1 su:1 lack:1 continuity:1 logistic:6 usa:2 attouch:1 succinctness:1 rk2:14 contain:1 y2:2 regularization:10 hence:1 goldfarb:2 deal:1 kyk2:3 shixiang:1 davis:2 noted:1 hong:2 performs:2 l1:1 variational:1 recently:3 funding:1 common:1 extend:5 kluwer:1 interpret:1 refer:1 ai:2 smoothness:2 unconstrained:1 fk:7 mathematics:2 peypouquet:1 han:1 gt:4 showed:2 recent:1 perspective:1 hxk:1 tikhonov:1 kxa:2 inequality:11 yi:1 seen:1 minimum:8 relaxed:1 r0:5 monotonically:1 signal:1 semi:3 ii:4 reduces:1 smooth:10 faster:4 academic:1 calculation:1 long:3 bland:1 kax:1 regression:12 basic:3 arxiv:2 iteration:7 c1:1 interval:2 else:4 publisher:1 rest:1 hz:1 seem:1 jordan:1 easy:5 tgt:2 hastie:1 suboptimal:3 idea:6 chebyshev:3 t0:5 gb:1 algebraic:1 returned:1 speaking:1 remark:1 matlab:2 tol:3 useful:1 clear:4 repeating:1 viscosity:1 argminz:1 r12:2 r22:2 rb:11 key:2 kci:1 libsvm:2 imaging:1 ma2:1 run:1 inverse:2 i5:1 throughout:2 ki:1 apg:31 guaranteed:1 convergent:1 quadratic:5 constraint:1 ri:5 aspect:1 min:9 attempting:1 rcv1:4 department:2 according:1 ball:24 kgt0:1 smaller:4 slightly:1 drusvyatskiy:3 making:1 intuitively:1 equation:2 scheinberg:1 needed:1 know:2 letting:1 end:7 adopted:1 operation:1 apply:1 away:2 liu3:1 alternative:1 rp:2 original:1 denotes:2 running:1 include:1 newton:5 calculating:1 chinese:1 especially:2 society:1 objective:3 question:3 strategy:1 gradient:15 topic:1 induction:1 code:1 ellipsoid:2 kk:1 minimizing:2 mini:1 akad:1 difficult:1 semismooth:2 favorably:1 enclosing:11 design:1 unknown:4 perform:2 observation:1 datasets:3 descent:7 extended:1 tmin:2 incorporated:1 incorporate:2 y1:2 rn:12 chen1:1 nip:2 address:2 suggested:1 dokl:1 program:2 max:5 memory:13 packard:1 royal:1 minimax:1 improve:5 axis:2 ready:3 review:1 geometric:12 acknowledgement:1 asymptotic:1 lecture:1 interesting:1 foundation:1 downloaded:1 sufficient:1 lessard:1 bk2:1 thresholding:1 course:1 repeat:1 supported:1 enjoys:1 side:1 absolute:1 tolerance:1 ghz:1 kz:1 author:1 commonly:1 far:1 cope:1 transaction:1 keep:1 monotonicity:1 confirm:1 global:4 conclude:1 xi:4 continuous:3 search:1 iterative:1 ca:1 elastic:10 tencent:1 hc:1 zou:1 microprocessor:1 main:2 terminated:2 gtk:2 x1:2 intel:1 wiley:1 lie:1 kxk2:3 weighting:2 rk:23 theorem:7 r2:2 exists:5 ci:2 conditioned:4 kx:7 flavor:1 boston:1 intersection:7 backtracking:16 univariate:1 bubeck:5 likely:1 kxk:2 contained:1 scalar:1 corresponds:1 minimizer:1 satisfies:2 ma:2 lipschitz:6 specifically:4 uniformly:1 averaging:1 lemma:19 attempted:1 brevity:1 accelerated:8 wibisono:1 constructive:1 tested:4 |
6,264 | 6,667 | Diffusion Approximations for Online Principal
Component Estimation and Global Convergence
Chris Junchi Li
Mengdi Wang
Princeton University
Department of Operations Research and Financial Engineering, Princeton, NJ 08544
{junchil,mengdiw}@princeton.edu
Tong Zhang
Tencent AI Lab
Shennan Ave, Nanshan District, Shenzhen, Guangdong Province 518057, China
[email protected]
Abstract
In this paper, we propose to adopt the diffusion approximation tools to study the
dynamics of Oja?s iteration which is an online stochastic gradient descent method
for the principal component analysis. Oja?s iteration maintains a running estimate
of the true principal component from streaming data and enjoys less temporal
and spatial complexities. We show that the Oja?s iteration for the top eigenvector
generates a continuous-state discrete-time Markov chain over the unit sphere. We
characterize the Oja?s iteration in three phases using diffusion approximation and
weak convergence tools. Our three-phase analysis further provides a finite-sample
error bound for the running estimate, which matches the minimax information
lower bound for principal component analysis under the additional assumption of
bounded samples.
1
Introduction
In the procedure of Principal Component Analysis (PCA) we aim at learning the principal leading
eigenvector of the covariance matrix of a d-dimensional random vector Z from its independent
and identically distributed realizations Z1 , . . . , Zn . Let E[Z] = 0, and let the eigenvalues of ? be
?1 > ?2 ? ? ? ? ? ?d > 0, then the PCA problem can be formulated as minimizing the expectation of
a nonconvex function:
minimize ? w> E ZZ > w,
(1.1)
subject to kwk = 1, w ? Rd ,
where k ? k denotes the Euclidean norm. Since the eigengap ?1 ? ?2 is nonzero, the solution to
(1.1) is unique, denoted by w? . The classical method of finding the estimator of the first leading
eigenvector w? can be formulated as the solution to the empirical covariance problem as
n
>
X
b (n) w,
b (n) ? 1
b (n) = argmin ?w> ?
w
where ?
Z (i) Z (i) .
n i=1
kwk=1
b (n) denotes the empirical covariance matrix for the first n samples. The estimator w
b (n)
In words, ?
b (n) . Precisely, [43] shows that the
produced via this process provides a statistical optimal solution w
e (n) that is a function of the first n samples and w? has the following
angle between any estimator w
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
Figure 1: Left: an objective function for the top-1 PCA, where we use both the radius and heatmap to
represent the function value at each point of the unit sphere. Right: A quiver plot on the unit sphere
denoting the directions of negative gradient of PCA objective.
minimax lower bound
h
i
d?1
e (n) , w? ) ? c ? ??2 ?
E sin2 ?(w
,
n
e (n) Z?M(? 2 ,d)
w
?
inf
sup
(1.2)
e (n) is taken over all principal eigenvector
where c is some positive constant. Here the infimum of w
2
estimators, and M(?? , d) is the collection of all d-dimensional subgaussian distributions with mean
zero and eigengap ?1 ? ?2 > 0 satisfying ?1 ?2 /(?1 ? ?2 )2 ? ??2 . Classical PCA method has time
complexity O(nd2 ) and space complexity O(d2 ). The drawback of this method is that, when the
data samples are high-dimensional, computing and storage of a large empirical covariance matrix can
be costly.
In this paper we concentrate on the streaming or online method for PCA that processes online data
and estimates the principal component sequentially without explicitly computing and storing the
b Over thirty years ago, Oja [30] proposed an online PCA iteration that
empirical covariance matrix ?.
can be regarded as a projected stochastic gradient descent method as
h
i
w(n) = ? w(n?1) + ?Z (n) (Z (n) )> w(n?1) .
(1.3)
Here ? is some positive learning rule or stepsize, and ? is defined as ?w = kwk?1 w for each
nonzero vector w, namely, ? projects any vector onto the unit sphere S d?1 = {w ? Rd | kwk = 1}.
Oja?s iteration enjoys a less expensive time complexity O(nd) and space complexity O(d) and
thereby has been used as an alternative method for PCA when both the dimension d and number of
samples n are large.
In this paper, we adopt the diffusion approximation method to characterize the stochastic algorithm
using Markov processes and its differential equation approximations. The diffusion process approximation is a fundamental and powerful analytic tool for analyzing complicated stochastic process. By
leveraging the tool of weak convergence, we are able to conduct a heuristic finite-sample analysis of
the Oja?s iteration and obtain a convergence rate which, by carefully choosing the stepsize ?, matches
the PCA minimax information lower bound. Our analysis involves the weak convergence theory for
Markov processes [11], which is believed to have a potential for a broader class of stochastic algorithms for nonconvex optimization, such as tensor decomposition, phase retrieval, matrix completion,
neural network, etc.
Our Contributions We provide a Markov chain characterization of the stochastic process {w(n) }
generated by the Oja?s iteration with constant stepsize. We show that upon appropriate scalings, the
iterates as a Markov process weakly converges to the solution of an ordinary differential equation
system, which is a multi-dimensional analogue to the logistic equations. Also locally around the
neighborhood of a stationary point, upon a different scaling the process weakly converges to the
multidimensional Ornstein-Uhlenbeck processes. Moreover, we identify from differential equation
approximations that the global convergence dynamics of the Oja?s iteration has three distinct phases:
2
Figure 2: A simulation plot of Oja?s method, marked with the three phases.
(i) The initial phase corresponds to escaping from unstable stationary points;
(ii) The second phase corresponds to fast deterministic crossing period;
(iii) The third phase corresponds to stable oscillation around the true principal component.
Lastly, this is the first work that analyze the global rate of convergence analysis of Oja?s iteration,
i.e., the convergence rate does not have any initialization requirements.
Related Literatures This paper is a natural companion to paper by the authors? recent work [23]
that gives explicit rate analysis using a discrete-time martingale-based approach. In this paper, we
provide a much simpler and more insightful heuristic analysis based on diffusion approximation
method under the additional assumption of bounded samples.
The idea of stochastic approximation for PCA problem can be traced back to Krasulina [19] published
almost fifty years ago. His work proposed an algorithm that is regarded as the stochastic gradient
descent method for the Rayleigh quotient. In contrast, Oja?s iteration can be regarded as a projected
stochastic gradient descent method. The method of using differential equation tools for PCA appeared
in the first papers [19, 31] to prove convergence result to the principal component, among which, [31]
also analyze the subspace learning for PCA. See also [16, Chap. 1] for a gradient flow dynamical
system perspective of Oja?s iteration.
The convergence rate analysis of the online PCA iteration has been very few until the recent big data
tsunami, when the need to handle massive amounts of data emerges. Recent works by [6, 10, 17, 34]
study the convergence of online PCA from different perspectives, and obtain some useful rate results.
Our analysis using the tools of diffusion approximations suggests a rate that is sharper than all existing
results, and our global convergence rate result poses no requirement for initialization.
More Literatures Our work is related to a very recent line of work [3, 13, 21, 33, 38?41] on
the global dynamics of nonconvex optimization with statistical structures. These works carefully
characterize the global geometry of the objective functions, and in special, around the unstable
stationary points including saddle points and local maximizers. To solve the optimization problem
various algorithms were used, including (stochastic) gradient method with random initialization or
noise injection as well as variants of Newton?s method. The unstable stationary points can hence be
avoided, enabling the global convergence to desirable local minimizers.
Our diffusion process-based characterization of SGD is also related to another line of work [8, 10, 24,
26, 37]. Among them, [10] uses techniques based on martingales in discrete time to quantify the global
3
convergence of SGD on matrix decomposition problems. In comparison, our techniques are based on
Stroock and Varadhan?s weak convergence of Markov chains to diffusion processes, which yield the
continuous-time dynamics of SGD. The rest of these results mostly focus on analyzing continuoustime dynamics of gradient descent or SGD on convex optimization problems. In comparison, we are
the first to characterize the global dynamics for nonconvex statistical optimization. In particular, the
first and second phases of our characterization, especially the unstable Ornstein-Uhlenbeck process,
are unique to nonconvex problems. Also, it is worth noting that, using the arguments of [26], we can
show that the diffusion process-based characterization admits a variational Bayesian interpretation of
nonconvex statistical optimization. However, we do not pursue this direction in this paper.
In the mathematical programming and statistics communities, the computational and statistical
aspects of PCA are often studied separately. From the statistical perspective, recent developments
have focused on estimating principal components for very high-dimensional data. When the data
dimension is much larger than the sample size, i.e., d n, classical method using decomposition of
the empirical convariance matrix produces inconsistent estimates [18, 29]. Sparsity-based methods
have been studied, such as the truncated power method studied by [45] and [44]. Other sparsity
regularization methods for high dimensional PCA has been studied in [2, 7, 9, 18, 25, 42, 43, 46], etc.
Note that in this paper we do not consider the high-dimensional regime and sparsity regularization.
From the computational perspective, power iterations or the Lanczos method are well studied. These
iterative methods require performing multiple products between vectors and empirical covariance
matrices. Such operation usually involves multiple passes over the data, whose complexity may scale
with the eigengap and dimensions [20, 28]. Recently, randomized algorithms have been developed to
reduce the computation complexity [12, 35, 36]. A critical trend today is to combine the computational
and statistical aspects and to develop algorithmic estimator that admits fast computation as well as
good estimation properties. Related literatures include [4, 5, 10, 14, 27].
Organization ?2 introduces the settings and distributional assumptions. ?3 briefly discusses the
Oja?s iteration from the Markov processes perspective and characterizes that it globally admits
ordinary differential equation approximation upon appropriate scaling, and also stochastic differential
equation approximation locally in the neighborhood of each stationary point. ?4 utilizes the weak
convergence results and provides a three-phase argument for the global convergence rate analysis,
which is near-optimal for the Oja?s iteration. Concluding remarks are provided in ?5.
2
Settings
In this section, we present the basic settings for the Oja?s iteration. The algorithm maintains a running
estimate w(n) of the true principal component w? , and updates it while receiving streaming samples
from exterior data source. We summarize our distributional assumptions.
Assumption 2.1. The random vectors Z ? Z (1) , . . . , Z (n) ? Rd are independent and identically
distributed and have the following properties:
(i) E[Z] = 0 and E ZZ > = ?;
(ii) ?1 > ?2 ? ? ? ? ? ?d > 0;
(iii) There is a constant B such that kZk2 ? B.
For the easiness of presentation, we transform the iterates w(n) and define the rescaled samples, as
follows. First we let the eigendecomposition of the covariance matrix be
? = E ZZ > = U?U> ,
where ? = diag(?1 , ?2 , . . . , ?d ) is a diagonal matrix with diagonal entries ?1 , ?2 , . . . , ?d , and U is
an orthogonal matrix consisting of column eigenvectors of ?. Clearly the first column of U is equal
to the principal component w? . Note that the diagonal decomposition might not be unique, in which
case we work with an arbitrary one. Second, let
Y (n) = U> Z (n) , v(n) = U> w(n) , v? = U> w? .
One can easily verify that
E[Y ] = 0,
E Y Y > = ?;
4
(2.1)
The principal component of the rescaled random variable Y , which we denote by v? , is equal to e1 ,
where {e1 , . . . , ed } is the canonical basis of Rd . By applying the orthonormal transformation U>
to the stochastic process {w(n) }, we obtain an iterative process {v(n) = U> w(n) } in the rescaled
space:
>
(n)
> (n)
> (n?1)
> (n)
(n)
> (n?1)
v =U w =? U w
+ ?U Z
Z
UU w
(2.2)
>
(n?1)
(n)
(n)
(n?1)
=? v
+ ?Y
Y
v
.
Moreover, the angle processes associated with {w(n) } and {v(n) } are equivalent, i.e.,
?(w(n) , w? ) = ?(v(n) , v? ).
Therefore it would be sufficient to study the rescaled iteration v
iteration Y (n) throughout the rest of this paper.
3
(2.3)
(n)
in (2.2) and the transformed
A Theory of Diffusion Approximation for PCA
In this section we show that the stochastic iterates generated by the Oja?s iteration can be approximated
by the solution of an ODE system upon appropriate scaling, as long as ? is small. To work on the
approximation we first observe that the iteration v(n) , n = 0, 1, . . . generated by (2.2) forms a
discrete-time, time-homogeneous Markov process that takes values on S d?1 . Furthermore, v(n)
holds strong Markov property.
3.1
Global ODE Approximation
To state our results on differential equation approximations, let us define a new process, which is
obtained by rescaling the time index n according to the stepsize ?
?1
Ve ? (t) ? v?,(bt? c) .
(3.1)
We add the superscript ? in the notation to emphasize the dependence of the process on ?. We will
show that Ve ? (t) converges weakly to a deterministic function V (t), as ? ? 0+ .
Furthermore, we can identify the limit V (t) as the closed-form solution to an ODE system. Under
Assumption 2.1 and using an infinitesimal generator analysis we have
?
Ve (t + ?) ? Ve ? (t) = O(B?).
It follows that, as ? ? 0+ , the infinitesimal conditional variance tends to 0:
h
i
? ?1 var Ve ? (t + ?) ? Ve ? (t) Ve ? (t) = v = O(B?),
and the infinitesimal
h mean is
i
?1
? E Ve ? (t + ?) ? Ve ? (t) Ve ? (t) = v = ? ? V > ?V V + O(B 2 ? 2 ).
Using the classical weak convergence to diffusion argument [11, Corollary 4.2 in ?7.4], we obtain
the following result.
Theorem 3.1. If v?,(0) converges weakly to some constant vector V o ? S d?1 as ? ? 0+ then the
?1
Markov process v?,(bt? c) converges weakly to the solution V = V (t) to the following ordinary
differential equation system
dV
= ? ? V > ?V V ,
(3.2)
dt
with initial values V (0) = V o .
We can straightforwardly check for sanity that the solution vector V (t) lies on the unit sphere
S d?1 , i.e., kV (t)k = 1 for all t ? 0. Written in coordinates V (t) = (V1 (t), . . . , Vd (t))> , the ODE
is expressed for k = 1, . . . , d
d
X
dVk
= Vk
(?k ? ?i )Vi2 .
dt
i=1
5
One can straightforwardly verify that the solution to (3.2) has
Vk (t) = (Z(t))
where Z(t) is the normalization function
Z(t) =
d
X
?1/2
Vk (0) exp(?k t),
(3.3)
2
(Vio ) exp(2?i t).
i=1
To understand the limit function given by (3.3), we note that in the special case where ?2 = ? ? ? = ?d
2
2
Z(t) = (V1o ) exp(2?1 t) + 1 ? (V1o ) exp(2?2 t),
and
2
2
(V1 (t)) =
(V1o ) exp(2?1 t)
.
2
2
(V1o ) exp(2?1 t) + 1 ? (V1o ) exp(2?2 t)
(3.4)
This is the formula of the logistic curve. Hence analogously, V (t) in (3.3) is namely the generalized
logistic curves.
3.2
Local Approximation by Diffusion Processes
The weak convergence to ODE theorem introduced in ?3.1 characterizes the global dynamics of the
Oja?s iteration. Such approximation explains many behaviors, but neglected the presence of noise that
plays a role in the algorithm. In this section we aim at understanding the Oja?s iteration via stochastic
differential equations (SDE). We refer the readers to [32] for more on basic concepts of SDE.
In this section, we instead show that under some scaling, the process admits an approximation
of multidimensional Ornstein-Uhlenbeck process within a neighborhood of each of the unstable
stationary points, both stable and unstable. Afterwards, we develop some weak convergence results
to give a rough estimate on the rate of convergence of the Oja?s iteration. For purposes of illustration
and brevity, we restrict ourselves to the case of starting point v(0) being the stationary point ek for
some k = 1, . . . , d, and denote an arbitrary vector xk to be a (d ? 1)-dimensional vector that keeps
all but the kth coordinate of x. Using theory from [11] we conclude the following theorem.
?,(0)
Theorem 3.2. Let k = 1, . . . , d be arbitrary. If ? ?1/2 vk
as ? ? 0+ , then the Markov process
converges weakly to some Uok ? Rd?1
?,(bt? ?1 c)
? ?1/2 vk
converges weakly to the solution of the multidimensional stochastic differential equation
1/2
dUk (t) = ?(?k Id?1 ? ?k )Uk dt + ?k ?k
dBk (t),
with initial values Uk (0) =
Uok .
(3.5)
Here Bk (t) is a standard (d ? 1)-dimensional Brownian motion.
1
The solution to (3.5) can be solved explicitly. We let for a matrix A ? Rn?n the matrix
expoP?
1/2
1/2
n
1/2
nentiation exp(A) as exp(A) = n=0 (1/n!)A . Also, let ?
= diag ?1 , . . . , ?d
for the
positive semidefinite diagonal matrix ? = diag(?1 , . . . , ?d ). The solution to (3.5) is hence
Z
1/2 t
exp (s ? t)(?k Id?1 ? ?k ) dBk (s),
Uk (t) = exp ?t(?k Id?1 ? ?k ) Uko + ?k ?k
0
which is known as the multidimensional Ornstein-Uhlenbeck process, whose behavior depends on
the matrix ?(?k Id?1 ? ?k ) and is discussed in details in ?4.
Before concluding this section, we emphasize that the weak convergence to diffusions results in
?3.1 and ?3.2 should be distinguished from the convergence of the Oja?s iteration. From a random
process theoretical perspective, the former one treats the weak convergence of finite dimensional
distributions of a sequence of rescaled processes as ? tends to 0, while the latter one charaterizes the
long-time behavior of a single realization of iterates generated by algorithm for a fixed ? > 0.
1
The reason we have a (d ? 1)-dimensional Ornstein-Uhlenbeck process is because the objective function
of PCA is defined on a (d ? 1)-dimensional manifold S d?1 and has d ? 1 independent variables.
6
4
Global Three-Phase Analysis of Oja?s Dynamics
Previously ?3.1 and ?3.2 develop the tools of weak convergence to diffusion under global and local
scalings. In this section, we apply these tools to analyze the dynamics of online PCA in three phases
in sequel. For purposes of illustration and brevity, we restrict ourselves to the case of starting point
v(0) that is near a saddle point ek . Let A? . B ? denotes lim sup??0+ A? /B ? ? 1, a.s., and
A? B ? when both A? . B ? and B ? . A? hold.
4.1
Phase I: Noise Initialization
In consideration of global convergence, we need to analyze the initial phase approximation by
considering restrictions on initialization: what happens if the starting point is in a small neighborhood
of the following set
Se = v ? S d?1 : v1 = 0 .
When thinking the sphere S d?1 as the globe with ?e1 as north and south poles, Se corresponds to
the equator of the globe. Therefore, all unstable stationary points (including saddle points and local
maximizers) e2 , . . . , ed lies on the equator Se .
4.2
Phase II: Deterministic Crossing
In Phase II, the iteration escapes from the neighborhood of equator Se and converges to a basin of
attraction of the local minimizer v? . From strong Markov property of the Oja?s iteration introduced
in the beginning of ?2, one can forget the iteration steps in Phase I and analyze the algorithm from its
(0)
output. Suppose we have an initial point v(0) that satisfies (v1 )2 = ?. If we fix ? ? (0, 1/2) and the
initial value V (0) = V o that satisfies V12 = ?. Theorem 3.1 concludes that the iteration moves in a
deterministic pattern.
4.3
Phase III: Convergence to Principal Component
In Phase III, the iteration fluctuates around the true principal component v? = e1 . We start our it(0)
eration from a neighborhood around the principal component, where v(0) has (v1 )2 = 1 ? ?.
Letting
(3.5) and taking the limit t ? ?,
we have the limit EkU1 (?)k2 =
k = 1 in
tr E U1 (t)U1 (t)> = (?1 /2) tr ?1 (?1 Id?1 ? ?1 )?1 . In terms of the iterations v(n) , rescaling
the Markov process along with some calculations gives as n ? ?, in very rough sense,
d
X
?1 ?k
?1
tr ?1 (?1 Id?1 ? ?1 )?1 = ? ?
.
2
2(?1 ? ?k )
k=2
(4.1)
The above display implies that there will be some nondiminishing fluctuations, variance being
proportional to the constant stepsize ?, as time goes to infinity or at stationarity. Therefore in terms
of angle, at stationarity the Markov process concentrates on O(? 1/2 ) scale around the origin.
E sin2 ?(v(n) , v? ) ? ? EkU1 (?)k2 = ? ?
4.4
Crossing Time Estimate
We turn to estimate the running time, namely the crossing time, that is required for the algorithm to
cross the corresponding region in different phases.
Phase I. In the first phase, the crossing time is the number of iterates required for the algorithm
to escape an O(1)-neighborhood of the equator Se . For illustrative purposes we only consider the
special case where v is ek the kth coordinate vector, which is a saddle point that has a negative
Hessian eigenvalue. In this situation, the SDE (3.5) in terms of the first coordinate U (t) of Uk
reduces to
1/2
dU (t) = (?1 ? ?k )U (t) dt + (?1 ?k )
dB(t),
(4.2)
with initial value U (0) = 0. Solution to (4.2) is known as unstable Ornstein-Uhlenbeck process [1]
and can be expressed explicitly in closed-form, as
Z t
1/2
U (t) = W ? exp ((?1 ? ?k )t) , where W ? ? (?1 ?k )
exp (?(?1 ? ?k )s) dB(s).
0
7
Rescaling the time back to the discrete algorithm, we let n = t? ?1 and obtain
(n)
v1
? 1/2 W ? exp (?(?1 ? ?k )n) .
(4.3)
?
In (4.3), the term W is approximately distributed as t = n? ? ?
1/2
?1 ?k
?
?,
W
2(?1 ? ?k )
where ? stands for a standard normal variable. We have
1/2
?1 ?k
(n)
1/2
?
1/2
v1 = ? W ?
? exp (?(?1 ? ?k )n) .
2(?1 ? ?k )
(4.4)
(n)
In order to have (v1 )2 = ? in (4.4), we have as ? ? 0+ the crossing time is approximately
?1/2
?1 ?d
?1 ?1
?1 ?1
?
?1
?1/2
?
N1 (?1 ? ?k ) ? log ?|?|
+ (?1 ? ?k ) ? log
.
2(?1 ? ?d )
(4.5)
Therefore we have whenever the smallest
eigenvalue
?
is
bounded
away
from
0,
then
asymptotically
d
?1
N1? 0.5 (?1 ? ?k ) ? ?1 log ? ?1 . This suggests that the noise helps initial location to diffuse
away from ek rapidly.
Phase II. We turn to estimate the crossing time N2? in Phase II. Simple calculation ensures the
existence of a constant T , that depends only on ? such that V12 (T ) ? 1 ? ?. Furthermore T has the
following bounds:
1??
1??
?
?1
?1
. N2 . (?1 ? ?2 ) log
.
(4.6)
(?1 ? ?d ) log
?
?
In special, the time it takes is asymptotically N2? . (?1 ? ?2 )?1 ? ?1 log ((1 ? ?)/?) iterates to have
(n)
v1 ? 1 ? ?. Theorem 3.1 indicates that when ? is positively small, the iterates needed for the first
coordinate squared to traverse from ? to 1 ? ? is of O(? ?1 ). This is substantiated by simulation
results [4] suggesting that the Oja?s iteration is fast from the warm initialization.
Phase III. To estimate the crossing time N3? or the number of iterates needed in Phase III, we restart
our counter and have from the approximation in Theorem 3.2 and (3.5) that
Z ?n
(n)
(0)
E(vk )2 = (vk )2 exp (?2(?1 ? ?k )?n) + ??1 ?k
exp (?2(?1 ? ?k )(t ? s)) ds
0
=??
d
X
k=2
??
d
X
k=2
?1 ?k
+
2(?1 ? ?k )
d
X
(0)
(vk )2 ? ? ?
k=2
?1 ?k
2(?1 ? ?k )
exp (?2?(?1 ? ?k )n)
?1 ?k
+ ? exp (?2?(?1 ? ?2 )n) .
2(?1 ? ?k )
Pd
In terms of the iterations v(n) , note the relationship E sin2 ?(v, e1 ) = k=2 vk2 = 1 ? v12 . The end
(0)
of Phase II implies that E sin2 ?(v(0) , e1 ) = 1 ? (v1 )2 = ?, and hence by setting
?
E sin2 ?(v(N3 ) , e1 ) = ? ?
d
X
k=2
?1 ?k
+ o(?),
2(?1 ? ?k )
+
we conclude that as ? ? 0
N3? (0.5 + o(1)) ? (?1 ? ?2 )?1 ? ?1 log ?? ?1 .
4.5
(4.7)
Finite-Sample Rate Bound
In this subsection we attempt to establish the global rate of convergence analysis using the crossing
time estimates in the previous subsection. Starting from v(0) = ek where k = 2, . . . , d is arbitrary,
the global convergence time N ? = N1? + N2? + N3? as ? ? 0+ such that, by choosing ? ? (0, 1/2)
as a small fixed constant,
?1
N ? (?1 ? ?2 ) ? ?1 log ? ?1 ,
8
with the following estimation on global convergence rate as in (4.1)
?
sin2 ?(v(N ) , v? ) = ? ?
d
X
k=2
?1 ?k
.
2(?1 ? ?k )
Given a fixed number of samples T , by choosing ? as
log T
(?1 ? ?2 )T
? )?
? = ?(T
(4.8)
? )
? )?1 log ?(T
? ) ?1 = N ?(T
we have T (?1 ? ?2 )?1 ?(T
. Plugging in ? in (4.8) we have, by the
angle-preserving property of coordinate transformation (2.3), that
E sin2 ?(w(N
?
?(T
)
)
, w? ) = E sin2 ?(v(N
?
?(T
)
)
, v? ) ?
d
X
k=2
?1 ?k
log T
?
.
2(?1 ? ?k ) (?1 ? ?2 )T
(4.9)
The finite sample bound in (4.9) is better than any existing results and matches the information lower
bound. Moreover, (4.9) implies that the rate in terms of sine-squared angle is sin2 ?(w(T ) , w? ) ?
C ? ?1 ?2 /(?1 ? ?2 )2 ? d log T /T, which matches the minimax information lower bound (up to a log T
factor), see for example, Theorem 3.1 of [43]. Limited by space, details about the rate comparison is
provided in the supplementary material.
5
Concluding Remarks
We make several concluding remarks on the global convergence rate estimations, as follows.
Crossing Time Comparison. From the crossing time estimates in (4.5), (4.6), (4.7) we conclude
(i) As ? ? 0+ we have N2? /N1? ? 0. This implies that the algorithm demonstrates the cutoff
phenomenon which frequently occur in discrete-time Markov processes [22]. In words,
the Phase II where the objective value in Rayleigh quotient drops from 1 ? ? to ? is an
asymptotically a phase of short time, compared to Phases I and III, so the convergence curve
occurs instead of an exponentially decaying curve.
(ii) As ? ? 0+ we have N3? /N1? 1. This suggests that for the high-d case that Phase I of
escaping from the equator consumes roughly the same iterations as in Phase III.
To summarize from above, the cold initialization iteration roughly takes twice the number of steps
than the warm initialization version which is consistent with the simulation discussions in [31].
Subspace Learning. In this work we primarily concentrates on the problem of finding the top1 eigenvector. It is believed that the problem of finding top-k eigenvectors, a.k.a. the subspace
PCA problem, can be analyzed using our approximation methods. This will involve a careful
characterization of subspace angles and is hence more complex and is left for further investigation.
References
[1] Aldous, D. (1989). Probability Approximations via the Poisson Clumping Heuristic, volume 77. Springer.
[2] Amini, A. & Wainwright, M. (2009). High-dimensional analysis of semidefinite relaxations for sparse
principal components. The Annals of Statistics, 37(5B), 2877?2921.
[3] Anandkumar, A. & Ge, R. (2016). Efficient approaches for escaping higher order saddle points in non-convex
optimization. arXiv preprint arXiv:1602.05908.
[4] Arora, R., Cotter, A., Livescu, K., & Srebro, N. (2012). Stochastic optimization for PCA and PLS. In 50th
Annual Allerton Conference on Communication, Control, and Computing (pp. 861?868).
[5] Arora, R., Cotter, A., & Srebro, N. (2013). Stochastic optimization of PCA with capped msg. In Advances
in Neural Information Processing Systems (pp. 1815?1823).
[6] Balsubramani, A., Dasgupta, S., & Freund, Y. (2013). The fast convergence of incremental PCA. In
Advances in Neural Information Processing Systems (pp. 3174?3182).
[7] Cai, T. T., Ma, Z., & Wu, Y. (2013). Sparse PCA: Optimal rates and adaptive estimation. The Annals of
Statistics, 41(6), 3074?3110.
9
[8] Darken, C. & Moody, J. (1991). Towards faster stochastic gradient search. In Advances in Neural Information
Processing Systems (pp. 1009?1016).
[9] d?Aspremont, A., Bach, F., & El Ghaoui, L. (2008). Optimal solutions for sparse principal component
analysis. Journal of Machine Learning Research, 9, 1269?1294.
[10] De Sa, C., Olukotun, K., & R?, C. (2015). Global convergence of stochastic gradient descent for some
non-convex matrix problems. In Proceedings of the 32nd International Conference on Machine Learning
(ICML-15) (pp. 2332?2341).
[11] Ethier, S. N. & Kurtz, T. G. (2005). Markov Processes: Characterization and Convergence, volume 282.
John Wiley & Sons.
[12] Garber, D. & Hazan, E. (2015). Fast and simple PCA via convex optimization. arXiv preprint
arXiv:1509.05647.
[13] Ge, R., Huang, F., Jin, C., & Yuan, Y. (2015). Escaping from saddle points ? online stochastic gradient for
tensor decomposition. In Proceedings of The 28th Conference on Learning Theory (pp. 797?842).
[14] Hardt, M. & Price, E. (2014). The noisy power method: A meta algorithm with applications. In Advances
in Neural Information Processing Systems (pp. 2861?2869).
[15] Hardt, Moritz & Price, Eric (2014). The Noisy Power Method: A Meta Algorithm with Applications.
NIPS, (pp. 2861?2869).
[16] Helmke, U. & Moore, J. B. (1994). Optimization and Dynamical Systems. Springer.
[17] Jain, P., Jin, C., Kakade, S. M., Netrapalli, P., & Sidford, A. (2016). Matching matrix bernstein with little
memory: Near-optimal finite sample guarantees for oja?s algorithm. arXiv preprint arXiv:1602.06929.
[18] Johnstone, I. M. & Lu, A. Y. (2009). On Consistency and Sparsity for Principal Components Analysis in
High Dimensions. Journal of the American Statistical Association, 104(486), 682?693.
[19] Krasulina, T. (1969). The method of stochastic approximation for the determination of the least eigenvalue
of a symmetrical matrix. USSR Computational Mathematics and Mathematical Physics, 9(6), 189?195.
[20] Kuczynski, J. & Wozniakowski, H. (1992). Estimating the largest eigenvalue by the power and lanczos
algorithms with a random start. SIAM journal on matrix analysis and applications, 13(4), 1094?1122.
[21] Lee, J. D., Simchowitz, M., Jordan, M. I., & Recht, B. (2016). Gradient descent only converges to
minimizers. In Conference on Learning Theory (pp. 1246?1257).
[22] Levin, D. A., Peres, Y., & Wilmer, E. L. (2009). Markov chains and mixing times. American Mathematical
Society.
[23] Li, C. J., Wang, M., Liu, H., & Zhang, T. (2016). Near-optimal stochastic approximation for online
principal component estimation. arXiv preprint arXiv:1603.05305.
[24] Li, Q., Tai, C., & E, W. (2015).
arXiv:1511.06251.
Dynamics of stochastic gradient algorithms.
arXiv preprint
[25] Ma, Z. (2013). Sparse principal component analysis and iterative thresholding. The Annals of Statistics,
41(2), 772?801.
[26] Mandt, S., Hoffman, M. D., & Blei, D. M. (2016). A variational analysis of stochastic gradient algorithms.
arXiv preprint arXiv:1602.02666.
[27] Mitliagkas, I., Caramanis, C., & Jain, P. (2013). Memory limited, streaming PCA. In Advances in Neural
Information Processing Systems (pp. 2886?2894).
[28] Musco, C. & Musco, C. (2015). Stronger approximate singular value decomposition via the block lanczos
and power methods. arXiv preprint arXiv:1504.05477.
[29] Nadler, B. (2008). Finite sample approximation results for principal component analysis: A matrix
perturbation approach. The Annals of Statistics, 41(2), 2791?2817.
[30] Oja, E. (1982). Simplified neuron model as a principal component analyzer. Journal of mathematical
biology, 15(3), 267?273.
10
[31] Oja, E. & Karhunen, J. (1985). On stochastic approximation of the eigenvectors and eigenvalues of the
expectation of a random matrix. Journal of mathematical analysis and applications, 106(1), 69?84.
[32] Oksendal, B. (2003). Stochastic Differential Equations. Springer.
[33] Panageas, I. & Piliouras, G. (2016). Gradient descent converges to minimizers: The case of non-isolated
critical points. arXiv preprint arXiv:1605.00405.
[34] Shamir, O. (2015a). Convergence of stochastic gradient descent for PCA. arXiv preprint arXiv:1509.09002.
[35] Shamir, O. (2015b). Fast stochastic algorithms for svd and PCA: Convergence properties and convexity.
arXiv preprint arXiv:1507.08788.
[36] Shamir, O. (2015c). A stochastic PCA and svd algorithm with an exponential convergence rate. In
Proceedings of the 32nd International Conference on Machine Learning (ICML-15) (pp. 144?152).
[37] Su, W., Boyd, S., & Candes, E. J. (2016). A differential equation for modeling Nesterov?s accelerated
gradient method: theory and insights. Journal of Machine Learning Research, 17(153), 1?43.
[38] Sun, J., Qu, Q., & Wright, J. (2015a). Complete dictionary recovery over the sphere i: Overview and the
geometric picture. arXiv preprint arXiv:1511.03607.
[39] Sun, J., Qu, Q., & Wright, J. (2015b). Complete dictionary recovery over the sphere ii: Recovery by
Riemannian trust-region method. arXiv preprint arXiv:1511.04777.
[40] Sun, J., Qu, Q., & Wright, J. (2015c). When are nonconvex problems not scary?
arXiv:1510.06096.
arXiv preprint
[41] Sun, J., Qu, Q., & Wright, J. (2016).
arXiv:1602.06664.
arXiv preprint
A geometric analysis of phase retrieval.
[42] Vu, V. Q. & Lei, J. (2012). Minimax Rates of Estimation for Sparse PCA in High Dimensions. AISTATS,
(pp. 1278?1286).
[43] Vu, V. Q. & Lei, J. (2013). Minimax sparse principal subspace estimation in high dimensions. The Annals
of Statistics, 41(6), 2905?2947.
[44] Wang, Z., Lu, H., & Liu, H. (2014). Nonconvex statistical optimization: Minimax-optimal sparse pca in
polynomial time. arXiv preprint arXiv:1408.5352.
[45] Yuan, X.-T. & Zhang, T. (2013). Truncated power method for sparse eigenvalue problems. Journal of
Machine Learning Research, 14(Apr), 899?925.
[46] Zou, H. (2006). The adaptive lasso and its oracle properties. Journal of the American statistical association,
101(476), 1418?1429.
11
| 6667 |@word version:1 briefly:1 polynomial:1 norm:1 stronger:1 nd:3 d2:1 simulation:3 covariance:7 decomposition:6 sgd:4 thereby:1 tr:3 initial:8 liu:2 denoting:1 existing:2 written:1 john:1 analytic:1 plot:2 drop:1 update:1 clumping:1 stationary:8 xk:1 beginning:1 short:1 blei:1 provides:3 characterization:6 iterates:8 location:1 traverse:1 allerton:1 district:1 org:1 zhang:3 simpler:1 mathematical:5 along:1 differential:12 yuan:2 prove:1 combine:1 roughly:2 behavior:3 frequently:1 multi:1 globally:1 chap:1 junchi:1 little:1 considering:1 project:1 estimating:2 bounded:3 moreover:3 provided:2 notation:1 what:1 sde:3 argmin:1 pursue:1 eigenvector:5 developed:1 finding:3 transformation:2 nj:1 guarantee:1 temporal:1 multidimensional:4 k2:2 demonstrates:1 uk:4 control:1 unit:5 kurtz:1 positive:3 before:1 engineering:1 local:6 treat:1 tends:2 limit:4 krasulina:2 analyzing:2 id:6 fluctuation:1 mandt:1 approximately:2 might:1 twice:1 initialization:8 china:1 studied:5 suggests:3 wozniakowski:1 limited:2 unique:3 thirty:1 vu:2 block:1 procedure:1 cold:1 empirical:6 matching:1 boyd:1 word:2 onto:1 storage:1 applying:1 restriction:1 equivalent:1 deterministic:4 go:1 starting:4 convex:4 focused:1 musco:2 recovery:3 estimator:5 rule:1 attraction:1 regarded:3 orthonormal:1 insight:1 financial:1 his:1 handle:1 coordinate:6 annals:5 shamir:3 today:1 play:1 massive:1 suppose:1 programming:1 homogeneous:1 us:1 livescu:1 origin:1 crossing:11 trend:1 satisfying:1 expensive:1 approximated:1 distributional:2 role:1 preprint:15 wang:3 solved:1 region:2 ensures:1 sun:4 counter:1 rescaled:5 consumes:1 pd:1 convexity:1 complexity:7 nesterov:1 dynamic:10 neglected:1 weakly:7 mengdi:1 upon:4 eric:1 basis:1 easily:1 various:1 caramanis:1 substantiated:1 distinct:1 fast:6 jain:2 choosing:3 neighborhood:7 sanity:1 whose:2 heuristic:3 larger:1 solve:1 fluctuates:1 supplementary:1 garber:1 statistic:6 transform:1 noisy:2 superscript:1 online:10 sequence:1 eigenvalue:7 cai:1 simchowitz:1 propose:1 product:1 realization:2 rapidly:1 mixing:1 kv:1 convergence:38 requirement:2 produce:1 incremental:1 converges:10 help:1 develop:3 completion:1 pose:1 sa:1 strong:2 netrapalli:1 scary:1 quotient:2 involves:2 implies:4 uu:1 quantify:1 direction:2 concentrate:3 radius:1 drawback:1 stochastic:29 material:1 explains:1 require:1 fix:1 convariance:1 investigation:1 hold:2 around:6 wright:4 normal:1 exp:19 nadler:1 algorithmic:1 dictionary:2 adopt:2 smallest:1 purpose:3 estimation:8 largest:1 tool:8 cotter:2 hoffman:1 rough:2 clearly:1 aim:2 broader:1 corollary:1 focus:1 nd2:1 vk:8 check:1 indicates:1 contrast:1 ave:1 sense:1 vk2:1 sin2:9 minimizers:3 el:1 streaming:4 bt:3 transformed:1 among:2 denoted:1 ussr:1 development:1 heatmap:1 spatial:1 special:4 equal:2 beach:1 zz:3 biology:1 icml:2 thinking:1 escape:2 few:1 primarily:1 oja:27 ve:10 phase:33 geometry:1 consisting:1 ourselves:2 n1:5 continuoustime:1 attempt:1 organization:1 stationarity:2 introduces:1 analyzed:1 semidefinite:2 chain:4 orthogonal:1 conduct:1 euclidean:1 v1o:5 isolated:1 theoretical:1 column:2 modeling:1 sidford:1 lanczos:3 zn:1 stroock:1 ordinary:3 pole:1 entry:1 levin:1 characterize:4 kuczynski:1 straightforwardly:2 st:1 recht:1 international:2 fundamental:1 randomized:1 siam:1 sequel:1 lee:1 physic:1 receiving:1 analogously:1 moody:1 squared:2 huang:1 ek:5 american:3 leading:2 rescaling:3 li:3 suggesting:1 potential:1 de:1 north:1 explicitly:3 kzk2:1 ornstein:6 depends:2 sine:1 lab:1 closed:2 kwk:4 sup:2 analyze:5 characterizes:2 start:2 maintains:2 complicated:1 decaying:1 hazan:1 candes:1 contribution:1 minimize:1 variance:2 yield:1 identify:2 shenzhen:1 weak:11 bayesian:1 produced:1 lu:2 worth:1 published:1 ago:2 whenever:1 ed:2 infinitesimal:3 pp:12 e2:1 associated:1 riemannian:1 hardt:2 lim:1 emerges:1 subsection:2 carefully:2 back:2 higher:1 dt:4 furthermore:3 lastly:1 until:1 d:1 trust:1 su:1 dvk:1 logistic:3 infimum:1 lei:2 usa:1 verify:2 true:4 concept:1 former:1 hence:5 regularization:2 moritz:1 guangdong:1 nonzero:2 moore:1 illustrative:1 generalized:1 complete:2 motion:1 variational:2 consideration:1 recently:1 overview:1 exponentially:1 volume:2 discussed:1 interpretation:1 association:2 refer:1 ai:1 rd:5 consistency:1 mathematics:1 ethier:1 varadhan:1 analyzer:1 dbk:2 stable:2 etc:2 add:1 brownian:1 recent:5 perspective:6 aldous:1 inf:1 nonconvex:8 top1:1 meta:2 preserving:1 additional:2 period:1 ii:10 multiple:2 desirable:1 afterwards:1 reduces:1 match:4 faster:1 calculation:2 believed:2 sphere:8 retrieval:2 long:3 cross:1 bach:1 determination:1 e1:7 plugging:1 variant:1 basic:2 globe:2 expectation:2 poisson:1 arxiv:30 iteration:35 represent:1 uhlenbeck:6 normalization:1 equator:5 separately:1 ode:5 singular:1 source:1 fifty:1 rest:2 oksendal:1 pass:1 south:1 subject:1 db:2 leveraging:1 flow:1 inconsistent:1 jordan:1 anandkumar:1 subgaussian:1 near:4 noting:1 presence:1 bernstein:1 iii:8 identically:2 restrict:2 escaping:4 lasso:1 reduce:1 idea:1 pca:31 eigengap:3 hessian:1 remark:3 useful:1 se:5 eigenvectors:3 involve:1 amount:1 locally:2 canonical:1 panageas:1 discrete:6 dasgupta:1 easiness:1 traced:1 cutoff:1 diffusion:15 v1:10 asymptotically:3 relaxation:1 olukotun:1 year:2 angle:6 powerful:1 almost:1 throughout:1 reader:1 v12:3 wu:1 utilizes:1 oscillation:1 scaling:6 bound:9 display:1 annual:1 oracle:1 msg:1 occur:1 precisely:1 infinity:1 n3:5 diffuse:1 generates:1 aspect:2 u1:2 argument:3 concluding:4 performing:1 injection:1 department:1 according:1 helmke:1 son:1 kakade:1 qu:4 happens:1 dv:1 ghaoui:1 tsunami:1 taken:1 equation:13 previously:1 tai:1 discus:1 turn:2 needed:2 letting:1 ge:2 end:1 operation:2 apply:1 observe:1 balsubramani:1 away:2 appropriate:3 amini:1 stepsize:5 distinguished:1 alternative:1 existence:1 top:3 running:4 denotes:3 include:1 charaterizes:1 newton:1 especially:1 establish:1 classical:4 society:1 tensor:2 objective:5 move:1 occurs:1 costly:1 dependence:1 diagonal:4 gradient:17 kth:2 subspace:5 vd:1 restart:1 chris:1 manifold:1 unstable:8 reason:1 index:1 relationship:1 illustration:2 minimizing:1 mostly:1 sharper:1 negative:2 neuron:1 darken:1 markov:17 finite:7 enabling:1 descent:9 jin:2 truncated:2 situation:1 peres:1 communication:1 rn:1 perturbation:1 arbitrary:4 community:1 introduced:2 bk:1 namely:3 required:2 z1:1 nip:2 capped:1 able:1 dynamical:2 usually:1 pattern:1 appeared:1 sparsity:4 regime:1 summarize:2 including:3 memory:2 vi2:1 analogue:1 power:7 critical:2 eration:1 natural:1 warm:2 wainwright:1 minimax:7 picture:1 arora:2 concludes:1 aspremont:1 literature:3 understanding:1 geometric:2 freund:1 proportional:1 srebro:2 var:1 generator:1 eigendecomposition:1 sufficient:1 basin:1 consistent:1 thresholding:1 storing:1 wilmer:1 enjoys:2 understand:1 johnstone:1 piliouras:1 taking:1 sparse:8 distributed:3 curve:4 dimension:6 stand:1 author:1 collection:1 adaptive:2 projected:2 avoided:1 simplified:1 approximate:1 emphasize:2 keep:1 ml:1 global:20 sequentially:1 vio:1 symmetrical:1 conclude:3 continuous:2 iterative:3 search:1 duk:1 ca:1 exterior:1 tencent:1 du:1 complex:1 zou:1 diag:3 uko:1 aistats:1 apr:1 big:1 noise:4 n2:5 positively:1 martingale:2 tong:1 wiley:1 explicit:1 exponential:1 lie:2 third:1 companion:1 theorem:8 formula:1 insightful:1 admits:4 maximizers:2 mitliagkas:1 province:1 karhunen:1 uok:2 forget:1 rayleigh:2 saddle:6 expressed:2 pls:1 springer:3 corresponds:4 minimizer:1 satisfies:2 ma:2 conditional:1 marked:1 formulated:2 presentation:1 careful:1 towards:1 price:2 principal:25 svd:2 latter:1 brevity:2 accelerated:1 princeton:3 phenomenon:1 |
6,265 | 6,668 | Avoiding Discrimination through Causal Reasoning
Niki Kilbertus??
[email protected]
Moritz Hardt?
[email protected]
Mateo Rojas-Carulla??
[email protected]
Dominik Janzing?
[email protected]
Giambattista Parascandolo??
[email protected]
Bernhard Sch?olkopf?
[email protected]
?
Max Planck Institute for Intelligent Systems
?
University of Cambridge
?
Max Planck ETH Center for Learning Systems
?
University of California, Berkeley
Abstract
Recent work on fairness in machine learning has focused on various statistical
discrimination criteria and how they trade off. Most of these criteria are observational: They depend only on the joint distribution of predictor, protected attribute,
features, and outcome. While convenient to work with, observational criteria have
severe inherent limitations that prevent them from resolving matters of fairness
conclusively.
Going beyond observational criteria, we frame the problem of discrimination
based on protected attributes in the language of causal reasoning. This viewpoint shifts attention from ?What is the right fairness criterion?? to ?What do we
want to assume about our model of the causal data generating process?? Through
the lens of causality, we make several contributions. First, we crisply articulate
why and when observational criteria fail, thus formalizing what was before a matter of opinion. Second, our approach exposes previously ignored subtleties and
why they are fundamental to the problem. Finally, we put forward natural causal
non-discrimination criteria and develop algorithms that satisfy them.
1
Introduction
As machine learning progresses rapidly, its societal impact has come under scrutiny. An important
concern is potential discrimination based on protected attributes such as gender, race, or religion.
Since learned predictors and risk scores increasingly support or even replace human judgment, there
is an opportunity to formalize what harmful discrimination means and to design algorithms that
avoid it. However, researchers have found it difficult to agree on a single measure of discrimination.
As of now, there are several competing approaches, representing different opinions and striking
different trade-offs. Most of the proposed fairness criteria are observational: They depend only
on the joint distribution of predictor R, protected attribute A, features X, and outcome Y. For
example, the natural requirement that R and A must be statistically independent is referred to as
demographic parity. Some approaches transform the features X to obfuscate the information they
contain about A [1]. The recently proposed equalized odds constraint [2] demands that the predictor
R and the attribute A be independent conditional on the actual outcome Y. All three are examples
of observational approaches.
A growing line of work points at the insufficiency of existing definitions. Hardt, Price and Srebro [2]
construct two scenarios with intuitively different social interpretations that admit identical joint dis31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
tributions over (R, A, Y, X). Thus, no observational criterion can distinguish them. While there
are non-observational criteria, notably the early work on individual fairness [3], these have not yet
gained traction. So, it might appear that the community has reached an impasse.
1.1
Our contributions
We assay the problem of discrimination in machine learning in the language of causal reasoning.
This viewpoint supports several contributions:
? Revisiting the two scenarios proposed in [2], we articulate a natural causal criterion that
formally distinguishes them. In particular, we show that observational criteria are unable
to determine if a protected attribute has direct causal influence on the predictor that is not
mitigated by resolving variables.
? We point out subtleties in fair decision making that arise naturally from a causal perspective, but have gone widely overlooked in the past. Specifically, we formally argue for the
need to distinguish between the underlying concept behind a protected attribute, such as
race or gender, and its proxies available to the algorithm, such as visual features or name.
? We introduce and discuss two natural causal criteria centered around the notion of interventions (relative to a causal graph) to formally describe specific forms of discrimination.
? Finally, we initiate the study of algorithms that avoid these forms of discrimination. Under
certain linearity assumptions about the underlying causal model generating the data, an
algorithm to remove a specific kind of discrimination leads to a simple and natural heuristic.
At a higher level, our work proposes a shift from trying to find a single statistical fairness criterion
to arguing about properties of the data and which assumptions about the generating process are
justified. Causality provides a flexible framework for organizing such assumptions.
1.2
Related work
Demographic parity and its variants have been discussed in numerous papers, e.g., [1, 4?6]. While
demographic parity is easy to work with, the authors of [3] already highlighted its insufficiency as
a fairness constraint. In an attempt to remedy the shortcomings of demographic parity [2] proposed
two notions, equal opportunity and equal odds, that were also considered in [7]. A review of various
fairness criteria can be found in [8], where they are discussed in the context of criminal justice.
In [9, 10] it has been shown that imperfect predictors cannot simultaneously satisfy equal odds and
calibration unless the groups have identical base rates, i.e. rates of positive outcomes.
A starting point for our investigation is the unidentifiability result of [2]. It shows that observedvational criteria are too weak to distinguish two intuitively very different scenarios. However, the
work does not provide a formal mechanism to articulate why and how these scenarios should be
considered different. Inspired by Pearl?s causal interpretation of Simpson?s paradox [11, Section 6],
we propose causality as a way of coping with this unidentifiability result.
An interesting non-observational fairness definition is the notion of individual fairness [3] that assumes the existence of a similarity measure on individuals, and requires that any two similar individuals should receive a similar distribution over outcomes. More recent work lends additional support
to such a definition [12]. From the perspective of causality, the idea of a similarity measure is akin
to the method of matching in counterfactual reasoning [13, 14]. That is, evaluating approximate
counterfactuals by comparing individuals with similar values of covariates excluding the protected
attribute.
Recently, [15] put forward one possible causal definition, namely the notion of counterfactual fairness. It requires modeling counterfactuals on a per individual level, which is a delicate task. Even
determining the effect of race at the group level is difficult; see the discussion in [16]. The goal of
our paper is to assay a more general causal framework for reasoning about discrimination in machine
learning without committing to a single fairness criterion, and without committing to evaluating individual causal effects. In particular, we draw an explicit distinction between the protected attribute
(for which interventions are often impossible in practice) and its proxies (which sometimes can be
intervened upon).
2
Moreover, causality has already been employed for the discovery of discrimination in existing data
sets by [14, 17]. Causal graphical conditions to identify meaningful partitions have been proposed
for the discovery and prevention of certain types of discrimination by preprocessing the data [18].
These conditions rely on the evaluation of path specific effects, which can be traced back all the
way to [11, Section 4.5.3]. The authors of [19] recently picked up this notion and generalized
Pearl?s approach by a constraint based prevention of discriminatory path specific effects arising
from counterfactual reasoning. Our research was done independently of these works.
1.3
Causal graphs and notation
Causal graphs are a convenient way of organizing assumptions about the data generating process.
We will generally consider causal graphs involving a protected attribute A, a set of proxy variables P,
features X, a predictor R and sometimes an observed outcome Y. For background on causal graphs
see [11]. In the present paper a causal graph is a directed, acyclic graph whose nodes represent
random variables. A directed path is a sequence of distinct nodes V1 , . . . , Vk , for k ? 2, such
that Vi ? Vi+1 for all i ? {1, . . . , k ? 1}. We say a directed path is blocked by a set of nodes Z,
where V1 , Vk ?
/ Z, if Vi ? Z for some i ? {2, . . . , k ? 1}.1
A structural equation model is a set of equations Vi = fi (pa(Vi ), Ni ), for i ? {1, . . . , n},
where pa(Vi ) are the parents of Vi , i.e. its direct causes, and the Ni are independent noise variables. We interpret these equations as assignments. Because we assume acyclicity, starting from
the roots of the graph, we can recursively compute the other variables, given the noise variables.
This leads us to view the structural equation model and its corresponding graph as a data generating model. The predictor R maps inputs, e.g., the features X, to a predicted output. Hence we
model it as a childless node, whose parents are its input variables. Finally, note that given the noise
variables, a structural equation model entails a unique joint distribution; however, the same joint
distribution can usually be entailed by multiple structural equation models corresponding to distinct
causal structures.
2
Unresolved discrimination and limitations of observational criteria
To bear out the limitations of observational criteria, we turn to
Pearl?s commentary on claimed gender discrimination in Berkeley college admissions [11, Section 4.5.3]. Bickel [20] had shown
earlier that a lower college-wide admission rate for women than
for men was explained by the fact that women applied in more
competitive departments. When adjusted for department choice,
women experienced a slightly higher acceptance rate compared
with men. From the causal point of view, what matters is the direct effect of the protected attribute (here, gender A) on the decision (here, college admission R) that cannot be ascribed to a resolving variable such as department choice X, see Figure 1. We
shall use the term resolving variable for any variable in the causal
graph that is influenced by A in a manner that we accept as nondiscriminatory. With this convention, the criterion can be stated as
follows.
A
X
R
Figure 1: The admission decision R does not only directly depend on gender A, but
also on department choice X,
which in turn is also affected
by gender A.
Definition 1 (Unresolved discrimination). A variable V in a causal graph exhibits unresolved discrimination if there exists a directed path from A to V that is not blocked by a resolving variable
and V itself is non-resolving.
Pearl?s commentary is consistent with what we call the skeptic viewpoint. All paths from the protected attribute A to R are problematic, unless they are justified by a resolving variable. The presence of unresolved discrimination in the predictor R is worrisome and demands further scrutiny.
In practice, R is not a priori part of a given graph. Instead it is our objective to construct it as a
function of the features X, some of which might be resolving. Hence we should first look for unresolved discrimination in the features. A canonical way to avoid unresolved discrimination in R is to
only input the set of features that do not exhibit unresolved discrimination. However, the remaining
1
As it is not needed in our work, we do not discuss the graph-theoretic notion of d-separation.
3
features might be affected by non-resolving and resolving variables. In Section 4 we investigate
whether one can exclusively remove unresolved discrimination from such features. A related notion
of ?explanatory features? in a non-causal setting was introduced in [21].
The definition of unresolved discrimination in
a predictor has some interesting special cases
R?
R?
worth highlighting. If we take the set of resolving variables to be empty, we intuitively get a
causal analog of demographic parity. No diX1
X2
A
Y
A
Y
rected paths from A to R are allowed, but A
and R can still be statistically dependent. SimiX2
X1
larly, if we choose the set of resolving variables
to be the singleton set {Y } containing the true
outcome, we obtain a causal analog of equal- Figure 2: Two graphs that may generate the same
ized odds where strict independence is not nec- joint distribution for? the Bayes optimal unconessary. The causal intuition implied by ?the strained? predictor R . If X1 is a resolving variprotected attribute should not affect the predic- able, R exhibits unresolved discrimination in the
tion?, and ?the protected attribute can only af- right graph (along the red paths), but not in the left
fect the prediction when the information comes one.
through the true label?, is neglected by (conditional) statistical independences A ?? R, and A ?? R | Y , but well captured by only considering
dependences mitigated along directed causal paths.
We will next show that observational criteria are fundamentally unable to determine whether a predictor exhibits unresolved discrimination or not. This is true even if the predictor is Bayes optimal.
In passing, we also note that fairness criteria such as equalized odds may or may not exhibit unresolved discrimination, but this is again something an observational criterion cannot determine.
Theorem 1. Given a joint distribution over the protected attribute A, the true label Y , and some
features X1 , . . . , Xn , in which we have already specified the resolving variables, no observational
criterion can generally determine whether the Bayes optimal unconstrained predictor or the Bayes
optimal equal odds predictor exhibit unresolved discrimination.
All proofs for the statements in this paper are in the supplementary material.
The two graphs in Figure 2 are taken from [2], which we here reinterpret in the causal context to
prove Theorem 1. We point out that there is an established set of conditions under which unresolved
discrimination can, in fact, be determined from observational data. Note that the two graphs are
not Markov equivalent. Therefore, to obtain the same joint distribution we must violate a condition
called faithfulness.2 We later argue that violation of faithfulness is by no means pathological, but
emerges naturally when designing predictors. In any case, interpreting conditional dependences can
be difficult in practice [22].
3
Proxy discrimination and interventions
We now turn to an important aspect of our framework. Determining causal effects in general requires
modeling interventions. Interventions on deeply rooted individual properties such as gender or
race are notoriously difficult to conceptualize?especially at an individual level, and impossible to
perform in a randomized trial. VanderWeele et al. [16] discuss the problem comprehensively in an
epidemiological setting. From a machine learning perspective, it thus makes sense to separate the
protected attribute A from its potential proxies, such as name, visual features, languages spoken at
home, etc. Intervention based on proxy variables poses a more manageable problem. By deciding on
a suitable proxy we can find an adequate mounting point for determining and removing its influence
on the prediction. Moreover, in practice we are often limited to imperfect measurements of A in any
case, making the distinction between root concept and proxy prudent.
As was the case with resolving variables, a proxy is a priori nothing more than a descendant of A in
the causal graph that we choose to label as a proxy. Nevertheless in reality we envision the proxy
2
If we do assume the Markov condition and faithfulness, then conditional independences determine the
graph up to its so called Markov equivalence class.
4
to be a clearly defined observable quantity that is significantly correlated with A, yet in our view
should not affect the prediction.
Definition 2 (Potential proxy discrimination). A variable V in a causal graph exhibits potential
proxy discrimination, if there exists a directed path from A to V that is blocked by a proxy variable
and V itself is not a proxy.
Potential proxy discrimination articulates a causal criterion that is in a sense dual to unresolved
discrimination. From the benevolent viewpoint, we allow any path from A to R unless it passes
through a proxy variable, which we consider worrisome. This viewpoint acknowledges the fact that
the influence of A on the graph may be complex and it can be too restraining to rule out all but a few
designated features. In practice, as with unresolved discrimination, we can naively build an unconstrained predictor based only on those features that do not exhibit potential proxy discrimination.
Then we must not provide P as input to R; unawareness, i.e. excluding P from the inputs of R,
suffices. However, by granting R access to P , we can carefully tune the function R(P, X) to cancel
the implicit influence of P on features X that exhibit potential proxy discrimination by the explicit
dependence on P . Due to this possible cancellation of paths, we called the path based criterion potential proxy discrimination. When building predictors that exhibit no overall proxy discrimination,
we precisely aim for such a cancellation.
Fortunately, this idea can be conveniently expressed by an intervention on P , which is denoted
by do(P = p) [11]. Visually, intervening on P amounts to removing all incoming arrows of P in
the graph; algebraically, it consists of replacing the structural equation of P by P = p, i.e. we put
point mass on the value p.
Definition 3 (Proxy discrimination). A predictor R exhibits no proxy discrimination based on a
proxy P if for all p, p0
P(R | do(P = p)) = P(R | do(P = p0 )) .
(1)
The interventional characterization of proxy discrimination leads to a simple procedure to remove
it in causal graphs that we will turn to in the next section. It also leads to several natural variants
of the definition that we discuss in Section 4.3. We remark that Equation (1) is an equality of
probabilities in the ?do-calculus? that cannot in general be inferred by an observational method,
because it depends on an underlying causal graph, see the discussion in [11]. However, in some
cases, we do not need to resort to interventions to avoid proxy discrimination.
Proposition 1. If there is no directed path from a proxy to a feature, unawareness avoids proxy
discrimination.
4
Procedures for avoiding discrimination
Having motivated the two types of discrimination that we distinguish, we now turn to building
predictors that avoid them in a given causal model. First, we remark that a more comprehensive
treatment requires individual judgement of not only variables, but the legitimacy of every existing
path that ends in R, i.e. evaluation of path-specific effects [18, 19], which is tedious in practice.
The natural concept of proxies and resolving variables covers most relevant scenarios and allows for
natural removal procedures.
4.1
Avoiding proxy discrimination
While presenting the general procedure, we illustrate each step in the example shown in Figure 3.
A protected attribute A affects a proxy P as well as a feature X. Both P and X have additional
unobserved causes NP and NX , where NP , NX , A are pairwise independent. Finally, the proxy also
has an effect on the features X and the predictor R is a function of P and X. Given labeled training
data, our task is to find a good predictor that exhibits no proxy discrimination within a hypothesis
class of functions R? (P, X) parameterized by a real valued vector ?.
We now work out a formal procedure to solve this task under specific assumptions and simultaneously illustrate it in a fully linear example, i.e. the structural equations are given by
P = ?P A + NP ,
X = ?X A + ?P + NX ,
R? = ?P P + ?X X .
Note that we choose linear functions parameterized by ? = (?P , ?X ) as the hypothesis class
for R? (P, X).
5
NP
NX
A
P
G?
NP
X
P
G
R
NX
A
NE
X
E
R
G?
Figure 3: A template graph G? for proxy discrimination (left) with its intervened version G
(right). While from the benevolent viewpoint we
do not generically prohibit any influence from A
on R, we want to guarantee that the proxy P has
no overall influence on the prediction, by adjusting P ? R to cancel the influence along P ?
X ? R in the intervened graph.
NX
A
X
R
NE
NX
A
E
G
X
R
Figure 4: A template graph G? for unresolved
discrimination (left) with its intervened version G (right). While from the skeptical
viewpoint we generically do not want A to
influence R, we first intervene on E interrupting all paths through E and only cancel
the remaining influence on A to R.
We will refer to the terminal ancestors of a node V in a causal graph D, denoted by taD (V ), which
are those ancestors of V that are also root nodes of D. Moreover, in the procedure we clarify the
notion of expressibility, which is an assumption about the relation of the given structural equations
and the hypothesis class we choose for R? .
Proposition 2. If there is a choice of parameters ?0 such that R?0 (P, X) is constant with respect
to its first argument and the structural equations are expressible, the following procedure returns a
predictor from the given hypothesis class that exhibits no proxy discrimination and is non-trivial in
the sense that it can make use of features that exhibit potential proxy discrimination.
1. Intervene on P by removing all incoming arrows and replacing the structural equation for P
by P = p. For the example in Figure 3,
P = p,
X = ?X A + ?P + NX ,
R? = ?P P + ?X X .
(2)
2. Iteratively substitute variables in the equation for R? from their structural equations until only
root nodes of the intervened graph are left, i.e. write R? (P, X) as R? (P, g(taG (X))) for some
function g. In the example, ta(X) = {A, P, NX } and
R? = (?P + ?X ?)p + ?X (?X A + NX ) .
(3)
3. We now require the distribution of R? in (3) to be independent of p, i.e. for all p, p0
P((?P + ?X ?)p + ?X (?X A + NX )) = P((?P + ?X ?)p0 + ?X (?X A + NX )) .
(4)
We seek to write the predictor as a function of P and all the other roots of G separately. If our
hypothesis class is such that there exists ?? such that R? (P, g(ta(X))) = R??(P, g?(ta(X) \ {P })),
we call the structural equation model and hypothesis class specified in (2) expressible. In our
example, this is possible with ?? = (?P + ?X ?, ?X ) and g? = ?X A + NX . Equation (4) then
yields the non-discrimination constraint ?? = ?0 . Here, a possible ?0 is ?0 = (0, ?X ), which
simply yields ?P = ??X ?.
4. Given labeled training data, we can optimize the predictor R? within the hypothesis class as given
in (2), subject to the non-discrimination constraint. In the example
R? = ??X ?P + ?X X = ?X (X ? ?P ) ,
with the free parameter ?X ? R.
In general, the non-discrimination constraint (4) is by construction just P(R | do(P = p)) =
P(R | do(P = p0 )), coinciding with Definition 3. Thus Proposition 2 holds by construction of
the procedure. The choice of ?0 strongly influences the non-discrimination constraint. However, as
the example shows, it allows R? to exploit features that exhibit potential proxy discrimination.
6
A
P
G?
R
X
A
P
G
DAG
R
X
DAG
Figure 5: Left: A generic graph G? to describe proxy discrimination. Right: The graph corresponding
to an intervention on P . The circle labeled ?DAG? represents any sub-DAG of G? and G containing
an arbitrary number of variables that is compatible with the shown arrows. Dashed arrows can, but
do not have to be present in a given scenario.
4.2
Avoiding unresolved discrimination
We proceed analogously to the previous subsection using the example graph in Figure 4. Instead of
the proxy, we consider a resolving variable E. The causal dependences are equivalent to the ones in
Figure 3 and we again assume linear structural equations
E = ?E A + NE ,
X = ?X A + ?E + NX ,
R? = ?E E + ?X X .
Let us now try to adjust the previous procedure to the context of avoiding unresolved discrimination.
1. Intervene on E by fixing it to a random variable ? with P(?) = P(E), the marginal distribution
? see Figure 4. In the example we find
of E in G,
E = ?,
X = ?X A + ?E + NX ,
R? = ?E E + ?X X .
(5)
2. By iterative substitution write R? (E, X) as R? (E, g(taG (X))) for some function g, i.e. in the
example
R? = (?E + ?X ?)? + ?X ?X A + ?X NX .
(6)
3. We now demand the distribution of R? in (6) be invariant under interventions on A, which coin? Hence, in the example, for all a, a0
cides with conditioning on A whenever A is a root of G.
P((?E + ?X ?)? + ?X ?X a + ?X NX )) = P((?E + ?X ?)? + ?X ?X a0 + ?X NX )) .
(7)
Here, the subtle asymmetry between proxy discrimination and unresolved discrimination becomes
apparent. Because R? is not explicitly a function of A, we cannot cancel implicit influences of A
through X. There might still be a ?0 such that R?0 indeed fulfils (7), but there is no principled way for us to construct it. In the example, (7) suggests the obvious non-discrimination constraint ?X = 0. We can then proceed as before and, given labeled training data, optimize R? = ?E E
by varying ?E . However, by setting ?X = 0, we also cancel the path A ? E ? X ? R, even
though it is blocked by a resolving variable. In general, if R? does not have access to A, we can not
adjust for unresolved discrimination without also removing resolved influences from A on R? .
If, however, R? is a function of A, i.e. we add the term ?A A to R? in (5), the non-discrimination
constraint is ?A = ??X ?X and we can proceed analogously to the procedure for proxies.
4.3
Relating proxy discriminations to other notions of fairness
Motivated by the algorithm to avoid proxy discrimination, we discuss some natural variants of the
notion in this section that connect our interventional approach to individual fairness and other proposed criteria. We consider a generic graph structure as shown on the left in Figure 5. The proxy P
and the features X could be multidimensional. The empty circle in the middle represents any number of variables forming a DAG that respects the drawn arrows. Figure 3 is an example thereof. All
dashed arrows are optional depending on the specifics of the situation.
Definition 4. A predictor R exhibits no individual proxy discrimination, if for all x and all p, p0
P(R | do(P = p), X = x) = P(R | do(P = p0 ), X = x) .
A predictor R exhibits no proxy discrimination in expectation, if for all p, p0
E[R | do(P = p)] = E[R | do(P = p0 )] .
7
Individual proxy discrimination aims at comparing examples with the same features X, for different
values of P . Note that this can be individuals with different values for the unobserved non-feature
variables. A true individual-level comparison of the form ?What would have happened to me, if I
had always belonged to another group? is captured by counterfactuals and discussed in [15, 19].
For an analysis of proxy discrimination, we need the structural equations for P, X, R in Figure 5
P = f?P (pa(P )) ,
X = f?X (pa(X)) = fX (P, taG (X) \ {P }) ,
R = f?R (P, X) = fR (P, taG (R) \ {P }) .
For convenience, we will use the notation taGP (X) := taG (X) \ {P }. We can find fX , fR
from f?X , f?R by first rewriting the functions in terms of root nodes of the intervened graph, shown
on the right side of Figure 5, and then assigning the overall dependence on P to the first argument.
We now compare proxy discrimination to other existing notions.
Theorem 2. Let the influence of P on X be additive and linear, i.e.
X = fX (P, taGP (X)) = gX (taGP (X)) + ?X P
for some function gX and ?X ? R. Then any predictor of the form
R = r(X ? E[X | do(P )])
for some function r exhibits no proxy discrimination.
Note that in general E[X | do(P )] 6= E[X | P ]. Since in practice we only have observational data
? one cannot simply build a predictor based on the ?regressed out features? X
? := X ?
from G,
E[X | P ] to avoid proxy discrimination. In the scenario of Figure 3, the direct effect of P on X
along the arrow P ? X in the left graph cannot be estimated by E[X | P ], because of the common
confounder A. The desired interventional expectation E[X | do(P )] coincides with E[X | P ] only
if one of the arrows A ? P or A ? X is not present. Estimating direct causal effects is a hard
problem, well studied by the causality community and often involves instrumental variables [23].
? as a ?fair representation? of X, as it implicitly
This cautions against the natural idea of using X
neglects that we often want to remove the effect of proxies and not the protected attribute. Nevertheless, the notion agrees with our interventional proxy discrimination in some cases.
Corollary 1. Under the assumptions of Theorem 2, if all directed paths from any ancestor of P
? :=
to X in the graph G are blocked by P , then any predictor based on the adjusted features X
X ? E[X | P ] exhibits no proxy discrimination and can be learned from the observational distribution P(P, X, Y ) when target labels Y are available.
Our definition of proxy discrimination in expectation (4) is motivated by a weaker notion proposed
in [24]. It asks for the expected outcome to be the same across the different populations E[R | P =
p] = E[R | P = p0 ]. Again, when talking about proxies, we must be careful to distinguish conditional
and interventional expectations, which is captured by the following proposition and its corollary.
Proposition 3. Any predictor of the form R = ?(X ? E[X | do(P )]) + c for ?, c ? R exhibits no
proxy discrimination in expectation.
From this and the proof of Corollary 1 we conclude the following Corollary.
Corollary 2. If all directed paths from any ancestor of P to X are blocked by P , any predictor of
the form R = r(X ? E[X | P ]) for linear r exhibits no proxy discrimination in expectation and can
be learned from the observational distribution P(P, X, Y ) when target labels Y are available.
5
Conclusion
The goal of our work is to assay fairness in machine learning within the context of causal reasoning.
This perspective naturally addresses shortcomings of earlier statistical approaches. Causal fairness
criteria are suitable whenever we are willing to make assumptions about the (causal) generating
8
process governing the data. Whilst not always feasible, the causal approach naturally creates an
incentive to scrutinize the data more closely and work out plausible assumptions to be discussed
alongside any conclusions regarding fairness.
Key concepts of our conceptual framework are resolving variables and proxy variables that play
a dual role in defining causal discrimination criteria. We develop a practical procedure to remove
proxy discrimination given the structural equation model and analyze a similar approach for unresolved discrimination. In the case of proxy discrimination for linear structural equations, the
procedure has an intuitive form that is similar to heuristics already used in the regression literature.
Our framework is limited by the assumption that we can construct a valid causal graph. The removal
of proxy discrimination moreover depends on the functional form of the causal dependencies. We
have focused on the conceptual and theoretical analysis, and experimental validations are beyond
the scope of the present work.
The causal perspective suggests a number of interesting new directions at the technical, empirical,
and conceptual level. We hope that the framework and language put forward in our work will be a
stepping stone for future investigations.
9
References
[1]
[2]
[3]
[4]
[5]
[6]
[7]
[8]
[9]
[10]
[11]
[12]
[13]
[14]
[15]
[16]
[17]
[18]
[19]
[20]
[21]
[22]
Richard S Zemel, Yu Wu, Kevin Swersky, Toniann Pitassi, and Cynthia Dwork. ?Learning
Fair Representations.? In: Proceedings of the International Conference of Machine Learning
28 (2013), pp. 325?333.
Moritz Hardt, Eric Price, Nati Srebro, et al. ?Equality of opportunity in supervised learning?.
In: Advances in Neural Information Processing Systems. 2016, pp. 3315?3323.
Cynthia Dwork, Moritz Hardt, Toniann Pitassi, Omer Reingold, and Richard Zemel. ?Fairness
Through Awareness?. In: Proceedings of the 3rd Innovations in Theoretical Computer Science
Conference. 2012, pp. 214?226.
Michael Feldman, Sorelle A Friedler, John Moeller, Carlos Scheidegger, and Suresh Venkatasubramanian. ?Certifying and removing disparate impact?. In: Proceedings of the 21th
ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 2015,
pp. 259?268.
Muhammad Bilal Zafar, Isabel Valera, Manuel G?omez Rogriguez, and Krishna P. Gummadi.
?Fairness Constraints: Mechanisms for Fair Classification?. In: Proceedings of the 20th International Conference on Artificial Intelligence and Statistics. 2017, pp. 962?970.
Harrison Edwards and Amos Storkey. ?Censoring Representations with an Adversary?. In:
(Nov. 18, 2015). arXiv: 1511.05897v3.
Muhammad Bilal Zafar, Isabel Valera, Manuel G?omez Rodriguez, and Krishna P. Gummadi.
?Fairness Beyond Disparate Treatment & Disparate Impact: Learning Classification Without
Disparate Mistreatment?. In: Proceedings of the 26th International Conference on World Wide
Web. 2017, pp. 1171?1180.
Richard Berk, Hoda Heidari, Shahin Jabbari, Michael Kearns, and Aaron Roth. ?Fairness
in Criminal Justice Risk Assessments: The State of the Art?. In: (Mar. 27, 2017). arXiv:
1703.09207v1.
Jon Kleinberg, Sendhil Mullainathan, and Manish Raghavan. ?Inherent Trade-Offs in the Fair
Determination of Risk Scores?. In: (Sept. 19, 2016). arXiv: 1609.05807v1.
Alexandra Chouldechova. ?Fair prediction with disparate impact: A study of bias in recidivism prediction instruments?. In: (Oct. 24, 2016). arXiv: 1610.07524v1.
Judea Pearl. Causality. Cambridge University Press, 2009.
Sorelle A. Friedler, Carlos Scheidegger, and Suresh Venkatasubramanian. ?On the
(im)possibility of fairness?. In: (Sept. 23, 2016). arXiv: 1609.07236v1.
Paul R Rosenbaum and Donald B Rubin. ?The central role of the propensity score in observational studies for causal effects?. In: Biometrika (1983), pp. 41?55.
Bilal Qureshi, Faisal Kamiran, Asim Karim, and Salvatore Ruggieri. ?Causal Discrimination
Discovery Through Propensity Score Analysis?. In: (Aug. 12, 2016). arXiv: 1608.03735.
Matt J. Kusner, Joshua R. Loftus, Chris Russell, and Ricardo Silva. ?Counterfactual Fairness?.
In: (Mar. 20, 2017). arXiv: 1703.06856v1.
Tyler J VanderWeele and Whitney R Robinson. ?On causal interpretation of race in regressions adjusting for confounding and mediating variables?. In: Epidemiology 25.4 (2014),
p. 473.
Francesco Bonchi, Sara Hajian, Bud Mishra, and Daniele Ramazzotti. ?Exposing the probabilistic causal structure of discrimination?. In: (Mar. 8, 2017). arXiv: 1510.00552v3.
Lu Zhang and Xintao Wu. ?Anti-discrimination learning: a causal modeling-based framework?. In: International Journal of Data Science and Analytics (2017), pp. 1?16.
Razieh Nabi and Ilya Shpitser. ?Fair Inference On Outcomes?. In: (May 29, 2017). arXiv:
1705.10378v1.
Peter J Bickel, Eugene A Hammel, J William O?Connell, et al. ?Sex bias in graduate admissions: Data from Berkeley?. In: Science 187.4175 (1975), pp. 398?404.
?
Faisal Kamiran, Indr?e Zliobait?
e, and Toon Calders. ?Quantifying explainable discrimination
and removing illegal discrimination in automated decision making?. In: Knowledge and information systems 35.3 (2013), pp. 613?644.
Nicholas Cornia and Joris M Mooij. ?Type-II errors of independence tests can lead to arbitrarily large errors in estimated causal effects: An illustrative example?. In: Proceedings of
the Workshop on Causal Inference (UAI). 2014, pp. 35?42.
10
[23]
Joshua Angrist and Alan B Krueger. Instrumental variables and the search for identification:
From supply and demand to natural experiments. Tech. rep. National Bureau of Economic
Research, 2001.
[24] Toon Calders and Sicco Verwer. ?Three naive Bayes approaches for discrimination-free classification?. In: Data Mining and Knowledge Discovery 21.2 (2010), pp. 277?292.
11
| 6668 |@word trial:1 middle:1 version:2 manageable:1 judgement:1 instrumental:2 justice:2 tedious:1 sex:1 calculus:1 willing:1 seek:1 p0:10 asks:1 recursively:1 venkatasubramanian:2 substitution:1 score:4 exclusively:1 sendhil:1 envision:1 bilal:3 past:1 existing:4 mishra:1 comparing:2 qureshi:1 manuel:2 yet:2 assigning:1 must:4 john:1 exposing:1 additive:1 partition:1 remove:5 unidentifiability:2 discrimination:90 mounting:1 intelligence:1 granting:1 provides:1 characterization:1 node:8 gx:2 zhang:1 admission:5 along:4 direct:5 supply:1 descendant:1 prove:1 consists:1 bonchi:1 manner:1 introduce:1 pairwise:1 ascribed:1 notably:1 expected:1 indeed:1 mpg:5 growing:1 terminal:1 expressibility:1 inspired:1 actual:1 considering:1 becomes:1 estimating:1 mitigated:2 formalizing:1 underlying:3 linearity:1 moreover:4 notation:2 what:7 mass:1 kind:1 caution:1 spoken:1 unobserved:2 whilst:1 guarantee:1 berkeley:4 every:1 reinterpret:1 multidimensional:1 biometrika:1 intervention:10 scrutiny:2 planck:2 appear:1 before:2 positive:1 insufficiency:2 path:20 might:4 mateo:1 studied:1 equivalence:1 suggests:2 sara:1 limited:2 discriminatory:1 gone:1 statistically:2 confounder:1 analytics:1 directed:9 unique:1 practical:1 arguing:1 graduate:1 practice:7 epidemiological:1 procedure:12 suresh:2 coping:1 empirical:1 eth:1 significantly:1 illegal:1 convenient:2 matching:1 donald:1 get:1 cannot:7 convenience:1 put:4 risk:3 influence:13 context:4 impossible:2 optimize:2 equivalent:2 map:1 center:1 roth:1 attention:1 starting:2 independently:1 focused:2 rule:1 population:1 notion:13 fx:3 construction:2 target:2 play:1 designing:1 larly:1 hypothesis:7 pa:4 storkey:1 tributions:1 labeled:4 mistreatment:1 observed:1 role:2 revisiting:1 verwer:1 trade:3 russell:1 deeply:1 principled:1 intuition:1 covariates:1 neglected:1 sicco:1 depend:3 upon:1 creates:1 eric:1 resolved:1 joint:8 isabel:2 various:2 distinct:2 committing:2 describe:2 shortcoming:2 artificial:1 equalized:2 zemel:2 kevin:1 outcome:9 whose:2 heuristic:2 widely:1 supplementary:1 valued:1 say:1 solve:1 apparent:1 plausible:1 statistic:1 transform:1 highlighted:1 itself:2 legitimacy:1 sequence:1 propose:1 skeptic:1 unresolved:22 fr:2 relevant:1 rapidly:1 moeller:1 organizing:2 omer:1 intervening:1 intuitive:1 olkopf:1 parent:2 empty:2 requirement:1 asymmetry:1 generating:6 illustrate:2 develop:2 depending:1 fixing:1 pose:1 progress:1 aug:1 edward:1 predicted:1 involves:1 come:2 rosenbaum:1 convention:1 direction:1 closely:1 attribute:18 centered:1 human:1 raghavan:1 observational:21 opinion:2 material:1 muhammad:2 require:1 suffices:1 investigation:2 articulate:3 proposition:5 im:1 adjusted:2 clarify:1 hold:1 around:1 considered:2 visually:1 deciding:1 tyler:1 scope:1 strained:1 bickel:2 early:1 friedler:2 label:5 expose:1 propensity:2 agrees:1 amos:1 hope:1 offs:2 clearly:1 always:2 aim:2 shahin:1 avoid:7 varying:1 corollary:5 vk:2 tech:1 sigkdd:1 sense:3 inference:2 dependent:1 accept:1 explanatory:1 a0:2 relation:1 ancestor:4 expressible:2 going:1 overall:3 dual:2 flexible:1 classification:3 prudent:1 priori:2 denoted:2 prevention:2 proposes:1 art:1 special:1 conceptualize:1 marginal:1 equal:5 construct:4 having:1 beach:1 identical:2 represents:2 look:1 cancel:5 fairness:24 jon:1 yu:1 future:1 np:5 intelligent:1 inherent:2 fundamentally:1 kilbertus:1 distinguishes:1 pathological:1 few:1 richard:3 simultaneously:2 national:1 comprehensive:1 individual:15 impasse:1 delicate:1 william:1 attempt:1 acceptance:1 mining:2 investigate:1 simpson:1 dwork:2 evaluation:2 severe:1 adjust:2 possibility:1 violation:1 entailed:1 generically:2 behind:1 mullainathan:1 unless:3 harmful:1 circle:2 desired:1 causal:56 theoretical:2 modeling:3 earlier:2 predic:1 cover:1 whitney:1 assignment:1 predictor:32 too:2 connect:1 dependency:1 fundamental:1 randomized:1 international:5 epidemiology:1 probabilistic:1 off:1 michael:2 analogously:2 ilya:1 fect:1 again:3 central:1 containing:2 choose:4 woman:3 admit:1 resort:1 shpitser:1 return:1 manish:1 ricardo:1 potential:10 de:5 singleton:1 matter:3 satisfy:2 explicitly:1 race:5 vi:7 depends:2 tion:1 root:7 picked:1 view:3 try:1 analyze:1 counterfactuals:3 reached:1 competitive:1 bayes:5 red:1 carlos:2 contribution:3 ni:2 judgment:1 identify:1 yield:2 weak:1 identification:1 lu:1 worth:1 researcher:1 notoriously:1 influenced:1 janzing:2 whenever:2 definition:12 against:1 pp:12 obvious:1 thereof:1 naturally:4 proof:2 judea:1 ruggieri:1 hardt:5 treatment:2 adjusting:2 counterfactual:4 subsection:1 emerges:1 knowledge:3 formalize:1 subtle:1 carefully:1 back:1 higher:2 ta:3 supervised:1 coinciding:1 done:1 though:1 strongly:1 mar:3 just:1 implicit:2 governing:1 heidari:1 until:1 web:1 replacing:2 assessment:1 rodriguez:1 alexandra:1 name:2 usa:1 effect:13 contain:1 concept:4 remedy:1 true:5 building:2 hence:3 equality:2 matt:1 moritz:3 iteratively:1 karim:1 assay:3 rooted:1 illustrative:1 prohibit:1 daniele:1 coincides:1 criterion:29 generalized:1 trying:1 stone:1 presenting:1 theoretic:1 interpreting:1 silva:1 reasoning:7 recently:3 fi:1 krueger:1 common:1 functional:1 stepping:1 conditioning:1 discussed:4 interpretation:3 analog:2 relating:1 interpret:1 measurement:1 blocked:6 refer:1 cambridge:2 feldman:1 dag:5 sorelle:2 rd:1 unconstrained:2 cancellation:2 language:4 had:2 calibration:1 entail:1 similarity:2 access:2 intervene:3 etc:1 base:1 add:1 something:1 pitassi:2 recent:2 restraining:1 perspective:5 confounding:1 scenario:7 claimed:1 certain:2 rep:1 arbitrarily:1 societal:1 joshua:2 captured:3 krishna:2 additional:2 commentary:2 fortunately:1 employed:1 determine:5 algebraically:1 v3:2 dashed:2 ii:1 resolving:19 multiple:1 violate:1 alan:1 technical:1 determination:1 af:1 long:1 gummadi:2 tad:1 impact:4 prediction:6 variant:3 involving:1 regression:2 expectation:6 arxiv:9 sometimes:2 represent:1 faisal:2 justified:2 receive:1 want:4 background:1 separately:1 scheidegger:2 harrison:1 sch:1 strict:1 pass:1 subject:1 reingold:1 parascandolo:1 odds:6 call:2 structural:15 presence:1 easy:1 automated:1 independence:4 affect:3 competing:1 imperfect:2 idea:3 regarding:1 economic:1 shift:2 whether:3 motivated:3 angrist:1 akin:1 explainable:1 peter:1 vanderweele:2 passing:1 cause:2 proceed:3 remark:2 adequate:1 ignored:1 generally:2 tune:1 amount:1 traction:1 kamiran:2 generate:1 problematic:1 canonical:1 happened:1 estimated:2 arising:1 per:1 write:3 shall:1 incentive:1 affected:2 group:3 key:1 nevertheless:2 traced:1 drawn:1 loftus:1 interventional:5 prevent:1 rewriting:1 v1:8 graph:37 parameterized:2 striking:1 swersky:1 wu:2 separation:1 home:1 draw:1 interrupting:1 decision:4 distinguish:5 constraint:10 precisely:1 x2:1 regressed:1 asim:1 fulfils:1 tag:5 certifying:1 aspect:1 kleinberg:1 argument:2 connell:1 recidivism:1 department:4 designated:1 unawareness:2 across:1 slightly:1 increasingly:1 kusner:1 b:1 making:3 intuitively:3 explained:1 invariant:1 taken:1 equation:20 agree:1 previously:1 calder:2 discus:5 turn:5 fail:1 mechanism:2 needed:1 initiate:1 instrument:1 demographic:5 end:1 jabbari:1 available:3 generic:2 nicholas:1 salvatore:1 coin:1 existence:1 substitute:1 hajian:1 assumes:1 remaining:2 bureau:1 graphical:1 opportunity:3 toon:2 neglect:1 joris:1 exploit:1 especially:1 build:2 implied:1 objective:1 already:4 quantity:1 dependence:5 exhibit:21 lends:1 unable:2 separate:1 tue:5 nx:18 me:1 chris:1 argue:2 trivial:1 bud:1 innovation:1 difficult:4 mediating:1 statement:1 stated:1 ized:1 disparate:5 design:1 perform:1 francesco:1 markov:3 anti:1 optional:1 situation:1 defining:1 excluding:2 paradox:1 frame:1 articulates:1 arbitrary:1 community:2 overlooked:1 inferred:1 criminal:2 introduced:1 namely:1 specified:2 faithfulness:3 california:1 learned:3 distinction:2 established:1 pearl:5 nip:1 robinson:1 address:1 beyond:3 able:1 alongside:1 usually:1 adversary:1 belonged:1 max:2 suitable:2 natural:11 rely:1 valera:2 representing:1 numerous:1 ne:3 acknowledges:1 naive:1 sept:2 niki:1 review:1 literature:1 discovery:5 removal:2 nati:1 eugene:1 determining:3 relative:1 mooij:1 toniann:2 fully:1 bear:1 interesting:3 limitation:3 men:2 worrisome:2 srebro:2 acyclic:1 validation:1 awareness:1 proxy:63 consistent:1 rubin:1 viewpoint:7 obfuscate:1 skeptical:1 compatible:1 censoring:1 parity:5 free:2 formal:2 allow:1 side:1 weaker:1 institute:1 wide:2 comprehensively:1 template:2 bias:2 xn:1 evaluating:2 avoids:1 valid:1 world:1 cornia:1 forward:3 author:2 preprocessing:1 social:1 approximate:1 observable:1 nov:1 implicitly:1 bernhard:1 conclusively:1 incoming:2 uai:1 conceptual:3 conclude:1 search:1 iterative:1 protected:16 why:3 reality:1 ca:1 complex:1 hoda:1 zafar:2 arrow:8 noise:3 arise:1 paul:1 nothing:1 fair:7 allowed:1 x1:3 causality:7 referred:1 experienced:1 sub:1 explicit:2 intervened:6 dominik:1 theorem:4 removing:6 specific:7 cynthia:2 concern:1 exists:3 naively:1 workshop:1 gained:1 nec:1 rected:1 demand:4 simply:2 forming:1 visual:2 conveniently:1 highlighting:1 expressed:1 religion:1 omez:2 subtlety:2 talking:1 chouldechova:1 gender:7 acm:1 oct:1 conditional:5 goal:2 rojas:1 quantifying:1 careful:1 replace:1 price:2 feasible:1 hard:1 specifically:1 determined:1 berk:1 kearns:1 lens:1 called:3 experimental:1 meaningful:1 acyclicity:1 aaron:1 formally:3 college:3 carulla:1 support:3 later:1 avoiding:5 correlated:1 |
6,266 | 6,669 | Nonparametric Online Regression
while Learning the Metric
Ilja Kuzborskij
EPFL
Switzerland
[email protected]
Nicol`o Cesa-Bianchi
Dipartimento di Informatica
Universit`a degli Studi di Milano
Milano 20135, Italy
[email protected]
Abstract
We study algorithms for online nonparametric regression that learn the directions
along which the regression function is smoother. Our algorithm learns the Mahalanobis metric based on the gradient outer product matrix G of the regression
function (automatically adapting to the effective rank of this matrix), while simultaneously bounding the regret ?on the same data sequence? in terms of the
spectrum of G. As a preliminary step in our analysis, we extend a nonparametric
online learning algorithm by Hazan and Megiddo enabling it to compete against
functions whose Lipschitzness is measured with respect to an arbitrary Mahalanobis
metric.
1
Introduction
An online learner is an agent interacting with an unknown and arbitrary environment over a sequence
of rounds. At each round t, the learner observes a data point (or instance) xt ? X ? Rd , outputs a
prediction ybt for the label yt ? R associated with that instance, and incurs some loss `t (b
yt ), which in
this paper is the square loss (b
yt ? yt )2 . At the end of the round, the label yt is given to the learner,
which he can use to reduce his loss in subsequent rounds. The performance of an online learner is
typically measured using the regret. This is defined as the amount by which the learner?s cumulative
loss exceeds the cumulative loss (on the same sequence of instances and labels) of any function f in
a given reference class F of functions,
T
X
RT (f ) =
`t (b
yt ) ? `t f (xt )
?f ? F .
(1)
t=1
Note that typical regret bounds apply to all f ? F and to all individual data sequences. However,
the bounds are allowed to scale with parameters arising from the interplay between f and the data
sequence.
In order to capture complex environments, the reference class of functions should be large. In this
work we focus on nonparametric classes F, containing all differentiable functions that are smooth
with respect to some metric on X . Our approach builds on the simple and versatile algorithm for
nonparametric online learning introduced in [6]. This algorithm has a bound on the regret RT (f ) of
order (ignoring logarithmic factors)
v
?
?
u d
uX
d
2
?1 + t
k?i f k ? T 1+d
?f ? F .
(2)
?
i=1
Here k?i f k? is the value of the partial derivative ?f (x) ?xi maximized over x ? X . The square
root term is the Lipschitz constant of f , measuring smoothness with respect to the Euclidean metric.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
However, in some directions f may be smoother than in others. Therefore, if we knew in advance the
set of directions along which the best performing reference functions f are smooth, we could use
this information to control regret better. In this paper we extend the algorithm from [6] and make it
adaptive to the Mahalanobis distance defined through an arbitrary positive definite matrix M with
d
spectrum (ui , ?i ) i=1 and unit spectral radius (?1 = 1). We prove a bound on the regret RT (f ) of
order (ignoring logarithmic factors)
v
?
?
u d
uX k?u f k2
?T
p
i
? ? 1+?T
? det? (M ) + t
?f ? F .
(3)
T
?i
i=1
Here ?T ? d is, roughly, the number of eigenvalues of M larger than a threshold shrinking
polynomially in T , and det? (M ) ? 1 is the determinant of M truncated at ?? (with ? ? ?T ). The
2
quantity k?ui f k? is defined like k?i f k? , but with the directional derivative ?f (x)> u instead of
the partial derivative. When the spectrum of M is light-tailed (so that ?T d and, simultaneously,
det? (M ) 1), with the smaller
eigenvalues ?i corresponding to eigenvectors in which f is smoother
2
(so that the ratios k?ui f k? ?i remain controlled), then our bound improves on (2). On the other
hand, when no preliminary knowledge about good f is available, we may run the algorithm with M
equal to the identity matrix and recover exactly the bound (2).
Given that the regret can be improved by informed choices of M , it is natural to ask whether some
kind of improvement is still possible when M is learned online, from the same data sequence on
which the regret is being measured. Of course, this question makes sense if the data tell us something
about the smoothness of the f against which we are measuring the regret. In the second part of the
paper we implement this idea by considering a scenario where instances are drawn i.i.d. from some
unknown distribution, labels are stochastically generated by some unknown regression function f0 ,
and we have no preliminary knowledge about the directions along which f0 is smoother.
In this stochastic scenario, the expected gradient outer product matrix G = E ?f0 (X)?f0 (X)>
2
provides a natural choice for the matrix M in our algorithm. Indeed, E ?f0 (X)> ui
= ?i
where u1 , . . . , ud are the eigenvectors of G while ?1 , . . . , ?d are the corresponding eigenvalues.
Thus, eigenvectors u1 , . . . ud capture the principal directions of variation for f . In fact, assuming
that the labels obey a statistical model Y = g(BX) + ? where ? is the noise and B ? Rk?d projects
X onto a k-dimensional subspace of X , one can show [21] that span(B) ? span(u1 , . . . , ud ). In
this sense, G is the ?best? metric, because it recovers the k-dimensional relevant subspace.
b of G.
When G is unknown, we run our algorithm in phases using a recently proposed estimator G
The estimator is trained on the same data sequence and is fed to the algorithm at the beginning of
each phase. Under mild assumptions on f0 , the noise in the labels, and the instance distribution, we
prove a high probability bound on the regret RT (f0 ) of order (ignoring logarithmic factors)
v
?
?
u d
?
eT
uX
?uj f0
+
?V f0
2
?
?
?1 + t
? T 1+?eT .
(4)
?j /?1
j=1
Observe that the rate at which the regret grows is the same as the one in (3), though now the effective
dimension parameter ?eT is larger than ?T by an amount related to the rate of convergence of the
b to those of G. The square root term is also similar to (3), but for the extra quantity
eigenvalues of G
k?V f0 k? , which accounts for the error in approximating the eigenvectors of G. More precisely,
k?V f0 k? is k?v f k? maximized over directions v in the span of V , where V contains those
eigenvectors of G that cannot be identified because their eigenvalues are too close to each other (we
come back to this issue shortly). Finally, we lose the dependence on the truncated determinant, which
is replaced here by its trivial upper bound 1.
The proof of (2) in [6] is based on the sequential construction of a sphere packing of X , where the
spheres are centered on adaptively chosen instances xt , and have radii shrinking polynomially with
time. Each sphere hosts an online learner, and each new incoming instance is predicted using the
learner hosted in the nearest sphere. Our variant of that algorithm uses an ellipsoid packing, and
computes distances using the Mahalanobis distance k?kM . The main new ingredient in the analysis
leading to (3) is our notion of effective dimension ?T (we call it the effective rank of M ), which
measures how fast the spectrum of M vanishes. The proof also uses an ellipsoid packing bound and
a lemma relating the Lipschitz constant to the Mahalanobis distance.
2
The proof of (4) is more intricate because G is only known up to a certain approximation. We use an
b recently proposed in [14], which is consistent under mild distributional assumptions
estimator G,
when f0 is continuously differentiable. The first source of difficulty is adjusting the notion of effective
rank (which the algorithm needs to compute) to compensate for the uncertainty in the knowledge of
the eigenvalues of G. A further problematic issue arises because we want to measure the smoothness
of f0 along the eigendirections of G, and so we need to control the convergence of the eigenvectors,
b converges to G in spectral norm. However, when two eigenvalues of G are close, then
given that G
b are strongly affected by the stochastic
the corresponding eigenvectors in the estimated matrix G
perturbation (a phenomenon known as hybridization or spectral leaking in matrix perturbation theory,
see [1, Section 2]). Hence, in our analysis we need to separate out the eigenvectors that correspond to
well spaced eigenvalues
from the
others. This lack of discrimination causes the appearance in the
regret of the extra term
?V f0
? .
2
Related works
Nonparametric estimation problems have been a long-standing topic of study in statistics, where one
is concerned with the recovery of an optimal function from a rich class under appropriate probabilistic
assumptions. In online learning, the nonparametric approach was investigated in [15, 16, 17] by
Vovk, who considered regression problems in large spaces and proved bounds on the regret. Minimax
rates for the regret were later derived in [13] using a non-constructive approach. The first explicit
online nonparametric algorithms for regression with minimax rates were obtained in [4].
The nonparametric online algorithm of [6] is known to have a suboptimal regret bound for Lipschitz
classes of functions. However, it is a simple and efficient algorithm, well suited to the design of
extensions that exhibit different forms of adaptivity to the data sequence. For example, the paper
[9] derived a variant that automatically adapts to the intrinsic dimension of the data manifold. Our
work explores an alternative direction of adaptivity, mainly aimed at taming the effect of the curse
of dimensionality in nonparametric prediction through the learning of an appropriate Mahalanobis
distance on the instance space. There is a rich literature on metric learning (see, e.g., the survey [2])
where the Mahalanobis metric k?kM is typically learned through minimization of the pairwise loss
function of the form `(M , x, x0 ). This loss is high whenever dissimilar pairs of x and x0 are close in
the Mahalanobis metric, and whenever similar ones are far apart in the same metric ?see, e.g., [19].
The works [5, 7, 18] analyzed generalization and consistency properties of online learning algorithms
employing pairwise losses.
In this work we are primarily interested in using a metric k?kM where M is close to the gradient
outer product matrix of the best model in the reference class of functions. As we are not aware
whether pairwise loss functions can indeed consistently recover such metrics, we directly estimate
the gradient outer product matrix. This approach to metric learning was mostly explored in statistics
?e.g., by locally-linear Empirical Risk Minimization on RKHS [12, 11], and through Stochastic
Gradient Descent [3]. Our learning approach combines ?in a phased manner? a Mahalanobis
metric extension of the algorithm by [6] with the estimator of [14]. Our work is also similar in spirit
to the ?gradient weights? approach of [8], which learns a distance based on a simpler diagonal matrix.
Preliminaries and notation. Let B(z, r) ? Rd be the ball of center z and radius r > 0 and let
B(r) = B(0, r). We assume instances x belong to X ? B(1) and labels y belong to Y ? [0, 1].
We consider the following online learning protocol with oblivious adversary. Given an unknown
sequence (x1 , y1 ), (x2 , y2 ), ? ? ? ? X ? Y of instances and labels, for every round t = 1, 2, . . .
1. the environment reveals instance xt ? X ;
2
2. the learner selects an action ybt ? Y and incurs the square loss `t ybt = ybt ? yt ;
3. the learner observes yt .
Given a positive definite
pd ? d matrix M , the norm kx ? zkM induced by M (a.k.a. Mahalanobis
distance) is defined by (x ? z)> M (x ? z).
Definition 1 (Covering and Packing Numbers). An ?-cover of a set S w.r.t. some metric ? is a set
{x01 , . . . , x0n } ? S such that for each x ? S there exists i ? {1, . . . , n} such that ?(x, x0i ) ? ?. The
covering number N (S, ?, ?) is the smallest cardinality of a ?-cover.
3
An ?-packing of a set S w.r.t. some metric ? is a set {x01 , . . . , x0m } ? S such that for any distinct
i, j ? {1, . . . , m}, we have ?(x0i , x0j ) > ?. The packing number M(S, ?, ?) is the largest cardinality
of a ?-packing.
It is well known that M(S, 2?, ?) ? N (S, ?, ?) ? M(S, ?, ?). For all differentiable f : X ? Y and
for any orthonormal basis V ? {u1 , . . . , uk } with k ? d we define
k?V f k? =
max
sup ?f (x)> v .
v ? span(V ) x?X
kvk = 1
If V = {u} we simply write k?u f k? .
In the following, M is a positive definite d?d matrix with eigenvalues ?1 ? ? ? ? ? ?d > 0 and eigenvectors u1 , . . . , ud . For each k = 1, . . . , d the truncated determinant is detk (M ) = ?1 ? ? ? ? ? ?k .
The kappa function for the matrix M is defined by
Figure 1: Quickly decreasing
n
o
2
spectrum of M implies slow
? 1+r
?(r, t) = max m : ?m ? t
, m = 1, . . . , d
(5)
growth of its effective rank in t.
for t ? 1 and r = 1, . . . , d.
Eigenvalues of M
1.0
?t = min {r : ?(r, t) ? r, r = 1, . . . , d} .
(6)
Since ?(d, t) ? d for all t ? 1, this is a well defined quantity.
Note that ?1 ? ?2 ? ? ? ? ? d. Also, ?t = d for all t ? 1
when M is the d ? d identity matrix. Note that the effective
rank ?t measures the number of eigenvalues that are larger than a
threshold that shrinks with t. Hence matrices M with extremely
light-tailed spectra cause ?t to remain small even when t grows
large. This behaviour is shown in Figure 1.
O
e
O
Throughout the paper, we use f = (g) and f = (g) to denote,
e
respectively, f = O(g) and f = O(g).
3
0.8
0.6
0.4
0.2
0.0
1
?t
Note that ?(r + 1, t) ? ?(r, t). Now define the effective rank of
M at horizon t by
10
9
8
7
6
5
4
3
2
1
0
2
3 4 5 6 7 8
Effective Rank of M
2000
4000
t
6000
8000
9
10
10000
Online nonparametric learning with ellipsoid packing
In this section we present a variant (Algorithm 1) of the online nonparametric regression algorithm
introduced in [6]. Since our analysis is invariant to rescalings of the matrix M , without loss of
generality we assume M has unit spectral radius (i.e., ?1 = 1). Algorithm 1 sequentially constructs
a packing of X using M -ellipsoids centered on a subset of the past observed instances. At each step
t, the label of the current instance xt is predicted using the average ybt of the labels of past instances
that fell inside the ellipsoid whose center xs is closest to xt in the Mahalanobis metric. At the end
of the step, if xt was outside of the closest ellipsoid, then a new ellipsoid is created with center xt .
The radii ?t of all ellipsoids are shrunk at rate t?1/(1+?t ) . Note that efficient (i.e., logarithmic in the
number of centers) implementations of approximate nearest-neighbor search for the active center xs
exist [10].
The core idea of the proof (deferred to the supplementary material) is to maintain a trade-off between
the regret contribution of the ellipsoids and an additional regret term due to the approximation of f
by the Voronoi partitioning. The regret contribution of each ellipsoid is logarithmic in the number of
predictions made. Since each instance is predicted by a single ellipsoid, if we ignore log factors the
overall regret contribution is equal to the number of ellipsoids, which is essentially controlled by the
packing number w.r.t. the metric defined by M . The second regret term is due to the fact that ?at
any point of time? the prediction of the algorithm is constant within the Voronoi cells of X induced
by the current centers (recall that we predict with nearest neighbor). Hence, we pay an extra term
equal to the radius of the ellipsoids times the Lipschitz constant which depends on the directional
Lipschitzness of f with respect to the eigenbasis of M .
Theorem 1 (Regret with Fixed Metric). Suppose Algorithm 1 is run with a positive definite matrix
M with eigenbasis u1 , . . . , ud and eigenvalues 1 = ?1 ? ? ? ? ? ?d > 0. Then, for any differentiable
4
Algorithm 1 Nonparametric online regression
Input: Positive definite d ? d matrix M .
1: S ? ?
2: for t = 1, 2, . . . do
1
3:
?t ? t? 1+?t
4:
Observe xt
5:
if S ? ? then
6:
S ? {t}, Tt ? ?
7:
end if
8:
s ? arg min kxt ? xs kM
. Centers
. Update radius
. Create initial ball
. Find active center
s?S
9:
10:
11:
12:
13:
14:
15:
16:
17:
18:
19:
20:
if Ts ? ? then
yt = 21
else
1 X
ybt ?
yt0
|Ts | 0
t ?Ts
end if
Observe yt
if kxt ? xs kM ? ?t then
Ts ? Ts ? {t}
else
S ? S ? {s}, Ts ? ?
end if
end for
. Predict using active center
. Update list for active center
. Create new center
f : X ? Y we have that
v
?
?
u d
uX k?u f k2
?T
p
e
O
i
? ? 1+?T
T
RT (f ) = ? det? (M ) + t
?i
i=1
where ? = ?(?T , T ) ? ?T ? d.
We first prove two technical lemmas about packings of ellipsoids.
0
Lemma 1 (Volumetric packing bound). Consider a pair of norms k ? k , k ? k and let B, B 0 ? Rd be
the corresponding unit balls. Then
vol B + 2? B 0
0
M(B, ?, k ? k ) ?
.
vol 2? B 0
Lemma 2 (Ellipsoid packing bound). If B is the unit Euclidean ball then
? !s s
n
o
Yp
p
8 2
?i
where
s = max i : ?i ? ?, i = 1, . . . , d .
M B, ?, k ? kM ?
?
i=1
The following lemma states that whenever f has bounded partial derivatives with respect to the
eigenbase of M , then f is Lipschitz with respect to k ? kM .
Lemma 3 (Bounded derivatives imply Lipschitzness in M -metric). Let f : X ? R be everywhere
differentiable. Then for any x, x0 ? X ,
v
u d
uX k?u f k2
i
?
f (x) ? f (x0 ) ? kx ? x0 k t
.
M
?
i
i=1
4
Learning while learning the metric
In this section, we assume instances xt are realizations of i.i.d. random variables X t drawn according
to some fixed and unknown distribution ? which has a continuous density on its support X . We also
5
assume labels yt are generated according to the noise model yt = f0 (xt ) + ?(xt ), where f0 is some
unknown regression function and ?(x) is a subgaussian zero-mean random variable for all x ? X .
We then simply write RT to denote the regret RT (f0 ). Note that RT is now a random variable which
we bound with high probability.
We now show how the nonparametric online learning algorithm (Algorithm 1) of Section 3 can be
combined with an algorithm that learns an estimate
n
X
bn = 1
b 0 (xt )?f
b 0 (xt )>
G
?f
(7)
n t=1
of the expected outer product gradient matrix G = E ?f0 (X)?f0 (X)> . The algorithm (described
in the supplementary material) is consistent under the following assumptions. Let X (? ) be X blown
up by a factor of 1 + ? .
Assumption 1.
1. There exists ?0 > 0 such that f0 is continuously differentiable on X (?0 ).
2. There exists G > 0 such that max k?f0 (x)k ? G.
x?X (?0 )
3. The distribution ?
is full-dimensional: there exists C? > 0 such that for all x ? X and
? > 0, ? B(x, ?) ? C? ?d .
b n is a consistent estimate of G.
In particular, the next lemma states that, under Assumption 1, G
Lemma 4 ([14, Theorem 1]). If Assumption 1 holds, then there exists a nonnegative and nonincreasing sequence {?n }n?1 such that for all n, the estimated gradient outerproduct (7) computed
b n ? G
? ?n with high probability with
with parameters ?n > 0, and 0 < ?n < ?0 satisfies
G
2
2
1
1/4
respect do the random draw of X 1 , . . . , X n . Moreover, if ?n = ? ?n , ?n = ? ln n d n? d ,
1
and ?n = O n? 2(d+1) then ?n ? 0 as n ? ?.
Our algorithm works in phases i = 1, 2, . . . where phase i has length n(i) = 2i . Let T (i) = 2i+1 ? 2
be the index of the last time step in phase i. The algorithm uses a nonincreasing regularization
c (0) = ? 0 I. During each phase i, the algorithm predicts the
sequence ? 0 ? ? 1 ? ? ? ? > 0. Let M
c (i ? 1)
M
c (i ? 1)k2 (where k ? k2 denotes the
data points by running Algorithm 1 with M = M
spectral norm). Simultaneously, the gradient outer product estimate (7) is trained over the same data
b
b T (i) is used
points. At the end of phase i, the current gradient outer product estimate G(i)
=G
c (i) = G(i)
b + ? T (i) I. Algorithm 1 is then restarted in phase i + 1 with
to form a new matrix M
c (i)
M
c (i)k2 . Note that the metric learning algorithm can be also implemented efficiently
M =M
through nearest-neighbor search as explained in [14].
Let ?1 ? ?2 ? ? ? ? ? ?d be the eigenvalues and u1 , . . . , ud be the eigenvectors of G. We define the
j-th eigenvalue separation ?j by
?j = min ?j ? ?k .
k6=j
For any ? > 0 define also V? ? uj : |?j ? ?k | ? ?, k 6= j and V?? = {u1 , . . . , ud } \ V? .
Our results are expressed in terms of the effective rank (6) of G at horizon T . However, in order to
account for the error in estimating the eigenvalues of G, we define the effective rank ?et with respect
to the following slight variant of the function kappa,
n
o
2
?
e(r, t) = max m : ?m + 2? t ? ?1 t? 1+r , m = 1, . . . , d
t ? 1 and r = 1, . . . , d.
c (i) be the estimated gradient outer product constructed at the end of phase i, and let ?
Let M
b1 (i) +
c (i), where
b 1 (i), . . . , u
b d (i) be the eigenvalues and eigenvectors of M
?(i) ? ? ? ? ? ?
bd (i) + ?(i) and u
we also write ?(i) to denote ? T (i) . We use ?
b to denote the kappa function with estimated eigenvalues
and ?b to denote the effective rank defined through ?
b. We start with a technical lemma.
?? 2
Lemma 5. Let ?d , ? > 0 and d ? 1. Then the derivative of F (t) = ?d + 2 T0 + t
t 1+d is
1/?
.
positive for all t ? 1 when T0 ? d+1
2?d
6
2(T0 +t)
?(d+1)
Proof. We have that F 0 (t) ? 0 if and only if t ?
t?
2?d (T0 + t)1+?
?(d + 1)
1 + (T0 + t)? ?d . This is implied by
or, equivalently, T0 ? A1/(1+?) t1/(1+?) ? t
where A = ?(d + 1)/(2?d ). The right-hand side A1/(1+?) t1/(1+?) ? t is a concave function of t.
Hence the maximum is found at the value of t where the derivative is zero, this value satisfies
A1/(1+?) ??/(1+?)
t
= 1 which solved for t gives t = A1/? (1 + ?)?(1+?)/? .
1+?
Substituting this value of t in A1/(1+?) t1/(1+?) ? t gives the condition T0 ? A1/? ?(1 + ?)?(1+?)/?
1/?
.
which is satisfied when T0 ? d+1
2?d
Theorem 2. Suppose Assumption 1 holds. If the algorithm is ran with a regularization sequence
1/?
? 0 = 1 and ? t = t?? for some ? > 0 such that ? t ? ?t for all t ? d + 1 2?d
and for
?1 ? ?2 ? ? ? ? > 0 satisfying Lemma 4, then for any given ? > 0
v
?
?
u d
?uj f0
+
?V ? f0
2
X
u
?
eT
e
O
?
?
?
? T 1+?eT
RT = ?1 + t
?j /?1
j=1
with high probability with respect to the random draw of X 1 , . . . , X T .
Note
that the asymptotic
notation is hiding terms that depend on 1/?, hence we can not zero out the
term
?V?? f0
? in the bound by taking ? arbitrarily small.
Proof. Pick the smallest i0 such that
T (i0 ) ?
d+1
2?d
1/?
(8)
(we need this condition in the proof). The total regret in phases 1, 2, . . . , i0 is bounded by
1/?
d + 1 2?d
= O(1). Let the value ?bT (i) at the end of phase i be denoted by ?b(i). By Theorem 1,
the regret RT (i + 1) of Algorithm 1 in each phase i + 1 > i0 is deterministically upper bounded by
v
?
2 ?
u d
uX
?u
?
?(i+1)
b
f
0
b
(i)
?
b
(i+1)
j
? ? (i+1) 1+?(i+1)
b
2
RT (i + 1) ? ?8 ln e2i+1 8 2
+ 4t
(9)
?
(i)
?
(i)
j
1
j=1
c (i)
M
c (i)k2 ? 1 for
where ?j (i) = ?
bj (i) + ?(i). Here we used the trivial upper bound det? M
2
all ? = 1, . . . , d. Now assume that ?
b1 (i) + ?(i) ? ?
bm (i) + ?(i) t 1+r for some m, r ? {1, . . . , d}
and for some t in phase i + 1. Hence, using Lemma 4 and ?t ? ? t , we have that
b ? G
? ?(i) with high probability.
max ?
bj (i) ? ?j ?
G(i)
(10)
2
j=1,...,d
where the first inequality is straightforward. Hence we may write
2
?1 ? ?1 ? ?(i) + ?(i) ? ?
b1 (i) + ?(i) ? ?
bm (i) + ?(i) t 1+r
2
? ?m + ?(i) + ?(i) t 1+r
2
? ?m + 2?(i) t 1+r .
Recall ?(i) = T (i)?? . Using Lemma 5, we observe that the derivative of
?? 2
F (t) = ?m + 2 T (i) + t
t 1+r
is positive for all t ? 1 when
T (i) ?
r+1
2?d
1/?
?
7
r+1
2?m
1/?
(using Lemma 4)
2
2
which is guaranteed by our choice (8). Hence, ?m + 2?(i) t 1+r ? ?m + 2? T ) T 1+r and so
2
?
b1 (i) + ?(i)
? t 1+r
?
bm (i) + ?(i)
implies
2
?1
? T 1+r .
?m + 2? T
Recalling the definitions of ?
e and ?
b, this in turn implies ?
b(r, t) ? ?
e(r, T ), which also gives ?bt ? ?eT
for all t ? T . Next, we bound the approximation error in each individual eigenvalue of G. By (10)
we obtain, for any phase i and for any j = 1, . . . , d,
?j + 2?(i) ? ?j + ?(i) + ?(i) ? ?
bj (i) + ?(i) ? ?j ? ?(i) + ?(i) ? ?j .
Hence, bound (9) implies
?
RT (i + 1) ? ?8 ln e2i+1 12?eT
v
2 ?
u
d
X
u
?
eT
?
f
0
b
u
j
? ? (i+1) 1+?eT
2
.
+ 4t ?1 + 2?(i)
?j
j=1
(11)
The error in approximating the eigenvectors of G is controlled via the following first-order eigenvector
approximation result from matrix perturbation theory [20, equation (10.2)], for any vector v of
constant norm,
c
X u>
k M (i) ? G uj >
>
c (i) ? G
2
b j (i) ? uj =
v u
v uk + o
M
2
?j ? ?k
k6=j
X 2?(i)
?
v > uk + o ?(i)2
?j ? ?k
(12)
k6=j
c
c
where we used u>
k M (i) ? G uj ? M (i) ? G 2 ? ?(i) + ?(i) ? 2?(i). Then for all j such
that uj ? V? ,
X 2?(i)
b j (i) ? uj =
?f0 (x)> u
?f0 (x)> uk + o ?(i)2
?j ? ?k
k6=j
?
2?(i) ?
d k?f0 (x)k2 + o ?(i)2 .
?
Note that the coefficients
c
u>
k M (i) ? G uj
?k =
+ o ?(i)2
?j ? ?k
k 6= j
b j (i) ? uj w.r.t. the orthonormal basis u1 , . . . , ud . Then,
are a subset of coordinate values of vector u
by Parseval?s identity,
X
2
4 ? kb
uj (i) ? uj k2 ?
?k2 .
k6=j
Therefore, it must be that
u> M
k c (i) ? G uj
max
? 2 + o ?(i)2 .
k6=j
?j ? ?k
For any j such that uj ? V?? , since ?j ? ?k ? ? for all uk ? V? , we may write
b j (i) ? uj
?f0 (x)> u
X
2?(i) X
?f0 (x)> uk + o ?(i)2
?
?f0 (x)> uk + 2 + o ?(i)2
?
?
uk ?V?
?
uk ?V?
?
2?(i) ?
d kP V? ?f0 (x)k2 + 2 + o ?(i)2
d
P V?? ?f0 (x)
2 + o ?(i)2
?
8
where P V? and P V?? are the orthogonal projections onto, respectively, V? and V?? . Therefore, we
have that
>
?u
b j (i) = sup ?f0 (x)> u
b j (i) ? uj + uj
b j f0 ? = sup ?f0 (x) u
x?X
x?X
b j (i) ? uj
? sup ?f0 (x)> uj + sup ?f0 (x)> u
x?X
x?X
?
2?(i) ?
?
?uj f0
? +
d k?V? f0 k? + 2 + o ?(i)2
d
?V?? f0
? + o ?(i)2
(13)
?
?
?
Letting ?? (i) = 2?(i)
d k?V? f0 k? + 2 + o ?(i)2
d
?V?? f0
? + o ?(i)2 we can upper
?
bound (11) as follows
v
?
?
u
d
?u f0
+ ?? (i) 2
X
?
eT
u
j
?
? 2(i+1) 1+?eT .
RT (i + 1) ? ?8 ln e2i+1 12?eT + 4t ?1 + 2?(i)
?j
j=1
Recall that, due to (10), the above holds at the end of each phase i + 1 with high probability. Now
O
observe that ?(i) = O 2??i and so ?? (i) = (k?V? f0 k? /?+
?V?? f0
? ). Hence, by summing
over phases i = 1, . . . , log2 T and applying the union bound,
dlog2 T e
RT =
X
RT (i)
i=1
v
?
u
d
?uj f0
+ ?? (i ? 1) 2 ?eT i
X
u
?
? 2 1+?eT
? ?8 ln eT 12d+ 4t ?1 + 2?(i ? 1)
?j
j=1
?
2 ?
d
?
f
+
?
f
?
X
uj 0 ?
V? 0 ?
e ?
O
? 1+?eT?eT
= ?1 +
?T
?j ?1
j=1
?
concluding the proof.
5
Conclusions and future work
We presented an efficient algorithm for online nonparametric regression which adapts to the directions
along which the regression function f0 is smoother. It does so by learning the Mahalanobis metric
through the estimation of the gradient outer product matrix E[?f0 (X)?f0 (X)> ]. As a preliminary
result, we analyzed the regret of a generalized version of the algorithm from [6], capturing situations
where one competes against functions with directional Lipschitzness with respect to an arbitrary
Mahalanobis metric. Our main result is then obtained through a phased algorithm that estimates the
gradient outer product matrix while running online nonparametric regression on the same sequence.
Both algorithms automatically adapt to the effective rank of the metric.
This work could be extended by investigating a variant of Algorithm 1 for classification, in which ball
radii shrink at a nonuniform rate, depending on the mistakes accumulated within each ball rather than
on time. This could lead to the ability of competing against functions f that are only locally Lipschitz.
In addition, it is conceivable that under appropriate assumptions, a fraction of the balls could stop
shrinking at a certain point when no more mistakes are made. This might yield better asymptotic
bounds than those implied by Theorem 1, because ?T would never attain the ambient dimension d.
Acknowledgments
Authors would like to thank S?ebastien Gerchinovitz and Samory Kpotufe for useful discussions on
this work. IK would like to thank Google for travel support. This work also was in parts funded
by the European Research Council (ERC) under the European Union?s Horizon 2020 research and
innovation programme (grant agreement no 637076).
9
References
[1] R. Allez and J.-P. Bouchaud. Eigenvector dynamics: general theory and some applications.
Physical Review E, 86(4):046202, 2012.
[2] A. Bellet, A. Habrard, and M. Sebban. A Survey on Metric Learning for Feature Vectors and
Structured Data. arXiv preprint arXiv:1306.6709, 2013.
[3] X. Dong and D.-X. Zhou. Learning Gradients by a Gradient Descent Algorithm. Journal of
Mathematical Analysis and Applications, 341(2):1018?1027, 2008.
[4] P. Gaillard and S. Gerchinovitz. A chaining Algorithm for Online Nonparametric Regression.
In Conference on Learning Theory (COLT), 2015.
[5] Z.-C. Guo, Y. Ying, and D.-X. Zhou. Online Regularized Learning with Pairwise Loss Functions.
Advances in Computational Mathematics, 43(1):127?150, 2017.
[6] E. Hazan and N. Megiddo. Online Learning with Prior Knowledge. In Learning Theory, pages
499?513. Springer, 2007.
[7] R. Jin, S. Wang, and Y. Zhou. Regularized Distance Metric Learning: Theory and Algorithm.
In Conference on Neural Information Processing Systems (NIPS), 2009.
[8] S. Kpotufe, A. Boularias, T. Schultz, and K. Kim. Gradients Weights Improve Regression and
Classification. Journal of Machine Learning Research, 17(22):1?34, 2016.
[9] S. Kpotufe and F. Orabona. Regression-Tree Tuning in a Streaming Setting. In Conference on
Neural Information Processing Systems (NIPS), 2013.
[10] R. Krauthgamer and J. R. Lee. Navigating nets: simple algorithms for proximity search. In
Proceedings of the 15th annual ACM-SIAM Symposium on Discrete algorithms, pages 798?807.
Society for Industrial and Applied Mathematics, 2004.
[11] S. Mukherjee and Q. Wu. Estimation of Gradients and Coordinate Covariation in Classification.
Journal of Machine Learning Research, 7(Nov):2481?2514, 2006.
[12] S. Mukherjee and D.-X. Zhou. Learning Coordinate Covariances via Gradients. Journal of
Machine Learning Research, 7(Mar):519?549, 2006.
[13] A. Rakhlin and K. Sridharan. Online Non-Parametric Regression. In Conference on Learning
Theory (COLT), 2014.
[14] S. Trivedi, J. Wang, S. Kpotufe, and G. Shakhnarovich. A consistent Estimator of the Expected
Gradient Outerproduct. In Conference on Uncertainty in Artificial Intelligence (UAI), 2014.
[15] V. Vovk. Metric entropy in competitive on-line prediction. arXiv preprint cs/0609045, 2006.
[16] V. Vovk. On-line regression competitive with reproducing kernel Hilbert spaces. In International
Conference on Theory and Applications of Models of Computation. Springer, 2006.
[17] V. Vovk. Competing with wild prediction rules. Machine Learning, 69(2):193?212, 2007.
[18] Y. Wang, R. Khardon, D. Pechyony, and R. Jones. Generalization Bounds for Online Learning
Algorithms with Pairwise Loss Functions. In Conference on Learning Theory (COLT), 2012.
[19] K. Q. Weinberger and L. K. Saul. Distance Metric Learning for Large Margin Nearest Neighbor
Classification. Journal of Machine Learning Research, 10:207?244, 2009.
[20] J. H. Wilkinson. The Algebraic Eigenvalue Problem, volume 87. Clarendon Press Oxford, 1965.
[21] Q. Wu, J. Guinney, M. Maggioni, and S. Mukherjee. Learning gradients: predictive models that infer geometry and statistical dependence. Journal of Machine Learning Research,
11(Aug):2175?2198, 2010.
10
| 6669 |@word mild:2 determinant:3 version:1 norm:5 km:7 bn:1 covariance:1 pick:1 incurs:2 versatile:1 initial:1 contains:1 rkhs:1 past:2 current:3 com:1 gmail:1 bd:1 must:1 subsequent:1 gerchinovitz:2 update:2 discrimination:1 intelligence:1 beginning:1 core:1 provides:1 simpler:1 outerproduct:2 mathematical:1 along:5 constructed:1 ilja:2 symposium:1 ik:1 prove:3 combine:1 wild:1 inside:1 manner:1 x0:5 pairwise:5 intricate:1 indeed:2 expected:3 roughly:1 decreasing:1 automatically:3 curse:1 considering:1 cardinality:2 hiding:1 project:1 estimating:1 notation:2 bounded:4 moreover:1 competes:1 kind:1 eigenvector:2 informed:1 lipschitzness:4 every:1 concave:1 growth:1 megiddo:2 exactly:1 universit:1 k2:11 x0m:1 control:2 unit:4 uk:9 partitioning:1 grant:1 positive:7 t1:3 mistake:2 oxford:1 might:1 phased:2 acknowledgment:1 union:2 regret:25 definite:5 implement:1 ybt:6 empirical:1 adapting:1 attain:1 projection:1 onto:2 cannot:1 close:4 risk:1 applying:1 yt:12 center:11 straightforward:1 survey:2 recovery:1 estimator:5 rule:1 orthonormal:2 his:1 notion:2 variation:1 coordinate:3 e2i:3 maggioni:1 construction:1 suppose:2 us:3 agreement:1 satisfying:1 kappa:3 mukherjee:3 distributional:1 predicts:1 observed:1 preprint:2 solved:1 capture:2 wang:3 trade:1 observes:2 ran:1 environment:3 vanishes:1 ui:4 pd:1 wilkinson:1 leaking:1 dynamic:1 trained:2 depend:1 shakhnarovich:1 predictive:1 learner:9 basis:2 packing:13 distinct:1 fast:1 effective:13 kp:1 artificial:1 tell:1 outside:1 whose:2 larger:3 supplementary:2 ability:1 statistic:2 online:24 interplay:1 sequence:13 differentiable:6 eigenvalue:19 kxt:2 net:1 product:10 relevant:1 realization:1 adapts:2 eigenbasis:2 convergence:2 converges:1 zkm:1 depending:1 measured:3 x0i:2 nearest:5 aug:1 implemented:1 predicted:3 c:1 come:1 implies:4 switzerland:1 direction:8 radius:8 stochastic:3 shrunk:1 centered:2 milano:2 kb:1 material:2 behaviour:1 generalization:2 preliminary:5 dipartimento:1 extension:2 hold:3 proximity:1 considered:1 predict:2 bj:3 substituting:1 smallest:2 estimation:3 travel:1 lose:1 label:11 council:1 gaillard:1 largest:1 create:2 minimization:2 rather:1 zhou:4 derived:2 focus:1 improvement:1 consistently:1 rank:11 mainly:1 industrial:1 kim:1 sense:2 voronoi:2 epfl:1 i0:4 typically:2 bt:2 accumulated:1 streaming:1 interested:1 selects:1 issue:2 overall:1 arg:1 classification:4 denoted:1 k6:6 colt:3 equal:3 aware:1 construct:1 never:1 beach:1 jones:1 future:1 others:2 primarily:1 oblivious:1 simultaneously:3 individual:2 replaced:1 phase:16 geometry:1 maintain:1 recalling:1 deferred:1 detk:1 analyzed:2 kvk:1 light:2 nonincreasing:2 ambient:1 partial:3 orthogonal:1 tree:1 euclidean:2 instance:16 cover:2 measuring:2 subset:2 habrard:1 too:1 combined:1 adaptively:1 st:1 density:1 explores:1 siam:1 international:1 standing:1 lee:1 probabilistic:1 off:1 dong:1 continuously:2 quickly:1 cesa:2 satisfied:1 containing:1 boularias:1 stochastically:1 derivative:8 leading:1 bx:1 yp:1 account:2 coefficient:1 depends:1 later:1 root:2 hazan:2 sup:5 start:1 recover:2 competitive:2 contribution:3 square:4 who:1 efficiently:1 maximized:2 correspond:1 spaced:1 yield:1 directional:3 pechyony:1 whenever:3 definition:2 volumetric:1 against:4 associated:1 di:2 recovers:1 proof:8 stop:1 proved:1 adjusting:1 ask:1 covariation:1 recall:3 knowledge:4 improves:1 dimensionality:1 hilbert:1 back:1 clarendon:1 improved:1 though:1 strongly:1 shrink:2 generality:1 mar:1 hand:2 lack:1 google:1 grows:2 bouchaud:1 usa:1 effect:1 y2:1 hence:10 regularization:2 mahalanobis:13 round:5 during:1 covering:2 chaining:1 generalized:1 tt:1 recently:2 sebban:1 physical:1 volume:1 extend:2 he:1 belong:2 relating:1 slight:1 smoothness:3 rd:3 tuning:1 consistency:1 mathematics:2 erc:1 funded:1 f0:49 nicolo:1 something:1 closest:2 italy:1 apart:1 scenario:2 certain:2 inequality:1 arbitrarily:1 additional:1 ud:8 smoother:5 full:1 infer:1 exceeds:1 smooth:2 technical:2 adapt:1 long:2 sphere:4 compensate:1 host:1 a1:6 controlled:3 prediction:6 variant:5 regression:18 essentially:1 metric:29 arxiv:3 kernel:1 cell:1 addition:1 want:1 else:2 source:1 rescalings:1 extra:3 fell:1 induced:2 hybridization:1 spirit:1 sridharan:1 call:1 subgaussian:1 concerned:1 identified:1 suboptimal:1 competing:2 reduce:1 idea:2 det:5 t0:8 whether:2 algebraic:1 cause:2 action:1 useful:1 eigenvectors:12 aimed:1 amount:2 nonparametric:17 locally:2 informatica:1 exist:1 problematic:1 blown:1 estimated:4 arising:1 write:5 discrete:1 vol:2 affected:1 threshold:2 kuzborskij:2 drawn:2 fraction:1 compete:1 run:3 everywhere:1 uncertainty:2 eigendirections:1 throughout:1 x0n:1 x0j:1 wu:2 separation:1 draw:2 capturing:1 bound:22 pay:1 guaranteed:1 nonnegative:1 annual:1 precisely:1 x2:1 u1:9 span:4 min:3 extremely:1 performing:1 concluding:1 structured:1 according:2 ball:7 smaller:1 remain:2 bellet:1 explained:1 invariant:1 ln:5 equation:1 turn:1 letting:1 fed:1 end:10 available:1 apply:1 obey:1 observe:5 spectral:5 appropriate:3 alternative:1 weinberger:1 shortly:1 denotes:1 running:2 krauthgamer:1 log2:1 build:1 uj:22 approximating:2 society:1 implied:2 question:1 quantity:3 parametric:1 rt:15 dependence:2 diagonal:1 exhibit:1 gradient:20 conceivable:1 subspace:2 distance:9 separate:1 thank:2 navigating:1 outer:10 topic:1 manifold:1 trivial:2 studi:1 assuming:1 length:1 index:1 ellipsoid:15 ratio:1 innovation:1 equivalently:1 ying:1 mostly:1 design:1 implementation:1 ebastien:1 unknown:7 kpotufe:4 bianchi:2 upper:4 enabling:1 descent:2 t:6 jin:1 truncated:3 situation:1 extended:1 y1:1 interacting:1 perturbation:3 nonuniform:1 reproducing:1 arbitrary:4 introduced:2 pair:2 learned:2 nip:3 adversary:1 max:7 natural:2 difficulty:1 regularized:2 minimax:2 improve:1 imply:1 created:1 taming:1 review:1 literature:1 prior:1 nicol:1 asymptotic:2 loss:13 parseval:1 adaptivity:2 ingredient:1 x01:2 agent:1 consistent:4 course:1 yt0:1 last:1 side:1 neighbor:4 saul:1 taking:1 dimension:4 cumulative:2 rich:2 computes:1 author:1 made:2 adaptive:1 bm:3 programme:1 far:1 polynomially:2 employing:1 schultz:1 approximate:1 nov:1 ignore:1 dlog2:1 sequentially:1 incoming:1 reveals:1 active:4 b1:4 summing:1 investigating:1 uai:1 knew:1 xi:1 degli:1 spectrum:6 search:3 continuous:1 tailed:2 learn:1 ca:1 ignoring:3 investigated:1 complex:1 european:2 protocol:1 main:2 bounding:1 noise:3 allowed:1 x1:1 hosted:1 slow:1 samory:1 shrinking:3 explicit:1 deterministically:1 khardon:1 learns:3 rk:1 theorem:5 xt:14 explored:1 x:4 list:1 rakhlin:1 intrinsic:1 exists:5 sequential:1 kx:2 horizon:3 trivedi:1 margin:1 suited:1 entropy:1 logarithmic:5 simply:2 appearance:1 expressed:1 ux:6 restarted:1 springer:2 satisfies:2 acm:1 identity:3 orabona:1 lipschitz:6 typical:1 unimi:1 vovk:4 principal:1 lemma:14 total:1 support:2 guo:1 arises:1 dissimilar:1 constructive:1 phenomenon:1 |
6,267 | 667 | Deriving Receptive Fields Using An
Optimal Encoding Criterion
Ralph Linsker
IBM T. J. Watson Research Center
P. O. Box 218, Yorktown Heights, NY 10598
Abstract
An information-theoretic optimization principle ('infomax') has
previously been used for unsupervised learning of statistical regularities in an input ensemble. The principle states that the inputoutput mapping implemented by a processing stage should be chosen so as to maximize the average mutual information between
input and output patterns, subject to constraints and in the presence of processing noise. In the present work I show how infomax,
when applied to a class of nonlinear input-output mappings, can
under certain conditions generate optimal filters that have additional useful properties: (1) Output activity (for each input pattern) tends to be concentrated among a relatively small number
of nodes. (2) The filters are sensitive to higher-order statistical
structure (beyond pairwise correlations). If the input features are
localized, the filters' receptive fields tend to be localized as well.
(3) Multiresolution sets of filters with subsampling at low spatial
frequencies - related to pyramid coding and wavelet representations
- emerge as favored solutions for certain types of input ensembles.
1
INTRODUCTION
In unsupervised network learning, the development of the connection weights is
influenced by statistical properties of the ensemble of input vectors, rather than by
the degree of mismatch between the network's output and some 'desired' output.
An implicit goal of such learning is that the network should transform the input
so that salient features present in the input are represented at the output in a
953
954
Linsker
more useful form. This is often done by reducing the input dimensionality in a way
that preserves the high-variance components of the input (e.g., principal component
analysis, Kohonen feature maps).
The principle of maximum information preservation ('infomax') is an unsupervised
learning strategy that states (Linsker 1988): From a set of allowed input-output
mappings (e.g., parametrized by the connection weights), choose a mapping that
maximizes the (ensemble-averaged) Shannon information that the output vector
conveys about the input vector, in the presence of noise. Such a mapping maximizes
the ensemble-averaged mutual information (MI) between input and output.
This paper (a) summarize earlier results on info max solutions for linear networks,
(b) identifies some limitations of these solutions (ways in which very different filter
sets are equally optimal from the infomax standpoint), and (c) shows how, by adding
a small nonlinearity to the network, one can remove these limitations and at the
same time improve the utility of the output representations. We show that infomax,
acting on the modified network, tends to favor sparsely coded representations and
(depending on the input ensemble) sets of filters that span multiple resolution scales
(related to wavelets and 'pyramid coding').
2
INFOMAX IN LINEAR NETWORKS
For definiteness and brevity, we consider a linear network having a particular type
of noise model and input statistical properties. For a more detailed discussion of
related models see (Linsker 1989).
Since the computation of the MI (which involves the output entropy) is in general
intractable for continuous-valued output vectors, previous work (and the present
paper) makes use of a surrogate MI, which we will call the 'as-if-Gaussian' MI. This
quantity is, by definition, computed as though the output vectors comprised a multivariate Gaussian distribution having the same mean and covariance as the actual
distribution of output vectors. Although expedient, this substitution has lacked a
principled justification. The Appendix shows that, under certain conditions, using
this 'surrogate MI' (and not the full MI) is indeed appropriate and justified.
=
Denote the input vector by S {Si} (Si is the activity at input node i), the output
vector by Z
{Zn}, the matrix of connection weights by C {Cni}, noise at the
input nodes by N
{Nd, and noise at the output nodes by v {vn }. Then our
processing model is, in matrix form, Z = C(S + N) + v. Assume that N and v are
Gaussian random variables, (S)
(N)
(v) 0, \fN T ) (Sv T ) (NvT) 0,
and, for the covariance matrices, (SsT)
Q, (N N ) TJI, (vv T ) {3I'. (Angle
brackets denote an ensemble average, superscript T denotes transpose, and I and
I' denote unit matrices on the input and output spaces, respectively.) In general,
MI = Hz - (Hzls) where Hz is the output entropy and HZls is the entropy of the
output for given S. Replacing MI by the 'as-if-Gaussian' MI means replacing Hz
by the expression for the entropy of a multivariate Gaussian distribution, which is
(apart from an irrelevant constant term) H~ = (1/2) lndet Q', where Q' (ZZT)
CQC T + TJCC T + (3I' is the output covariance. Note that, when S is fixed, Z =
CS+(CN +v) is a Gaussian distribution centered on CS, so that we have (Hzls) =
(1/2)lndetQII where Q" = ((CN + v)(CN + v)T) = TJCC T + {3I'. Therefore the
=
=
=
=
= =
=
=
=
=
=
=
=
=
=
Deriving Receptive Fields Using An Optimal Encoding Criterion
'as-if-Gaussian' MI is
(1)
MI' = (1/2)[lndetQ' -lndetQ"].
The variance of the output at node n (prior to adding noise lin) is Vn = ([C(S +
N)]~) = (CQC T + TJCCT)nn. We will constrain the dynamic range of each output
node (limiting the number of output values that can be discriminated from one
another in the presence of output noise) by requiring that Vn = 1 for each n.
Subject to this constra.int, we are to find a matrix C that maximizes MI'. For a
local Hebbian algorithm that accomplishes this maximization, see (Linsker 1992).
Here, in order to proceed analytically, we consider a special case of interest.
Suppose that the input statistics are shift-invariant, so that the covariance (SiSj)
is a function of (j - i). We then use a shift-invariant filter Ansatz, Cni C(i - n).
Infomax then determines the optimal filter gain as a function of spatial frequency;
i.e., the magnitude of the Fourier components c(k) of C(i - n). The derivation is
summarized below.
=
Denote by q(k), q'(k), and q"(k) the Fourier transforms of QU - i), Q'(m - n),
and Q"(m - n) respectively. Since Q' = CQCT + TJccf1' + (3I', therefore
q'(k) = [q(k) + TJ) I c(k) 12 +(3. Similarly, q"(k) = TJ I c(k) 12 +(3. We obtain
MI' = (1/2)~dlnq'(k) - Inq"(k)]. Each node's output variance Vn is equal to
V = (I/K)~dq(k)+TJ] I c(k) 12 where K is the number of terms in the sum over k.
To maximize MI' subject to the constraint on V we use the Lagrange multiplier
method; that is, we maximize MI" MI' + J.L(V - 1) with respect to each I c(k) 12.
!his yields an equation for each k that is quadratic in I c(k) 12. The unique solution
=
IS
2_
(TJ/{3) I c(k) I - -1
q(k)
+ 2[q(k) + TJ]{1 + [1-
2TJK 1/2
J.L{3q(k)] }
(2)
if the RHS is positive, and zero otherwise. The Lagrange multiplier J.L( < 0) is chosen
so that the {I c( k) I} satisfy V = 1.
Starting from a differently-stated goal (that of reducing redundancy subject to a
limit on information loss), which turns out to be closely related to infomax, (Atick
& Redlich 1990a) found an expression for the optimal filter gain that is the same
as that of Eq. 2 except for the choice of constraint.
Filter properties found using this approach are related to those found in early stages
of biological sensory processing. Smoothing and bandpass (contrast-enhancing)
filters emerge as infomax solutions (Linsker 1989, Atick & Redlich 1990a) in certain
cases, and good agreement with retinal contrast sensitivity measurements has been
found (Atick & Redlich 1990b).
Nonetheless, the value of the infomax solution Eq. 2 is limited in two important
ways. First, the phases of the {c(k)} are left undetermined. Any choice of phases is
equally good at maximizing MI' in a linear network. Thus the real-space response
function C(i - n), which determines the receptive field properties of the output
nodes, is nonunique (and indeed may be highly nonlocalized in space).
Second, it is useful to extend the solution Ansatz to allow a number of different filter
types a = 1, ... , A at each output site, while continuing to require that each type
955
956
Linsker
satisfy the shift-invariance condition Cni(a) == C(i - n;a). For example, one may
want to model a topographic 'retinocortical' mapping in which each patch of cortex
(each 'site') contains multiple filter types, yet each patch carries out the same set of
processing functions on its input. For this Ansatz, one again obtains Eq. 2 (derivation omitted here), but with 1 c(k) 12 on the LHS replaced by ~ap(a)1 c(k; a) 12,
where c(k; a) is the F.T. of C(i - n; a), and p(a) is the fraction of the total number
of filters (at each site) that are of type a. The partitioning of the overall (sumsquared) gain among the multiple filter types is thus left undetermined.
The higher-order statistical structure of the input (beyond covariance) is not being
exploited by infomaxin the above analysis, because (1) the network is linear and (2)
only pairwise correlations among the output activities enter into MI'. We shall show
that if we make the network even mildly nonlinear, MI' is no longer independent of
the choice of phases or of the partitioning of gain among multiple filter types.
3
NETWORK WITH WEAK NONLINEARITY
We consider the weakly nonlinear input-output relation Zn = Un + fU~ + ~iCniNi +
lin, where Un ~iCniSi, for small f. This differs from the linear network analyzed
above by the term in U~. (For simplicity, terms nonlinear in the noise are not
included.) The cubic term increases the signal-to-noise ratio selectively when Un is
large in absolute value. We maximize MI' as defined in Eq. l.
=
Heuristically, the new term will cause infomax to favor solutions in which some
output nodes have large (absolute) activity values, over solutions in which all output
nodes have moderate activities. The output layer can thus encode information
about the input vector (e.g., signal the presence of a feature) via the high activity
of a small number of nodes, rather than via the particular activity values of many
nodes. This has several (interrelated) potential advantages. (1) The concentration
of activity among fewer nodes is a type of sparse coding. (2) The resulting output
representation may be more resistant to noise. (3) The presence of a feature can be
signaled to a later processing stage using fewer connections. (4) Since the particular
nodes that have high activity depend upon the input vector, this type of mapping
transforms a set of continuous-valued inputs at each site into a partially place-coded
representation. A model of this sort may thus be useful for understanding better
the formation of place-coded representations in biological systems.
3.1
MATHEMATICAL DETAILS
This section may be skipped without loss of continuity. In matrix form, U CS,
Wn - U~ for each n, and Z = U + f W + C N + II. Keeping terms through first
{ZZT} = CQCT + 1]CCT + (3I' + fF,
order in f, the output covariance is Q'
where F {WUT)+{UWT} . [As an aside, Fnm = (UnUm(U~+U~)} resembles the
covariance (UnUm), except that presentations having large U~ +U~ are given greater
weight in the ensemble average.] For shift-invariant input statistics and one filter
type Cni C( i-n), taking the Fourier transform yields q'(k) = [q(k)+1]]1 c(k; a) 12+
(3 + ff(k) where f(k) is the F.T. of F(m - n)
Fnm. So In det Q' = ~k lnq'(k) =
~ln{[q(k)+1]] 1c(k) 12 +(3}+f~g(k) where g(k) _ [f(k)/{[q(k) +1]]1 c(k;a) 12+{3}] .
Using a Lagrange multiplier as before, the quantity to be maximized is MI"
=
=
=
=
=
Deriving Receptive Fields Using An Optimal Encoding Criterion
~~J[~
~~ -+-11~
Figure 1: Breaking of phase degeneracy. See text for discussion.
MI"(e = 0)
+ (e/2)~g(k).
Now suppose there are multiple filter types a = 1, ... , A at each output site. For
each k define d(k) to be the A x A matrix whose elements are: d(k)a.b
[q(k) +
TJ]c(k; a)c*(k; b) + [f3/ p(a)]c5a.b where c5a.b is the Kronecker delta. Also define f(k) to
be the A x A matrix each of whose elements f(k)ab is the F.T. of F(m - n;a,b)
where F(m - n; a, b) = (Un(a)Wm(b)) + (Wn(a)Um(b)). Then the O{e) part of MI"
is: (e/2)~kTr{[d(k)]-lf(k)}. Note that [d(k)]-l is the inverse of the matrix d(k),
and that 'Tr' denotes the trace. [Outline of derivation: In the basis defined by the
Fourier harmonics, Q' is block diagonal (one A x A block for each k). So In det Q' =
~k Indetq'(k) where each q'(k) is an A x A matrix of the form q~(k) + eq~(k).
Expanding In det q' (k) through O( e) yields the stated result.]
=
The infomax calculation to lowest order in e [i.e., O( eO)] is the same as for the linear
network. Here, for simplicity, we determine the sum-squared gain, ~ap(a)1 c(k; a) 12 ,
as in the linear case; then seek to maximize the new term, of O(e), subject to
this constraint on the value of the sum-squared gain. How the nonlinear term
breaks phase and gain-apportionment degeneracies is of interest here; a small O( e)
correction to the sum-squared gain is not.
4
ILLUSTRATIVE RESULTS
Two examples will show how adding the nonlinear perturbative term to the network's output breaks a degeneracy among different filter solutions. In each case the
input space is a one-dimensional 'retina' with wraparound.
4.1
BREAKING THE PHASE DEGENERACY
In this example (see Figure 1) there is one filter type at each output site. We consider
two types of input ensembles: (1) Each input vector (Fig. la shows one example)
is drawn from a multivariate Gaussian distribution (so there is no higher-order
statistical structure beyond pairwise correlations). The input covariance matrix
Q(j - i) is a Gaussian function of the distance between the sites. (2) Each input
957
958
Linsker
vector is a random sum of Gaussian 'bumps': Si = E j aj [s( i - j) - so] where s( i - j)
is a Gaussian (shown in Fig. 1b for j=20j there are 64 nodes in all); So is the mean
value of s( i - j); and each aj is independently and randomly chosen (with constant
probability) to be 1 or O. This ensemble does have higher-order structure, with
each input presentation being characterized by the presence of localized features
(the bumps) at particular locations.
The infomax solution for I c(k) 12 is plotted versus spatial frequency k in Fig. 1c
for a particular choice of noise parameters (1], (3). As stated earlier, MI' for a linear
network is indifferent to the phases of the Fourier components {c(k)}. A particular
random choice of phases produces the real-space filter C(i - n) shown in Fig. Id,
which spans the entire 'retina.' Setting all phases to zero produces the localized
filter shown in Fig. 1?. If the Gaussian 'bump' of Fig. 1b is presented as input to a
network of filters each of which is a shifted version of Fig. 1d, the linear response of
the network (i.e., the convolution of the 'bump' with the filter) is shown in Fig. Ie.
Replacing the filter of Fig. 1d by that of Fig. 1?, but keeping the input the same,
produces the output response shown in Fig. 19.
The cubic nonlinearity causes MI' to be larger for the filter of Fig. 1? than for that of
Fig. 1d. Heuristically, if we focus on the diagonal elements of the output covariance
Q', the nonlinear term is 2?(U~). Maximizing MI' favors increasing this term (subject to a constraint on output variance) hence favors filter solutions for which the
Un distribution is non-Gaussian with a preponderance of large values. Projection
pursuit methods also use a measure of the non-Gaussianity of the output distribution to construct filters that extract 'interesting' features from high-dimensional
data (cf. Intrator 1992).
4.2
BREAKING THE PARTITIONING DEGENERACY FOR
MULTIPLE FILTER TYPES
In this example (see Fig. 2), the input ensemble comprises a set of self-similar
patterns (each is a sine-Gabor 'ripple' as in Fig. 2a) that are related by translation
and dilation (scale change over a factor of 80). Figure 2b shows the input power
spectrum vs. k; the scaling region goes as 11k. Figure 2c shows the infomax solution
for the gain I c(k;a) I vs. k when there is just one filter type. When the input SNR
is large (as in the scaling region) the infomax filters 'whiten' the output; note the
flat portion of the output power spectrum (Fig. 2d). [We modify the infomax
solution by extending the power-law form of I c(k) I to low k (dotted line in Figs.
2c,d). This avoids artifacts resulting from the rapid increase in I c(k) I, which is
in turn caused by our having omitted low-k patterns from the input ensemble for
reasons of numerical efficiency.] The dotted envelope curve in Figure 2e shows the
sum-squared gain Eap(a) I c(k) 12 when multiple filter types a are allowed. The
quantity plotted is just the square of that shown in Fig. 2c, but on a linear rather
than log-log plot (note values greater than 5 are cut off to save space).
The network nonlinearity has the following effect. We first allow two filter types
to share the overall gain. Optimizing MI' over various partitionings, we find that
info max favors a crossover between filter types at k ~ 400. Allowing three, then four,
filter types produces additional crossovers at lower k. For an Ansatz in which each
filter's share of the sum-squared gain is tapered linearly near its cutoff frequencies,
Deriving Receptive Fields Using An Optimal Encoding Criterion
the best solution found for each p( a) 1c( k) 12 is shown in Fig. 2e (semilog plot vs.
k ). Figure 2f plots the corresponding 1 c( k; a) 1 vs. k on a linear scale. Note that the
three lower-k filters appear roughly self-similar. (The peak in the highest-k filter is
an artifact due to the cutoff of the input ensemble at high k.) The four real-space
filters C(i-n; a) are plotted vs. (i-n) in Fig. 2g [phases chosen to make C(i-n; a)
antisymmetric] .
The resulting filters span multiple resolution scales. The density p( a) is less for the
lower-frequency filters (spatial subsampling). When more filter types are allowed,
the increase in MI' becomes progressively less. Although in our model the filters are
present with density p at each output site, a similar MIl is obtained if one spaces
adjacent filters of type a by a distance <X 1/ p( a). The resulting arrangement of
filters resembles the 'tiling' of the joint space and spatial-frequency domain that is
used in wavelet and 'pyramid coding' approaches to image processing. [The infomax
filters overlap, rather than disjointly tiling the (:I;, k) domain.]
Using the infomax method, the region of (:1;, k) space spanned by an optimal filter
has an aspect ratio that depends upon the relative distances - along the :I; and k axes
- over which the input feature is 'coherent' (possesses higher-order correlations).
One may thus be able to use infomax to predict relationships between statistical
measures of coherence in natural scenes and observed (:1;, k) aspect ratios for, e.g.,
orientation-selective cells. See (Field 1989) for a discussion of this issue that is not
based on infomax.
5
APPENDIX: HEURISTIC JUSTIFICATION FOR
USING A SURROGATE, 'AS-IF-GAUSSIAN,'
MUTUAL INFORMATION
The mutual information between input S and output Z is MI
!dSdZPszln(Psz/PsPz)
!dSPsKD(Pzls;Pz) where KD(Pzls;PZ)
! dZPzlsln(Pzls/PZ) is a Kullback divergence. So, maximizing MI means maximizing the average (over S) of KD(Pzls; Pz).
What does the KD represent? Suppose that the network has somehow learned the
distribution Pz. Before being presented with a particular input S, the network 'expects' an output vector drawn from Pz. The actual output response to S, however,
is a vector drawn from PZls. The KD measures the 'surprise' (i.e., the amount of
information gained) upon seeing the actual distribution PZls when one expected
Pz. Infomax maximizes this average 'surprise.'
However, the network cannot in general have access to the full distribution Pz,
which contains far too much information (including all higher-order statistics) to
be stored in the connections and nodes of the network. Let us suppose for definiteness that the system remembers only the mean and the covariance matrix of
Z. Define pff to be the multivariate Gaussian distribution that has the same mean
and covariance as Pz. Then we may think of the system as a priori 'expecting' the
output vector to be drawn from the distribution pff.
We accordingly modify the principle so that we maximize the averThis equals
age (over S) of KD(PzIS; pff) (note the superscript G).
959
960
Linsker
1000
100
10
0.1
10
10
1000
10
100
1000
(d)
0.1 I
o
100
10
100
1000
~
__-
(9)
,...=;;
'
80
=
Figure 2: Partitioning among multiple filter types. See text.
J dSPs JdZ PZls In( P Z1s / pi) = (- H zls) s - JdZ Pz In pi
(where H denotes entropy). Using a property of the Gaussian distribution, we have - J dZPz In pi =
- J dZpi InPff = He;. We conclude that the average of KD equals He; - (HZls)s,
which is exactly equal to the surrogate ' as-if-Gaussian' MI defined preceding Eq. 1.
This argument provides a principled justification for using the surrogate MI, when
the system has stored information about the output vectors' mean and covariance,
but not about higher-order statistics.
References
J. J. Atick & A. N. Redlich. (1990a) Towards a theory of early visual processing.
Neural Computation 2:308-320.
J. J. Atick & A. N. Redlich. (1990b) Quantitative tests of a theory of retinal
processing: contrast sensitivity curves. Inst. Adv. Study IASSNS-HEP-90/51.
D. J. Field. (1989) What the statistics of natural images tell us about visual coding.
In Proc. SPIE 1077:269-276.
N.lntrator. (1992) Feature extraction using an unsupervised neural network. Neural Computation 4:98-107.
R. Linsker. (1988) Self-organization in a perceptual network. Computer 21(3):105117.
R. Linsker. (1989) An application of the principle of maximum information preservation to linear systems. In D. S. Touretzky (ed.), Advances in Neural Information
Processing Systems 1, 186-194. San Mateo, CA: Morgan Kaufmann.
R. Linsker. (1992) Local synaptic learning rules suffice to maximize mutual information in a linear network. Neural Computation 4(5 ):691-702.
| 667 |@word version:1 nd:1 heuristically:2 seek:1 covariance:12 tr:1 carry:1 substitution:1 contains:2 si:3 yet:1 perturbative:1 fn:1 numerical:1 remove:1 plot:3 progressively:1 aside:1 v:5 fewer:2 accordingly:1 provides:1 node:16 location:1 height:1 mathematical:1 along:1 pairwise:3 expected:1 indeed:2 rapid:1 roughly:1 cct:1 eap:1 actual:3 increasing:1 becomes:1 suffice:1 maximizes:4 lowest:1 what:2 quantitative:1 exactly:1 um:1 partitioning:5 unit:1 appear:1 positive:1 before:2 local:2 modify:2 tends:2 limit:1 encoding:4 id:1 ap:2 resembles:2 mateo:1 sumsquared:1 limited:1 range:1 averaged:2 unique:1 block:2 differs:1 lf:1 lndet:1 crossover:2 gabor:1 projection:1 seeing:1 cannot:1 disjointly:1 map:1 center:1 maximizing:4 go:1 starting:1 independently:1 resolution:2 simplicity:2 rule:1 deriving:4 spanned:1 his:1 justification:3 limiting:1 zzt:2 suppose:4 agreement:1 element:3 sparsely:1 cut:1 observed:1 region:3 adv:1 highest:1 principled:2 expecting:1 dynamic:1 weakly:1 depend:1 upon:3 efficiency:1 basis:1 expedient:1 joint:1 differently:1 represented:1 various:1 derivation:3 tell:1 formation:1 whose:2 heuristic:1 larger:1 valued:2 otherwise:1 favor:5 statistic:5 topographic:1 think:1 transform:2 superscript:2 advantage:1 kohonen:1 multiresolution:1 ktr:1 inputoutput:1 regularity:1 extending:1 ripple:1 produce:4 depending:1 eq:6 implemented:1 c:3 involves:1 closely:1 tji:1 filter:48 centered:1 require:1 biological:2 correction:1 mapping:7 predict:1 bump:4 tjk:1 early:2 omitted:2 proc:1 sensitive:1 gaussian:17 modified:1 rather:4 sisj:1 mil:1 encode:1 ax:1 focus:1 contrast:3 skipped:1 inst:1 nn:1 entire:1 relation:1 selective:1 ralph:1 overall:2 among:7 orientation:1 issue:1 favored:1 priori:1 development:1 spatial:5 special:1 smoothing:1 mutual:5 field:8 equal:4 f3:1 having:4 construct:1 extraction:1 unsupervised:4 linsker:12 retina:2 randomly:1 preserve:1 divergence:1 replaced:1 phase:10 ab:1 organization:1 interest:2 highly:1 indifferent:1 analyzed:1 bracket:1 tj:6 fu:1 lh:1 continuing:1 desired:1 signaled:1 plotted:3 earlier:2 hep:1 zn:2 maximization:1 expects:1 undetermined:2 snr:1 comprised:1 too:1 stored:2 sv:1 density:2 peak:1 sensitivity:2 ie:1 off:1 ansatz:4 infomax:21 again:1 squared:5 choose:1 potential:1 retinal:2 coding:5 summarized:1 gaussianity:1 int:1 satisfy:2 caused:1 depends:1 later:1 break:2 sine:1 portion:1 wm:1 sort:1 square:1 variance:4 kaufmann:1 ensemble:13 yield:3 uwt:1 maximized:1 cni:4 weak:1 influenced:1 touretzky:1 ed:1 synaptic:1 definition:1 nonetheless:1 frequency:6 conveys:1 mi:32 spie:1 degeneracy:5 gain:12 dimensionality:1 higher:7 response:4 done:1 box:1 though:1 just:2 stage:3 implicit:1 atick:5 correlation:4 replacing:3 nonlinear:7 somehow:1 continuity:1 artifact:2 aj:2 effect:1 requiring:1 multiplier:3 preponderance:1 analytically:1 hence:1 inq:1 adjacent:1 self:3 illustrative:1 yorktown:1 whiten:1 criterion:4 c5a:2 theoretic:1 outline:1 image:2 harmonic:1 discriminated:1 extend:1 he:2 measurement:1 enter:1 similarly:1 nonlinearity:4 resistant:1 access:1 cortex:1 longer:1 multivariate:4 optimizing:1 irrelevant:1 apart:1 moderate:1 certain:4 watson:1 exploited:1 morgan:1 additional:2 greater:2 preceding:1 eo:1 accomplishes:1 determine:1 maximize:7 signal:2 preservation:2 ii:1 multiple:9 full:2 hebbian:1 characterized:1 calculation:1 lin:2 equally:2 coded:3 enhancing:1 represent:1 pyramid:3 cell:1 justified:1 want:1 standpoint:1 envelope:1 semilog:1 posse:1 subject:6 tend:1 hz:3 call:1 near:1 presence:6 wn:2 cn:3 det:3 shift:4 expression:2 utility:1 lnq:1 proceed:1 cause:2 retinocortical:1 useful:4 detailed:1 sst:1 transforms:2 amount:1 concentrated:1 generate:1 shifted:1 dotted:2 delta:1 shall:1 redundancy:1 salient:1 four:2 drawn:4 tapered:1 cutoff:2 fraction:1 sum:7 angle:1 inverse:1 cqct:2 place:2 vn:4 patch:2 coherence:1 appendix:2 scaling:2 layer:1 quadratic:1 activity:9 constraint:5 kronecker:1 constrain:1 scene:1 flat:1 fourier:5 aspect:2 argument:1 span:3 relatively:1 zls:1 kd:6 qu:1 invariant:3 ln:1 equation:1 previously:1 turn:2 tiling:2 pursuit:1 lacked:1 intrator:1 appropriate:1 save:1 denotes:3 subsampling:2 cf:1 fnm:2 arrangement:1 quantity:3 receptive:6 strategy:1 concentration:1 diagonal:2 surrogate:5 distance:3 parametrized:1 reason:1 relationship:1 ratio:3 info:2 trace:1 stated:3 allowing:1 convolution:1 nonunique:1 dsps:1 wraparound:1 connection:5 coherent:1 learned:1 beyond:3 able:1 below:1 pattern:4 mismatch:1 summarize:1 max:2 including:1 power:3 overlap:1 natural:2 improve:1 identifies:1 extract:1 remembers:1 text:2 prior:1 understanding:1 relative:1 law:1 loss:2 interesting:1 limitation:2 versus:1 localized:4 age:1 degree:1 principle:5 dq:1 share:2 pi:3 ibm:1 translation:1 transpose:1 keeping:2 allow:2 vv:1 taking:1 emerge:2 absolute:2 sparse:1 curve:2 avoids:1 sensory:1 san:1 far:1 obtains:1 kullback:1 conclude:1 spectrum:2 continuous:2 un:5 dilation:1 expanding:1 ca:1 domain:2 antisymmetric:1 linearly:1 rh:1 noise:11 allowed:3 constra:1 site:8 redlich:5 fig:20 ff:2 cubic:2 definiteness:2 ny:1 comprises:1 bandpass:1 perceptual:1 breaking:3 cqc:2 wavelet:3 jdz:2 pz:10 intractable:1 adding:3 gained:1 magnitude:1 pff:3 mildly:1 surprise:2 entropy:5 interrelated:1 visual:2 lagrange:3 partially:1 determines:2 goal:2 presentation:2 towards:1 lntrator:1 change:1 included:1 except:2 reducing:2 acting:1 principal:1 total:1 invariance:1 la:1 shannon:1 selectively:1 brevity:1 |
6,268 | 6,670 | Recycling Privileged Learning
and Distribution Matching for Fairness
Novi Quadrianto?
Predictive Analytics Lab (PAL)
University of Sussex
Brighton, United Kingdom
[email protected]
Viktoriia Sharmanska
Department of Computing
Imperial College London
London, United Kingdom
[email protected]
Abstract
Equipping machine learning models with ethical and legal constraints is a serious
issue; without this, the future of machine learning is at risk. This paper takes a step
forward in this direction and focuses on ensuring machine learning models deliver
fair decisions. In legal scholarships, the notion of fairness itself is evolving and
multi-faceted. We set an overarching goal to develop a unified machine learning
framework that is able to handle any definitions of fairness, their combinations,
and also new definitions that might be stipulated in the future. To achieve our
goal, we recycle two well-established machine learning techniques, privileged
learning and distribution matching, and harmonize them for satisfying multi-faceted
fairness definitions. We consider protected characteristics such as race and gender
as privileged information that is available at training but not at test time; this
accelerates model training and delivers fairness through unawareness. Further, we
cast demographic parity, equalized odds, and equality of opportunity as a classical
two-sample problem of conditional distributions, which can be solved in a general
form by using distance measures in Hilbert Space. We show several existing
models are special cases of ours. Finally, we advocate returning the Pareto frontier
of multi-objective minimization of error and unfairness in predictions. This will
facilitate decision makers to select an operating point and to be accountable for it.
1
Introduction
Machine learning technologies have permeated everyday life and it is common nowadays that an
automated system makes decisions for/about us, such as who is going to get bank credit. As
more decisions in employment, housing, and credit become automated, there is a pressing need
for addressing ethical and legal aspects, including fairness, accountability, transparency, privacy,
and confidentiality, posed by those machine learning technologies [1, 2]. This paper focuses on
enforcing fairness in the decisions made by machine learning models. A decision is fair if [3, 4,
5]: i) it is not based on a protected characteristic [6] such as gender, marital status, or age (fair
treatment), ii) it does not disproportionately benefit or hurt individuals sharing a certain value of
their protected characteristic (fair impact), and iii) given the target outcomes, it enforces equal
discrepancies between decisions and target outcomes across groups of individuals based on their
protected characteristic (fair supervised performance).
The above three fairness definitions have been studied before, and several machine learning frameworks for addressing each one or a combination of them are available. We first note that one could
ensure fair treatment by simply ignoring protected characteristic features, i.e. fairness through
unawareness. However this poses a risk of unfairness by proxy as there are ways of predicting
?
Also with National Research University Higher School of Economics, Moscow, Russia.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
protected characteristic features from other features [7, 8]. Existing models guard against unfairness
by proxy by enforcing fair impact or fair supervised performance constraints in addition to the fair
treatment constraint. An example of the fair impact constraint is the 80% rule (see e.g. [3, 9, 10]) in
which positive decisions must be in favour of group B individuals at least 80% as often as in favour
of group A individuals for the case of a binary protected characteristic and a binary decision. Another
example of the fair impact constraint is a demographic parity in which positive decisions of group B
individuals must be at the same rate as positive decisions of group A individuals (see e.g. [11] and
earlier works [12, 13, 14]).
In contrast to the fair impact that only concerns about decisions of an automated system, the
fair supervised performance takes into account, when enforcing fairness, the discrepancy between
decisions (predictions) and target outcomes, which is compatible to the standard supervised learning
setting. Kleinberg et al. [15] show that fair impact and fair supervised performance are indeed
mutually exclusive measures of fairness. Examples of the fair supervised performance constraint
are equality of opportunity [4] in which the true positive rates (false negative rates) across groups
must match, and equalized odds [4] in which both the true positive rates and false positive rates must
match. Hardt et al. [4] enforce equality of opportunity or equalized odds by post-processing the
soft-outputs of an unfair classifier. The post-processing step consists of learning a different threshold
for a different group of individuals. The utilization of an unfair classifier as a building block of the
model is deliberate as the main goal of supervised machine learning models is to perform prediction
tasks for future data as accurately as possible. Suppose the target outcome is correlated with the
protected characteristic, Hardt et al.?s model will be able to learn the ideal predictor, which is not
unfair as it represents the target outcome [4]. However, Hardt et al.?s model needs to access the
value of the protected characteristic for future data. Situations where the protected characteristic
is unavailable due to confidentiality or is prohibited to be accessed due to the fair treatment law
requirement will make the model futile [5]. Recent work of Zafar et al. [5] propose de-correlation
constraints between supervised performance, e.g. true positive rate, and protected characteristics as a
way to achieve fair supervised performance. Zafar et al.?s model, however, will not be able to learn
the ideal predictor when the target outcome is indeed correlated with a protected characteristic.
This paper combines the benefits of Hardt et al.?s model [4] in its ability to learn the ideal predictor
and of Zafar et al.?s model [5] in not requiring the availability of protected characteristic for future
data at prediction time. To achieve this, we will be building upon recent advances in the use of
privileged information for training machine learning models [16, 17, 18, 19]. Privileged information
refers to features that can be used at training time but will not be available for future data at prediction
time. We propose to consider protected characteristics such as race, gender, or marital status as
privileged information. The privileged learning framework is remarkably suitable for incorporating
fairness, as it learns the ideal predictor and does not require protected characteristics for future data.
Therefore, this paper recycles the overlooked privileged learning framework, which is designed for
accelerating learning and improving prediction performance, for building a fair classification model.
Enforcing fairness using the privileged learning framework alone, however, might increase the risk
of unfairness by proxy. Our proposed model guards against this by explicitly adding fair impact
and/or fair supervised performance constraints into the privileged learning model. We recycle a
distribution matching measure for fairness. This measure can be instantiated for both fair impact
(e.g. demographic parity) and fair supervised performance (e.g. equalized odds and equality of
opportunity) constraints. Matching a distribution between function outputs (decisions) across different
groups will deliver fair impact, and matching a distribution between errors (discrepancies between
decisions and target outcomes) across different groups will deliver fair supervised performance. We
further show several existing methods are special cases of ours.
2
Related Work
There is much work on the topic of fairness in the machine learning context in addition to those that
have been embedded in the introduction. One line of research can be described in terms of learning
fair models by modifying feature representations of the data (e.g. [20, 10, 21]), class label annotations
([22]), or even the data itself ([23]). Another line of research is to develop classifier regularizers that
penalize unfairness (e.g. [13, 14, 24, 11, 5]). Our method falls into this second line of research. It
has also been emphasized that fair models could enforce group fairness definitions (covered in the
introduction) as well as individual fairness definitions. Dwork et al. and Joseph et al. [25, 26] define
2
an individual fairness as a non-preferential treatment towards an individual A if this individual is not
as qualified as another individual B; this is a continuous analog of fairness through unawareness [23].
On privileged learning Vapnik et al. [16] introduce privileged learning in the context of Support
Vector Machines (SVM) and use the privileged features to predict values of the slack variables. It
was shown that this procedure can provably reduce the amount of data needed for learning an optimal
hyperplane [16, 27, 19]. Additional features for training a classifier that will not necessarily be
available at prediction time, privileged information, are widespread. As an example, features from 3D
cameras and laser scanners are slow to acquire and expensive to store but have the potential to boost
the predictive capability of a trained 2D system. Many variants of privileged learning methods and
settings have been proposed such as, structured prediction [28], margin transfer [17], and Bayesian
privileged learning [18, 29]. Privileged learning has also been shown [30] to be intimately related
to Hinton et al.?s knowledge distillation [31] and Bucila et al.?s [32] model compression in which a
complex model is learnt and is then replicated by a simpler model.
On distribution matching Distribution matching has been explored in the context of domain adaptation (e.g. [33, 34]), transduction learning (e.g. [35]), and recently in privileged learning [36],
among others. The empirical Maximum Mean Discrepancy (MMD) [37] is commonly used as the
nonparametric metric that captures discrepancy between two distributions. In the domain adaptation
setting, Pan et al. [38] use the MMD metric to project data from target and related source domain into
a common subspace such that the difference between the distributions of source and target domain
data is reduced. A similar idea has been explored in the context of deep neural networks by Zhang
et al. [34], where they use the MMD metric to match both the distribution of the features and the
distribution of the labels given features in the source and target domains. In the transduction setting,
Quadrianto et al. [35] propose to minimize the mismatch between the distribution of function outputs
on the training data and on the target test data. Recently, Sharmanska et al. [36] devise a cross-dataset
transfer learning method by matching the distribution of classifier errors across datasets.
3
The Fairness Model
In this section, we will formalize the setup of a supervised binary classification task subject to fairness
constraints. Assume that we are given a set of N training examples, represented by feature vectors
X = {x1 , . . . , xN } ? X = Rd , their label annotation, Y = {y1 , . . . , yN } ? Y = {+1, ?1}, and
protected characteristic information also in the form of feature vectors, Z = {z1 , . . . , zN } ? Z,
where zn encodes the protected characteristics of sample xn . The task of interest is to infer a predictor
f for the label ynew of an un-seen instance xnew , given Y , X and Z. However, f cannot use the
protected characteristic Z at decision (prediction) time, as it will constitute an unfair treatment.
The availability of protected characteristic at training time can be used to enforce fair impact
and/or fair supervised performance constraints. We first describe how to deliver fair treatment via
privileged learning. We then detail distribution matching viewpoint of fair impact and fair supervised
performance. Frameworks of privileged learning and distribution matching are suitable for protected
characteristics with binary/multi-class/continuous values. In this paper, however, we focus on a single
protected characteristic admitting binary values as in existing work (e.g. [20, 4, 5]).
3.1
Fairness through Unawareness: Privileged Learning
In the privileged learning setting [16], we are given training triplets (x1 , x?1 , y1 ), . . . , (xN , x?N , yN )
where (xn , yn ) ? X ? Y is the standard training input-output pair and x?n ? X ? is additional
information about a training instance xn . This additional (privileged) information is only available
during training. In our earlier illustrative example in the related work, xn is for example a colour
feature from a 2D image while x?n is a feature from 3D cameras and laser scanners. There is no direct
limitation on the form of privileged information, i.e. it could be yet another feature representation
like shape features from the 2D image, or a completely different modality like 3D cameras in addition
to the 2D image, that is specific for each training instance. The goal of privileged learning is to use
x?n to accelerate the learning process of inferring an optimal (ideal) predictor in the data space X ,
i.e. f : X ? Y. The difference between accelerated and non-accelerated methods is in the rate of
?
convergence to the optimal predictor, e.g. 1/N cf.1/ N for margin-based classifiers [16, 19].
From the description above, it is apparent that both privileged learning model and fairness model aim
to use data, privileged feature x?n and protected characteristic zn respectively, that are available at
3
training time only. We propose to recycle privileged learning model for achieving fairness through
unawareness by taking protected characteristics as privileged information. For a single binary
protected characteristic zn , x?n is formed by concatenating xn and zn . This is because privileged
information has to be instance specific and richer than xn alone, and this is not the case when only a
single binary protected characteristic is used. By using privileged learning framework, the predictor
f is unaware of protected characteristic zn as this information is not used as an input to the predictor.
Instead, zn , together with xn , is used to distinguish between easy-to-classify and difficult-to-classify
data instances and subsequently to use this knowledge to accelerate the learning process of a predictor
f [16, 17]. Easiness and hardness can be defined, for example, based on the distance of data instance
to the decision boundary (margin) [16, 17, 19] or based on the steepness of the logistic likelihood
function [18]. Our specific choice of easiness and hardness definition is detailed in Section 3.3.
A direct advantage of approaching fairness from the privileged lens is the learning acceleration can
be used to limit the performance degradation of the fair model as it now has to trade-off two goals:
good prediction performance and respecting fairness constraints. An obvious disadvantage is an
increased risk of unfairness by proxy as knowledge of easy-to-classify and difficult-to-classify data
instances is based on protected characteristics. The next section describes a way to alleviate this
based on a distribution matching principle.
3.2
Demographic Parity, Equalized Odds, Equality of opportunity, and Beyond: Matching
Conditional Distributions
We have the following definitions for several fairness criteria [25, 4, 5]:
Definition A Demographic parity (fair impact): A binary decision model is fair if its decision
{+1, ?1} are independent of the protected characteristic z ? {0, 1}. A decision f? satisfies this
definition if
P (sign(f?(x)) = +1|z = 0) = P (sign(f?(x)) = +1|z = 1).
Definition B Equalized odds (fair supervised performance): A binary decision model is fair if its
decisions {+1, ?1} are conditionally independent of the protected characteristic z ? {0, 1} given
the target outcome y. A decision f? satisfies this definition if
P (sign(f?(x)) = +1|z = 0, y) = P (sign(f?(x)) = +1|z = 1, y), for y ? {+1, ?1}.
For the target outcome y = +1, the definition above requires that f? has equal true positive rates
across two different values of protected characteristic. It requires f? to have equal false positive rates
for the target outcome y = ?1.
Definition C Equality of opportunity (fair supervised performance): A binary decision model is fair
if its decisions {+1, ?1} are conditionally independent of the protected characteristic z ? {0, 1}
given the positive target outcome y. A decision f? satisfies this definition if
P (sign(f?(x)) = +1|z = 0, y = +1) = P (sign(f?(x)) = +1|z = 1, y = +1).
Equality of opportunity only constrains equal true positive rates across the two demographics.
All three fairness criteria rely on the definition that data across the two demographics should exhibit
similar behaviour, i.e. matching positive predictions, matching true positive rates, and matching false
positive rates. A natural pathway to inject these into any learning model is to use a distribution
matching framework. This matching assumption is well founded if we assume that both data
z=1
Xz=0 = {xz=0
, . . . , xz=0
, . . . , xz=1
1
Nz=0 } ? X and another data Xz=1 = {x1
Nz=1 } ? X are drawn
independently and identically distributed from the same distribution p(x) on a domain X . It therefore
follows that for any function (or set of functions) f the distribution of f (x) where x ? p(x) should
also behave in the same way across the two demographics. We know that this is not automatically
true if we get to choose f after seeing Xz=0 and Xz=1 . In order to allow us to draw on a rich body
of literature for comparing distributions, we cast the goal of enforcing distributional similarity across
two demographics as a two-sample problem.
3.2.1
Distribution matching
First, we denote the applications of our predictor f? : X ? R to data having protected
?
characteristic value zero by f?(XZ=0 ) := {f?(xz=0
), . . . , f?(xz=0
1
Nz=0 )}, likewise by f (XZ=1 ) :=
4
{f?(xz=1
), . . . , f?(xz=1
1
Nz=1 )} for value one. For enforcing the demographic parity criterion, we can
enforce the closeness between the distributions of f?(x). We can achieve this by minimizing:
D(f?(XZ=0 ), f?(XZ=1 )), the distance between the two distributions f?(XZ=0 ) and f?(XZ=1 ). (1)
For enforcing the equalized odds criterion, we need to minimize both
D(I[Y = +1]f?(XZ=0 ), I[Y = +1]f?(XZ=1 )) and D(I[Y = ?1]f?(XZ=0 ), I[Y = ?1]f?(XZ=1 )).
(2)
We make use of Iverson?s bracket notation: I[P ] = 1 when condition P is true and 0 otherwise.
The first will match true positive rates (and also false negative rates) across the two demographics
and the latter will match false positive rates (and also true negative rates). For enforcing equality of
opportunity, we just need to minimize
D(I[Y = +1]f?(XZ=0 ), I[Y = +1]f?(XZ=1 )).
(3)
To go beyond true positive rates and false positive rates, Zafar et al. [5] raise the potential of removing
unfairness by enforcing equal misclassification rates, false discovery rates, and false omission rates
across two demographics. False discovery and false omission rates, however, with their fairness
model are difficult to encode. In the distribution matching sense, those can be easily enforced by
minimizing
D(1 ? Y f?(XZ=0 ), 1 ? Y f?(XZ=1 )),
D(I[Y = +1] max(0, ?f?(XZ=0 )), I[Y = +1] max(0, ?f?(XZ=1 ))), and
D(I[Y = ?1] max(0, f?(XZ=0 )), I[Y = ?1] max(0, f?(XZ=1 )))
(4)
(5)
(6)
for misclassification, false omission, and false discovery rates, respectively.
Maximum mean discrepancy To avoid a parametric assumption on the distance estimate between
distributions, we use the Maximum Mean Discrepancy (MMD) criterion [37], a non-parametric
distance estimate. Denote by H a Reproducing Kernel Hilbert Space with kernel k defined on X . In
this case one can show [37] that whenever k is characteristic (or universal), the map
2
? :p ? ?[p] := Ex?p(x) [k(f?(x), ?)] with associated distance MMD2 (p, p0 ) := k?[p] ? ?[p0 ]k
characterizes a distribution uniquely. Examples of characteristic kernels [39] are Gaussian RBF, Laplacian and B2n+1 -splines.
With a this choice of kernel functions, the
MMD criterion matches infinitely many moments in the Reproducing Kernel Hilbert Space
(RKHS). We use an unbiased linear-time estimate of MMD as follows [37, Lemma 14]:
\2 = 1 PN k(f?(x2i?1 ), f?(x2i )) ? k(f?(x2i?1 ), f?(x2i )) ? k(f?(x2i ), f?(x2i?1 )) +
MMD
z=0
z=1
z=0
z=0
z=0
z=1
i
N
? 2i
k(f?(x2i?1
z=1 ), f (xz=1 )), with N := bmin(Nz=1 , Nz=0 )c.
3.2.2
Special cases
Before discussing a specific composition of privileged learning and distribution matching to achieve
fairness, we consider a number of special cases of matching constraint to show that many of existing
methods use this basic idea.
Mean matching for demographic parity Zemel et al. [20] balance the mapping from data to
one of C latent prototypes across the two demographics by imposing the following constraint:
PNz=0 ? z=0
PNz=1 ? z=1
1
1
? z=0
n=1 f (xn ; c) = Nz=1
n=1 f (xn ; c); ?c = 1, . . . , C, where f (xn ) is a softmax
Nz=0
function with C prototypes. Assuming a linear kernel k on this constraint is equivalent to requiring
that for each c
?[f?(xz=0
n ; c)] =
1
N
z=0
X
Nz=0
n=1
D
E
f?(xz=0
n ; c), ? =
1
N
z=1
X
Nz=1
n=1
D
E
? z=1
f?(xz=1
n ; c), ? = ?[f (xn ; c)].
Mean matching for equalized odds and equality of opportunity To ensure equal false positive
rates across the two demographics, Zafar et al. [5] add the following constraint to the training objective
5
PNz=0
PNz=1
min(0, I[yn = ?1]f?(xz=0
of a linear classifier f?(x) = hw, xi: n=1
n )) =
n=1 min(0, I[yn =
?1]f?(xz=1
)).
Again,
assuming
a
linear
kernel
k
on
this
constraint
is
equivalent
to requiring that
n
N
z=0 D
E
1 X
min(0, I[yn = ?1]f?(xz=0
?[min(0, I[yn = ?1]f?(xz=0
n ))] =
n )), ?
Nz=0 n=1
=
1
N
z=1
X
Nz=1
n=1
D
E
? z=1
min(0, I[yn = ?1]f?(xz=1
n )), ? = ?[min(0, I[yn = ?1]f (xn ))].
The min(?) function ensures that we only match false positive rates as without it both false positive
and true negative rates will be matched. Relying on means for matching both false positive and true
negative is not sufficient as the underlying distributions are multi-modal; it motivates the need for
distribution matching.
3.3
Privileged learning with fairness constraints
Here we describe the proposed model that recycles two established frameworks, privileged learning
and distribution matching, and subsequently harmonizes them for addressing fair treatment, fair
impact, fair supervised performance and beyond in a unified fashion. We use SVM? + [19], an
SVM-based classification method for privileged learning, as a building block. SVM? + modifies
the required distance of data instance to the decision boundary based on easiness/hardness of
that data instance in the privileged space X ? , a space that contains protected characteristic Z.
Easiness/hardness is reflected in the negative of the confidence, ?yn (hw? , x?n i + b? ) where w? and
b? are some parameters; the higher this value, the harder this data instance to be classified correctly
even in the rich privileged space. Injecting the distribution matching constraint, the final Distribution
Matching+ (DM+) optimization problem is now:
minimize 1/2
2
kwk`2
| {z }
+1/2?
w?Rd ,b?R
?
w? ?Rd ,b? ?R regularisation on model without
protected characteristic
+C
N
X
2
kw? k`2
| {z }
+C?
N
X
max (0, ?yn [hw? , x?n i + b? ]) +
n=1
regularisation on model with
protected characteristic
|
{z
}
hinge loss on model with protected characteristic
max (0, 1 ? yn [hw? , x?n i + b? ] ? yn [hw, xn i + b])
(7a)
n=1
|
subject to
{z
hinge loss on model without protected characteristic
but with margin dependent on protected characteristic
\2 (p , p ) ? ,
MMD
z=0 z=1
|
{z
}
}
(7b)
constraint for removing unfairness by proxy
where C, ?, ? and an upper-bound are hyper-parameters. Terms pz=0 and pz=1 are distributions over
appropriately defined fairness variables across the two demographics, e.g. f?(XZ=0 ) and f?(XZ=1 )
with f?(?) = hw, ?i + b for demographic parity and I[Y = +1]f?(XZ=0 ) and I[Y = +1]f?(XZ=1 )
for equality of opportunity. We have the following observations of the knowledge transfer from the
privileged space to the space X without protected characteristic (refer to the last term in (7a)):
? Very large positive value of the negative of the confidence in the space that includes protected
characteristic, ?yn [hw? , x?n i + b? ] >> 0 means xn , without protected characteristic, is expected
to be a hard-to-classify instance therefore its margin distance to the decision boundary is increased.
? Very large negative value of the negative of the confidence in the space that includes protected
characteristic, ?yn [hw? , x?n i + b? ] << 0 means xn , without protected characteristic, is expected
to be an easy-to-classify instance therefore its margin distance to the decision boundary is reduced.
The formulation in (7) is a multi-objective optimization with three competing goals: minimizing
empirical error (hinge loss), minimizing model complexity (`2 regularisation), and minimizing
prediction discrepancy across the two demographics (MMD). Each goal corresponds to a different
optimal solution and we have to accept a compromise in the goals. While solving a single-objective
optimization implies to search for a single best solution, a collection of solutions at which no goal
can be improved without damaging one of the others (Pareto frontier) [40] is sought when solving a
multi-objective optimization.
6
Multi-objective optimization We first note that the MMD fairness criteria will introduce nonconvexity to our optimization problem. For a non-convex multi-objective optimization, the Pareto
frontier may have non-convex portions. However, any Pareto optimal solution of a multi-objective
optimization can be obtained by solving the constraint problem for an upper bound (as in (7b))
regardless of the non-convexity of the Pareto frontier [40].
Alternatively, the Convex Concave Procedure (CCP) [41], can be used to find an approximate solution
of the problem in (7) by solving a succession of convex programs. CCP has been used in several
other algorithms enforcing fair impact and fair supervised performance to deal with non-convexity of
the objective function (e.g. [24, 5]). However, it was noted in [35] that for an objective function that
has an additive structure as in our DM+ model, it is better to use the non-convex objective directly.
4
Experiments
We experiment with two datasets: The ProPublica COMPAS dataset and the Adult income dataset.
ProPublica COMPAS (Correctional Offender Management Profiling for Alternative Sanctions) has a
total of 5,278 data instances, each with 5 features (e.g., count of prior offences, charge for which the
person was arrested, race). The binary target outcome is whether or not the defendant recidivated
within two years. For this dataset, we follow the setting in [5] and consider race, which is binarized
as either black or white, as a protected characteristic. We use 4, 222 instances for training and 1, 056
instances for test. The Adult dataset has a total of 45, 222 data instances, each with 14 features (e.g.,
gender, educational level, number of work hours per week). The binary target outcome is whether
or not income is larger than 50K dollars. For this dataset, we follow [20] and consider gender as a
binary protected characteristic. We use 36, 178 instances for training and 9, 044 instances for test.
Methods We have two variants of our distribution matching framework: DM that uses SVM as the
base classifier coupled with the constraint in (7b), and DM+ ((7a) and (7b)). We compare our methods
with several baselines: support vector machine (SVM), logistic regression (LR), mean matching with
logistic regression as the base classifier (Zafar et al.) [5], and a threshold classifier method with
protected characteristic-specific thresholds on the output of a logistic regression model (Hardt et
al.) [4]. All methods but Hardt et al. do not use protected characteristics at prediction time.
Optimization procedure For our DM and DM+ methods, we identify at least three options on how
to optimize the multi-objective optimization problem in (7): using Convex Concave Procedure (CCP),
using Broyden-Fletcher-Goldfarb-Shanno gradient descent method with limited-memory variation
(L-BFGS), and using evolutionary multi-objective optimization (EMO). We discuss those options
\2 (p , p ) fairness constraint (7b)
in turn. First, we can express each additive term in the MMD
z=0 z=1
as a difference of two convex functions, find the convex upper bound of each term, and place the
convexified fairness constraint as part of the objective function. In our initial experiments, solving
(7) with CCP tends to ignore the fairness constraint, therefore we do not explore this approach
further. As mentioned earlier, the convex upper bounds on each of the additive terms in the MMD
constraint become increasingly loose as we move away from the current point of approximation. This
leads to the second optimization approach. We turn the constrained optimization problem into an
\2 (p , p ) term.
unconstrained one by introducing a non-negative weight CMMD to scale the MMD
z=0 z=1
We then solve this unconstrained problem using L-BFGS. The main challenge with this procedure
is the need to trade-off multiple competing goals by tuning several hyper-parameters, which will
be discussed in the next section. The CCP and L-BFGS procedures will only return one optimal
solution from the Pareto frontier. Third, to approximate the Pareto-optimal set, we can instead use
EMO procedures (e.g. Non-dominated Sorting Genetic Algorithm (NSGA) ? II and Strength Pareto
Evolutionary Algorithm (SPEA) - II). For the EMO, we also solve the unconstrained problem as
in the second approach, but we do not need to introduce a trade-off parameter for each term in the
objective function. We use the DEAP toolbox [42] for experimenting with EMO.
Model selection For the baseline Zafar et al., as in [5], we set the hyper-parameters ? and
? corresponding to the Penalty Convex Concave procedure to 5.0 and 1.2, respectively. Gaussian RBF kernel with a kernel width ? 2 is used for the MMD term. When solving DM (and
DM+) optimization problems with L-BFGS, the hyper-parameters C, CMMD , ? 2 , (and ?) are set
to 1000., 5000., 10., (and 1.) for both datasets. For DM+, we select ? over the range {1., 2., . . . , 10.}
7
Table 1: Results on multi-objective optimization which balances two main objectives: performance
accuracies and fairness criteria. Equal true positive rates are required for ProPublica COMPAS
dataset, and equal accuracies between two demographics z = 0 and z = 1 are required for Adult
dataset. The solver of Zafar et al. fails on the Adult dataset while enforcing equal accuracies
across the two demographics. Hardt et al.?s method does not enforce equal accuracies. SVM and
LR only optimize performance accuracies. The terms |Acc.z=0 - Acc.z=1 |, |TPRz=0 - TPRz=1 |, and
|FPRz=0 - FPRz=1 | denote accuracy, true positive rate, and false positive rate discrepancies in an
absolute term between the two demographics (the smaller the fairer). For ProPublica COMPAS
dataset, we boldface |TPRz=0 - TPRz=1 | since we enforce the equality of opportunity criterion on
this dataset. For Adult dataset, we boldface |Acc.z=0 - Acc.z=1 | since this is the fairness criterion.
ProPublica COMPAS dataset (Fairness Constraint on equal TPRs)
|Acc.z=0 - Acc.z=1 | |TPRz=0 - TPRz=1 |
LR
0.0151?0.0116
0.2504?0.0417
SVM
0.0172?0.0102
0.2573?0.0158
Zafar et al.
0.0174?0.0142
0.1144?0.0482
Hardt et al.?
0.0219?0.0191
0.0463?0.0185
DM (L-BFGS)
0.0457?0.0289
0.1169?0.0690
DM+ (L-BFGS)
0.0608?0.0259
0.1065?0.0413
DM (EMO Usr1)
0.0537?0.0121
0.1346?0.0360
DM (EMO Usr2)
0.0535?0.0213
0.1248?0.0509
?
use protected characteristics at prediction time.
|FPRz=0 - FPRz=1 |
0.1618?0.0471
0.1603?0.0490
0.1914?0.0314
0.0518?0.0413
0.0791?0.0395
0.0973?0.0272
0.1028?0.0481
0.0906?0.0507
Acc.
0.6652?0.0139
0.6367?0.0212
0.6118?0.0198
0.6547?0.0128
0.5931?0.0599
0.6089?0.0398
0.6261?0.0133
0.6148?0.0137
Adult dataset (Fairness Constraint on equal accuracies)
SVM
DM (L-BFGS)
DM+ (L-BFGS)
DM (EMO Usr1)
DM (EMO Usr2)
|Acc.z=0 - Acc.z=1 |
0.1136?0.0064
0.0640?0.0280
0.0459?0.0372
0.0388?0.0179
0.0482?0.0143
|TPRz=0 - TPRz=1 |
0.0964?0.0289
0.0804?0.0659
0.0759?0.0738
0.0398?0.0284
0.0302?0.0212
|FPRz=0 - FPRz=1 |
0.0694?0.0109
0.0346?0.0343
0.0368?0.0349
0.0398?0.0284
0.0135?0.0056
Acc.
0.8457?0.0034
0.8152?0.0068
0.8127?0.0134
0.8057?0.0108
0.8111?0.0122
using 5-fold cross validation. The selection process goes as follow: we first sort ? values according
to how well they satisfy the fairness criterion, we then select a ? value at a point before it yields
a lower incremental classification accuracy. As stated earlier, we do not need C, CMMD , ? 2 , ?, ?
hyper-parameters for balancing multiple terms in the objective function when using EMO for DM
and DM+. There are however several free parameters related to the evolutionary algorithm itself.
We use the NSGA ? II selection strategy with a polynomial mutation operator as in the the original
implementation [43], and the mutation probability is set to 0.5. We do not use any mating operator.
We use 500 individuals in a loop of 50 iterations (generations).
Results Experimental results over 5 repeats are presented in Table 1. In the ProPublica COMPAS
dataset, we enforce equality of opportunity |TPRz=0 - TPRz=1 |, i.e. equal true positive rates (Equation
(3)), as the fairness criterion (refer to the ProPublica COMPAS dataset in Table 1). Additionally, our
distribution matching methods, DM+ and DM also deliver a reduction in discrepancies between false
positive rates. We experiment with both L-BFGS and EMO optimization procedures. For EMO, we
simulate two decision makers choosing an operating point based on the visualization of Pareto frontier
in Figure 1 ? Right (shown as DM (EMO Usr1) and DM (EMO Usr2) in Table 1). For this dataset,
Usr1 has an inclination to be more lenient in being fair for a gain in accuracy in comparison to the
Usr2. This is actually reflected in the selection of the operating point (see supplementary material).
The EMO is run on the 60% of the training data, the selection is done on the remaining 40%, and the
reported results are on the separate test set based on the model trained on the 60% of the training data.
The method Zafar et al. achieves similar reduction rate in the fairness criterion to our distribution
matching methods. As a reference, we also include results of Hardt et al.?s method; it achieves
the best equality of opportunity measure with only a slight drop in accuracy performance w.r.t. the
unfair LR. It is important to note that Hardt et al.?s method requires protected characteristics at
test time. If we allow the usage of protected characteristics at test time, we should expect similar
reduction rate in fairness and accuracy measures for other methods [5].
8
SVM
0.25
|T P RZ=0 ? T P RZ=1 |
0.06
0.04
0.20
0.02
MMD0.00
0.15
0.02
0.04
0.06
0.08
0.10
0.05
0
Re 5 10
gu 15
lar 20 25
iza 30
tio 35
n 40
2.50
2.25
1.50
1.25
1.00
0.00
0.75
Loss
Hinge
2.00
1.75
0.35
0.40
0.45
0.50
Classification Error
0.55
0.60
Figure 1: Visualization of a Pareto frontier of our DM method for the ProPublica COMPAS dataset.
Left: In a 3D criterion space corresponding to the three objective functions: hinge loss, i.e.
2
\2 (p , p ).
max (0, 1 ? yn [hw, xn i + b]), regularization, i.e. kwk`2 , and MMD, i.e. MMD
z=0 z=1
Fairer models (smaller MMD values) are gained at the expense of model complexity (higher regularization and/or hinge loss values). Note that the unbiased estimate of MMD may be negative [37].
Right: The same Pareto frontier but in a 2D space of error and unfairness in predictions. Only the
first repeat is visualized; please refer to the supplementary material for the other four repeats, for the
Adult dataset, and of the DM+ method.
In the Adult dataset, we enforce equal accuracies |Acc.z=0 - Acc.z=1 | (Equation (4)) as the fairness
criterion (refer to the Adult dataset in Table 1). The method whereby a decision maker uses a
Pareto frontier visualization for choosing the operating point (DM (EMO Usr1)) reaches the smallest
discrepancy between the two demographics. In addition to equal accuracies (Equation (4)), our
distribution matching methods, DM+ and DM, also deliver a reduction in discrepancies between true
positive and false positive rates w.r.t. SVM (second and third column). In this dataset, Zafar et
al. falls into numerical problems when enforcing equal accuracies (vide our earlier discussion on
different optimization procedures, especially related to CCP). As observed in prior work [5, 20],
the methods that do not enforce fairness (equal accuracies or equal true positive rates), SVM and LR,
achieve higher classification accuracy compared to the methods that do enforce fairness: Zafar et
al., DM+, and DM. This can be seen in the last column of Table 1.
5
Discussion and Conclusion
We have proposed a unified machine learning framework that is able to handle any definitions of
fairness, e.g. fairness through unawareness, demographic parity, equalized odds, and equality of opportunity. Our framework is based on learning using privileged information and matching conditional
distributions using a two-sample problem. By using distance measures in Hilbert Space to solve the
two-sample problem, our framework is general and will be applicable for protected characteristics
with binary/multi-class/continuous values. The current work focuses on a single binary protected
characteristic. This corresponds to conditional distribution matching with a binary conditioning
variable. To generalize this to any type and multiple dependence of protected characteristics, we can
use the Hilbert Space embedding of conditional distributions framework of [44, 45].
We note that there are important factors external to machine learning models that are relevant to
fairness. However, this paper adopts the established approach of existing work on fair machine
learning. In particular, it is taken as given that one typically does not have any control over the data
collection process because there is no practical way of enforcing truth/un-biasedness in datasets that
are generated by others, such as banks, police forces, and companies.
Acknowledgments
NQ is supported by the UK EPSRC project EP/P03442X/1 ?EthicalML: Injecting Ethical and Legal
Constraints into Machine Learning Models? and the Russian Academic Excellence Project ?5-100?.
We gratefully acknowledge NVIDIA for GPU donation and Amazon for AWS Cloud Credits. We
also thank Kristian Kersting and Oliver Thomas for helpful discussions, Muhammad Bilal Zafar for
sharing his implementations of [4] and [5], and Sienna Quadrianto for supporting the work.
9
References
[1] Executive Office of the President. Big data: A report on algorithmic systems, opportunity, and
civil rights. Technical report, 2016.
[2] The Royal Society Working Group. Machine learning: the power and promise of computers
that learn by example. Technical report, 2017.
[3] Solon Barocas and Andrew D. Selbst. Big data?s disparate impact. California Law Review,
104:671?732, 2016.
[4] Moritz Hardt, Eric Price, and Nati Srebro. Equality of opportunity in supervised learning. In
D. D. Lee, M. Sugiyama, U. V. Luxburg, I. Guyon, and R. Garnett, editors, Advances in Neural
Information Processing Systems (NIPS) 29, pages 3315?3323, 2016.
[5] Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez-Rodriguez, and Krishna P. Gummadi.
Fairness beyond disparate treatment & disparate impact: Learning classification without disparate mistreatment. In International Conference on World Wide Web (WWW), pages 1171?1180,
2017.
[6] Equality Act. London: HMSO, 2010.
[7] Salvatore Ruggieri, Dino Pedreschi, and Franco Turini. Data mining for discrimination discovery.
ACM Transactions on Knowledge Discovery from Data (TKDD), 4:9:1?9:40, 2010.
[8] Philip Adler, Casey Falk, Sorelle A Friedler, Gabriel Rybeck, Carlos Scheidegger, Brandon
Smith, and Suresh Venkatasubramanian. Auditing black-box models for indirect influence. In
ICDM, 2016.
[9] Andrea Romei and Salvatore Ruggieri. A multidisciplinary survey on discrimination analysis.
The Knowledge Engineering Review, pages 582?638, 2014.
[10] Michael Feldman, Sorelle A Friedler, John Moeller, Carlos Scheidegger, and Suresh Venkatasubramanian. Certifying and removing disparate impact. In International Conference on
Knowledge Discovery and Data Mining (KDD), pages 259?268, 2015.
[11] Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez Rodriguez, and Krishna P. Gummadi.
Fairness Constraints: Mechanisms for Fair Classification. In Aarti Singh and Jerry Zhu, editors,
International Conference on Artificial Intelligence and Statistics (AISTATS), volume 54 of
Proceedings of Machine Learning Research, pages 962?970. PMLR, 2017.
[12] Toon Calders, Faisal Kamiran, and Mykola Pechenizkiy. Building classifiers with independency
constraints. In International Conference on Data Mining Workshops (ICDMW), pages 13?18,
2009.
[13] Toshihiro Kamishima, Shotaro Akaho, and Jun Sakuma. Fairness-aware learning through
regularization approach. In International Conference on Data Mining Workshops (ICDMW),
pages 643?650, 2011.
[14] Toshihiro Kamishima, Shotaro Akaho, Hideki Asoh, and Jun Sakuma. Fairness-aware classifier with prejudice remover regularizer. In European conference on Machine Learning and
Knowledge Discovery in Databases (ECML PKDD), pages 35?50, 2012.
[15] Jon M. Kleinberg, Sendhil Mullainathan, and Manish Raghavan. Inherent trade-offs in the fair
determination of risk scores. CoRR, abs/1609.05807, 2016.
[16] Vladimir Vapnik and Akshay Vashist. A new learning paradigm: Learning using privileged
information. Neural Networks, pages 544?557, 2009.
[17] Viktoriia Sharmanska, Novi Quadrianto, and Christoph H. Lampert. Learning to rank using
privileged information. In International Conference on Computer Vision (ICCV), pages 825?832,
2013.
10
[18] Daniel Hern?ndez-Lobato, Viktoriia Sharmanska, Kristian Kersting, Christoph H Lampert, and
Novi Quadrianto. Mind the nuisance: Gaussian process classification using privileged noise.
In Z. Ghahramani, M. Welling, C. Cortes, N. D. Lawrence, and K. Q. Weinberger, editors,
Advances in Neural Information Processing Systems (NIPS) 27, pages 837?845, 2014.
[19] Vladimir Vapnik and Rauf Izmailov. Learning using privileged information: similarity control
and knowledge transfer. Journal of Machine Learning Research (JMLR), 16:2023?2049, 2015.
[20] Richard Zemel, Yu Wu, Kevin Swersky, Toniann Pitassi, and Cynthia Dwork. Learning fair
representations. In S. Dasgupta and D. McAllester, editors, International Conference on
Machine Learning (ICML), Proceedings of Machine Learning Research, pages 325?333. PMLR,
2013.
[21] Christos Louizos, Kevin Swersky, Yujia Li, Max Welling, and Richard Zemel. The variational
fair autoencoder. CoRR, abs/1511.00830, 2015.
[22] Binh Thanh Luong, Salvatore Ruggieri, and Franco Turini. k-NN as an implementation of
situation testing for discrimination discovery and prevention. In International Conference on
Knowledge Discovery and Data Mining (KDD), pages 502?510, 2011.
[23] Matt J. Kusner, Joshua R. Loftus, Chris Russell, and Ricardo Silva. Counterfactual fairness.
In U. V. Luxburg, I. Guyon, S. Bengio, H. Wallach, R. Fergus, S.V.N. Vishwanathan, and
R. Garnett, editors, Advances in Neural Information Processing Systems (NIPS) 30, 2017.
[24] Gabriel Goh, Andrew Cotter, Maya Gupta, and Michael P Friedlander. Satisfying real-world
goals with dataset constraints. In D. D. Lee, M. Sugiyama, U. V. Luxburg, I. Guyon, and
R. Garnett, editors, Advances in Neural Information Processing Systems (NIPS) 29, pages
2415?2423, 2016.
[25] Cynthia Dwork, Moritz Hardt, Toniann Pitassi, Omer Reingold, and Richard Zemel. Fairness
through awareness. In Innovations in Theoretical Computer Science (ITCS), pages 214?226.
ACM, 2012.
[26] Matthew Joseph, Michael Kearns, Jamie H Morgenstern, and Aaron Roth. Fairness in learning:
Classic and contextual bandits. In D. D. Lee, M. Sugiyama, U. V. Luxburg, I. Guyon, and
R. Garnett, editors, Advances in Neural Information Processing Systems (NIPS) 29, pages
325?333, 2016.
[27] Dmitry Pechyony and Vladimir Vapnik. On the theory of learnining with privileged information.
In John D. Lafferty, Christopher K. I. Williams, John Shawe-Taylor, Richard S. Zemel, and
Aron Culotta, editors, Advances in Neural Information Processing Systems (NIPS) 23, pages
1894?1902, 2010.
[28] Jan Feyereisl, Suha Kwak, Jeany Son, and Bohyung Han. Object localization based on structural
svm using privileged information. In Zoubin Ghahramani, Max Welling, Corinna Cortes, Neil D.
Lawrence, and Kilian Q. Weinberger, editors, Advances in Neural Information Processing
Systems (NIPS) 27, pages 208?216, 2014.
[29] Viktoriia Sharmanska, Daniel Hern?ndez-Lobato, Jos? Miguel Hern?ndez-Lobato, and Novi
Quadrianto. Ambiguity helps: Classification with disagreements in crowdsourced annotations.
In Computer Vision and Pattern Recognition (CVPR), pages 2194?2202. IEEE Computer
Society, 2016.
[30] D. Lopez-Paz, B. Sch?lkopf, L. Bottou, and V. Vapnik. Unifying distillation and privileged
information. In International Conference on Learning Representations (ICLR), 2016.
[31] Geoffrey E. Hinton, Oriol Vinyals, and Jeffrey Dean. Distilling the knowledge in a neural
network. CoRR, abs/1503.02531, 2015.
[32] Cristian Bucila, Rich Caruana, and Alexandru Niculescu-Mizil. Model compression. In
International Conference on Knowledge Discovery and Data Mining (KDD), pages 535?541,
2006.
11
[33] Wen Li, Lixin Duan, Dong Xu, and Ivor W Tsang. Learning with augmented features for
supervised and semi-supervised heterogeneous domain adaptation. IEEE Transactions on
Pattern Analysis and Machine Intelligence (TPAMI), pages 1134?1148, 2014.
[34] Xu Zhang, Felix Xinnan Yu, Shih-Fu Chang, and Shengjin Wang. Deep transfer network:
Unsupervised domain adaptation. CoRR, abs/1503.00591, 2015.
[35] Novi Quadrianto, James Petterson, and Alex J. Smola. Distribution matching for transduction.
In Y. Bengio, D. Schuurmans, J. D. Lafferty, C. K. I. Williams, and A. Culotta, editors, Advances
in Neural Information Processing Systems (NIPS) 22, pages 1500?1508. 2009.
[36] Viktoriia Sharmanska and Novi Quadrianto. Learning from the mistakes of others: Matching
errors in cross-dataset learning. In Computer Vision and Pattern Recognition (CVPR), pages
3967?3975. IEEE Computer Society, 2016.
[37] Arthur Gretton, Karsten M. Borgwardt, Malte J. Rasch, Bernhard Sch?lkopf, and Alexander
Smola. A kernel two-sample test. Journal of Machine Learning Research (JMLR), 13:723?773,
2012.
[38] Sinno Jialin Pan, Ivor W. Tsang, James T. Kwok, and Qiang Yang. Domain adaptation via
transfer component analysis. In Craig Boutilier, editor, International Joint Conference on
Artifical Intelligence (IJCAI), pages 1187?1192, 2009.
[39] Bharath K. Sriperumbudur, Kenji Fukumizu, and Gert R. G. Lanckriet. Universality, characteristic kernels and RKHS embedding of measures. Journal of Machine Learning Research
(JMLR), 12:2389?2410, 2011.
[40] Yann Collette and Patrick Siarry. Multiobjective Optimization: Principles and Case Studies.
Springer, 2003.
[41] Thomas Lipp and Stephen Boyd. Variations and extension of the convex?concave procedure.
Optimization and Engineering, 17(2):263?287, 2016.
[42] F?lix-Antoine Fortin, Fran?ois-Michel De Rainville, Marc-Andr? Gardner Gardner, Marc
Parizeau, and Christian Gagn?. DEAP: Evolutionary algorithms made easy. Journal of Machine
Learning Research (JMLR), 13(1):2171?2175, 2012.
[43] Kalyanmoy Deb, Samir Agrawal, Amrit Pratap, and T. Meyarivan. A fast elitist non-dominated
sorting genetic algorithm for multi-objective optimisation: NSGA-II. In International Conference on Parallel Problem Solving from Nature (PPSN), pages 849?858, 2000.
[44] Le Song, Jonathan Huang, Alex Smola, and Kenji Fukumizu. Hilbert space embeddings of
conditional distributions with applications to dynamical systems. In International Conference
on Machine Learning (ICML), pages 961?968, 2009.
[45] L. Song, K. Fukumizu, and A. Gretton. Kernel embeddings of conditional distributions: A
unified kernel framework for nonparametric inference in graphical models. IEEE Signal
Processing Magazine, 30(4):98?111, 2013.
12
| 6670 |@word compression:2 polynomial:1 fairer:2 p0:2 harder:1 reduction:4 initial:1 venkatasubramanian:2 ndez:3 score:1 united:2 moment:1 daniel:2 contains:1 rkhs:2 sendhil:1 genetic:2 ours:2 bilal:3 existing:6 current:2 comparing:1 contextual:1 manuel:2 com:1 yet:1 gmail:1 must:4 gpu:1 universality:1 john:3 additive:3 numerical:1 permeated:1 shape:1 christian:1 kdd:3 designed:1 propublica:8 drop:1 discrimination:3 alone:2 intelligence:3 nq:1 smith:1 lr:5 simpler:1 zhang:2 accessed:1 harmonize:1 guard:2 iverson:1 direct:2 become:2 lopez:1 consists:1 advocate:1 pathway:1 combine:1 vide:1 privacy:1 introduce:3 excellence:1 expected:2 indeed:2 hardness:4 pkdd:1 xz:42 faceted:2 multi:15 karsten:1 andrea:1 relying:1 company:1 automatically:1 duan:1 solver:1 project:3 matched:1 notation:1 underlying:1 sinno:1 unfairness:9 morgenstern:1 unified:4 binarized:1 act:1 concave:4 charge:1 returning:1 classifier:12 uk:2 control:2 utilization:1 yn:16 before:3 felix:1 engineering:2 multiobjective:1 positive:31 tends:1 mistake:1 limit:1 black:2 might:2 nz:12 studied:1 accountability:1 wallach:1 christoph:2 limited:1 analytics:1 range:1 confidentiality:2 acknowledgment:1 camera:3 enforces:1 practical:1 testing:1 block:2 procedure:11 suresh:2 jan:1 universal:1 evolving:1 empirical:2 matching:37 boyd:1 confidence:3 refers:1 seeing:1 hmso:1 zoubin:1 get:2 ccp:6 cannot:1 selection:5 operator:2 context:4 risk:5 influence:1 optimize:2 www:1 dean:1 equivalent:2 thanh:1 roth:1 lobato:3 educational:1 economics:1 regardless:1 independently:1 williams:2 survey:1 go:2 modifies:1 amazon:1 overarching:1 convex:11 rule:1 his:1 classic:1 handle:2 notion:1 embedding:2 hurt:1 president:1 gert:1 variation:2 target:17 suppose:1 magazine:1 us:2 lanckriet:1 satisfying:2 recognition:2 expensive:1 distributional:1 database:1 mistreatment:1 observed:1 epsrc:1 ep:1 cloud:1 lix:1 wang:1 capture:1 tsang:2 solved:1 culotta:2 ensures:1 kilian:1 trade:4 russell:1 learnining:1 mentioned:1 convexity:2 respecting:1 constrains:1 complexity:2 employment:1 trained:2 raise:1 solving:7 singh:1 compromise:1 predictive:2 deliver:6 localization:1 upon:1 eric:1 pedreschi:1 completely:1 gu:1 accelerate:2 easily:1 indirect:1 joint:1 isabel:2 represented:1 arrested:1 regularizer:1 laser:2 instantiated:1 fast:1 describe:2 london:3 artificial:1 equalized:9 zemel:5 hyper:5 choosing:2 sanction:1 outcome:13 kevin:2 apparent:1 richer:1 posed:1 larger:1 cvpr:2 supplementary:2 solve:3 otherwise:1 ability:1 statistic:1 remover:1 neil:1 itself:3 final:1 cristian:1 housing:1 advantage:1 pressing:1 biasedness:1 agrawal:1 tpami:1 propose:4 jamie:1 adaptation:5 relevant:1 loop:1 moeller:1 omer:1 achieve:6 marital:2 description:1 everyday:1 ijcai:1 convergence:1 requirement:1 incremental:1 object:1 help:1 donation:1 andrew:2 develop:2 pose:1 ac:1 miguel:1 school:1 ex:1 kenji:2 ois:1 implies:1 distilling:1 rasch:1 direction:1 alexandru:1 modifying:1 stipulated:1 subsequently:2 raghavan:1 mcallester:1 material:2 muhammad:3 disproportionately:1 require:1 behaviour:1 recidivated:1 alleviate:1 frontier:9 extension:1 scanner:2 brandon:1 credit:3 prohibited:1 fletcher:1 algorithmic:1 lawrence:2 mapping:1 predict:1 matthew:1 week:1 sought:1 achieves:2 smallest:1 friedler:2 aarti:1 injecting:2 applicable:1 label:4 maker:3 cotter:1 minimization:1 fukumizu:3 offs:1 gaussian:3 aim:1 pn:1 asoh:1 avoid:1 kersting:2 office:1 encode:1 focus:4 casey:1 rank:1 likelihood:1 kwak:1 experimenting:1 contrast:1 baseline:2 dollar:1 sense:1 helpful:1 inference:1 deap:2 dependent:1 nn:1 niculescu:1 typically:1 accept:1 bandit:1 going:1 provably:1 issue:1 classification:10 among:1 prevention:1 constrained:1 softmax:1 special:4 equal:18 aware:2 having:1 beach:1 qiang:1 kw:1 represents:1 yu:2 novi:6 icml:2 fairness:57 unsupervised:1 jon:1 discrepancy:12 future:7 others:4 spline:1 report:3 serious:1 richard:4 barocas:1 wen:1 inherent:1 national:1 petterson:1 falk:1 individual:13 jeffrey:1 ab:4 interest:1 dwork:3 mining:6 bracket:1 admitting:1 regularizers:1 jialin:1 jeany:1 oliver:1 nowadays:1 auditing:1 fu:1 arthur:1 mullainathan:1 preferential:1 ynew:1 taylor:1 re:1 goh:1 theoretical:1 increased:2 classify:6 instance:18 column:2 earlier:5 xinnan:1 soft:1 disadvantage:1 caruana:1 zn:7 introducing:1 addressing:3 predictor:11 paz:1 pal:1 reported:1 learnt:1 adler:1 st:1 borgwardt:1 shanno:1 international:13 person:1 lee:3 dong:1 off:3 jos:1 michael:3 together:1 again:1 ambiguity:1 management:1 choose:1 huang:1 russia:1 external:1 luong:1 inject:1 return:1 michel:1 manish:1 li:2 ricardo:1 potential:2 account:1 de:2 bfgs:9 includes:2 availability:2 satisfy:1 explicitly:1 race:4 aron:1 lab:1 kwk:2 characterizes:1 portion:1 sort:1 option:2 parallel:1 capability:1 annotation:3 carlos:2 crowdsourced:1 mutation:2 minimize:4 formed:1 accuracy:16 who:1 likewise:1 succession:1 characteristic:56 yield:1 identify:1 generalize:1 lkopf:2 bayesian:1 itcs:1 vashist:1 accurately:1 craig:1 pechyony:1 classified:1 acc:12 bharath:1 mmd2:1 reach:1 whenever:1 sharing:2 mating:1 definition:17 sriperumbudur:1 against:2 james:2 obvious:1 dm:30 associated:1 gain:1 ruggieri:3 dataset:24 hardt:12 treatment:9 counterfactual:1 knowledge:12 feyereisl:1 hilbert:6 formalize:1 actually:1 higher:4 supervised:22 follow:3 reflected:2 modal:1 improved:1 formulation:1 done:1 box:1 just:1 smola:3 equipping:1 correlation:1 working:1 web:1 christopher:1 rodriguez:2 widespread:1 logistic:4 multidisciplinary:1 lar:1 russian:1 usage:1 facilitate:1 usa:1 requiring:3 true:18 building:5 matt:1 unbiased:2 equality:16 regularization:3 jerry:1 moritz:2 goldfarb:1 deal:1 conditionally:2 white:1 during:1 width:1 nuisance:1 uniquely:1 please:1 illustrative:1 noted:1 whereby:1 sussex:2 criterion:15 brighton:1 bmin:1 delivers:1 silva:1 image:3 variational:1 recently:2 common:2 conditioning:1 volume:1 discussed:1 slight:1 analog:1 louizos:1 refer:4 distillation:2 composition:1 feldman:1 imposing:1 broyden:1 sorelle:2 rd:3 unconstrained:3 tuning:1 akaho:2 sugiyama:3 gratefully:1 dino:1 shawe:1 convexified:1 access:1 han:1 similarity:2 operating:4 pitassi:2 add:1 base:2 patrick:1 recent:2 store:1 certain:1 nvidia:1 parizeau:1 binary:16 discussing:1 life:1 joshua:1 devise:1 krishna:2 seen:2 additional:3 paradigm:1 signal:1 ii:5 semi:1 stephen:1 multiple:3 gretton:2 transparency:1 infer:1 recycles:2 technical:2 match:7 academic:1 determination:1 pechenizkiy:1 long:1 profiling:1 cross:3 icdm:1 post:2 gummadi:2 privileged:47 laplacian:1 impact:17 ensuring:1 prediction:15 basic:1 regression:3 heterogeneous:1 vision:3 metric:3 optimisation:1 variant:2 iteration:1 kernel:13 faisal:1 mmd:18 penalize:1 addition:4 remarkably:1 scheidegger:2 compas:8 aws:1 source:3 modality:1 appropriately:1 sch:2 subject:2 reingold:1 bohyung:1 lafferty:2 odds:9 structural:1 nsga:3 yang:1 ideal:5 bengio:2 easy:4 embeddings:2 automated:3 iii:1 identically:1 approaching:1 competing:2 reduce:1 idea:2 prototype:2 favour:2 whether:2 colour:1 accelerating:1 penalty:1 song:2 constitute:1 deep:2 boutilier:1 gabriel:2 detailed:1 covered:1 amount:1 nonparametric:2 kamiran:1 visualized:1 reduced:2 deliberate:1 andr:1 sign:6 accountable:1 correctly:1 per:1 dasgupta:1 promise:1 steepness:1 express:1 group:11 shih:1 four:1 easiness:4 threshold:3 independency:1 drawn:1 loftus:1 achieving:1 imperial:1 nonconvexity:1 year:1 enforced:1 luxburg:4 run:1 defendant:1 selbst:1 sakuma:2 place:1 swersky:2 guyon:4 wu:1 yann:1 fran:1 draw:1 decision:31 accelerates:1 bound:4 maya:1 gomez:2 distinguish:1 xnew:1 fold:1 strength:1 constraint:32 vishwanathan:1 deb:1 alex:2 encodes:1 certifying:1 dominated:2 kleinberg:2 aspect:1 simulate:1 franco:2 min:7 department:1 structured:1 according:1 combination:2 unawareness:6 recycle:3 across:17 describes:1 smaller:2 increasingly:1 son:1 kusner:1 pan:2 joseph:2 intimately:1 iccv:1 taken:1 legal:4 equation:3 visualization:3 mutually:1 calder:1 hern:3 loose:1 mechanism:1 count:1 slack:1 emo:15 mind:1 discus:1 turn:2 needed:1 demographic:23 know:1 available:6 kwok:1 away:1 enforce:10 disagreement:1 salvatore:3 pmlr:2 alternative:1 corinna:1 weinberger:2 shotaro:2 rz:2 original:1 moscow:1 remaining:1 ensure:2 cf:1 include:1 graphical:1 opportunity:16 hinge:6 recycling:1 unifying:1 lixin:1 toon:1 lenient:1 scholarship:1 especially:1 ghahramani:2 classical:1 society:3 move:1 objective:20 parametric:2 strategy:1 dependence:1 map:1 exclusive:1 antoine:1 exhibit:1 gradient:1 iclr:1 subspace:1 evolutionary:4 separate:1 thank:1 distance:10 philip:1 chris:1 topic:1 enforcing:13 boldface:2 assuming:2 balance:2 minimizing:5 innovation:1 vladimir:3 kingdom:2 difficult:3 setup:1 acquire:1 expense:1 stated:1 negative:11 disparate:5 implementation:3 motivates:1 perform:1 upper:4 cmmd:3 observation:1 b2n:1 datasets:4 acknowledge:1 descent:1 behave:1 ecml:1 supporting:1 situation:2 hinton:2 y1:2 reproducing:2 omission:3 police:1 sharmanska:7 overlooked:1 cast:2 required:3 pair:1 toolbox:1 z1:1 california:1 inclination:1 hideki:1 established:3 hour:1 boost:1 nip:9 adult:9 beyond:4 able:4 dynamical:1 pattern:3 mismatch:1 yujia:1 challenge:1 program:1 royal:1 memory:1 max:9 including:1 power:1 suitable:2 misclassification:2 malte:1 rely:1 force:1 gagn:1 predicting:1 valera:2 natural:1 mizil:1 zhu:1 technology:2 x2i:7 offender:1 fortin:1 gardner:2 jun:2 elitist:1 coupled:1 autoencoder:1 prior:2 literature:1 review:2 discovery:10 friedlander:1 nati:1 regularisation:3 law:2 embedded:1 toniann:2 expect:1 loss:6 generation:1 limitation:1 srebro:1 geoffrey:1 viktoriia:5 age:1 validation:1 executive:1 awareness:1 solon:1 sufficient:1 proxy:5 principle:2 editor:11 viewpoint:1 bank:2 pareto:12 balancing:1 compatible:1 repeat:3 last:2 parity:9 free:1 qualified:1 supported:1 allow:2 fall:2 wide:1 taking:1 akshay:1 absolute:1 distributed:1 benefit:2 boundary:4 xn:18 world:2 rich:3 unaware:1 suha:1 forward:1 made:2 commonly:1 replicated:1 icdmw:2 adopts:1 founded:1 collection:2 turini:2 welling:3 transaction:2 income:2 approximate:2 ignore:1 status:2 dmitry:1 bernhard:1 xi:1 fergus:1 alternatively:1 un:2 latent:1 continuous:3 triplet:1 search:1 protected:53 table:6 additionally:1 toshihiro:2 learn:4 transfer:6 nature:1 ca:1 ignoring:1 unavailable:1 improving:1 schuurmans:1 futile:1 bottou:1 complex:1 european:1 necessarily:1 domain:9 zafar:15 garnett:4 marc:2 ppsn:1 tkdd:1 aistats:1 main:3 big:2 noise:1 lampert:2 quadrianto:9 fair:50 body:1 x1:3 xu:2 augmented:1 transduction:3 fashion:1 slow:1 binh:1 christos:1 fails:1 inferring:1 concatenating:1 samir:1 unfair:5 jmlr:4 third:2 learns:1 hw:9 removing:3 specific:5 emphasized:1 cynthia:2 shengjin:1 explored:2 pz:2 svm:13 gupta:1 thomas:2 closeness:1 cortes:2 incorporating:1 bucila:2 concern:1 false:20 adding:1 corr:4 gained:1 workshop:2 vapnik:5 tio:1 margin:6 sorting:2 civil:1 simply:1 explore:1 infinitely:1 ivor:2 vinyals:1 ethical:3 chang:1 kristian:2 gender:5 springer:1 truth:1 satisfies:3 corresponds:2 kamishima:2 acm:2 conditional:7 goal:12 acceleration:1 pnz:4 rbf:2 towards:1 price:1 hard:1 hyperplane:1 prejudice:1 degradation:1 lemma:1 total:2 lens:1 kearns:1 experimental:1 aaron:1 select:3 college:1 damaging:1 support:2 latter:1 jonathan:1 alexander:1 accelerated:2 oriol:1 artifical:1 correlated:2 |
6,269 | 6,671 | Safe and Nested Subgame Solving for
Imperfect-Information Games
Noam Brown
Computer Science Department
Carnegie Mellon University
Pittsburgh, PA 15217
[email protected]
Tuomas Sandholm
Computer Science Department
Carnegie Mellon University
Pittsburgh, PA 15217
[email protected]
Abstract
In imperfect-information games, the optimal strategy in a subgame may depend
on the strategy in other, unreached subgames. Thus a subgame cannot be solved
in isolation and must instead consider the strategy for the entire game as a whole,
unlike perfect-information games. Nevertheless, it is possible to first approximate
a solution for the whole game and then improve it in individual subgames. This
is referred to as subgame solving. We introduce subgame-solving techniques that
outperform prior methods both in theory and practice. We also show how to adapt
them, and past subgame-solving techniques, to respond to opponent actions that
are outside the original action abstraction; this significantly outperforms the prior
state-of-the-art approach, action translation. Finally, we show that subgame solving
can be repeated as the game progresses down the game tree, leading to far lower
exploitability. These techniques were a key component of Libratus, the first AI to
defeat top humans in heads-up no-limit Texas hold?em poker.
1
Introduction
Imperfect-information games model strategic settings that have hidden information. They have a
myriad of applications including negotiation, auctions, cybersecurity, and physical security.
In perfect-information games, determining the optimal strategy at a decision point only requires
knowledge of the game tree?s current node and the remaining game tree beyond that node (the
subgame rooted at that node). This fact has been leveraged by nearly every AI for perfect-information
games, including AIs that defeated top humans in chess [6] and Go [28]. In checkers, the ability to
decompose the game into smaller independent subgames was even used to solve the entire game [26].
However, it is not possible to determine a subgame?s optimal strategy in an imperfect-information
game using only knowledge of that subgame, because the game tree?s exact node is typically unknown.
Instead, the optimal strategy may depend on the value an opponent could have received in some other,
unreached subgame. Although this is counter-intuitive, we provide a demonstration in Section 2.
Rather than rely on subgame decomposition, past approaches for imperfect-information games
typically solved the game as a whole upfront. For example, heads-up limit Texas hold?em, a relatively
simple form of poker with 1013 decision points, was essentially solved without decomposition [2].
However, this approach cannot extend to larger games, such as heads-up no-limit Texas hold?em?the
primary benchmark in imperfect-information game solving?which has 10161 decision points [15].
The standard approach to computing strategies in such large games is to first generate an abstraction
of the game, which is a smaller version of the game that retains as much as possible the strategic
characteristics of the original game [23, 25, 24]. For example, a continuous action space might
be discretized. This abstract game is solved and its solution is used when playing the full game
by mapping states in the full game to states in the abstract game. We refer to the solution of an
abstraction (or more generally any approximate solution to a game) as a blueprint strategy.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
In heavily abstracted games, a blueprint strategy may be far from the true solution. Subgame solving
attempts to improve upon the blueprint strategy by solving in real time a more fine-grained abstraction
for an encountered subgame, while fitting its solution within the overarching blueprint strategy.
2
Coin Toss
In this section we provide intuition for why an imperfect-information subgame cannot be solved in
isolation. We demonstrate this in a simple game we call Coin Toss, shown in Figure 1a, which will
be used as a running example throughout the paper.
Coin Toss is played between players P1 and P2 . The figure shows rewards only for P1 ; P2 always
receives the negation of P1 ?s reward. A coin is flipped and lands either Heads or Tails with equal
probability, but only P1 sees the outcome. P1 then chooses between actions ?Sell? and ?Play.? The
Sell action leads to a subgame whose details are not important, but the expected value (EV) of
choosing the Sell action will be important. (For simplicity, one can equivalently assume in this
section that Sell leads to an immediate terminal reward, where the value depends on whether the
coin landed Heads or Tails). If the coin lands Heads, it is considered lucky and P1 receives an EV of
$0.50 for choosing Sell. On the other hand, if the coin lands Tails, it is considered unlucky and P1
receives an EV of ?$0.50 for action Sell. (That is, P1 must on average pay $0.50 to get rid of the
coin). If P1 instead chooses Play, then P2 may guess how the coin landed. If P2 guesses correctly,
then P1 receives a reward of ?$1. If P2 guesses incorrectly, then P1 receives $1. P2 may also forfeit,
which should never be chosen but will be relevant in later sections. We wish to determine the optimal
strategy for P2 in the subgame S that occurs after P1 chooses Play, shown in Figure 1a.
Figure 1: (a) The example game of Coin Toss. ?C? represents a chance node. S is a Player 2 (P2 ) subgame.
The dotted line between the two P2 nodes means that P2 cannot distinguish between them. (b) The public game
tree of Coin Toss. The two outcomes of the coin flip are only observed by P1 .
Were P2 to always guess Heads, P1 would receive $0.50 for choosing Sell when the coin lands Heads,
and $1 for Play when it lands Tails. This would result in an average of $0.75 for P1 . Alternatively,
were P2 to always guess Tails, P1 would receive $1 for choosing Play when the coin lands Heads,
and ?$0.50 for choosing Sell when it lands Tails. This would result in an average reward of $0.25 for
P1 . However, P2 would do even better by guessing Heads with 25% probability and Tails with 75%
probability. In that case, P1 could only receive $0.50 (on average) by choosing Play when the coin
lands Heads?the same value received for choosing Sell. Similarly, P1 could only receive ?$0.50 by
choosing Play when the coin lands Tails, which is the same value received for choosing Sell. This
would yield an average reward of $0 for P1 . It is easy to see that this is the best P2 can do, because
P1 can average $0 by always choosing Sell. Therefore, choosing Heads with 25% probability and
Tails with 75% probability is an optimal strategy for P2 in the ?Play? subgame.
Now suppose the coin is considered lucky if it lands Tails and unlucky if it lands Heads. That is,
the expected reward for selling the coin when it lands Heads is now ?$0.50 and when it lands Tails
is now $0.50. It is easy to see that P2 ?s optimal strategy for the ?Play? subgame is now to guess
Heads with 75% probability and Tails with 25% probability. This shows that a player?s optimal
strategy in a subgame can depend on the strategies and outcomes in other parts of the game. Thus,
one cannot solve a subgame using information about that subgame alone. This is the central challenge
of imperfect-information games as opposed to perfect-information games.
2
3
Notation and Background
In a two-player zero-sum extensive-form game there are two players, P = {1, 2}. H is the set of
all possible nodes, represented as a sequence of actions. A(h) is the actions available in a node and
P (h) ? P ? c is the player who acts at that node, where c denotes chance. Chance plays an action
a ? A(h) with a fixed probability. If action a ? A(h) leads from h to h0 , then we write h ? a = h0 . If
a sequence of actions leads from h to h0 , then we write h @ h0 . The set of nodes Z ? H are terminal
nodes. For each player i ? P, there is a payoff function ui : Z ? < where u1 = ?u2 .
Imperfect information is represented by information sets (infosets). Every node h ? H belongs
to exactly one infoset for each player. For any infoset Ii , nodes h, h0 ? Ii are indistinguishable to
player i. Thus the same player must act at all the nodes in an infoset, and the same actions must be
available. Let P (Ii ) and A(Ii ) be such that all h ? Ii , P (Ii ) = P (h) and A(Ii ) = A(h).
A strategy ?i (Ii ) is a probability vector over A(Ii ) for infosets where P (Ii ) = i. The probability of
action a is denoted by ?i (Ii , a). For all h ? Ii , ?i (h) = ?i (Ii ). A full-game strategy ?i ? ?i defines
a strategy for each player i infoset. A strategy profile ? is a tuple of strategies, one for each player.
The expected payoff for player i if all players play the strategy profile h?i , ??i i is ui (?i , ??i ), where
??i denotes the strategies in ? of all players other than i.
Q
Let ? ? (h) = h0 ?avh ?P (h0 ) (h0 , a) denote the probability of reaching h if all players play according
to ?. ?i? (h) is the contribution of player i to this probability (that is, the probability of reaching h if
?
chance and all players other than i always chose actions leading to h). ??i
(h) is the contribution of
?
0
all players, and chance, other than i. ? (h, h ) is the probability of reaching h0 given that h has been
reached, and 0 if h 6@ h0 . This papers focuses on perfect-recall games, where a player never forgets
past information. Thus, for every Ii , ?h, h0 ? Ii , ?i? (h) = ?i? (h0 ). We define ?i? (Ii ) = ?i? (h) for
h ? Ii . Also, Ii0 @ Ii if for some h0 ? Ii0 and some h ? Ii , h0 @ h. Similarly, Ii0 ? a @ Ii if h0 ? a @ h.
A Nash equilibrium [21] is a strategy profile ? ? where no player can improve by shifting to a different
?
?
strategy, so ? ? satisfies ?i, ui (?i? , ??i
) = max?i0 ??i ui (?i0 , ??i
). A best response BR(??i ) is a
strategy for player i that is optimal against ??i . Formally, BR(??i ) satisfies ui (BR(??i ), ??i ) =
max?i0 ??i ui (?i0 , ??i ). In a two-player zero-sum game, the exploitability exp(?i ) of a strategy ?i is
how much worse ?i does against an opponent best response than a Nash equilibrium strategy would
do. Formally, exploitability of ?i is ui (? ? ) ? ui (?i , BR(?i )), where ? ? is a Nash equilibrium.
P
The expected value of a node h when players play according to ? is vi? (h) = z?Z ? ? (h, z)ui (z) .
An infoset?s value is the weighted average of the values of the nodes in the infoset, where a node
P
is weighed by the player?s belief that she is in that node. Formally, vi? (Ii ) =
P
and vi? (Ii , a) =
h?Ii
?
??i
(h)vi? (h?a)
?
h?I ??i (h)
P
h?Ii
P
?
??i
(h)vi? (h)
? (h)
??i
h?Ii
. A counterfactual best response [20] CBR(??i ) is a best
i
response that also maximizes value in unreached infosets. Specifically, a counterfactual best response is a best response ?i with the additional condition that if ?i (Ii , a) > 0 then vi? (Ii , a) =
maxa0 vi? (Ii , a0 ). We further define counterfactual best response value CBV ??i (Ii ) as the value
player i expects to achieve by playing according to CBR(??i ), having already reached infoset Ii .
hCBR(??i ),??i i
hCBR(??i ),??i i
Formally, CBV ??i (Ii ) = vi
(Ii ) and CBV ??i (Ii , a) = vi
(Ii , a).
An imperfect-information subgame, which we refer to simply as a subgame in this paper, can in
most cases (but not all) be described as including all nodes which share prior public actions (that is,
actions viewable to both players). In poker, for example, a subgame is uniquely defined by a sequence
of bets and public board cards. Figure 1b shows the public game tree of Coin Toss. Formally, an
imperfect-information subgame is a set of nodes S ? H such that for all h ? S, if h @ h0 , then
h0 ? S, and for all h ? S and all i ? P, if h0 ? Ii (h) then h0 ? S. Define Stop as the set of
earliest-reachable nodes in S. That is, h ? Stop if h ? S and h0 6? S for any h0 @ h.
4
Prior Approaches to Subgame Solving
This section reviews prior techniques for subgame solving in imperfect-information games, which we
build upon. Throughout this section, we refer to the Coin Toss game shown in Figure 1a.
As discussed in Section 1, a standard approach to dealing with large imperfect-information games is
to solve an abstraction of the game. The abstract solution is a (probably suboptimal) strategy profile
3
in the full game. We refer to this full-game strategy profile as the blueprint. The goal of subgame
solving is to improve upon the blueprint by changing the strategy only in a subgame.
Figure 2: The blueprint strategy we refer to in the game of Coin Toss. The Sell action leads to a subgame that is
not displayed. Probabilities are shown for all actions. The dotted line means the two P2 nodes share an infoset.
The EV of each P1 action is also shown.
Assume that a blueprint strategy profile ? (shown in Figure 2) has already been computed for Coin
Toss in which P1 chooses Play 34 of the time with Heads and 12 of the time with Tails, and P2 chooses
Heads 12 of the time, Tails 14 of the time, and Forfeit 14 of the time after P1 chooses Play. The details
of the blueprint strategy in the Sell subgame are not relevant in this section, but the EV for choosing
the Sell action is relevant. We assume that if P1 chose the Sell action and played optimally thereafter,
then she would receive an expected payoff of 0.5 if the coin is Heads, and ?0.5 if the coin is Tails.
We will attempt to improve P2 ?s strategy in the subgame S that follows P1 choosing Play.
4.1 Unsafe Subgame Solving
We first review the most intuitive form of subgame solving, which we refer to as Unsafe subgame
solving [1, 11, 12, 9]. This form of subgame solving assumes both players played according to the
blueprint strategy prior to reaching the subgame. That defines a probability distribution over the
nodes at the root of the subgame S, representing the probability that the true game state matches that
node. A strategy for the subgame is then calculated which assumes that this distribution is correct.
In all subgame solving algorithms, an augmented subgame containing S and a few additional nodes
is solved to determine the strategy for S. Applying Unsafe subgame solving to the blueprint strategy
in Coin Toss (after P1 chooses Play) means solving the augmented subgame shown in Figure 3a.
Specifically, the augmented subgame consists of only an initial chance node and S. The initial chance
?
node reaches h ? Stop with probability P 0 ? (h)
? ? (h0 ) . The augmented subgame is solved and its
h ?Stop
strategy for P2 is used in S rather than the blueprint strategy.
Unsafe subgame solving lacks theoretical solution quality guarantees and there are many situations
where it performs extremely poorly. Indeed, if it were applied to the blueprint strategy of Coin Toss
then P2 would always choose Heads?which P1 could exploit severely by only choosing Play with
Tails. Despite the lack of theoretical guarantees and potentially bad performance, Unsafe subgame
solving is simple and can sometimes produce low-exploitability strategies, as we show later.
We now move to discussing safe subgame-solving techniques, that is, ones that ensure that the
exploitability of the strategy is no higher than that of the blueprint strategy.
(a) Unsafe subgame solving
(b) Resolve subgame solving
Figure 3: The augmented subgames solved to find a P2 strategy in the Play subgame of Coin Toss.
4
4.2 Subgame Resolving
In subgame Resolving [5], a safe strategy is computed for P2 in the subgame by solving the augmented
subgame shown in Figure 3b, producing an equilibrium strategy ? S . This augmented subgame differs
from Unsafe subgame solving by giving P1 the option to ?opt out? from entering S and instead
receive the EV of playing optimally against P2 ?s blueprint strategy in S.
Specifically, the augmented subgame for Resolving differs from unsafe subgame solving as follows.
For each htop ? Stop we insert a new P1 node hr , which exists only in the augmented subgame,
between the initial chance node and htop . The set of these hr nodes is Sr . The initial chance node
connects to each node hr ? Sr in proportion to the probability that player P1 could reach htop if P1
?
tried to do so (that is, in proportion to ??1
(htop )). At each node hr ? Sr , P1 has two possible actions.
0
0
Action aS leads to htop , while action aT leads to a terminal payoff that awards the value of playing
optimally against P2 ?s blueprint strategy, which is CBV ?2 (I1 (htop )). In the blueprint strategy of
Coin Toss, P1 choosing Play after the coin lands Heads results in an EV of 0, and 12 if the coin is
Tails. Therefore, a0T leads to a terminal payoff of 0 for Heads and 12 for Tails. After the equilibrium
strategy ? S is computed in the augmented subgame, P2 plays according to the computed subgame
strategy ?2S rather than the blueprint strategy when in S. The P1 strategy ?1S is not used.
Clearly P1 cannot do worse than always picking action a0T (which awards the highest EV P1 could
achieve against P2 ?s blueprint). But P1 also cannot do better than always picking a0T , because P2
could simply play according to the blueprint in S, which means action a0S would give the same EV to
P1 as action a0T (if P1 played optimally in S). In this way, the strategy for P2 in S is pressured to be
no worse than that of the blueprint. In Coin Toss, if P2 were to always choose Heads (as was the case
in Unsafe subgame solving), then P1 would always choose a0T with Heads and a0S with Tails.
Resolving guarantees that P2 ?s exploitability will be no higher than the blueprint?s (and may be
better). However, it may miss opportunities for improvement. For example, if we apply Resolving to
the example blueprint in Coin Toss, one solution to the augmented subgame is the blueprint itself, so
P2 may choose Forfeit 25% of the time even though Heads and Tails dominate that action. Indeed,
the original purpose of Resolving was not to improve upon a blueprint strategy in a subgame, but
rather to compactly store it by keeping only the EV at the root of the subgame and then reconstructing
the strategy in real time when needed rather than storing the whole subgame strategy.
Maxmargin subgame solving [20], discussed in Appendix A, can improve performance by definS
S
ing a margin M ? (I1 ) = CBV ?2 (I1 ) ? CBV ?2 (I1 ) for each I1 ? Stop and maximizing
S
minI1 ?Stop M ? (I1 ). Resolving only makes all margins nonnegative. However, Maxmargin does
worse in practice when using estimates of equilibrium values as discussed in Appendix C.
5
Reach Subgame Solving
All of the subgame-solving techniques described in Section 4 only consider the target subgame in
isolation, which can lead to suboptimal strategies. For example, Maxmargin solving applied to S
in Coin Toss results in P2 choosing Heads with probability 58 and Tails with 38 in S. This results in
P1 receiving an EV of ? 14 by choosing Play in the Heads state, and an EV of 14 in the Tails state.
However, P1 could simply always choose Sell in the Heads state (earning an EV of 0.5) and Play in
the Tails state and receive an EV of 38 for the entire game. In this section we introduce Reach subgame
solving, an improvement to past subgame-solving techniques that considers what the opponent could
have alternatively received from other subgames.1 For example, a better strategy for P2 would be
to choose Heads with probability 34 and Tails with probability 14 . Then P1 is indifferent between
choosing Sell and Play in both cases and overall receives an expected payoff of 0 for the whole game.
However, that strategy is only optimal if P1 would indeed achieve an EV of 0.5 for choosing Sell
in the Heads state and ?0.5 in the Tails state. That would be the case if P2 played according to the
blueprint in the Sell subgame (which is not shown), but in reality we would apply subgame solving to
the Sell subgame if the Sell action were taken, which would change P2 ?s strategy there and therefore
P1 ?s EVs. Applying subgame solving to any subgame encountered during play is equivalent to
applying it to all subgames independently; ultimately, the same strategy is played in both cases. Thus,
we must consider that the EVs from other subgames may differ from what the blueprint says because
subgame solving would be applied to them as well.
1
Other subgame-solving methods have also considered the cost of reaching a subgame [30, 14]. However,
those approaches are not correct in theory when applied in real time to any subgame reached during play.
5
Figure 4: Left: A modified game of Coin Toss with two subgames. The nodes C1 and C2 are public chance
nodes whose outcomes are seen by both P1 and P2 . Right: An augmented subgame for one of the subgames
according to Reach subgame solving. If only one of the subgames is being solved, then the alternative payoff
for Heads can be at most 1. However, if both are solved independently, then the gift must be split among the
subgames and must sum to at most 1. For example, the alternative payoff in both subgames can be 0.5.
As an example of this issue, consider the game shown in Figure 4 which contains two identical
subgames S1 and S2 where the blueprint has P2 pick Heads and Tails with 50% probability. The Sell
action leads to an EV of 0.5 from the Heads state, while Play leads to an EV of 0. If we were to solve
just S1 , then P2 could afford to always choose Tails in S1 , thereby letting P1 achieve an EV of 1
for reaching that subgame from Heads because, due to the chance node C1 , S1 is only reached with
50% probability. Thus, P1 ?s EV for choosing Play would be 0.5 from Heads and ?0.5 from Tails,
which is optimal. We can achieve this strategy in S1 by solving an augmented subgame in which the
alternative payoff for Heads is 1. In that augmented subgame, P2 always choosing Tails would be a
solution (though not the only solution).
However, if the same reasoning were applied independently to S2 as well, then P2 might always
choose Tails in both subgames and P1 ?s EV for choosing Play from Heads would become 1 while the
EV for Sell would only be 0.5. Instead, we could allow P1 to achieve an EV of 0.5 for reaching each
subgame from Heads (by setting the alternative payoff for Heads to 0.5). In that case, P1 ?s overall
EV for choosing Play could only increase to 0.5, even if both S1 and S2 were solved independently.
We capture this intuition by considering for each I1 ? Stop all the infosets and actions I10 ? a0 @ I1
that P1 would have taken along the path to I1 . If, at some I10 ? a0 @ I1 where P1 acted, there was a
different action a? ? A(I10 ) that leads to a higher EV, then P1 would have taken a suboptimal action
if they reached I1 . The difference in value between a? and a0 is referred to as a gift. We can afford
to let P1 ?s value for I1 increase beyond the blueprint value (and in the process lower P1 ?s value in
some other infoset in Stop ), so long as the increase to I1 ?s value is small enough that choosing actions
leading to I1 is still suboptimal for P1 . Critically, we must ensure that the increase in value is small
enough even when the potential increase across all subgames is summed together, as in Figure 4.2
A complicating factor is that gifts we assumed were present may actually not exist. For example, in
Coin Toss, suppose applying subgame solving to the Sell subgame results in P1 ?s value for Sell from
the Heads state decreasing from 0.5 to 0.25. If we independently solve the Play subgame, we have
no way of knowing that P1 ?s value for Sell is lower than the blueprint suggested, so we may still
assume there is a gift of 0.5 from the Heads state based on the blueprint. Thus, in order to guarantee a
theoretical result on exploitability that is as strong as possible, we use in our theory and experiments
a lower bound on what gifts could be after subgame solving was applied to all other subgames.
Formally, let ?2 be a P2 blueprint and let ?2?S be the P2 strategy that results from applying subgame solving independently to a set of disjoint subgames other than S. Since we do not want
?S
to compute ?2?S in order to apply subgame solving to S, let bg ?2 (I10 , a0 )c be a lower bound of
?S
?S
CBV ?2 (I10 ) ? CBV ?2 (I10 , a0 ) that does not require knowledge of ?2?S . In our experiments we
2
In this paper and in our experiments, we allow any infoset that descends from a gift to increase by the size
of the gift (e.g., in Figure 4 the gift from Heads is 0.5, so we allow P1 ?s value for Heads in both S1 and S2
to increase by 0.5). However, any division of the gift among subgames is acceptable so long as the potential
increase across all subgames (multiplied by the probability of P1 reaching that subgame) does not exceed the
original gift. For example in Figure 4 if we only apply Reach subgame solving to S1 , then we could allow the
Heads state in S1 to increase by 1 rather than just by 0.5. In practice, some divisions may do better than others.
The division we use in this paper (applying gifts equally to all subgames) did well in practice.
6
?S
use bg ?2 (I10 , a0 )c = maxa?Az (I10 )?{a0 } CBV ?2 (I10 , a) ? CBV ?2 (I10 , a0 ) where Az (I10 ) ? A(I10 )
is the set of actions leading immediately to terminal nodes. Reach subgame solving modifies the
augmented subgame in Resolving and Maxmargin by increasing the alternative payoff for infoset
P
?S
I1 ? Stop by I 0 ?a0 vI1 |P (I 0 )=P1 bg ?2 (I10 , a0 )c. Formally, we define a reach margin as
1
1
X
?S
S
S
Mr? (I1 ) = M ? (I1 ) +
bg ?2 (I10 , a0 )c
(1)
I10 ?a0 vI1 |P (I10 )=P1
?S
This margin is larger than or equal to the one for Maxmargin, because bg ?2 (I 0 , a0 )c is nonnegative.
We refer to the modified algorithms as Reach-Resolve and Reach-Maxmargin.
Using a lower bound on gifts is not necessary to guarantee safety. So long as we use a gift value
0
g ? (I10 , a0 ) ? CBV ?2 (I10 ) ? CBV ?2 (I10 , a0 ), the resulting strategy will be safe. However, using
a lower bound further guarantees a reduction to exploitability when a P1 best response reaches
with positive probability an infoset I1 ? Stop that has positive margin, as proven in Theorem 1. In
?S
practice, it may be best to use an accurate estimate of gifts. One option is to use g??2 (I10 , a0 ) =
?S
?
?2 0
?2 0
? 2 is the closest P1 can get to
?
?
CBV
(I1 ) ? CBV
(I1 , a0 ) in place of bg ?2 (I10 , a0 )c, where CBV
the value of a counterfactual best response while P1 is constrained to playing within the abstraction
that generated the blueprint. Using estimates is covered in more detail in Appendix C.
Theorem 1 shows that when subgames are solved independently and using lower bounds on gifts,
Reach-Maxmargin solving has exploitability lower than or equal to past safe techniques. The theorem
statement is similar to that of Maxmargin [20], but the margins are now larger (or equal) in size.
Theorem 1. Given a strategy ?2 in a two-player zero-sum game, a set of disjoint subgames S,
and a strategy ?2S for each subgame S ? S produced via Reach-Maxmargin solving using lower
bounds for gifts, let ?20 be the strategy that plays according to ?2S for each subgame S ? S, and ?2
elsewhere. Moreover, let ?2?S be the strategy that plays according to ?20 everywhere except for P2
BR(?20 )
nodes in S, where it instead plays according to ?2 . If ?1
(I1 ) > 0 for some I1 ? Stop , then
P
S
?2
exp(?20 ) ? exp(?2?S ) ? h?I1 ??1
(h)Mr? (I1 ).
So far the described techniques have guaranteed a reduction in exploitability over the blueprint by
setting the value of a0T equal to the value of P1 playing optimally to P2 ?s blueprint. Relaxing this
guarantee by instead setting the value of a0T equal to an estimate of P1 ?s value when both players
play optimally leads to far lower exploitability in practice. We discuss this approach in Appendix C.
6
Nested Subgame Solving
As we have discussed, large games must be abstracted to reduce the game to a tractable size. This is
particularly common in games with large or continuous action spaces. Typically the action space is
discretized by action abstraction so that only a few actions are included in the abstraction. While
we might limit ourselves to the actions we included in the abstraction, an opponent might choose
actions that are not in the abstraction. In that case, the off-tree action can be mapped to an action that
is in the abstraction, and the strategy from that in-abstraction action can be used. For example, in an
auction game we might include a bid of $100 in our abstraction. If a player bids $101, we simply
treat that as a bid of $100. This is referred to as action translation [13, 27, 7]. Action translation is
the state-of-the-art prior approach to dealing with this issue. It has been used, for example, by all the
leading competitors in the Annual Computer Poker Competition (ACPC).
In this section, we develop techniques for applying subgame solving to calculate responses to
opponent off-tree actions, thereby obviating the need for action translation. That is, rather than simply
treat a bid of $101 as $100, we calculate in real time a unique response to the bid of $101. This can
also be done in a nested fashion in response to subsequent opponent off-tree actions. Additionally,
these techniques can be used to solve finer-grained models as play progresses down the game tree.
We refer to the first method as the inexpensive method.3 When P1 chooses an off-tree action a,
a subgame S is generated following that action such that for any infoset I1 that P1 might be in,
I1 ? a ? Stop . This subgame may itself be an abstraction. A solution ? S is computed via subgame
solving, and ? S is combined with ? to form a new blueprint ? 0 in the expanded abstraction that now
includes action a. The process repeats whenever P1 again chooses an off-tree action.
3
Following our study, the AI DeepStack used a technique similar to this form of nested subgame solving [19].
7
To conduct safe subgame solving in response to off-tree action a, we could calculate CBV ?2 (I1 , a)
by defining, via action translation, a P2 blueprint following a and best responding to it [4]. However,
that could be computationally expensive and would likely perform poorly in practice because, as we
show later, action translation is highly exploitable. Instead, we relax the guarantee of safety and use
? ?2 (I1 ) for the alternative payoff, where CBV
? ?2 (I1 ) is P1 ?s counterfactual best response value
CBV
in I1 when constrained to playing in the blueprint abstraction (which excludes action a). In this case,
? ?2 (I1 ) approximates CBV ?2? (I1 ), where ? ? is an optimal
exploitability depends on how well CBV
2
4
P2 strategy (see Appendix C). In general, we find that only a small number of near-optimal actions
? ?2 (I1 ) to be close to CBV ?2? (I1 ). We can
need to be included in the blueprint abstraction for CBV
then approximate a near-optimal response to any opponent action, even in a continuous action space.
The ?inexpensive? approach cannot be combined with Unsafe subgame solving because the probability
of reaching an action outside of a player?s abstraction is undefined. Nevertheless, a similar approach
is possible with Unsafe subgame solving (as well as all the other subgame-solving techniques) by
starting the subgame solving at h rather than at h ? a. In other words, if action a taken in node h is
not in the abstraction, then Unsafe subgame solving is conducted in the smallest subgame containing
h (and action a is added to that abstraction). This increases the size of the subgame compared to the
inexpensive method because a strategy must be recomputed for every action a0 ? A(h) in addition to
a. We therefore call this method the expensive method. We present experiments with both methods.
7
Experiments
Our experiments were conducted on heads-up no-limit Texas hold?em, as well as two smaller-scale
poker games we call No-Limit Flop Hold?em (NLFH) and No-Limit Turn Hold?em (NLTH). The
description for these games can be found in Appendix G. For equilibrium finding, we used CFR+ [29].
Our first experiment compares the performance of the subgame-solving techniques when applied
to information abstraction (which is card abstraction in the case of poker). Specifically, we solve
NLFH with no information abstraction on the preflop. On the flop, there are 1,286,792 infosets for
each betting sequence; the abstraction buckets them into 200, 2,000, or 30,000 abstract ones (using a
leading information abstraction algorithm [8]). We then apply subgame solving immediately after the
flop community cards are dealt. We experiment with two versions of the game, one small and one
large, which include only a few of the available actions in each infoset. We also experimented on
abstractions of NLTH. In that case, we solve NLTH with no information abstraction on the preflop or
flop. On the turn, there are 55,190,538 infosets for each betting sequence; the abstraction buckets
them into 200, 2,000, or 20,000 abstract ones. We apply subgame solving immediately after the
turn community card is dealt. Table 1 shows the performance of each technique when using 30,000
buckets (20,000 for NLTH). The full results are presented in Appendix E. In all our experiments,
exploitability is measured in the standard units used in this field: milli big blinds per hand (mbb/h).
Small Flop Holdem Large Flop Holdem
Blueprint Strategy
91.28
41.41
Unsafe
5.514
396.8
Resolve
54.07
23.11
Maxmargin
43.43
19.50
Reach-Maxmargin
41.47
18.80
Reach-Maxmargin (no split)
25.88
16.41
Estimate
24.23
30.09
Estimate+Distributional
34.30
10.54
Reach-Estimate+Distributional
22.58
9.840
Reach-Estimate+Distributional (no split) 17.33
8.777
Table 1: Exploitability of the various subgame-solving techniques in nested subgame solving.
of the pseudo-harmonic action translation is also shown.
Turn Holdem
345.5
79.34
251.8
234.4
233.5
175.5
76.44
74.35
72.59
70.68
The performance
Estimate and Estimate+Distributional are techniques introduced in Appendix C. We use a normal
distribution in the Distributional subgame solving experiments, with standard deviation determined
by the heuristic presented in Appendix C.1.
Since subgame solving begins immediately after a chance node with an extremely high branching
factor (1, 755 in NLFH), the gifts for the Reach algorithms are divided among subgames inefficiently.
?
?
?
?
We estimate CBV ?2 (I1 ) rather than CBV ?2 (I1 , a) because CBV ?2 (I1 ) ? CBV ?2 (I1 , a) is a gift that
may be added to the alternative payoff anyway.
4
8
Many subgames do not use the gifts at all, while others could make use of more. In the experiments
we show results both for the theoretically safe splitting of gifts, as well as a more aggressive version
where gifts are scaled up by the branching factor of the chance node (1, 755). This weakens the
theoretical guarantees of the algorithm, but in general did better than splitting gifts in a theoretically
correct manner. However, this is not universally true. Appendix F shows that in at least one case,
exploitability increased when gifts were scaled up too aggressively. In all cases, using Reach subgame
solving in at least the theoretical safe method led to lower exploitability.
Despite lacking theoretical guarantees, Unsafe subgame solving did surprisingly well in most games.
However, it did substantially worse in Large NLFH with 30,000 buckets. This exemplifies its
variability. Among the safe methods, all of the changes we introduce show improvement over
past techniques. The Reach-Estimate + Distributional algorithm generally resulted in the lowest
exploitability among the various choices, and in most cases beat unsafe subgame solving.
The second experiment evaluates nested subgame solving, and compares it to action translation. In
order to also evaluate action translation, in this experiment, we create an NLFH game that includes 3
bet sizes at every point in the game tree (0.5, 0.75, and 1.0 times the size of the pot); a player can also
decide not to bet. Only one bet (i.e., no raises) is allowed on the preflop, and three bets are allowed on
the flop. There is no information abstraction anywhere in the game. We also created a second, smaller
abstraction of the game in which there is still no information abstraction, but the 0.75? pot bet is
never available. We calculate the exploitability of one player using the smaller abstraction, while
the other player uses the larger abstraction. Whenever the large-abstraction player chooses a 0.75?
pot bet, the small-abstraction player generates and solves a subgame for the remainder of the game
(which again does not include any subsequent 0.75? pot bets) using the nested subgame-solving
techniques described above. This subgame strategy is then used as long as the large-abstraction player
plays within the small abstraction, but if she chooses the 0.75? pot bet again later, then the subgame
solving is used again, and so on.
Table 2 shows that all the subgame-solving techniques substantially outperform action translation.
We did not test distributional alternative payoffs in this experiment, since the calculated best response
values are likely quite accurate. These results suggest that nested subgame solving is preferable to
action translation (if there is sufficient time to solve the subgame).
mbb/h
Randomized Pseudo-Harmonic Mapping 1,465
Resolve
150.2
Reach-Maxmargin (Expensive)
149.2
Unsafe (Expensive)
148.3
Maxmargin
122.0
Reach-Maxmargin
119.1
Table 2: Exploitability of the various subgame-solving techniques in nested subgame solving. The performance
of the pseudo-harmonic action translation is also shown.
We used the techniques presented in this paper to develop Libratus, an AI that competed against
four top human specialists in heads-up no-limit Texas hold?em. Heads-up no-limit Texas hold?em
has been the primary benchmark challenge for AI in imperfect-information games. The competition
involved 120,000 hands of poker and a prize pool of $200,000 split among the humans to incentivize
strong play. The AI decisively defeated the human team by 147 mbb / hand, with 99.98% statistical
significance. This was the first, and so far only, time an AI defeated top humans in no-limit poker.
8
Conclusion
We introduced a subgame-solving technique for imperfect-information games that has stronger
theoretical guarantees and better practical performance than prior subgame-solving methods. We
presented results on exploitability of both safe and unsafe subgame-solving techniques. We also
introduced a method for nested subgame solving in response to the opponent?s off-tree actions, and
demonstrated that this leads to dramatically better performance than the usual approach of action
translation. This is, to our knowledge, the first time that exploitability of subgame-solving techniques
has been measured in large games.
Finally, we demonstrated the effectiveness of these techniques in practice in heads-up no-limit Texas
Hold?em poker, the main benchmark challenge for AI in imperfect-information games. We developed
the first AI to reach the milestone of defeating top humans in heads-up no-limit Texas Hold?em.
9
9
Acknowledgments
This material is based on work supported by the National Science Foundation under grants IIS1718457, IIS-1617590, and CCF-1733556, and the ARO under award W911NF-17-1-0082, as well
as XSEDE computing resources provided by the Pittsburgh Supercomputing Center. The Brains vs.
AI competition was sponsored by Carnegie Mellon University, Rivers Casino, GreatPoint Ventures,
Avenue4Analytics, TNG Technology Consulting, Artificial Intelligence, Intel, and Optimized Markets,
Inc. We thank Kristen Gardner, Marcelo Gutierrez, Theo Gutman-Solo, Eric Jackson, Christian Kroer,
Tim Reiff, and the anonymous reviewers for helpful feedback.
References
[1] Darse Billings, Neil Burch, Aaron Davidson, Robert Holte, Jonathan Schaeffer, Terence
Schauenberg, and Duane Szafron. Approximating game-theoretic optimal strategies for fullscale poker. In Proceedings of the 18th International Joint Conference on Artificial Intelligence
(IJCAI), 2003.
[2] Michael Bowling, Neil Burch, Michael Johanson, and Oskari Tammelin. Heads-up limit
hold?em poker is solved. Science, 347(6218):145?149, January 2015.
[3] Noam Brown, Christian Kroer, and Tuomas Sandholm. Dynamic thresholding and pruning
for regret minimization. In AAAI Conference on Artificial Intelligence (AAAI), pages 421?429,
2017.
[4] Noam Brown and Tuomas Sandholm. Simultaneous abstraction and equilibrium finding in
games. In Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI),
2015.
[5] Neil Burch, Michael Johanson, and Michael Bowling. Solving imperfect information games
using decomposition. In AAAI Conference on Artificial Intelligence (AAAI), pages 602?608,
2014.
[6] Murray Campbell, A Joseph Hoane, and Feng-Hsiung Hsu. Deep Blue. Artificial intelligence,
134(1-2):57?83, 2002.
[7] Sam Ganzfried and Tuomas Sandholm. Action translation in extensive-form games with large
action spaces: axioms, paradoxes, and the pseudo-harmonic mapping. In Proceedings of the
Twenty-Third international joint conference on Artificial Intelligence, pages 120?128. AAAI
Press, 2013.
[8] Sam Ganzfried and Tuomas Sandholm. Potential-aware imperfect-recall abstraction with earth
mover?s distance in imperfect-information games. In AAAI Conference on Artificial Intelligence
(AAAI), 2014.
[9] Sam Ganzfried and Tuomas Sandholm. Endgame solving in large imperfect-information games.
In International Conference on Autonomous Agents and Multi-Agent Systems (AAMAS), pages
37?45, 2015.
[10] Andrew Gilpin, Javier Pe?a, and Tuomas Sandholm. First-order algorithm with O(ln(1/))
convergence for -equilibrium in two-person zero-sum games. Mathematical Programming,
133(1?2):279?298, 2012. Conference version appeared in AAAI-08.
[11] Andrew Gilpin and Tuomas Sandholm. A competitive Texas Hold?em poker player via automated abstraction and real-time equilibrium computation. In Proceedings of the National
Conference on Artificial Intelligence (AAAI), pages 1007?1013, 2006.
[12] Andrew Gilpin and Tuomas Sandholm. Better automated abstraction techniques for imperfect
information games, with application to Texas Hold?em poker. In International Conference on
Autonomous Agents and Multi-Agent Systems (AAMAS), pages 1168?1175, 2007.
10
[13] Andrew Gilpin, Tuomas Sandholm, and Troels Bjerre S?rensen. A heads-up no-limit texas
hold?em poker player: discretized betting models and automatically generated equilibriumfinding programs. In Proceedings of the Seventh International Joint Conference on Autonomous
Agents and Multiagent Systems-Volume 2, pages 911?918. International Foundation for Autonomous Agents and Multiagent Systems, 2008.
[14] Eric Jackson. A time and space efficient algorithm for approximately solving large imperfect
information games. In AAAI Workshop on Computer Poker and Imperfect Information, 2014.
[15] Michael Johanson. Measuring the size of large no-limit poker games. Technical report,
University of Alberta, 2013.
[16] Michael Johanson, Nolan Bard, Neil Burch, and Michael Bowling. Finding optimal abstract
strategies in extensive-form games. In Proceedings of the Twenty-Sixth AAAI Conference on
Artificial Intelligence, pages 1371?1379. AAAI Press, 2012.
[17] Christian Kroer, Kevin Waugh, Fatma K?l?n?-Karzan, and Tuomas Sandholm. Theoretical
and practical advances on smoothing for extensive-form games. In Proceedings of the ACM
Conference on Economics and Computation (EC), 2017.
[18] Nick Littlestone and M. K. Warmuth. The weighted majority algorithm. Information and
Computation, 108(2):212?261, 1994.
[19] Matej Morav?c?k, Martin Schmid, Neil Burch, Viliam Lis?, Dustin Morrill, Nolan Bard, Trevor
Davis, Kevin Waugh, Michael Johanson, and Michael Bowling. Deepstack: Expert-level
artificial intelligence in heads-up no-limit poker. Science, 2017.
[20] Matej Moravcik, Martin Schmid, Karel Ha, Milan Hladik, and Stephen Gaukrodger. Refining
subgames in large imperfect information games. In AAAI Conference on Artificial Intelligence
(AAAI), 2016.
[21] John Nash. Equilibrium points in n-person games. Proceedings of the National Academy of
Sciences, 36:48?49, 1950.
[22] Yurii Nesterov. Excessive gap technique in nonsmooth convex minimization. SIAM Journal of
Optimization, 16(1):235?249, 2005.
[23] Tuomas Sandholm. The state of solving large incomplete-information games, and application
to poker. AI Magazine, pages 13?32, Winter 2010. Special issue on Algorithmic Game Theory.
[24] Tuomas Sandholm. Abstraction for solving large incomplete-information games. In AAAI
Conference on Artificial Intelligence (AAAI), pages 4127?4131, 2015. Senior Member Track.
[25] Tuomas Sandholm. Solving imperfect-information games. Science, 347(6218):122?123, 2015.
[26] Jonathan Schaeffer, Neil Burch, Yngvi Bj?rnsson, Akihiro Kishimoto, Martin M?ller, Robert
Lake, Paul Lu, and Steve Sutphen. Checkers is solved. Science, 317(5844):1518?1522, 2007.
[27] David Schnizlein, Michael Bowling, and Duane Szafron. Probabilistic state translation in
extensive games with large action sets. In Proceedings of the Twenty-First International Joint
Conference on Artificial Intelligence, pages 278?284, 2009.
[28] David Silver, Aja Huang, Chris J Maddison, Arthur Guez, Laurent Sifre, George Van Den Driessche, Julian Schrittwieser, Ioannis Antonoglou, Veda Panneershelvam, Marc Lanctot, et al. Mastering the game of go with deep neural networks and tree search. Nature, 529(7587):484?489,
2016.
[29] Oskari Tammelin, Neil Burch, Michael Johanson, and Michael Bowling. Solving heads-up limit
texas hold?em. In Proceedings of the International Joint Conference on Artificial Intelligence
(IJCAI), pages 645?652, 2015.
[30] Kevin Waugh, Nolan Bard, and Michael Bowling. Strategy grafting in extensive games. In
Proceedings of the Annual Conference on Neural Information Processing Systems (NIPS), 2009.
[31] Martin Zinkevich, Michael Johanson, Michael H Bowling, and Carmelo Piccione. Regret
minimization in games with incomplete information. In Proceedings of the Annual Conference
on Neural Information Processing Systems (NIPS), pages 1729?1736, 2007.
11
| 6671 |@word version:4 proportion:2 szafron:2 vi1:2 stronger:1 tried:1 decomposition:3 pick:1 thereby:2 reduction:2 initial:4 contains:1 past:6 outperforms:1 current:1 guez:1 must:10 john:1 subsequent:2 christian:3 sponsored:1 v:1 alone:1 intelligence:15 guess:6 warmuth:1 libratus:2 prize:1 aja:1 consulting:1 node:41 mathematical:1 along:1 c2:1 become:1 consists:1 fitting:1 manner:1 introduce:3 theoretically:2 indeed:3 market:1 expected:6 p1:73 multi:2 brain:1 discretized:3 terminal:5 decreasing:1 alberta:1 automatically:1 resolve:4 considering:1 increasing:1 gift:23 begin:1 provided:1 notation:1 moreover:1 maximizes:1 lowest:1 what:3 substantially:2 maxa:1 developed:1 finding:3 guarantee:11 pseudo:4 every:5 act:2 exactly:1 preferable:1 scaled:2 milestone:1 unit:1 grant:1 producing:1 safety:2 positive:2 treat:2 limit:17 severely:1 despite:2 laurent:1 path:1 approximately:1 might:6 chose:2 relaxing:1 unique:1 practical:2 acknowledgment:1 carmelo:1 practice:8 regret:2 differs:2 subgame:141 axiom:1 lucky:2 significantly:1 word:1 suggest:1 get:2 cannot:8 close:1 xsede:1 applying:7 weighed:1 zinkevich:1 equivalent:1 demonstrated:2 center:1 blueprint:40 maximizing:1 go:2 overarching:1 modifies:1 independently:7 starting:1 reviewer:1 economics:1 convex:1 simplicity:1 splitting:2 immediately:4 ii0:3 dominate:1 jackson:2 anyway:1 autonomous:4 target:1 play:39 heavily:1 suppose:2 exact:1 programming:1 magazine:1 us:1 pa:2 expensive:4 particularly:1 gutman:1 distributional:7 observed:1 akihiro:1 yngvi:1 solved:14 capture:1 calculate:4 counter:1 highest:1 intuition:2 nash:4 ui:9 reward:7 nesterov:1 dynamic:1 ultimately:1 depend:3 solving:86 raise:1 myriad:1 upon:4 division:3 eric:2 selling:1 compactly:1 joint:6 milli:1 represented:2 various:3 artificial:15 kevin:3 outside:2 outcome:4 choosing:24 h0:22 whose:2 heuristic:1 larger:4 solve:9 quite:1 say:1 relax:1 nolan:3 ability:1 neil:7 itself:2 sequence:5 aro:1 remainder:1 relevant:3 poorly:2 achieve:6 academy:1 intuitive:2 description:1 venture:1 competition:3 az:2 milan:1 ijcai:3 mbb:3 convergence:1 produce:1 perfect:5 silver:1 tim:1 weakens:1 develop:2 andrew:4 measured:2 received:4 progress:2 strong:2 solves:1 descends:1 c:2 pot:5 p2:47 differ:1 safe:10 correct:3 human:7 public:5 material:1 require:1 maxa0:1 landed:2 decompose:1 kristen:1 opt:1 anonymous:1 insert:1 hold:15 considered:4 normal:1 exp:3 equilibrium:11 mapping:3 algorithmic:1 bj:1 smallest:1 purpose:1 earth:1 deepstack:2 gutierrez:1 create:1 defeating:1 karel:1 weighted:2 minimization:3 clearly:1 always:14 modified:2 rather:9 reaching:9 johanson:7 bet:9 earliest:1 exemplifies:1 focus:1 refining:1 competed:1 she:3 improvement:3 helpful:1 waugh:3 abstraction:42 i0:4 entire:3 typically:3 a0:20 hidden:1 troels:1 i1:38 overall:2 among:6 issue:3 denoted:1 negotiation:1 art:2 summed:1 constrained:2 smoothing:1 special:1 equal:6 aware:1 never:3 having:1 beach:1 field:1 identical:1 flipped:1 sell:26 represents:1 excessive:1 nearly:1 others:2 tammelin:2 report:1 nonsmooth:1 viewable:1 few:3 winter:1 mover:1 resulted:1 national:3 individual:1 connects:1 ourselves:1 negation:1 attempt:2 cbr:2 decisively:1 highly:1 subgames:25 kroer:3 indifferent:1 unlucky:2 undefined:1 accurate:2 solo:1 tuple:1 necessary:1 arthur:1 tree:16 conduct:1 incomplete:3 littlestone:1 theoretical:8 increased:1 w911nf:1 retains:1 measuring:1 strategic:2 cost:1 deviation:1 expects:1 conducted:2 seventh:1 too:1 optimally:6 holdem:3 chooses:11 combined:2 st:1 person:2 international:9 randomized:1 river:1 siam:1 probabilistic:1 off:7 receiving:1 picking:2 pool:1 together:1 terence:1 michael:15 oskari:2 again:4 central:1 aaai:16 opposed:1 leveraged:1 containing:2 choose:9 huang:1 worse:5 expert:1 leading:6 li:1 aggressive:1 potential:3 casino:1 ioannis:1 includes:2 inc:1 depends:2 vi:9 bg:6 later:4 root:2 blind:1 reached:5 competitive:1 option:2 contribution:2 marcelo:1 characteristic:1 who:1 yield:1 dealt:2 critically:1 produced:1 lu:1 finer:1 simultaneous:1 reach:23 whenever:2 trevor:1 sixth:1 against:6 competitor:1 inexpensive:3 evaluates:1 involved:1 fatma:1 schaeffer:2 stop:13 hsu:1 schauenberg:1 counterfactual:5 recall:2 knowledge:4 javier:1 actually:1 campbell:1 matej:2 steve:1 higher:3 response:17 done:1 though:2 just:2 anywhere:1 hand:4 receives:6 ganzfried:3 lack:2 defines:2 quality:1 usa:1 brown:3 true:3 ccf:1 aggressively:1 entering:1 indistinguishable:1 game:94 during:2 uniquely:1 branching:2 rooted:1 bowling:8 davis:1 theoretic:1 demonstrate:1 performs:1 auction:2 reasoning:1 harmonic:4 common:1 physical:1 defeat:1 volume:1 extend:1 tail:30 discussed:4 approximates:1 karzan:1 mellon:3 refer:8 ai:12 similarly:2 reachable:1 closest:1 belongs:1 store:1 discussing:1 seen:1 holte:1 additional:2 george:1 mr:2 determine:3 ller:1 ii:36 resolving:8 full:6 stephen:1 ing:1 technical:1 match:1 adapt:1 defeated:3 long:5 divided:1 equally:1 award:3 essentially:1 cmu:2 sometimes:1 c1:2 receive:7 background:1 want:1 fine:1 addition:1 hoane:1 unlike:1 checker:2 sr:3 probably:1 member:1 effectiveness:1 call:3 near:2 exceed:1 split:4 forfeit:3 easy:2 enough:2 bid:5 automated:2 isolation:3 suboptimal:4 imperfect:25 reduce:1 billing:1 knowing:1 br:5 texas:12 whether:1 a0t:7 veda:1 afford:2 action:78 deep:2 dramatically:1 generally:2 covered:1 generate:1 outperform:2 exist:1 rensen:1 dotted:2 upfront:1 disjoint:2 correctly:1 per:1 track:1 blue:1 carnegie:3 write:2 key:1 thereafter:1 recomputed:1 nevertheless:2 four:1 changing:1 incentivize:1 excludes:1 sum:5 everywhere:1 respond:1 preflop:3 place:1 throughout:2 decide:1 lake:1 earning:1 decision:3 appendix:10 acceptable:1 lanctot:1 bound:6 pay:1 guaranteed:1 played:6 distinguish:1 encountered:2 nonnegative:2 annual:3 burch:7 generates:1 u1:1 endgame:1 extremely:2 expanded:1 relatively:1 betting:3 martin:4 acted:1 department:2 according:11 sandholm:15 smaller:5 em:15 reconstructing:1 across:2 sam:3 mastering:1 joseph:1 s1:9 chess:1 maxmargin:15 den:1 bucket:4 taken:4 computationally:1 resource:1 ln:1 discus:1 turn:4 needed:1 flip:1 letting:1 tractable:1 antonoglou:1 yurii:1 available:4 panneershelvam:1 opponent:9 multiplied:1 apply:6 i10:21 alternative:8 coin:35 specialist:1 original:4 inefficiently:1 top:5 remaining:1 running:1 denotes:2 assumes:2 ensure:2 opportunity:1 include:3 responding:1 exploit:1 giving:1 build:1 murray:1 approximating:1 feng:1 move:1 already:2 added:2 occurs:1 strategy:79 primary:2 usual:1 guessing:1 poker:18 distance:1 thank:1 card:4 mapped:1 majority:1 chris:1 maddison:1 considers:1 cfr:1 bard:3 tuomas:14 julian:1 demonstration:1 schrittwieser:1 equivalently:1 robert:2 potentially:1 statement:1 noam:3 darse:1 unknown:1 perform:1 twenty:3 benchmark:3 schnizlein:1 beat:1 displayed:1 incorrectly:1 immediate:1 payoff:14 situation:1 head:52 defining:1 flop:7 variability:1 team:1 paradox:1 acpc:1 community:2 introduced:3 david:2 extensive:6 optimized:1 security:1 nick:1 kishimoto:1 unsafe:17 nip:3 beyond:2 suggested:1 ev:26 appeared:1 challenge:3 program:1 including:3 max:2 belief:1 shifting:1 rely:1 hr:4 representing:1 improve:7 technology:1 gardner:1 created:1 schmid:2 prior:8 review:2 determining:1 lacking:1 multiagent:2 piccione:1 proven:1 foundation:2 agent:6 sufficient:1 thresholding:1 playing:7 share:2 land:14 translation:15 storing:1 elsewhere:1 repeat:1 surprisingly:1 keeping:1 supported:1 theo:1 allow:4 senior:1 van:1 feedback:1 calculated:2 complicating:1 universally:1 sifre:1 supercomputing:1 far:5 ec:1 viliam:1 approximate:3 pruning:1 grafting:1 dealing:2 abstracted:2 rid:1 pittsburgh:3 assumed:1 davidson:1 alternatively:2 continuous:3 search:1 why:1 reality:1 additionally:1 table:4 nature:1 exploitability:21 ca:1 marc:1 did:5 significance:1 cbv:26 main:1 whole:5 s2:4 big:1 profile:6 paul:1 repeated:1 obviating:1 allowed:2 aamas:2 exploitable:1 augmented:15 referred:3 intel:1 board:1 fashion:1 wish:1 january:1 pe:1 forgets:1 third:1 dustin:1 grained:2 down:2 theorem:4 bad:1 experimented:1 exists:1 workshop:1 margin:6 gap:1 led:1 simply:5 likely:2 u2:1 hladik:1 driessche:1 duane:2 nested:10 chance:13 satisfies:2 acm:1 goal:1 a0s:2 toss:18 change:2 included:3 specifically:4 except:1 determined:1 miss:1 cybersecurity:1 player:41 aaron:1 formally:7 gilpin:4 jonathan:2 evaluate:1 |
6,270 | 6,672 | Unsupervised Image-to-Image Translation Networks
Ming-Yu Liu,
Thomas Breuel, Jan Kautz
NVIDIA
{mingyul,tbreuel,jkautz}@nvidia.com
Abstract
Unsupervised image-to-image translation aims at learning a joint distribution of
images in different domains by using images from the marginal distributions in
individual domains. Since there exists an infinite set of joint distributions that
can arrive the given marginal distributions, one could infer nothing about the joint
distribution from the marginal distributions without additional assumptions. To
address the problem, we make a shared-latent space assumption and propose an
unsupervised image-to-image translation framework based on Coupled GANs.
We compare the proposed framework with competing approaches and present
high quality image translation results on various challenging unsupervised image
translation tasks, including street scene image translation, animal image translation,
and face image translation. We also apply the proposed framework to domain
adaptation and achieve state-of-the-art performance on benchmark datasets. Code
and additional results are available in https://github.com/mingyuliutw/unit.
1
Introduction
Many computer visions problems can be posed as an image-to-image translation problem, mapping
an image in one domain to a corresponding image in another domain. For example, super-resolution
can be considered as a problem of mapping a low-resolution image to a corresponding high-resolution
image; colorization can be considered as a problem of mapping a gray-scale image to a corresponding
color image. The problem can be studied in supervised and unsupervised learning settings. In the
supervised setting, paired of corresponding images in different domains are available [8, 15]. In the
unsupervised setting, we only have two independent sets of images where one consists of images
in one domain and the other consists of images in another domain?there exist no paired examples
showing how an image could be translated to a corresponding image in another domain. Due to
lack of corresponding images, the UNsupervised Image-to-image Translation (UNIT) problem is
considered harder, but it is more applicable since training data collection is easier.
When analyzing the image translation problem from a probabilistic modeling perspective, the key
challenge is to learn a joint distribution of images in different domains. In the unsupervised setting,
the two sets consist of images from two marginal distributions in two different domains, and the task is
to infer the joint distribution using these images. The coupling theory [16] states there exist an infinite
set of joint distributions that can arrive the given marginal distributions in general. Hence, inferring
the joint distribution from the marginal distributions is a highly ill-posed problem. To address the
ill-posed problem, we need additional assumptions on the structure of the joint distribution.
To this end we make a shared-latent space assumption, which assumes a pair of corresponding images
in different domains can be mapped to a same latent representation in a shared-latent space. Based on
the assumption, we propose a UNIT framework that are based on generative adversarial networks
(GANs) and variational autoencoders (VAEs). We model each image domain using a VAE-GAN. The
adversarial training objective interacts with a weight-sharing constraint, which enforces a sharedlatent space, to generate corresponding images in two domains, while the variational autoencoders
relate translated images with input images in the respective domains. We applied the proposed
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
E1
Z : shared latent space
x
?1!1
1
x1
z
E1
D1
G1
E2
G1
T/F
x
?2!1
2
G2
x2
x1
X2
X1
x
?1!2
1
x2
(a)
z
E2
T/F
x
?2!2
2
G2
(b)
D2
Figure 1: (a) The shared latent space assumption. We assume a pair of corresponding images (x1 , x2 ) in
two different domains X1 and X2 can be mapped to a same latent code z in a shared-latent space Z. E1 and
E2 are two encoding functions, mapping images to latent codes. G1 and G2 are two generation functions,
mapping latent codes to images. (b) The proposed UNIT framework. We represent E1 E2 G1 and G2 using
CNNs and implement the shared-latent space assumption using a weight sharing constraint where the connection
weights of the last few layers (high-level layers) in E1 and E2 are tied (illustrated using dashed lines) and the
connection weights of the first few layers (high-level layers) in G1 and G2 are tied. Here, x
?1!1
and x
?2!2
1
2
1!2
2!1
are self-reconstructed images, and x
?1
and x
?2
are domain-translated images. D1 and D2 are adversarial
discriminators for the respective domains, in charge of evaluating whether the translated images are realistic.
Table 1: Interpretation of the roles of the subnetworks in the proposed framework.
Networks
Roles
{E1 , G1 }
{E1 , G2 }
{G1 , D1 }
{E1 , G1 , D1 }
VAE for X1 Image Translator X1 ! X2 GAN for X1 VAE-GAN [14]
{G1 , G2 , D1 , D2 }
CoGAN [17]
framework to various unsupervised image-to-image translation problems and achieved high quality
image translation results. We also applied it to the domain adaptation problem and achieved state-ofthe-art accuracies on benchmark datasets. The shared-latent space assumption was used in Coupled
GAN [17] for joint distribution learning. Here, we extend the Coupled GAN work for the UNIT
problem. We also note that several contemporary works propose the cycle-consistency constraint
assumption [29, 10], which hypothesizes the existence of a cycle-consistency mapping so that an
image in the source domain can be mapped to an image in the target domain and this translated image
in the target domain can be mapped back to the original image in the source domain. In the paper, we
show that the shared-latent space constraint implies the cycle-consistency constraint.
2
Assumptions
Let X1 and X2 be two image domains. In supervised image-to-image translation, we are given
samples (x1 , x2 ) drawn from a joint distribution PX1 ,X2 (x1 , x2 ). In unsupervised image-to-image
translation, we are given samples drawn from the marginal distributions PX1 (x1 ) and PX2 (x2 ). Since
an infinite set of possible joint distributions can yield the given marginal distributions, we could infer
nothing about the joint distribution from the marginal samples without additional assumptions.
We make the shared-latent space assumption. As shown Figure 1, we assume for any given pair
of images x1 and x2 , there exists a shared latent code z in a shared-latent space, such that we
can recover both images from this code, and we can compute this code from each of the two
images. That is, we postulate there exist functions E1? , E2? , G?1 , and G?2 such that, given a pair of
corresponding images (x1 , x2 ) from the joint distribution, we have z = E1? (x1 ) = E2? (x2 ) and
?
conversely x1 = G?1 (z) and x2 = G?2 (z). Within this model, the function x2 = F1!2
(x1 ) that
?
?
?
maps from X1 to X2 can be represented by the composition F1!2 (x1 ) = G2 (E1 (x1 )). Similarly,
?
?
x1 = F2!1
(x2 ) = G?1 (E2? (x2 )). The UNIT problem then becomes a problem of learning F1!2
?
?
?
and F2!1
. We note that a necessary condition for F1!2
and F2!1
to exist is the cycle-consistency
?
?
?
?
constraint [29, 10]: x1 = F2!1
(F1!2
(x1 )) and x2 = F1!2
(F2!1
(x2 )). We can reconstruct
the input image from translating back the translated input image. In other words, the proposed
shared-latent space assumption implies the cycle-consistency assumption (but not vice versa).
To implement the shared-latent space assumption, we further assume a shared intermediate representation h such that the process of generating a pair of corresponding images admits a form of
2
x1
.
(1)
x2
Consequently, we have G?1 ? G?L,1 G?H and G?2 ? G?L,2 G?H where G?H is a common high-level
generation function that maps z to h and G?L,1 and G?L,2 are low-level generation functions that map
h to x1 and x2 , respectively. In the case of multi-domain image translation (e.g., sunny and rainy
image translation), z can be regarded as the compact, high-level representation of a scene ("car in
front, trees in back"), and h can be considered a particular realization of z through G?H ("car/tree
occupy the following pixels"), and G?L,1 and G?L,2 would be the actual image formation functions
in each modality ("tree is lush green in the sunny domain, but dark green in the rainy domain").
?
?
?
?
Assuming h also allow us to represent E1? and E2? by E1? ? EH
EL,1
and E2? ? EH
EL,2
.
z!h
%
&
In the next section, we discuss how we realize the above ideas in the proposed UNIT framework.
3
Framework
Our framework, as illustrated in Figure 1, is based on variational autoencoders (VAEs) [13, 22, 14]
and generative adversarial networks (GANs) [6, 17]. It consists of 6 subnetworks: including two
domain image encoders E1 and E2 , two domain image generators G1 and G2 , and two domain
adversarial discriminators D1 and D2 . Several ways exist to interpret the roles of the subnetworks,
which we summarize in Table 1. Our framework learns translation in both directions in one shot.
VAE. The encoder?generator pair {E1 , G1 } constitutes a VAE for the X1 domain, termed VAE1 . For
an input image x1 2 X1 , the VAE1 first maps x1 to a code in a latent space Z via the encoder E1 and
then decodes a random-perturbed version of the code to reconstruct the input image via the generator
G1 . We assume the components in the latent space Z are conditionally independent and Gaussian with
unit variance. In our formulation, the encoder outputs a mean vector E?,1 (x1 ) and the distribution
of the latent code z1 is given by q1 (z1 |x1 ) ? N (z1 |E?,1 (x1 ), I) where I is an identity matrix. The
reconstructed image is x
?1!1
= G1 (z1 ? q1 (z1 |x1 )). Note that here we abused the notation since
1
we treated the distribution of q1 (z1 |x1 ) as a random vector of N (E?,1 (x1 ), I) and sampled from it.
Similarly, {E2 , G2 } constitutes a VAE for X2 : VAE2 where the encoder E2 outputs a mean vector
E?,2 (x2 ) and the distribution of the latent code z2 is given by q2 (z2 |x2 ) ? N (z2 |E?,2 (x2 ), I). The
reconstructed image is x
?2!2
= G2 (z2 ? q2 (z2 |x2 )).
2
Utilizing the reparameterization trick [13], the non-differentiable sampling operation can be reparameterized as a differentiable operation using auxiliary random variables. This reparameterization trick
allows us to train VAEs using back-prop. Let ? be a random vector with a multi-variate Gaussian
distribution: ? ? N (?|0, I). The sampling operations of z1 ? q1 (z1 |x1 ) and z2 ? q2 (z2 |x2 ) can be
implemented via z1 = E?,1 (x1 ) + ? and z2 = E?,2 (x2 ) + ?, respectively.
Weight-sharing. Based on the shared-latent space assumption discussed in Section 2, we enforce a
weight-sharing constraint to relate the two VAEs. Specifically, we share the weights of the last few
layers of E1 and E2 that are responsible for extracting high-level representations of the input images
in the two domains. Similarly, we share the weights of the first few layers of G1 and G2 responsible
for decoding high-level representations for reconstructing the input images.
Note that the weight-sharing constraint alone does not guarantee that corresponding images in two
domains will have the same latent code. In the unsupervised setting, no pair of corresponding images
in the two domains exists to train the network to output a same latent code. The extracted latent
codes for a pair of corresponding images are different in general. Even if they are the same, the same
latent component may have different semantic meanings in different domains. Hence, the same latent
code could still be decoded to output two unrelated images. However, we will show that through
adversarial training, a pair of corresponding images in the two domains can be mapped to a common
latent code by E1 and E2 , respectively, and a latent code will be mapped to a pair of corresponding
images in the two domains by G1 and G2 , respectively.
The shared-latent space assumption allows us to perform image-to-image translation. We can
translate an image x1 in X1 to an image in X2 through applying G2 (z1 ? q1 (z1 |x1 )). We term such
an information processing stream as the image translation stream. Two image translation streams exist
in the proposed framework: X1 ! X2 and X2 ! X1 . The two streams are trained jointly with the
two image reconstruction streams from the VAEs. Once we could ensure that a pair of corresponding
3
images are mapped to a same latent code and a same latent code is decoded to a pair of corresponding
images, (x1 , G2 (z1 ? q1 (z1 |x1 ))) would form a pair of corresponding images. In other words, the
?
composition of E1 and G2 functions approximates F1!2
for unsupervised image-to-image translation
?
discussed in Section 2, and the composition of E2 and G1 function approximates F2!1
.
GANs. Our framework has two generative adversarial networks: GAN1 = {D1 , G1 } and GAN2 =
{D2 , G2 }. In GAN1 , for real images sampled from the first domain, D1 should output true, while
for images generated by G1 , it should output false. G1 can generate two types of images: 1) images
from the reconstruction stream x
?1!1
= G1 (z1 ? q1 (z1 |x1 )) and 2) images from the translation
1
stream x
?2!1
=
G
(z
?
q
(z
|x
)).
Since
the reconstruction stream can be supervisedly trained, it
1 2
2 2 2
2
is suffice that we only apply adversarial training to images from the translation stream, x
?2!1
. We
2
apply a similar processing to GAN2 where D2 is trained to output true for real images sampled from
the second domain dataset and false for images generated from G2 .
Cycle-consistency (CC). Since the shared-latent space assumption implies the cycle-consistency
constraint (See Section 2), we could also enforce the cycle-consistency constraint in the proposed
framework to further regularize the ill-posed unsupervised image-to-image translation problem. The
resulting information processing stream is called the cycle-reconstruction stream.
Learning. We jointly solve the learning problems of the VAE1 , VAE2 , GAN1 and GAN2 for the
image reconstruction streams, the image translation streams, and the cycle-reconstruction streams:
min
max LVAE1 (E1 , G1 ) + LGAN1 (E1 , G1 , D1 ) + LCC1 (E1 , G1 , E2 , G2 )
E1 ,E2 ,G1 ,G2 D1 ,D2
LVAE2 (E2 , G2 ) + LGAN2 (E2 , G2 , D2 ) + LCC2 (E2 , G2 , E1 , G1 ).
(2)
VAE training aims for minimizing a variational upper bound In (2), the VAE objects are
LVAE1 (E1 , G1 ) =
LVAE2 (E2 , G2 ) =
1 KL(q1 (z1 |x1 )||p? (z))
1 KL(q2 (z2 |x2 )||p? (z))
2 Ez1 ?q1 (z1 |x1 ) [log pG1 (x1 |z1 )]
2 Ez2 ?q2 (z2 |x2 ) [log pG2 (x2 |z2 )].
(3)
(4)
where the hyper-parameters 1 and 2 control the weights of the objective terms and the KL
divergence terms penalize deviation of the distribution of the latent code from the prior distribution.
The regularization allows an easy way to sample from the latent space [13]. We model pG1 and pG2
using Laplacian distributions, respectively. Hence, minimizing the negative log-likelihood term is
equivalent to minimizing the absolute distance between the image and the reconstructed image. The
prior distribution is a zero mean Gaussian p? (z) = N (z|0, I).
In (2), the GAN objective functions are given by
LGAN1 (E1 , G1 , D1 ) =
LGAN2 (E2 , G2 , D2 ) =
0 Ex1 ?PX1 [log D1 (x1 )]
+
0 Ex2 ?PX2 [log D2 (x2 )] +
0 Ez2 ?q2 (z2 |x2 ) [log(1
0 Ez1 ?q1 (z1 |x1 ) [log(1
D1 (G1 (z2 )))]
(5)
D2 (G2 (z1 )))].
(6)
The objective functions in (5) and (6) are conditional GAN objective functions. They are used to
ensure the translated images resembling images in the target domains, respectively. The hyperparameter 0 controls the impact of the GAN objective functions.
We use a VAE-like objective function to model the cycle-consistency constraint, which is given by
LCC1 (E1 , G1 , E2 , G2 ) =
LCC2 (E2 , G2 , E1 , G1 ) =
+ 3 KL(q2 (z2 |x1!2
))||p? (z))
1
[log
p
(x
|z
)]
4 Ez2 ?q2 (z2 |x1!2
G
1
2
)
1
1
3 KL(q1 (z1 |x1 )||p? (z))
3 KL(q2 (z2 |x2 )||p? (z))
2!1
))||p? (z))
3 KL(q1 (z1 |x2
+
4 Ez1 ?q1 (z1 |x2!1
) [log pG2 (x2 |z1 )].
2
(7)
(8)
where the negative log-likelihood objective term ensures a twice translated image resembles the
input one and the KL terms penalize the latent codes deviating from the prior distribution in the
cycle-reconstruction stream (Therefore, there are two KL terms). The hyper-parameters 3 and 4
control the weights of the two different objective terms.
Inheriting from GAN, training of the proposed framework results in solving a mini-max problem
where the optimization aims to find a saddle point. It can be seen as a two player zero-sum game.
The first player is a team consisting of the encoders and generators. The second player is a team
consisting of the adversarial discriminators. In addition to defeating the second player, the first player
has to minimize the VAE losses and the cycle-consistency losses. We apply an alternating gradient
4
Accuracy
0.7
0.7
0.56
0.56
0.42
0.42
0.28
0.28
0.14
0
1
6 Dis
5 Dis
4 Dis
3 Dis
2
3
0.14
4
# of shared layers in gen.
(a)
Method
(b)
0
1000
= 10
11=10
=1
11=1
= 0.1
11=0.1
= 0.01
11=0.01
100
10
1
Accuracy
Weight 0.569?0.029
Sharing
Cycle 0.568?0.010
Consistenc
Proposed 0.600?0.015
22
(c)
(d)
Figure 2: (a) Illustration of the Map dataset. Left: satellite image. Right: map. We translate holdout satellite
images to maps and measure the accuracy achieved by various configurations of the proposed framework.
(b) Translation accuracy versus different network architectures. (c) Translation accuracy versus different
hyper-parameter values. (d) Impact of weight-sharing and cycle-consistency constraints on translation accuracy.
update scheme similar to the one described in [6] to solve (2). Specifically, we first apply a gradient
ascent step to update D1 and D2 with E1 , E2 , G1 , and G2 fixed. We then apply a gradient descent
step to update E1 , E2 , G1 , and G2 with D1 and D2 fixed.
Translation: After learning, we obtain two image translation functions by assembling a subset of the
subnetworks. We have F1!2 (x1 ) = G2 (z1 ? q1 (z1 |x1 )) for translating images from X1 to X2 and
F2!1 (x2 ) = G1 (z2 ? q2 (z2 |x2 )) for translating images from X2 to X1 .
4
Experiments
We first analyze various components of the proposed framework. We then present visual results on
challenging translation tasks. Finally, we apply our framework to the domain adaptation tasks.
Performance Analysis. We used ADAM [11] for training where the learning rate was set to 0.0001
and momentums were set to 0.5 and 0.999. Each mini-batch consisted of one image from the first
domain and one image from the second domain. Our framework had several hyper-parameters. The
default values were 0 = 10, 3 = 1 = 0.1 and 4 = 2 = 100. For the network architecture,
our encoders consisted of 3 convolutional layers as the front-end and 4 basic residual blocks [7] as
the back-end. The generators consisted of 4 basic residual blocks as the front-end and 3 transposed
convolutional layers as the back-end. The discriminators consisted of stacks of convolutional layers.
We used LeakyReLU for nonlinearity. The details of the networks are given in Appendix A.
We used the map dataset [8] (visualized in Figure 2), which contained corresponding pairs of images
in two domains (satellite image and map) useful for quantitative evaluation. Here, the goal was to
learn to translate between satellite images and maps. We operated in an unsupervised setting where
we used the 1096 satellite images from the training set as the first domain and 1098 maps from the
validation set as the second domain. We trained for 100K iterations and used the final model to
translate 1098 satellite images in the test set. We then compared the difference between a translated
satellite image (supposed to be maps) and the corresponding ground truth maps pixel-wisely. A pixel
translation was counted correct if the color difference was within 16 of the ground truth color value.
We used the average pixel accuracy over the images in the test set as the performance metric. We
could use color difference for measuring translation accuracy since the target translation function
was unimodal. We did not evaluate the translation from maps to images since the translation was
multi-modal, which was difficult to construct a proper evaluation metric.
In one experiment, we varied the number of weight-sharing layers in the VAEs and paired each
configuration with discriminator architectures of different depths during training. We changed the
number of weight-sharing layers from 1 to 4. (Sharing 1 layer in VAEs means sharing 1 layer for
E1 and E2 and, at the same time, sharing 1 layer for G1 and G2 .) The results were reported in
Figure 2(b). Each curve corresponded to a discriminator architecture of a different depth. The x-axis
denoted the number of weigh-sharing layers in the VAEs. We found that the shallowest discriminator
architecture led to the worst performance. We also found that the number of weight-sharing layer
had little impact. This was due to the use of the residual blocks. As tying the weight of one layer, it
effectively constrained the other layers since the residual blocks only updated the residual information.
In the rest of the experiments, we used VAEs with 1 sharing layer and discriminators of 5 layers.
5
We analyzed impact of the hyper-parameter values to the translation accuracy. For different weight
values on the negative log likelihood terms (i.e., 2 , 4 ), we computed the achieved translation
accuracy over different weight values on the KL terms (i.e., 1 , 3 ). The results were reported in
Figure 2(c). We found that, in general, a larger weight value on the negative log likelihood terms
yielded a better translation accuracy. We also found setting the weights of the KL terms to 0.1 resulted
in consistently good performance. We hence set 1 = 3 = 0.1 and 2 = 4 = 100.
We performed an ablation study measuring impact of the weight-sharing and cycle-consistency
constraints to the translation performance and showed the results in Figure 2(d). We reported average
accuracy over 5 trials (trained with different initialized weights.). We note that when we removed
the weight-sharing constraint (as a consequence, we also removed the reconstruction streams in the
framework), the framework was reduced to the CycleGAN architecture [29, 10]. We found the model
achieved an average pixel accuracy of 0.569. When we removed the cycle-consistency constraint
and only used the weight-sharing constraint1 , it achieved 0.568 average pixel accuracy. But when we
used the full model, it achieved the best performance of 0.600 average pixel accuracy. This echoed
our point that for the ill-posed joint distribution recovery problem, more constraints are beneficial.
Qualitative results. Figure 3 to 6 showed results of the proposed framework on various UNIT tasks.
Street images. We applied the proposed framework to several unsupervised street scene image
translation tasks including sunny to rainy, day to night, summery to snowy, and vice versa. For each
task, we used a set of images extracted from driving videos recorded at different days and cities. The
numbers of the images in the sunny/day, rainy, night, summery, and snowy sets are 86165, 28915,
36280, 6838, and 6044. We trained the network to translate street scene image of size 640?480. In
Figure 3, we showed several example translation results . We found that our method could generate
realistic translated images. We also found that one translation was usually harder than the other.
Specifically, the translation that required adding more details to the image was usually harder (e.g.
night to day). Additional results are available in https://github.com/mingyuliutw/unit.
Synthetic to real. In Figure 3, we showed several example results achieved by applying the proposed
framework to translate images between the synthetic images in the SYNTHIA dataset [23] and the
real images in the Cityscape dataset [2]. For the real to synthetic translation, we found our method
made the cityscape images cartoon like. For the synthetic to real translation, our method achieved
better results in the building, sky, road, and car regions than in the human regions.
Dog breed conversion. We used the images of Husky, German Shepherd, Corgi, Samoyed, and Old
English Sheep dogs in the ImageNet dataset to learn to translate dog images between different breeds.
We only used the head regions, which were extracted by a template matching algorithm. Several
example results were shown in Figure 4. We found our method translated a dog to a different breed.
Cat species conversion. We also used the images of house cat, tiger, lion, cougar, leopard, jaguar,
and cheetah in the ImageNet dataset to learn to translate cat images between different species. We
only used the head regions, which again were extracted by a template matching algorithm. Several
example results were shown in Figure 5. We found our method translated a cat to a different specie.
Face attribute. We used the CelebA dataset [18] for attribute-based face images translation. Each face
image in the dataset had several attributes, including blond hair, smiling, goatee, and eyeglasses. The
face images with an attribute constituted the 1st domain, while those without the attribute constituted
the 2nd domain. In Figure 6, we visualized the results where we translated several images that do not
have blond hair, eye glasses, goatee, and smiling to corresponding images with each of the individual
attributes. We found that the translated face images were realistic.
Domain Adaptation. We applied the proposed framework to the problem for adapting a classifier
trained using labeled samples in one domain (source domain) to classify samples in a new domain
(target domain) where labeled samples in the new domain are unavailable during training. Early
works have explored ideas from subspace learning [4] to deep feature learning [5, 17, 26].
We performed multi-task learning where we trained the framework to 1) translate images between
the source and target domains and 2) classify samples in the source domain using the features
extracted by the discriminator in the source domain. Here, we tied the weights of the high-level
layers of D1 and D2 . This allows us to adapt a classifier trained in the source domain to the target
domain. Also, for a pair of generated images in different domains, we minimized the L1 distance
1
We used this architecture in an earlier version of the paper.
6
Figure 3: Street scene image translation results. For each pair, left is input and right is the translated image.
Input
Old Eng.
Sheep Dog
Input
Cougar
Husky
German
Shepherd
Corgi
Input
Husky
Corgi
Input
Leopard
Figure 4: Dog breed translation results.
Input
Cheetah
+Blond Hair +Eyeglasses
Leopard
Lion
Tiger
Figure 5: Cat species translation results.
+Goatee
+Smiling
Input
+Blond Hair +Eyeglasses
Figure 6: Attribute-based face translation results.
7
+Goatee
+Smiling
Table 2: Unsupervised domain adaptation performance. The reported numbers are classification accuracies.
Method
SA [4]
DANN [5]
DTN [26]
CoGAN
UNIT (proposed)
SVHN! MNIST
MNIST! USPS
USPS! MNIST
0.5932
-
0.7385
-
0.8488
-
0.9565
0.9315
0.9053
0.9597
0.9358
between the features extracted by the highest layer of the discriminators, which further encouraged
D1 and D2 to interpret a pair of corresponding images in the same way. We applied the approach to
several tasks including adapting from the Street View House Number (SVHN) dataset [20] to the
MNIST dataset and adapting between the MNIST and USPS datasets. Table 2 reported the achieved
performance with comparison to the competing approaches. We found that our method achieved a
0.9053 accuracy for the SVHN!MNIST task, which was much better than 0.8488 achieved by the
previous state-of-the-art method [26]. We also achieved better performance for the MNIST$SVHN
task than the Coupled GAN approach, which was the state-of-the-art. The digit images had a small
resolution. Hence, we used a small network. We also found that the cycle-consistency constraint was
not necessary for this task. More details about the experiments are available in Appendix B.
5
Related Work
Several deep generative models were recently proposed for image generation including GANs [6],
VAEs [13, 22], and PixelCNN [27]. The proposed framework was based on GANs and VAEs but it
was designed for the unsupervised image-to-image translation task, which could be considered as a
conditional image generation model. In the following, we first review several recent GAN and VAE
works and then discuss related image translation works.
GAN learning is via staging a zero-sum game between the generator and discriminator. The quality
of GAN-generated images had improved dramatically since the introduction. LapGAN [3] proposed
a Laplacian pyramid implementation of GANs. DCGAN [21] used a deeper convolutional network.
Several GAN training tricks were proposed in [24]. WGAN [1] used the Wasserstein distance.
VAEs optimize a variational bound. By improving the variational approximation, better image
generation results were achieved [19, 12]. In [14], a VAE-GAN architecture was proposed to improve
image generation quality of VAEs. VAEs were applied to translate face image attribute in [28].
Conditional generative model is a popular approach for mapping an image from one domain to
another. Most of the existing works were based on supervised learning [15, 8, 9]. Our work differed
to the previous works in that we do not need corresponding images. Recently, [26] proposed the
domain transformation network (DTN) and achieved promising results on translating small resolution
face and digit images. In addition to faces and digits, we demonstrated that the proposed framework
can translate large resolution natural images. It also achieved a better performance in the unsupervised
domain adaptation task. In [25], a conditional generative adversarial network-based approach was
proposed to translate a rendering images to a real image for gaze estimation. In order to ensure
the generated real image was similar to the original rendering image, the L1 distance between
the generated and original image was minimized. We note that two contemporary papers [29, 10]
independently introduced the cycle-consistency constraint for the unsupervised image translation.
We showed that that the cycle-consistency constraint is a natural consequence of the proposed
shared-latent space assumption. From our experiment, we found that cycle-consistency and the
weight-sharing (a realization of the shared-latent space assumption) constraints rendered comparable
performance. When the two constraints were jointed used, the best performance was achieved.
6
Conclusion and Future Work
We presented a general framework for unsupervised image-to-image translation. We showed it
learned to translate an image from one domain to another without any corresponding images in two
domains in the training dataset. The current framework has two limitations. First, the translation
model is unimodal due to the Gaussian latent space assumption. Second, training could be unstable
due to the saddle point searching problem. We plan to address these issues in the future work.
8
References
[1] M. Arjovsky, S. Chintala, and L. Bottou. Wasserstein gan. arXiv preprint arXiv:1701.07875, 2017.
[2] M. Cordts, M. Omran, S. Ramos, T. Scharw?chter, M. Enzweiler, R. Benenson, U. Franke, S. Roth, and
B. Schiele. The cityscapes dataset. Conference on Computer Vision and Pattern Recognition Workshop,
2015.
[3] E. L. Denton, S. Chintala, R. Fergus, et al. Deep generative image models using a laplacian pyramid of
adversarial networks. Advances in Neural Information Processing Systems, 2015.
[4] B. Fernando, A. Habrard, M. Sebban, and T. Tuytelaars. Unsupervised visual domain adaptation using
subspace alignment. International Conference on Computer Vision, 2013.
[5] Y. Ganin, E. Ustinova, H. Ajakan, P. Germain, H. Larochelle, F. Laviolette, M. Marchand, and V. Lempitsky.
Domain-adversarial training of neural networks. Journal of Machine Learning Research, 2016.
[6] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio.
Generative adversarial nets. Advances in Neural Information Processing Systems, 2014.
[7] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. Computer Vision and
Pattern Recognition, 2016.
[8] P. Isola, J.-Y. Zhu, T. Zhou, and A. A. Efros. Image-to-image translation with conditional adversarial
networks. Conference on Computer Vision and Pattern Recognition, 2017.
[9] J. Johnson, A. Alahi, and L. Fei-Fei. Perceptual losses for real-time style transfer and super-resolution.
European Conference in Computer Vision, 2016.
[10] T. Kim, M. Cha, H. Kim, J. Lee, and J. Kim. Learning to discover cross-domain relations with generative
adversarial networks. International Conference on Machine Learning, 2017.
[11] D. Kingma and J. Ba. Adam: A method for stochastic optimization. International Conference on Learning
Representations, 2015.
[12] D. P. Kingma, T. Salimans, and M. Welling. Improving variational inference with inverse autoregressive
flow. Advances in Neural Information Processing Systems, 2016.
[13] D. P. Kingma and M. Welling. Auto-encoding variational bayes. International Conference on Learning
Representations, 2014.
[14] A. B. L. Larsen, S. K. S?nderby, H. Larochelle, and O. Winther. Autoencoding beyond pixels using a
learned similarity metric. International Conference on Machine Learning, 2016.
[15] C. Ledig, L. Theis, F. Husz?r, J. Caballero, A. Cunningham, A. Acosta, A. Aitken, A. Tejani, J. Totz,
Z. Wang, et al. Photo-realistic single image super-resolution using a generative adversarial network.
Conference on Computer Vision and Pattern Recognition, 2017.
[16] T. Lindvall. Lectures on the coupling method. Courier Corporation, 2002.
[17] M.-Y. Liu and O. Tuzel. Coupled generative adversarial networks. Advances in Neural Information
Processing Systems, 2016.
[18] Z. Liu, P. Luo, X. Wang, and X. Tang. Deep learning face attributes in the wild. International Conference
on Computer Vision, 2015.
[19] L. Maal?e, C. K. S?nderby, S. K. S?nderby, and O. Winther. Auxiliary deep generative models. International Conference on Machine Learning, 2016.
[20] Y. Netzer, T. Wang, A. Coates, A. Bissacco, B. Wu, and A. Y. Ng. Reading digits in natural images with
unsupervised feature learning. Advances in Neural Information Processing Systems workshop, 2011.
[21] A. Radford, L. Metz, and S. Chintala. Unsupervised representation learning with deep convolutional
generative adversarial networks. International Conference on Learning Representations, 2016.
[22] D. J. Rezende, S. Mohamed, and D. Wierstra. Stochastic backpropagation and variational inference in
deep latent gaussian models. International Conference on Machine Learning, 2014.
[23] G. Ros, L. Sellart, J. Materzynska, D. Vazquez, and A. Lopez. The SYNTHIA Dataset: A large collection
of synthetic images for semantic segmentation of urban scenes. Conference on Computer Vision and
Pattern Recognition, 2016.
[24] T. Salimans, I. Goodfellow, W. Zaremba, V. Cheung, A. Radford, and X. Chen. Improved techniques for
training gans. Advances in Neural Information Processing Systems, 2016.
[25] A. Shrivastava, T. Pfister, O. Tuzel, J. Susskind, W. Wang, and R. Webb. Learning from simulated and
unsupervised images through adversarial training. Conference on Computer Vision and Pattern Recognition,
2017.
[26] Y. Taigman, A. Polyak, and L. Wolf. Unsupervised cross-domain image generation. International
Conference on Learning Representations, 2017.
[27] A. van den Oord, N. Kalchbrenner, L. Espeholt, O. Vinyals, A. Graves, et al. Conditional image generation
with pixelcnn decoders. Advances in Neural Information Processing Systems, 2016.
[28] X. Yan, J. Yang, K. Sohn, and H. Lee. Attribute2image: Conditional image generation from visual
attributes. European Conference in Computer Vision, 2016.
[29] J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros. Unpaired image-to-image translation using cycle-consistent
adversarial networks. International Conference on Computer Vision, 2017.
9
| 6672 |@word trial:1 version:2 nd:1 cha:1 d2:15 eng:1 q1:14 shot:1 harder:3 liu:3 configuration:2 existing:1 current:1 com:3 z2:18 luo:1 realize:1 realistic:4 designed:1 update:3 alone:1 generative:13 gan2:3 bissacco:1 zhang:1 wierstra:1 qualitative:1 consists:3 lopez:1 mingyuliutw:2 wild:1 ex2:1 aitken:1 cheetah:2 multi:4 ming:1 actual:1 little:1 becomes:1 discover:1 notation:1 unrelated:1 suffice:1 snowy:2 tying:1 q2:10 transformation:1 corporation:1 guarantee:1 quantitative:1 sky:1 charge:1 pg2:3 alahi:1 zaremba:1 ro:1 classifier:2 ustinova:1 control:3 unit:11 consequence:2 encoding:2 analyzing:1 ajakan:1 twice:1 studied:1 resembles:1 conversely:1 challenging:2 responsible:2 enforces:1 block:4 implement:2 backpropagation:1 digit:4 susskind:1 jan:1 tuzel:2 yan:1 adapting:3 matching:2 courier:1 word:2 road:1 applying:2 franke:1 optimize:1 equivalent:1 map:14 demonstrated:1 roth:1 resembling:1 independently:1 resolution:8 recovery:1 pouget:1 utilizing:1 regarded:1 regularize:1 reparameterization:2 searching:1 updated:1 target:7 goodfellow:2 trick:3 recognition:7 nderby:3 labeled:2 role:3 preprint:1 wang:4 worst:1 region:4 ensures:1 cycle:22 sun:1 contemporary:2 removed:3 highest:1 weigh:1 jaguar:1 schiele:1 warde:1 trained:9 solving:1 f2:7 usps:3 translated:15 joint:14 various:5 represented:1 cat:5 train:2 supervisedly:1 abused:1 corresponded:1 formation:1 hyper:5 kalchbrenner:1 posed:5 solve:2 larger:1 reconstruct:2 encoder:4 g1:34 tuytelaars:1 breed:4 jointly:2 final:1 autoencoding:1 differentiable:2 net:1 breuel:1 propose:3 reconstruction:8 adaptation:7 realization:2 ablation:1 gen:1 translate:13 achieve:1 supposed:1 satellite:7 generating:1 adam:2 object:1 coupling:2 ganin:1 sa:1 auxiliary:2 implemented:1 implies:3 larochelle:2 direction:1 correct:1 attribute:10 cnns:1 stochastic:2 human:1 translating:4 espeholt:1 f1:8 leopard:3 considered:5 ground:2 caballero:1 mapping:7 echoed:1 driving:1 efros:2 early:1 estimation:1 applicable:1 goatee:4 vice:2 defeating:1 city:1 shallowest:1 gaussian:5 aim:3 super:3 husz:1 zhou:1 vae:12 rezende:1 consistently:1 likelihood:4 adversarial:20 kim:3 glass:1 inference:2 el:2 cunningham:1 relation:1 pixel:8 issue:1 classification:1 ill:4 denoted:1 animal:1 art:4 constrained:1 plan:1 marginal:9 once:1 construct:1 beach:1 sampling:2 cartoon:1 encouraged:1 ng:1 park:1 yu:1 denton:1 unsupervised:25 constitutes:2 celeba:1 future:2 minimized:2 mirza:1 few:4 rainy:4 divergence:1 resulted:1 individual:2 summery:2 deviating:1 wgan:1 consisting:2 highly:1 evaluation:2 sheep:2 alignment:1 analyzed:1 operated:1 farley:1 staging:1 necessary:2 netzer:1 respective:2 tree:3 old:2 initialized:1 classify:2 modeling:1 earlier:1 measuring:2 deviation:1 subset:1 habrard:1 johnson:1 front:3 reported:5 encoders:3 perturbed:1 synthetic:5 lush:1 st:2 international:11 winther:2 oord:1 probabilistic:1 lee:2 decoding:1 gaze:1 gans:8 again:1 postulate:1 recorded:1 cougar:2 style:1 hypothesizes:1 dann:1 stream:16 eyeglass:3 performed:2 view:1 analyze:1 recover:1 bayes:1 metz:1 kautz:1 minimize:1 accuracy:18 convolutional:5 variance:1 yield:1 ofthe:1 translator:1 tejani:1 decodes:1 ren:1 cc:1 vazquez:1 sharing:19 larsen:1 mohamed:1 e2:28 chintala:3 transposed:1 sampled:3 ledig:1 dataset:14 holdout:1 popular:1 color:4 car:3 segmentation:1 back:6 supervised:4 day:4 totz:1 modal:1 improved:2 formulation:1 cyclegan:1 autoencoders:3 night:3 lack:1 quality:4 gray:1 usa:1 building:1 smiling:4 consisted:4 true:2 hence:5 regularization:1 alternating:1 semantic:2 illustrated:2 conditionally:1 ex1:1 game:2 self:1 during:2 leakyrelu:1 l1:2 svhn:4 image:189 variational:9 meaning:1 recently:2 common:2 sebban:1 enzweiler:1 extend:1 interpretation:1 discussed:2 approximates:2 interpret:2 assembling:1 he:1 composition:3 versa:2 consistency:17 px1:3 similarly:3 nonlinearity:1 had:5 pixelcnn:2 similarity:1 cityscape:3 showed:6 recent:1 perspective:1 termed:1 nvidia:2 seen:1 arjovsky:1 additional:5 wasserstein:2 isola:2 fernando:1 dashed:1 full:1 unimodal:2 infer:3 px2:2 adapt:1 cross:2 long:1 e1:31 ez1:3 paired:3 laplacian:3 impact:5 basic:2 hair:4 vision:12 metric:3 arxiv:2 iteration:1 represent:2 pyramid:2 achieved:17 penalize:2 addition:2 source:7 modality:1 benenson:1 rest:1 ascent:1 shepherd:2 flow:1 extracting:1 yang:1 intermediate:1 bengio:1 easy:1 rendering:2 variate:1 architecture:8 competing:2 polyak:1 idea:2 whether:1 deep:8 dramatically:1 useful:1 dark:1 visualized:2 sohn:1 unpaired:1 reduced:1 http:2 generate:3 occupy:1 exist:6 cordts:1 wisely:1 coates:1 hyperparameter:1 key:1 drawn:2 urban:1 sum:2 taigman:1 inverse:1 arrive:2 wu:1 appendix:2 jointed:1 comparable:1 layer:23 bound:2 courville:1 marchand:1 yielded:1 constraint:21 fei:2 scene:6 x2:46 min:1 rendered:1 sunny:4 pg1:2 beneficial:1 reconstructing:1 acosta:1 den:1 lapgan:1 discus:2 german:2 end:5 subnetworks:4 photo:1 maal:1 available:4 operation:3 apply:7 salimans:2 enforce:2 batch:1 existence:1 thomas:1 original:3 assumes:1 ensure:3 gan:16 laviolette:1 objective:9 interacts:1 gradient:3 subspace:2 distance:4 mapped:7 simulated:1 street:6 decoder:1 unstable:1 ozair:1 assuming:1 code:21 colorization:1 mini:2 illustration:1 minimizing:3 difficult:1 webb:1 relate:2 negative:4 ba:1 implementation:1 proper:1 perform:1 upper:1 cogan:2 conversion:2 datasets:3 benchmark:2 descent:1 reparameterized:1 team:2 husky:3 head:2 varied:1 stack:1 introduced:1 pair:17 required:1 kl:11 dog:6 connection:2 discriminator:11 z1:26 imagenet:2 germain:1 learned:2 kingma:3 nip:1 address:3 beyond:1 usually:2 lion:2 pattern:6 reading:1 challenge:1 summarize:1 including:6 green:2 max:2 video:1 treated:1 eh:2 natural:3 ramos:1 residual:6 zhu:2 scheme:1 improve:1 github:2 eye:1 axis:1 coupled:5 auto:1 omran:1 prior:3 review:1 theis:1 graf:1 ez2:3 loss:3 lecture:1 generation:10 limitation:1 versus:2 generator:6 validation:1 consistent:1 share:2 translation:59 changed:1 last:2 english:1 dis:4 allow:1 deeper:1 template:2 face:11 gan1:3 absolute:1 van:1 curve:1 default:1 depth:2 evaluating:1 dtn:2 autoregressive:1 collection:2 made:1 counted:1 attribute2image:1 welling:2 reconstructed:4 compact:1 fergus:1 latent:41 table:4 promising:1 learn:4 transfer:1 ca:1 shrivastava:1 unavailable:1 improving:2 bottou:1 european:2 domain:70 inheriting:1 did:1 constituted:2 nothing:2 x1:57 xu:1 differed:1 inferring:1 decoded:2 momentum:1 house:2 tied:3 perceptual:1 learns:1 tang:1 chter:1 showing:1 explored:1 abadie:1 admits:1 exists:3 consist:1 mnist:7 false:2 adding:1 effectively:1 workshop:2 chen:1 easier:1 led:1 saddle:2 visual:3 vinyals:1 contained:1 dcgan:1 g2:32 radford:2 wolf:1 truth:2 extracted:6 prop:1 conditional:7 lempitsky:1 identity:1 goal:1 cheung:1 consequently:1 shared:21 tiger:2 infinite:3 specifically:3 called:1 specie:4 blond:4 pfister:1 player:5 vaes:14 evaluate:1 d1:17 |
6,271 | 6,673 | Coded Distributed Computing for Inverse Problems
Yaoqing Yang, Pulkit Grover and Soummya Kar
Carnegie Mellon University
{yyaoqing, pgrover, soummyak}@andrew.cmu.edu
Abstract
Computationally intensive distributed and parallel computing is often bottlenecked
by a small set of slow workers known as stragglers. In this paper, we utilize the
emerging idea of ?coded computation? to design a novel error-correcting-code
inspired technique for solving linear inverse problems under specific iterative
methods in a parallelized implementation affected by stragglers. Example machinelearning applications include inverse problems such as personalized PageRank and
sampling on graphs. We provably show that our coded-computation technique can
reduce the mean-squared error under a computational deadline constraint. In fact,
the ratio of mean-squared error of replication-based and coded techniques diverges
to infinity as the deadline increases. Our experiments for personalized PageRank
performed on real systems and real social networks show that this ratio can be
as large as 104 . Further, unlike coded-computation techniques proposed thus far,
our strategy combines outputs of all workers, including the stragglers, to produce
more accurate estimates at the computational deadline. This also ensures that the
accuracy degrades ?gracefully? in the event that the number of stragglers is large.
1
Introduction
The speed of distributed computing is often affected by a few slow workers known as the ?stragglers?
[1?4]. This issue is often addressed by replicating tasks across workers and using this redundancy to
ignore some of the stragglers. Recently, methods from error-correcting codes (ECC) have been used
for speeding up distributed computing [5?15], which build on classical works on algorithm-based
fault-tolerance [16]. The key idea is to treat stragglers as ?erasures? and use ECC to retrieve the result
after a subset of fast workers have finished. In some cases, (e.g. [6, 8] for matrix multiplications),
techniques that utilize ECC achieve scaling-sense speedups in average computation time compared to
replication. In this work, we propose a novel coding-inspired technique to deal with stragglers in
distributed computing of linear inverse problems using iterative solvers [17].
Existing techniques that use coding to deal with stragglers treat straggling workers as ?erasures?, that
is, they ignore computation results of the stragglers. In contrast, when using iterative methods for
linear inverse problems, even if the computation result at a straggler has not converged, the proposed
algorithm does not ignore the result, but instead combines it (with appropriate weights) with results
from other workers. This is in part because the results of iterative methods often converge gradually
to the true solutions. We use a small example shown in Fig. 1 to illustrate this idea. Suppose we
want to solve two linear inverse problems with solutions x?1 and x?2 . We ?encode the computation? by
adding an extra linear inverse problem with solution x?1 + x?2 (see Section 3), and distribute these
three problems to three workers. Using this method, the solutions x?1 and x?2 can be obtained from
the results of any combination of two fast workers that first return their solutions.
But what if we have a computational deadline, Tdl , by which only one worker converges? The natural
extension of existing strategies (e.g., [6]) will declare a failure because it needs at least two workers
to respond. However, our strategy does not require convergence: even intermediate results can be
utilized to estimate solutions. In other words, our strategy degrades gracefully as the number of
stragglers increases, or as the deadline is pulled earlier. Indeed, we show that it is suboptimal to
ignore stragglers as erasures, and design strategies that treat the difference from the optimal solution
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
General coded computation
r1
r2
E
n
c
o
d
e
r1
r2
r1+r2
Processor1
Processor2
(slow)
Processor3
(slow)
waitEforEtwoE
fastEworkers
*1
D
e
c
o
ignore!
d
x1*+x2*
e
x
x*2
Proposed coded method
fail
E
n
c
o
d
e
r1
r2
r1
r2
r1+r2
DeadlineETdl
Processor1
x1* +e1
Processor2
(slow)
x2* +e2
Processor3
(slow)
x1* +x2* +e3
D
e weightedE
c combination
o
d
e
Figure 1: A comparison between the existing scheme in [6] and the proposed algorithm.
as ?soft? additive noise (see Section 3). We use an algorithm that is similar to weighted least-squares
for decoding, giving each worker a weight based on its proximity to convergence. In this way, we can
expect to fully utilize the computation results from all workers and obtain better speedup.
Theoretically, we show that for a specified deadline time Tdl , under certain conditions on worker
speed distributions, the coded linear inverse solver using structured codes has smaller mean squared
error than the replication-based linear solver (Theorem 4.4). In fact, under more relaxed conditions
on worker speed distributions, when the computation time Tdl increases, the ratio of the mean-squared
error (MSE) of replication-based and coded linear solvers can get arbitrarily large (Theorem 4.5)!
For validation of our theory, we performed experiments to compare coded and replication-based
computation for a graph mining problem, namely personalized PageRank [18] using the classical
power-iteration method [19]. We conduct experiments on the Twitter and Google Plus social networks
under a deadline on computation time using a given number of workers on a real computation cluster
(Section 6). We observe that the MSE of coded PageRank is smaller than that of replication by
a factor of 104 at Tdl = 2 seconds. From an intuitive perspective, the advantage of coding over
replication is that coding utilizes the diversity of all heterogeneous workers, whereas replication
cannot (see section 7 for details). To compare with existing coded technique in [6], we adapt it to
inverse problems by inverting only the partial results from the fast workers. However, from our
experiments, if only the results from the fast workers are used, the error amplifies due to inverting an
ill-conditioned submatrix during decoding (Section 6). This ill-conditioning issue of real-number
erasure codes has also been recognized in a recent communication problem [20]. In contrast, our
novel way of combining all the partial results including those from the stragglers helps bypass the
difficulty of inverting an ill-conditioned matrix.
The focus of this work is on utilizing computations to deliver the minimal MSE in solving linear
inverse problems. Our algorithm does not reduce the communication cost. However, because each
worker performs sophisticated iterative computations in our problem, such as the power-iteration
computations, the time required for computation dominates that of communication (Section 5.2).
This is unlike some recent works (e.g.[21?24]) where communication costs are observed to dominate
because the per-processor computation is smaller.
Finally, we summarize our main contributions in this paper:
? We propose a coded computing algorithm for multiple instances of a linear inverse problem;
? We theoretically analyze the mean-squared error of coded, uncoded and replication-based
iterative linear solvers under a deadline constraint, and show scaling sense advantage of
coded solvers in theory and orders of magnitude smaller error in data experiments.
? This is the first work that treats stragglers as soft errors instead of erasures, which leads to
graceful degradation in the event that the number of stragglers is large.
2
2.1
System Model and Problem Formulation
Preliminaries on Solving Linear Systems using Iterative Methods
Consider the problem of solving k inverse problems with the same linear transform matrix M and
different inputs ri : Mxi = ri , i = 1, 2, . . . k. When M is a square matrix, the closed-form solution
is xi = M?1 ri . When M is a non-square matrix, the regularized least-square solution is xi =
(M> M + ?I)?1 M> ri , i = 1, 2, . . . k, with an appropriate regularization parameter ?. Since matrix
inversion is hard, iterative methods are often used. We now look at two ordinary iterative methods,
namely the Jacobian method [17] and the gradient descent method. For a square matrix M = D + L,
(l+1)
(l)
where D is diagonal, the Jacobian iteration is written as xi
= D?1 (ri ? Lxi ). Under certain
conditions of D and L ([17, p.115]), the computation result converges to the true solution. One
example is the PageRank algorithm discussed in Section 2.2. For the `2 -minimization problem with a
(l+1)
(l)
non-square M, the gradient descent method has the form xi
= ((1??)I?M> M)xi +M> ri ,
2
where is an appropriate step-size. We can see that both the Jacobian iteration and the gradient
descent iteration mentioned above have the form
(l+1)
(l)
xi
= Bxi + Kri , i = 1, 2, . . . k,
(1)
?
for two appropriate matrices B and K, which solves the following equation with true solution xi :
x?i = Bx?i + Kri , i = 1, 2, . . . k.
(2)
(l)
(l)
Therefore, subtracting (2) from (1), we have that the computation error ei = xi ? x?i satisfies
(l+1)
(l)
ei
= Bei .
(3)
For the iterative method to converge, we always assume the spectral radius ?(B) < 1 (see [17,
p.115]). We will study iterative methods that have the form (1) throughout this paper.
2.2
Motivating Applications of Linear Inverse Problems
Our coded computation technique requires solving multiple inverse problems with the same linear
transform matrix M. One such problem is personalized PageRank. For a directed graph, the
PageRank algorithm [19] aims to measure the nodes? importance by solving the linear problem
x = Nd 1N + (1 ? d)Ax, where d = 0.15 is called the ?teleport? probability, N is the number of
nodes and A is the column-normalized adjacency matrix. The personalized PageRank problem [18]
considers a more general equation x = dr + (1 ? d)Ax, for any possible vector r ? RN that satisfies
1> r = 1. Compared to PageRank [19], personalized PageRank [18] incorporates r as the preference
of different users or topics. A classical method to solve PageRank is power-iteration, which iterates
the computation x(l+1) = dr + (1 ? d)Ax(l) until convergence. This iterative method is the same
as (1), which is essentially the Jacobian method mentioned above. Another example application is
the sampling and recovery problem in the emerging field of graph signal processing [25, 26] as a
non-square system, which is discussed in Supplementary section 8.1.
2.3
Problem Formulation: Distributed Computing and the Straggler Effect
Consider solving k linear inverse problems Mxi = ri , i = 1, 2, . . . k in n > k workers using the
iterative method (1), where each worker solves one inverse problem. Due to the straggler effect, the
computation at different workers can have different speeds. The goal is to obtain minimal MSE in
solving linear inverse problems before a deadline time Tdl . Suppose after Tdl , the i-th worker has
completed li iterations in (1). Then, from (3), the residual error at the i-th worker is
(l )
(0)
ei i = Bli ei .
(4)
For our theoretical results, we sometimes need the following assumption.
Assumption 1. We assume that the optimal solutions x?i , i = 1, 2, . . . k, are i.i.d.
Denote by ?E and CE respectively the mean and the covariance of each x?i . Note that Assumption 1
is equivalent to the assumption that the inputs ri , i = 1, 2, . . . k are i.i.d., because ri and x?i are
related by the linear equation (2). For the personalized PageRank problem discussed above, this
assumption is reasonable because queries from different users or topics are unrelated. Assume we
(0)
have estimated the mean ?E beforehand and we start with the initial estimate xi = ?E . Then,
(0)
(0)
ei = xi ? x?i has mean 0N and covariance CE . We also try to extend our results for the case
when x?i ?s (or equivalently, ri ?s) are correlated. Since the extension is rather long and may hinder the
understanding of the main paper, we provide it in supplementary section 8.2 and section 8.5.
2.4
Preliminaries on Error Correcting Codes
We will use ?encode? and ?decode? to denote preprocessing and post-processing before and after
parallel computation. In this paper, the encoder multiplies the inputs to the parallel workers with a
?generator matrix? G and the decoder multiplies the outputs of the workers with a ?decoding matrix?
L (see Algorithm 1). We call a code an (n, k) code if the generator matrix has size k ? n. We often
use generator matrices G with orthonormal rows, which means Gk?n G>
n?k = Ik . An example of
such a matrix is the submatrix formed by any k rows of an n ? n orthonormal matrix (e.g., a Fourier
matrix). Under this assumption, Gk?n can be augmented toform an n ?
n orthonormal matrix using
Gk?n
another matrix H(n?k)?n , i.e. the square matrix Fn?n =
satisfies F> F = In .
H(n?k)?n
3
3
Coded Distributed Computing of Linear Inverse Problems
The proposed coded linear inverse algorithm (Algorithm 1) has three stages: (1) preprocessing
(encoding) at the central controller, (2) parallel computing at n > k parallel workers, and (3) postprocessing (decoding) at the central controller. As we show later in the analysis of computing error,
the entries trace(C(li )) in the diagonal matrix ? are the expected MSE at each worker prior to
decoding. The decoding matrix Lk?n in the decoding step (7) is chosen to be (G??1 G> )?1 G??1
to reduce the mean-squared error of the estimates of linear inverse solutions by assigning different
weights to different workers based on the estimated accuracy of their computation (which is what ?
provides). This particular choice of ? is inspired from the weighted least-square solution.
Algorithm 1 Coded Distributed Linear Inverse
Input: Input vectors [r1 , r2 , . . . , rk ], generator matrix Gk?n , the linear system matrices B and K
defined in (1).
Initialize (Encoding): Encode the input vectors and the initial estimates by multiplying G:
[s1 , s2 , . . . , sn ] = [r1 , r2 , . . . , rk ] ? G.
(0)
(0)
[y1 , y2 , . . . , yn(0) ]
(5)
(0)
(0)
(0)
[x1 , x2 , . . . , xk ]
=
? G.
(6)
Parallel Computing:
for i = 1 to n (in parallel) do
(0)
(0)
Send si and yi to the i-th worker. Execute the iterative method (1) with initial estimate yi
and input si at each worker.
end for
(l )
After a deadline time Tdl , collect all linear inverse results yi i from these n workers. The
(li )
superscript li in yi represents that the i-th worker finished li iterations. Denote by Y(Tdl ) the
(Tdl )
(l )
(l )
(l )
collection of all results YN ?n
= [y1 1 , y2 2 , . . . , yn n ].
Post Processing (decoding at the central controller):
Compute an estimate of the linear inverse solutions using the following matrix multiplication:
? > = L ? (Y(Tdl ) )> := (G??1 G> )?1 G??1 (Y(Tdl ) )> ,
X
(7)
? N ?k = [?
?2, . . . , x
? k ], the matrix ? is
where the estimate X
x1 , x
? = diag [trace(C(l1 )), . . . , trace(C(ln ))] ,
(8)
where the matrices C(li ), i = 1, . . . , n are defined as
C(li ) = Bli CE (B> )li .
(9)
In computation of ?, if trace(C(li )) are not available, one can use precomputed estimates of this
trace as discussed in Supplementary Section 8.9 with negligible computational complexity and
theoretically guaranteed accuracy.
3.1
Bounds on Performance of the Coded Linear Inverse Algorithm
Define l = [l1 , l2 , . . . ln ] as the vector of the number of iterations at all workers. E[?|l] denotes the
conditional expectation taken with respect to the randomness of the optimal solution x?i (see Assumption 1) conditioned on fixed iteration number li at each worker, i.e., E[X|l] = E[X|l1 , l2 , . . . ln ].
Define X?N ?k = [x?1 , x?2 , . . . x?k ] as the matrix composed of all the true solutions.
? ? X? , i.e., the error of the decoding result (7). Assuming that the
Theorem 3.1. Define E = X
solutions for each linear inverse problem are chosen i.i.d. (across all problems) according to a
distribution with covariance CE . Then, the error covariance of E satisfies
2
E[kEk |l] ? ?max (G> G)trace (G??1 G> )?1 ,
(10)
where the norm k?k is the Frobenius norm, ?max (G> G) is the maximum eigenvalue of G> G and
the matrix ? is defined in (8). Further, when G has orthonormal rows,
2
E[kEk |l] ? trace (G??1 G> )?1 ,
(11)
4
Proof overview. See supplementary Section 8.3 for the complete proof. Here we provide the main
intuition by analyzing a ?scalar version? of the linear inverse problem, in which case the matrix B is
equal to a scalar a. For B = a, the inputs and the initial estimates in (5) and (6) are vectors instead
of matrices. As we show in Supplementary Section 8.3, if we encode both the inputs and the initial
estimates using (5) and (6), we also ?encode? the error
(0)
(0)
(0)
(0)
(0)
[1 , 2 , . . . , (0)
n ] = [e1 , e2 , . . . , ek ] ? G =: E0 G,
(0)
(0)
(0)
(12)
(0)
where i = yi ? yi? is the initial error at the i-th worker, ei = xi ? x?i is the initial error of
(0) (0)
(0)
(0)
the i-th linear inverse problem, and E0 := [e1 , e2 , . . . ek ]. Suppose var[ei ] = ce , which is a
scalar version of CE after Assumption 1. From (4), the error satisfies:
(l )
(0)
i i = ali i , i = 1, 2, . . . n.
(13)
Denote by D = diag{al1 , al2 , . . . aln }. Therefore, from (12) and (13), the error before the decoding
step (7) can be written as
(l )
(l )
(0)
(0)
(0)
n)
[1 1 , 2 2 , . . . (l
n ] =[1 , 2 , . . . n ] ? D = E0 GD.
(14)
We can show (see Supplementary Section 8.3 for details) that after the decoding step (7), the error
vector is also multiplied by the decoding matrix L = (G??1 G> )?1 G??1 :
h
i>
(l ) (l )
n)
(15)
E> = L 1 1 , 2 2 , . . . (l
= LD> G> E>
n
0.
Thus,
2
>
E[kEk |l] =E[trace[E> E]|l] = trace[LD> G> E[E>
0 E0 |l]GDL ]
(a)
= trace[LD> G> ce Ik GDL> ] = ce trace[LD> G> GDL> ]
(16)
(b)
?ce ?max (G> G)trace[LD> DL> ] = ?max (G> G)trace[L(ce D> D)L> ]
(c)
(d)
= ?max (G> G)trace[L?L> ] = ?max (G> G)trace[(G??1 G> )?1 ],
(0)
(0)
(0)
(0)
where (a) holds because E0 := [e1 , e2 , . . . ek ] and var[ei ] = ce , (b) holds because G> G
?max (G> G)In , (c) holds because ce D> D = ?, which is from the fact that for a scalar linear system
matrix B = a, the entries in the ? matrix in (8) satisfy
trace(C(li )) = ali ce (a> )li = ce a2li ,
(17)
>
which is the same as the entries in the diagonal matrix ce D D. Finally, (d) is obtained by directly plugging in L := (G??1 G> )?1 G??1 . Finally, inequality 11 holds because when G has
orthonormal rows, ?(G> G) = 1.
Additionally, we note that in (10), the term trace (G??1 G> )?1 resembles the MSE of ordinary
weighted least-square solution, and the term ?max (G> G) represents the ?inaccuracy? due to using
the weighted least-square solution as the decoding result, because the inputs to different workers
become correlated by multiplying the i.i.d. inputs with matrix G (see (5)).
4
Comparison with Uncoded Schemes and Replication-based Schemes
Here, we often assume (we will state explicitly in the theorem) that the number of iterations li at
different workers are i.i.d.. We use Ef [?] to denote expectation on randomness of both the linear
inverse solutions x?i and the number of iterations li (this is different from the notation E[?|l]).
Assumption 2. Within time Tdl , the number of iterations of linear inverse computations (see (1)) at
each worker follows an i.i.d. distribution li ? f (l).
4.1
Comparison between the coded and uncoded linear inverse before a deadline
First, we compare the coded linear inverse scheme with an uncoded scheme, in which case we use
the first k workers to solve k linear inverse problems in (2) without coding. The following theorem
quantifies the overall mean-squared error of the uncoded scheme given l1 , l2 , . . . , lk . The proof is in
Supplementary Section 8.6.
5
2
h
i
(l )
(l )
2
Theorem 4.1. In the uncoded scheme, the error E kEuncoded k |l = E
[e1 1 . . . , ek k ]
l =
Pk
i=1 trace (C(li )). Further, when the i.i.d. Assumption 2 holds,
h
i
2
Ef kEuncoded k = kEf [trace(C(l1 ))].
(18)
Then, we compare the overall mean-squared error of coded and uncoded linear inverse algorithms.
Note that this comparison is not fair because the coded algorithm uses more workers than uncoded.
However, we still include Theorem 4.2 because we need it for the fair comparison between coded and
replication-based linear inverse. The proof is in Supplementary section 8.4.
Theorem 4.2. (Coded linear inverse beats uncoded) Suppose the i.i.d. Assumptions 1 and
2 hold and
Gk?n
suppose G is a k ? n submatrix of an n ? n Fourier transform matrix F, i.e., Fn?n =
.
H(n?k)?n
Then, expected error of the coded linear inverse is strictly less than that of uncoded:
h
i
h
i
2
2
>
Ef kEuncoded k ? Ef kEcoded k ? Ef [trace(J2 J?1
(19)
4 J2 )],
J1 J2
where J2 and J4 are the submatrices of F?F> := >
and the matrix ? is defined in
J2 J4 n?n
(8). That is, (J1 )k?k is G?G> , (J2 )k?(n?k) is G?H> , and (J4 )(n?k)?(n?k) is H?H> .
4.2
Comparison between the replication-based and coded linear inverse before a deadline
Consider an alternative way of doing linear inverse using n > k workers. In this paper, we only
consider the case when n ? k < k, i.e., the number of extra workers is only slightly bigger than
the number of problems (both in theory and in experiments). Since we have n ? k extra workers,
a natural way is to pick any (n ? k) linear inverse problems and replicate them using these extra
(n ? k) workers. After we obtain two computation results for the same equation, we use two
natural ?decoding? strategies for this replication-based linear inverse: (i) choose the worker with
1
2
higher number of iterations; (ii) compute the weighted average using weights w1w+w
and w1w+w
,
2
2
p
p
where w1 = 1/ trace(C(l1 )) and w2 = 1/ trace(C(l2 )), and l1 and l2 are the number of iterations
completed at the two workers (recall that trace(C(li )) represents the residual MSE at the i-th worker).
Theorem 4.3. The replication-based schemes satisfy the following lower bound on the MSE:
h
i
h
i
2
2
Ef kErep k >Ef kEuncoded k ? (n ? k)Ef [trace(C(l1 ))].
(20)
Proof overview. Here the goal is to obtain a lower bound on the MSE of replication-based linear
inverse and compare it with an upper bound on the MSE of coded linear inverse.
Note that if an extra worker is used to replicate the computation at the i-th worker, i.e., the linear
inverse problem with input ri is solved on two workers, the expected error of the result of the i-th
problem could at best reduced from Ef [trace(C(l1 ))] (see Thm. 4.1) to zero1 . Therefore, (n?k) extra
workers make the error decrease by at most (and strictly smaller than) (n ? k)Ef [trace(C(l1 ))].
Using this lower bound, we can provably show that coded linear inverse beats replication-based linear
inverse when certain conditions are satisfied. One crucial condition is that the distribution of the
random variable trace(C(l)) (i.e., the expected MSE at each worker) satisfies a ?variance heavy-tail?
property defined as follows.
Definition 1. The random variable trace(C(l)) is said to have a ??-variance heavy-tail? property if
varf [trace(C(l))] > ?E2f [trace(C(l))],
(21)
for some constant ? > 1. Notice that the term trace(C(l)) is essentially the remaining MSE after l
iterations at a single machine. Therefore, this property simply means the remaining error at a single
machine has large variance. For the coded linear inverse, we will use a ?Fourier code?, the generator
matrix G of which is a submatrix of a Fourier matrix. This particular choice of code is only for ease
of analysis in comparing coded linear inverse and replication-based linear inverse. In practice, the
code that minimizes mean-squared error should be chosen.
1
Although this is clearly a loose bound, it makes for convenient comparison with coded linear inverse.
6
Theorem 4.4. (Coded linear inverse beats replication) Suppose the i.i.d. Assumptions 1 and 2 hold?and
G is a k ? n submatrix of k rows of an n ? n Fourier matrix F. Further, suppose (n ? k) = o( n).
Then, the expected error of the coded linear inverse satisfies
i
h
ii var [trace(C(l ))]
1 h h
f
1
2
2
lim
Ef kEuncoded k ? Ef kEcoded k
?
.
(22)
n?? n ? k
Ef [trace(C(l1 ))]
Moreover, if the random variable trace(C(l)) satisfies the ?-variance heavy-tail property for ? > 1,
coded linear inverse outperforms replication-based linear inverse in the following sense,
lim
n??
1
1
1
lim
Ef kEuncoded k2 ? Ef kErep k2 <
Ef kEuncoded k2 ? Ef kEcoded k2 .
(n ? k)
? n?? (n ? k)
(23)
Proof overview. See Supplementary Section 8.7 for a complete and rigorous proof.
h Here we only
i
2
provide the main intuition behind the proof. From Theorem 4.2, we have Ef kEuncoded k ?
h
i
2
>
Ef kEcoded k ? Ef [trace(J2 J?1
4 J2 )]. Therefore, to prove (22), the main technical difficulty is
?1 >
to simplifythe term trace(J
2 J4 J2 ). For a Fourier matrix F, we are able to show that the matrix
J1 J2
F?F> = >
(see Theorem 4.2) is a Toeplitz matrix, which provides a good structure for
J2 J 4
us to study its behavior. Then, we use the Gershgorin circle theorem [27] (with some algebraic
manipulations) to show that the maximum eigenvalue of J4 satisfies ?max (J4 ) ? Ef [trace(C(l1 ))],
and separately using some algebraic manipulations, we show
trace(J2 J>
2 ) ? (n ? k)varf [trace(C(l1 ))],
>
?1 >
for large matrix size n. Since trace(J2 J?1
J2 ) =
4 J2 ) ? trace(J2 (?max (J4 ))
(24)
1
>
?max (J4 ) trace(J2 J2 ),
(n ? k)varf [trace(C(l1 ))]
,
(25)
Ef [trace(C(l1 ))]
for large n. Then, (22) can be proved by plugging (25) into (19). After that, we can combine (22),
(20) and the variance heavy-tail property to prove (23).
>
trace(J2 J?1
4 J2 ) ?
4.3
Asymptotic Comparison between Coded, Uncoded and Replication-based linear inverse
as the Deadline Tdl ? ?
Assumption 3. We assume the computation time of one power iteration is fixed at each worker for
each linear inverse computation, i.e., there exist n independent (not necessarily identically distributed)
random variables v1 , v2 , . . . vn such that li = d Tvdli e, i = 1, 2, . . . n.
The above assumption is validated in experiments in Supplementary Section 8.13.
The k-th order statistic of a sample is equal to its k-th smallest value. Suppose the order statistics
of the sequence v1 , v2 , . . . vn are vi1 < vi2 < . . . vin , where {i1 , i2 , . . . in } is a permutation of
{1, 2, . . . n}. Denote by [k] the set {1, 2, . . . k} and [n] the set {1, 2, . . . n}.
Theorem 4.5. (Error exponent comparison when Tdl ? ?) Suppose the i.i.d. Assumption 1 and
Assumption 3 hold. Suppose n ? k < k. Then, the error exponents of the coded and uncoded
computation schemes satisfy
1
2
1
2
log E[kEcoded k |l] ?
log
,
(26)
Tdl ??
Tdl
vi k
1?d
1
1
1
2
2
2
lim ?
log E[kEuncoded k |l] = lim ?
log E[kErep k |l] =
log
. (27)
Tdl ??
Tdl ??
Tdl
Tdl
maxi?[k] vi
1?d
lim ?
The error exponents satisfy coded>replication=uncoded. Here the expectation E[?|l] is only taken
with respect to the randomness of the linear inverse sequence xi , i = 1, 2, . . . k.
Proof overview. See Supplementary Section 8.8 for a detailed proof. The main intuition behind this
result is the following: when Tdl approaches infinity, the error of uncoded computation is dominated
7
by the slowest worker among the first k workers, which has per-iteration time maxi?[k] vi . For the
replication-based scheme, since the number of extra workers n?k < k, there is a non-zero probability
(which does not change with Tdl ) that the n ? k extra workers do not replicate the computation in
the slowest one among the first worker. Therefore, replication when n ? k < k does not improve
the error exponent, because the error is dominated by this slowest worker. For coded computation,
we show in Supplementary Section 8.8 that the slowest n ? k workers among the overall n workers
do not affect the error exponent, which means that the error is dominated by the k-th fastest worker,
which has per-iteration time vik . Since the k-th fastest worker among all n workers can not be slower
than the slowest one among the first (unordered) k workers, the error exponent of coded linear inverse
is larger than that of the uncoded and the replication-based linear inverse.
5
5.1
Analyzing the Computational Complexity
Encoding and decoding complexity
We first show that the encoding and decoding complexity of Algorithm 1 are in scaling-sense smaller
than that of the computation at each worker. This ensures that straggling comes from the parallel
workers, not the encoder or decoder. The proof of Theorem 5.1 is in Supplementary Section 8.10. In
our experiment on the Google Plus graph (See Section 6) for computing PageRank, the computation
time at each worker is 30 seconds and the encoding and decoding time at the central controller is
about 1 second.
Theorem 5.1. The computational complexity for the encoding and decoding is ?(nkN ), where N is
the number of rows in the matrix B and k, n depend on the number of available workers assuming
that each worker performs a single linear inverse computation. For a general dense matrix B, the
computational complexity of computing linear inverse at each worker is ?(N 2 l), where l is the
number of iterations in the specified iterative algorithm. The complexity of encoding and decoding is
smaller than that of the computation at each user for large B matrices (large N ).
5.2
Analysis on the cost of communication versus computation
In this work, we focus on optimizing the computation cost. However, what if the computation cost is
small compared to the overall cost, including the communication cost? If this is true, optimizing the
computation cost is not very useful. In Theorem 5.2 (proof appears in Supplementary Section 8.11),
we show that the computation cost is larger than the communication cost in the scaling-sense.
Theorem 5.2. The ratio between the number of operations (computation) and the number of bits trans? operations
mitted (communication) at the i-th worker is COSTcomputation /COSTcommunication = ?(li d)
?
per integer, where li is the number of iterations at the i-th worker, and d is the average number of
non-zeros in each row of the B matrix.
6
Experiments on Real Systems
We test the performance of the coded linear inverse algorithm for the PageRank problem on the
Twitter graph and the Google Plus graph from the SNAP datasets [28]. The Twitter graph has 81,306
nodes and 1,768,149 edges, and the Google Plus graph has 107,614 nodes and 13,673,453 edges. We
use the HT-condor framework in a cluster to conduct the experiments. The task is to solve k = 100
personalized PageRank problems in parallel using n = 120 workers. The uncoded algorithm picks
the first k workers and uses one worker for each PageRank problem. The two replication-based
schemes replicate the computation of the first n ? k PageRank problems in the extra n ? k workers
(see Section 4.2). The coded PageRank uses n workers to solve these k = 100 equations using
Algorithm 1. We use a (120, 100) code where the generator matrix is the submatrix composed of the
first 100 rows in a 120 ? 120 DFT matrix. The computation results are shown in the left two figures in
Fig. 2. Note that the two graphs are of different sizes so the computation in the two experiments take
different time. From Fig. 2, we can see that the mean-squared error of uncoded and replication-based
schemes is larger than that of coded computation by a factor of 104 for large deadlines.
We also compare Algorithm 1 with the coded computing algorithm proposed in [6]. As we discussed
in the Figure 1, the original coded technique in [6] ignores partial results and is suboptimal even in the
toy example of three workers. However, it has a natural extension to iterative methods, which will be
8
Average mean-squared error
Google Plus graph
10 0
10 0
Repetition-2
Google Plus graph
Twitter graph
Uncoded
Uncoded
Coded
Repetition-1
10 -5
Twitter graph
Binary
10 0
10 0
10 -5
Repetition-2
Extension of coded
Method in [Lee et.al.]
Gaussian
>10 4
DFT
Original Coded
Method in [Lee et.al.]
Repetition-1
10 -10
10 -10
Coded
10 -5
Sparse
Algorithm 1
0
10
20
30
Deadline Tdl (sec)
0.5
1
1.5
Deadline Tdl (sec)
2
0
10
20
30
Deadline Tdl (sec)
0
1
Deadline Tdl (sec)
2
Figure 2: From left to right: (1,2) Experimentally computed overall MSE of uncoded, replicationbased and coded personalized PageRank on the Twitter and Google Plus graph on a cluster with 120
workers. The ratio of MSE for repetition-based schemes and coded linear inverse increase as Tdl
increases. (3) Comparison between an extended version of the algorithm in [6] and Algorithm 1 on
the Google Plus graph. The figure shows that naively extending the general coded method using
matrix inverse introduces error amplification. (4) Comparison of different codes. In this experiment
the DFT-code out-performs the other candidates in MSE.
discussed in details later. The third figure in Fig. 2 shows the comparison between the performance
of Algorithm 1 and this extension of the algorithm from [6]. This extension uses the (unfinished)
partial results from the k fastest workers to retrieve the required PageRank solutions. More concretely,
suppose S ? [n] is the index set of the k fastest workers. Then, this extension retrieves the solutions to
the original k PageRank problems by solving the equation YS = [x?1 , x?2 , . . . , x?k ] ? GS , where YS is
composed of the (partial) computation results obtained from the fastest k workers and GS is the k ? k
submatrix composed of the columns in the generator matrix G with indexes in S. However, since
there is some remaining error at each worker (i.e., the computation results YS have not converged
yet), when conducting the matrix-inverse-based decoding from [6], the error is magnified due to the
large condition number of GS . This is why the algorithm in [6] should not be naively extended in the
coded linear inverse problem.
One question remains: what is the best code design for the coded linear inverse algorithm? Although
we do not have a concrete answer to this question, we have tested different codes (with different
generator matrices G) in the Twitter graph experiment, all using Algorithm 1. The results are shown
in the fourth figure in Fig. 2. The generator matrix used for the ?binary? curve has i.i.d. binary entries
in {?1, 1}. The generator matrix used for the ?sparse? curve has random binary sparse entries. The
generator matrix for the ?Gaussian? curve has i.i.d. standard Gaussian entries. In this experiment, the
DFT-code performs the best. However, finding the best code in general is a meaningful future work.
7
Conclusions
By studying coding for iterative algorithms designed for distributed inverse problems, we aim to
introduce new applications and analytical tools to the problem of coded computing with stragglers.
Since these iterative algorithms designed for inverse problems commonly have decreasing error
with time, the partial computation results at stragglers can provide useful information for the final
outputs. Note that this is unlike recent works on coding for multi-stage computing problems [29, 30],
where the computation error can accumulate with time and coding has to be applied repeatedly to
suppress this error accumulation. An important connection worth discussing is the diversity gain in
this coded computing problem. The distributed computing setting in this work resembles random
fading channels, which means coding can be used to exploit straggling diversity just as coding is
used in communication channels to turn diverse channel fading into an advantage. What makes
coding even more suitable in our setting is that the amount of diversity gain achieved here through
replication is actually smaller than that can be achieved by replication in fading channels. This
is because for two computers that solve the same equation Mxi = ri , the remaining error at the
slow worker is a deterministic multiple of the remaining error at the fast worker (see equation (3)).
Therefore, taking a weighted average of the two computation results through replication does not
reduce error as in independent fading channels. How diversity gain can be achieved here optimally is
worth deep investigation. Our next goals are two-fold: (1) extend the current method to solving a
single large-scale inverse problem, such as graph mining with graphs that exceed the memory of a
single machine; (2) carry out experiments on faster distributed systems such as Amazon EC2.
9
References
[1] J. Dean and L. A. Barroso. The tail at scale. Communications of the ACM, 56(2):74?80, 2013.
[2] G. Joshi, Y. Liu, and E. Soljanin. On the delay-storage trade-off in content download from
coded distributed storage systems. IEEE Journal on Selected Areas in Communications, 32(5):
989?997, 2014.
[3] D. Wang, G. Joshi, and G. Wornell. Efficient task replication for fast response times in parallel
computation. In ACM SIGMETRICS Performance Evaluation Review, volume 42, pages
599?600. ACM, 2014.
[4] D. Wang, G. Joshi, and G. Wornell. Using straggler replication to reduce latency in large-scale
parallel computing. ACM SIGMETRICS Performance Evaluation Review, 43(3):7?11, 2015.
[5] L. Huang, S. Pawar, H. Zhang, and K. Ramchandran. Codes can reduce queueing delay in data
centers. In IEEE International Symposium on Information Theory Proceedings (ISIT), pages
2766?2770. IEEE, 2012.
[6] K. Lee, M. Lam, R. Pedarsani, D. Papailiopoulos, and K. Ramchandran. Speeding up distributed
machine learning using codes. In IEEE International Symposium on Information Theory (ISIT),
pages 1143?1147. IEEE, 2016.
[7] R. Tandon, Q. Lei, A. G. Dimakis, and N. Karampatziakis. Gradient coding. 2016.
[8] S. Dutta, V. Cadambe, and P. Grover. Short-dot: Computing large linear transforms distributedly
using coded short dot products. In Advances In Neural Information Processing Systems, pages
2092?2100, 2016.
[9] N. S. Ferdinand and S. C. Draper. Anytime coding for distributed computation. In 54th Annual
Allerton Conference on Communication, Control, and Computing (Allerton), pages 954?960.
IEEE, 2016.
[10] S. Li, M. A. Maddah-Ali, and A. S. Avestimehr. A unified coding framework for distributed
computing with straggling servers. In IEEE Globecom Workshops (GC Wkshps), pages 1?6.
IEEE, 2016.
[11] A. Reisizadehmobarakeh, S. Prakash, R. Pedarsani, and S. Avestimehr. Coded computation
over heterogeneous clusters. In IEEE International Symposium on Information Theory (ISIT),
pages 2408?2412. IEEE, 2017.
[12] S. Li, M. A. Maddah-Ali, and A. S. Avestimehr. Coding for distributed fog computing. IEEE
Communications Magazine, 55(4):34?40, 2017.
[13] Q. Yu, M. A. Maddah-Ali, and A. S. Avestimehr. Polynomial codes: an optimal design for
high-dimensional coded matrix multiplication. In Advances In Neural Information Processing
Systems, 2017.
[14] K. Lee, C. Suh, and K. Ramchandran. High-dimensional coded matrix multiplication. In IEEE
International Symposium on Information Theory (ISIT), pages 2418?2422. IEEE, 2017.
[15] K. Lee, R. Pedarsani, D. Papailiopoulos, and K. Ramchandran. Coded computation for multicore
setups. In IEEE International Symposium on Information Theory (ISIT), pages 2413?2417.
IEEE, 2017.
[16] K.-H. Huang et al. Algorithm-based fault tolerance for matrix operations. IEEE transactions on
computers, 100(6):518?528, 1984.
[17] Y. Saad. Iterative methods for sparse linear systems. SIAM, 2003.
[18] T. H. Haveliwala. Topic-sensitive pagerank. In Proceedings of the 11th international conference
on World Wide Web, pages 517?526. ACM, 2002.
[19] L. Page, S. Brin, R. Motwani, and T. Winograd. The pagerank citation ranking: Bringing order
to the web. Technical report, Stanford InfoLab, 1999.
10
[20] M. Haikin and R. Zamir. Analog coding of a source with erasures. In IEEE International
Symposium on Information Theory, pages 2074?2078. IEEE, 2016.
[21] A. G. Dimakis, P. B. Godfrey, Y. Wu, M. J. Wainwright, and K. Ramchandran. Network coding
for distributed storage systems. IEEE Transactions on Information Theory, 56(9):4539?4551,
2010.
[22] M. Sathiamoorthy, M. Asteris, D. Papailiopoulos, A. G. Dimakis, R. Vadali, S. Chen, and
D. Borthakur. Xoring elephants: Novel erasure codes for big data. In Proceedings of the VLDB
Endowment, volume 6, pages 325?336. VLDB Endowment, 2013.
[23] M. A. Maddah-Ali and U. Niesen. Decentralized coded caching attains order-optimal memoryrate tradeoff. IEEE/ACM Transactions on Networking, 23(4):1029?1040, 2015.
[24] S. Li, M. A. Maddah-Ali, and A. S. Avestimehr. Coded mapreduce. In Communication, Control,
and Computing (Allerton), 2015 53rd Annual Allerton Conference on, pages 964?971. IEEE,
2015.
[25] D. I. Shuman, S. K. Narang, P. Frossard, A. Ortega, and P. Vandergheynst. The emerging field
of signal processing on graphs: Extending high-dimensional data analysis to networks and other
irregular domains. IEEE Signal Processing Magazine, 30(3):83?98, 2013.
[26] A. Sandryhaila and J. M. F. Moura. Discrete signal processing on graphs. IEEE transactions on
signal processing, 61(7):1644?1656, 2013.
[27] G. H. Golub and C. F. van Loan. Matrix computations, volume 3. JHU Press, 2012.
[28] J. Leskovec and J. J. Mcauley. Learning to discover social circles in ego networks. In Advances
in neural information processing systems, pages 539?547, 2012.
[29] Y. Yang, P. Grover, and S. Kar. Computing linear transformations with unreliable components.
IEEE Transactions on Information Theory, 2017.
[30] Y. Yang, P. Grover, and S. Kar. Rate distortion for lossy in-network linear function computation
and consensus: Distortion accumulation and sequential reverse water-filling. IEEE Transactions
on Information Theory, 2017.
[31] X. Wang, P. Liu, and Y. Gu. Local-set-based graph signal reconstruction. IEEE Transactions on
Signal Processing, 63(9):2432?2444, 2015.
[32] S. K. Narang, A. Gadde, E. Sanou, and A. Ortega. Localized iterative methods for interpolation
in graph structured data. In 2013 IEEE Global Conference on Signal and Information Processing
(GlobalSIP), pages 491?494. IEEE, 2013.
[33] S. Chen, R. Varma, A. Sandryhaila, and J. Kova?cevi?c. Discrete signal processing on graphs:
Sampling theory. IEEE Transactions on Signal Processing, 63(24):6510?6523, 2015.
[34] S. Chen, Y. Yang, C. Faloutsos, and J. Kovacevic. Monitoring manhattan?s traffic at 5 intersections? In IEEE 2016 GlobalSIP Conference on Signal and Information Processing (GlobalSIP),
2016.
[35] A. M. Mood, F. A. Graybill, and D. C. Boes. Introduction to the theory of statistics, 3rd edition.
1974.
[36] H. Zhang and F. Ding. On the kronecker products and their applications. Journal of Applied
Mathematics, 2013, 2013.
11
| 6673 |@word version:3 inversion:1 polynomial:1 norm:2 replicate:4 nd:1 vi1:1 vldb:2 covariance:4 pick:2 carry:1 mcauley:1 ld:5 liu:2 initial:7 outperforms:1 existing:4 current:1 comparing:1 si:2 assigning:1 yet:1 written:2 fn:2 additive:1 j1:3 designed:2 selected:1 xk:1 short:2 iterates:1 provides:2 node:4 preference:1 allerton:4 zhang:2 become:1 symposium:6 ik:2 replication:31 tdl:27 prove:2 combine:3 introduce:1 theoretically:3 expected:5 indeed:1 frossard:1 behavior:1 multi:1 inspired:3 decreasing:1 solver:6 discover:1 unrelated:1 notation:1 moreover:1 what:5 aln:1 minimizes:1 emerging:3 dimakis:3 unified:1 finding:1 magnified:1 transformation:1 prakash:1 k2:4 control:2 yn:3 before:5 ecc:3 declare:1 negligible:1 treat:4 local:1 encoding:7 analyzing:2 interpolation:1 plus:8 resembles:2 collect:1 ease:1 fastest:5 directed:1 practice:1 erasure:7 area:1 asteris:1 submatrices:1 jhu:1 convenient:1 word:1 get:1 cannot:1 storage:3 accumulation:2 equivalent:1 deterministic:1 dean:1 center:1 send:1 distributedly:1 amazon:1 recovery:1 correcting:3 machinelearning:1 utilizing:1 dominate:1 orthonormal:5 varma:1 retrieve:2 papailiopoulos:3 suppose:11 tandon:1 user:3 decode:1 magazine:2 us:4 ego:1 utilized:1 winograd:1 observed:1 ding:1 solved:1 wang:3 wornell:2 zamir:1 ensures:2 decrease:1 trade:1 mentioned:2 intuition:3 complexity:7 hinder:1 straggler:21 solving:10 depend:1 ali:7 deliver:1 gu:1 retrieves:1 fast:6 query:1 supplementary:14 solve:6 larger:3 snap:1 stanford:1 elephant:1 narang:2 encoder:2 toeplitz:1 statistic:3 distortion:2 transform:3 superscript:1 final:1 mood:1 advantage:3 eigenvalue:2 sequence:2 analytical:1 propose:2 subtracting:1 lam:1 product:2 reconstruction:1 j2:20 combining:1 achieve:1 straggling:4 intuitive:1 frobenius:1 amplification:1 amplifies:1 ferdinand:1 convergence:3 cluster:4 motwani:1 diverges:1 r1:8 produce:1 extending:2 converges:2 help:1 illustrate:1 andrew:1 kef:1 multicore:1 solves:2 come:1 radius:1 brin:1 adjacency:1 require:1 preliminary:2 investigation:1 isit:5 extension:7 strictly:2 hold:8 proximity:1 teleport:1 al1:1 smallest:1 sensitive:1 repetition:5 tool:1 weighted:6 minimization:1 clearly:1 always:1 gaussian:3 aim:2 sigmetrics:2 rather:1 caching:1 nkn:1 encode:5 ax:3 focus:2 validated:1 karampatziakis:1 slowest:5 contrast:2 rigorous:1 attains:1 sense:5 twitter:7 i1:1 provably:2 issue:2 overall:5 ill:3 among:5 exponent:6 multiplies:2 godfrey:1 initialize:1 field:2 equal:2 beach:1 sampling:3 represents:3 look:1 yu:1 filling:1 future:1 report:1 e2f:1 simplify:1 few:1 composed:4 mining:2 evaluation:2 golub:1 introduces:1 fog:1 behind:2 accurate:1 beforehand:1 edge:2 condor:1 partial:6 worker:88 pulkit:1 conduct:2 circle:2 e0:5 theoretical:1 minimal:2 leskovec:1 instance:1 column:2 earlier:1 soft:2 ordinary:2 cost:10 subset:1 entry:6 delay:2 haveliwala:1 motivating:1 optimally:1 answer:1 gd:1 st:1 international:7 ec2:1 siam:1 lee:5 off:1 decoding:20 concrete:1 w1:1 squared:11 central:4 satisfied:1 choose:1 huang:2 dr:2 ek:4 return:1 bx:1 li:23 toy:1 distribute:1 diversity:5 unordered:1 coding:17 sec:4 satisfy:4 explicitly:1 ranking:1 vi:3 performed:2 try:1 later:2 closed:1 analyze:1 doing:1 traffic:1 start:1 parallel:11 vin:1 contribution:1 square:11 dutta:1 accuracy:3 formed:1 kek:3 variance:5 conducting:1 barroso:1 infolab:1 shuman:1 multiplying:2 worth:2 globalsip:3 gershgorin:1 monitoring:1 processor:1 converged:2 randomness:3 moura:1 networking:1 definition:1 failure:1 e2:4 gdl:3 proof:12 gain:3 proved:1 recall:1 lim:6 anytime:1 sophisticated:1 actually:1 appears:1 higher:1 varf:3 response:1 formulation:2 execute:1 just:1 stage:2 until:1 web:2 ei:8 google:8 lei:1 lossy:1 usa:1 effect:2 normalized:1 true:5 y2:2 regularization:1 i2:1 deal:2 during:1 ortega:2 complete:2 performs:4 l1:15 postprocessing:1 novel:4 recently:1 ef:22 al2:1 overview:4 bxi:1 conditioning:1 volume:3 discussed:6 extend:2 tail:5 analog:1 accumulate:1 mellon:1 kri:2 dft:4 rd:2 mathematics:1 replicating:1 dot:2 j4:8 recent:3 perspective:1 optimizing:2 reverse:1 manipulation:2 certain:3 server:1 kar:3 inequality:1 arbitrarily:1 binary:4 fault:2 discussing:1 yi:6 relaxed:1 parallelized:1 recognized:1 converge:2 signal:11 ii:2 multiple:3 technical:2 faster:1 adapt:1 long:2 deadline:18 e1:5 post:2 coded:65 plugging:2 bigger:1 y:3 heterogeneous:2 essentially:2 cmu:1 controller:4 expectation:3 iteration:21 sometimes:1 achieved:3 avestimehr:5 irregular:1 whereas:1 want:1 separately:1 addressed:1 source:1 boes:1 crucial:1 extra:9 w2:1 unlike:3 saad:1 bringing:1 incorporates:1 call:1 integer:1 joshi:3 maddah:5 yang:4 intermediate:1 exceed:1 identically:1 affect:1 suboptimal:2 reduce:6 idea:3 vik:1 tradeoff:1 intensive:1 sandryhaila:2 algebraic:2 e3:1 bli:2 repeatedly:1 deep:1 useful:2 latency:1 detailed:1 amount:1 transforms:1 bottlenecked:1 reduced:1 exist:1 notice:1 estimated:2 per:4 diverse:1 carnegie:1 discrete:2 affected:2 redundancy:1 key:1 queueing:1 ce:15 ht:1 utilize:3 draper:1 v1:2 cevi:1 graph:24 inverse:69 fourth:1 respond:1 throughout:1 reasonable:1 wu:1 vn:2 utilizes:1 scaling:4 submatrix:7 bit:1 bound:6 cadambe:1 guaranteed:1 fold:1 g:3 annual:2 fading:4 constraint:2 infinity:2 kronecker:1 x2:4 ri:12 personalized:9 dominated:3 fourier:6 speed:4 graceful:1 speedup:2 structured:2 according:1 combination:2 across:2 smaller:8 slightly:1 globecom:1 s1:1 gradually:1 taken:2 computationally:1 equation:8 ln:3 remains:1 turn:1 precomputed:1 fail:1 loose:1 end:1 studying:1 available:2 operation:3 decentralized:1 multiplied:1 observe:1 v2:2 appropriate:4 spectral:1 alternative:1 faloutsos:1 lxi:1 slower:1 original:3 denotes:1 remaining:5 include:2 completed:2 exploit:1 giving:1 build:1 classical:3 question:2 strategy:6 degrades:2 diagonal:3 said:1 gradient:4 decoder:2 gracefully:2 topic:3 considers:1 consensus:1 water:1 assuming:2 code:21 index:2 ratio:5 equivalently:1 setup:1 kova:1 gk:5 trace:45 suppress:1 design:4 implementation:1 upper:1 datasets:1 descent:3 beat:3 extended:2 communication:14 y1:2 rn:1 gc:1 thm:1 download:1 inverting:3 namely:2 required:2 specified:2 connection:1 inaccuracy:1 nip:1 trans:1 able:1 summarize:1 pagerank:23 including:3 max:11 vi2:1 memory:1 wainwright:1 power:4 event:2 suitable:1 natural:4 difficulty:2 regularized:1 residual:2 scheme:13 improve:1 mxi:3 finished:2 uncoded:20 lk:2 speeding:2 sn:1 prior:1 understanding:1 l2:5 review:2 mapreduce:1 multiplication:4 asymptotic:1 manhattan:1 fully:1 expect:1 permutation:1 grover:4 var:3 versus:1 vandergheynst:1 generator:11 validation:1 localized:1 pedarsani:3 unfinished:1 bypass:1 heavy:4 endowment:2 row:8 pulled:1 wide:1 taking:1 sparse:4 distributed:18 tolerance:2 curve:3 van:1 world:1 ignores:1 concretely:1 collection:1 commonly:1 preprocessing:2 far:1 social:3 transaction:8 citation:1 ignore:5 unreliable:1 global:1 xi:12 suh:1 iterative:20 quantifies:1 why:1 additionally:1 channel:5 ca:1 mse:15 necessarily:1 domain:1 diag:2 pk:1 main:6 dense:1 s2:1 noise:1 big:1 edition:1 fair:2 x1:5 augmented:1 fig:5 slow:7 candidate:1 jacobian:4 bei:1 third:1 theorem:18 rk:2 specific:1 maxi:2 r2:8 dominates:1 dl:1 naively:2 workshop:1 adding:1 sequential:1 importance:1 magnitude:1 ramchandran:5 conditioned:3 chen:3 intersection:1 simply:1 gadde:1 w1w:2 scalar:4 satisfies:9 acm:6 conditional:1 goal:3 content:1 hard:1 change:1 experimentally:1 loan:1 degradation:1 called:1 meaningful:1 tested:1 correlated:2 |
6,272 | 6,674 | A Screening Rule for `1-Regularized
Ising Model Estimation
Zhaobin Kuang1 , Sinong Geng2 , David Page3
University of Wisconsin
[email protected] , [email protected] , [email protected]
Abstract
We discover a screening rule for `1 -regularized Ising model estimation. The simple
closed-form screening rule is a necessary and sufficient condition for exactly
recovering the blockwise structure of a solution under any given regularization
parameters. With enough sparsity, the screening rule can be combined with various
optimization procedures to deliver solutions efficiently in practice. The screening
rule is especially suitable for large-scale exploratory data analysis, where the
number of variables in the dataset can be thousands while we are only interested
in the relationship among a handful of variables within moderate-size clusters for
interpretability. Experimental results on various datasets demonstrate the efficiency
and insights gained from the introduction of the screening rule.
1
Introduction
While the field of statistical learning with sparsity [Hastie et al., 2015] has been steadily rising to
prominence ever since the introduction of the lasso (least absolute shrinkage and selection operator)
at the end of the last century [Tibshirani, 1996], it was not until the recent decade that various
screening rules debuted to further equip the ever-evolving optimization arsenals for some of the
most fundamental problems in sparse learning such as `1 -regularized generalized linear models
(GLMs, Friedman et al. 2010) and inverse covariance matrix estimation [Friedman et al., 2008].
Screening rules, usually in the form of an analytic formula or an optimization procedure that is
extremely fast to solve, can accelerate learning drastically by leveraging the inherent sparsity of many
high-dimensional problems. Generally speaking, screening rules can identify a significant portion of
the zero components of an optimal solution beforehand at the cost of minimal computational overhead,
and hence substantially reduce the dimension of the parameterization, which makes possible efficient
computation for large-scale sparse learning problems.
Pioneered by Ghaoui et al. 2010, various screening rules have emerged to speed up learning for
generative models (e.g. Gaussian graphical models) as well as for discriminative models (e.g. GLMs),
and for continuous variables (e.g. lasso) as well as for discrete variables (e.g. logistic regression,
support vector machines). Table 1 summarizes some of the iconic work in the literature, where, to the
best of our knowledge, screening rules for generative models with discrete variables are still notably
absent.
Contrasted with this notable absence is the ever stronger craving in the big data era for scaling
up the learning of generative models with discrete variables, especially in a blockwise structure
identification setting. For example, in gene mutation analysis [Wan et al., 2015, 2016], among tens of
thousands of sparse binary variables representing mutations of genes, we are interested in identifying
a handful of mutated genes that are connected into various blocks and exert synergistic effects on the
phenotype. While a sparse Ising model is a desirable choice, for such an application the scalability of
the model fails due to the innate N P-hardness [Karger and Srebro, 2001] of inference, and hence
learning, owing to the partition function. To date, even with modern approximation techniques, a
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
Table 1: Screening rules in the literature at a glance
Discriminative Models
Generative Models
Continuous
Variables
Ghaoui et al. 2010, Tibshirani et al. 2012
Wang et al. 2013, Liu et al. 2013
Fercoq et al. 2015, Xiang et al. 2016
Banerjee et al. 2008 Honorio and Samaras 2010
Witten et al. 2011,Mazumder and Hastie 2012
Danaher et al. 2014, Luo et al. 2014
Yang et al. 2015
Discrete
Variables
Ghaoui et al. 2010, Tibshirani et al. 2012
Wang et al. 2014
?
typical application with sparse discrete graphical models usually involves only hundreds of variables
[Viallon et al., 2014, Barber et al., 2015, Vuffray et al., 2016].
Between the need for the scalability of high-dimensional Ising models and the absence of screening
rules that are deemed crucial to accelerated and scalable learning, we have a technical gap to bridge:
can we identify screening rules that can speed up the learning of `1 -regularized Ising models? The
major contribution of this paper is to give an affirmative answer to this question. Specifically, we
show the following.
? The screening rule is a simple closed-form formula that is a necessary and sufficient condition for
exact blockwise structure recovery of the solution with a given regularization parameter. Upon the
identification of blockwise structures, different blocks of variables can be considered as different
Ising models and can be solved separately. The various blocks can even be solved in parallel to
attain further efficiency. Empirical results on both simulated and real-world datasets demonstrate
the tremendous efficiency, scalability, and insights gained from the introduction of the screening
rule. Efficient learning of `1 -regularized Ising models from thousands of variables on a single
machine is hence readily attainable.
? As an initial attempt to fill in the vacancy illustrated in Table 1, our work is instructive to further
exploration of screening rules for other graphical models with discrete random variables, and
to combining screening rules with various optimization methods to facilitate better learning.
Furthermore, compared with its Gaussian counterpart, where screening rules are available (Table 1)
and learning is scalable [Hsieh et al., 2013], the proposed screening rule is especially valuable and
desperately needed to address the more challenging learning problem of sparse Ising models.
We defer all the proofs in the paper to the supplement and focus on providing intuition and interpretation of the technical results in the paper.
2
2.1
Notation and Background
Ising Models
>
Let X = [X1 , X2 , ? ? ? , Xp ] be a p ? 1 binary random vector, with Xi ? {?1, 1}, and i ?
{1, 2, ? ? ? , p} , V . Let
there be a dataset X with n independent and identically distributed samples
of X, denoted as X = x(1) , x(2) , ? ? ? , x(n) . Here, x(k) is a p?1 vector of assignments that realizes
(k)
X, where k ? {1, 2, ? ? ? , n}. We further use xi to denote the ith component of the k th sample in
the dataset. Let ? ? ? be a p ? p symmetric matrix whose diagonal entries are zeros. An Ising model
[Wan et al., 2016] with the parameterization ? is:
?
?
p?1 X
p
X
1
P? (x) =
exp ?
?ij xi xj ? ,
(1)
Z(?)
i=1 j>i
where ?ij represents the component of ? at the ith row and the j th column, and xi and xj represent
the ith and the j th components of x, respectively. Z(?) is a normalization constant, partition
function, that ensures
probabilitiessum up to one. The partition function is given as Z(?) =
P theP
P
p?1
p
x?{?1,1}p exp
i=1
j>i ?ij xi xj . Note that for ease of presentation, we consider Ising
models with only pairwise interaction/potential here. Generalization to Ising models with unary
potentials is given in Section 6.
2
2.2
Graphical Interpretation
With the notion of the probability given by an Ising model in (1), estimating an `1 -regularized Ising
? the penalized maximum likelihood estimator (MLE) under the lasso
model is defined as finding ?,
penalty:
n
?
1X
?? = arg max
log P? x(k) ? k?k1
? n
2
k=1
(2)
n p?1 p
?
1 XXX
(k) (k)
?ij xi xj + A(?) + k?k1 .
= arg min ?
?
n
2
i=1 j>i
k=1
Pp Pp
Here, A(?) = log Z(?) is the log-partition function; k?k1 = i=1 j=1 |?ij | is the lasso penalty
that encourages a sparse parameterization. ? ? 0 is a given regularization parameter. Using ?2 is
Pp?1 Pp
suggestive of the symmetry of ? so that ?2 k?k1 = ? i=1 j>i |?ij |, which echoes the summations
in the negative log-likelihood function. Note that ? corresponds to the adjacency matrix constructed
by the p components of X as nodes, and ?ij 6= 0 indicates that there is an edge between Xi and
Xj . We further denote a partition of V into L blocks as {C1 , C2 , ? ? ? , CL }, where Cl , Cl0 ? V ,
SL
Cl ? Cl0 = ?, l=1 Cl = V , l 6= l0 , and for all l, l0 ? {1, 2, ? ? ? , L}. Without loss of generality, we
assume that the nodes in different blocks are ordered such that if i ? Cl , j ? Cl0 , and l < l0 , then
i < j.
2.3
Blockwise Solutions
We introduce the definition of a blockwise parameterization:
Definition 1. We call ? blockwise with respect to the partition {C1 , C2 , ? ? ? , CL } if ?l and l0 ?
{1, 2, ? ? ? , L}, where l 6= l0 , and ?i ? Cl , ?j ? Cl0 , we have ?ij = 0.
When ? is blockwise, we can represent ? in a block diagonal fashion:
? = diag (?1 , ?2 , ? ? ? , ?L ) ,
(3)
where ?1 , ?2 , ? ? ? , and ?L are symmetric matrices that correspond to C1 , C2 , ? ? ? , and CL , respectively.
Note that if we can identify the blockwise structure of ?? in advance, we can solve each block
independently (See A.1). Since the size of each block could be much smaller than the size of the
original problem, each block could be much easier to learn compared with the original problem.
Therefore, efficient identification of blockwise structure could lead to substantial speedup in learning.
3
3.1
The Screening Rule
Main Results
The preparation in Section 2 leads to the discovery of the following strikingly simple screening rule
presented in Theorem 1.
Theorem 1. Let a partition of V, {C1 , C2 , ? ? ? , CL }, be given. Let the dataset X =
(1) (2)
Pn
(k) (k)
x , x , ? ? ? , x(n) be given. Define EX Xi Xj = n1 k=1 xi xj . A necessary and sufficient
condition for ?? to be blockwise with respect to the given partition is that
|EX Xi Xj | ? ?,
(4)
for all l and l0 ? {1, 2, ? ? ? , L}, where l 6= l0 , and for all i ? Cl , j ? Cl0 .
In terms of exact blockwise structure identification, Theorem 1 provides a foolproof (necessary and
sufficient) and yet easily checkable result by comparing the absolute second empirical moments
|EX Xi Xj |?s with the regularization parameter ?. We also notice the remarkable similarity between
the proposed screening rule and the screening rule for Gaussian graphical model blockwise structure
identification in Witten et al. 2011, Mazumder and Hastie 2012. In the Gaussian case, the screening
rule can be attained by simply replacing the second empirical moment matrix in (4) with the sample
3
Algorithm 1 Blockwise Minimization
Input: dataset X, regularization parameter ?.
?
Output: ?.
?i, j ? V such that j > i, compute the second empirical moments EX Xi Xj ?s .
Identify the partition {C1 , C2 , ? ? ? , CL } using the second empirical moments from the previous
step and according to Witten et al. [2011], Mazumder and Hastie [2012].
5: ?l ? L, perform blockwise optimization over Cl for ??l .
?
6: Ensemble ??l ?s according to (3) for ?.
?
7: Return ?.
1:
2:
3:
4:
covariance matrix. While the exact solution in the Gaussian case can be computed in polynomial time,
estimating an Ising model exactly in general is N P-hard. However, as a consequence of applying the
screening rule, the blockwise structure of an `1 -regularized Ising model can be determined as easily
as the blockwise structure of a Gaussian graphical model, despite the fact that within each block,
exact learning of a sparse Ising model could still be challenging.
Furthermore, the screening rule also provides us a principal approach to leverage sparsity for the gain
of efficiency: by increasing ?, the nodes of the Ising model will be shattered into smaller and smaller
blocks, according to the screening rule. Solving many Ising models with small blocks of variables is
amenable to both estimation algorithm and parallelism.
3.2
Regularization Parameters
The screening rule also leads to a significant implication to the range of regularization parameters in
which ?? 6= 0. Specifically, we have the following theorem.
Theorem 2. Let the dataset X = x(1) , x(2) , ? ? ? , x(n) be given, and let ? = ?max represent the
smallest regularization parameter such that ?? = 0 in (2). Then ?max = maxi,j?V,i6=j |EX Xi Xj | ? 1.
With ?max , one can decide the range of regularization parameters, [0, ?max ], that generates graphs
with nonempty edge sets, which is an important first step for pathwise optimization algorithms
(a.k.a. homotopy algorithms) that learn the solutions to the problem under a range of ??s. Furthermore,
the fact that ?max ? 1 for any given dataset X suggests that comparison across different networks
generated by different datasets is comprehensible. Finally, in Section 4, ?max will also help to
establish the connection between the screening rule for exact learning and some of the popular inexact
(alternative) learning algorithms in the literature.
3.3
Fully Disconnected Nodes
Another consequence of the screening rule is the necessary and sufficient condition that determines
the regularization parameter with which a node is fully disconnected from the remaining nodes:
Corollary 1. Let the dataset X = x(1) , x(2) , ? ? ? , x(n) be given. Xi is fully disconnected from
? where i ? V (i.e., ??ij = ??ji = 0, ?j ? V \ {i}), if and only if
the remaining nodes in ?,
? ? maxj?V \{i} |EX Xi Xj |.
In high-dimensional exploratory data analysis, it is usually the case that most of the variables are
fully disconnected [Danaher et al., 2014, Wan et al., 2016]. In this scenario, Corollary 1 provides a
regularization parameter threshold with which we can identify exactly the subset of fully disconnected
nodes. Since we can choose a threshold large enough to make any nodes fully disconnected, we can
discard a significant portion of the variables efficiently and flexibly at will with exact optimization
guarantees due to Corollary 1. By discarding the large portion of fully disconnected variables, the
learning algorithm can focus on only a moderate number of connected variables, which potentially
results in a substantial efficiency gain.
3.4
Blockwise Minimization
We conclude this section by providing the blockwise minimization algorithm in Algorithm 1 due
to the screening rule. Note that both the second empirical moments and the partition of V in the
4
algorithm can be computed in O(p2 ) operations [Witten et al., 2011, Mazumder and Hastie, 2012].
On the contrary, the complexity of the exact optimization of a block of variables grows exponentially
with respect to the maximal clique size of that block. Therefore, by encouraging enough sparsity,
the blockwise minimization due to the screening rule can provide remarkable speedup by not only
shrinking the size of the blocks in general but also potentially reducing the size of cliques within each
block via eliminating enough edges.
4
Applications to Inexact (Alternative) Methods
We now discuss the interplay between the screening rule and two popular inexact (alternative)
estimation methods: node-wise (NW) logistic regression [Wainwright et al., 2006, Ravikumar et al.,
2010] and the pseudolikelihood (PL) method [H?fling and Tibshirani, 2009]. In what follows, we
use ??NW and ??PL to denote the solutions given by the node-wise logistic regression method and the
pseudolikelihood method, respectively. NW can be considered as an asymmetric pseudolikelihood
NW
NW
method (i.e., ?i,j ? V such that i 6= j and ??ij
), while PL is a pseudolikelihood method that
6= ??ji
is similar to NW but imposes additional symmetric constraints on the parameterization (i.e., ?i,j ? V
PL
PL
where i 6= j, we have ??ij
= ??ji
).
Our incorporation of the screening rule to the inexact methods is straightforward: after using the
screening rule to identify different blocks in the solution, we use inexact methods to solve each block
for the solution. As shown in Section 3, when combined with exact optimization, the screening
rule is foolproof for blockwise structure identification. However, in general, when combined with
inexact methods, the proposed screening rule is not foolproof any more because the screening rule is
derived from the exact problem in (2) instead of the approximate problems such as NW and PL. We
provide a toy example in A.6 to illustrate mistakes made by the screening rule when combined with
inexact methods. Nonetheless, as we will show in this section, NW and PL are deeply connected to
the screening rule, and when given a large enough regularization parameter, the application of the
screening rule to NW and PL can be lossless in practice (see Section 5). Therefore, when applied to
NW and PL, the proposed screening rule can be considered as a strong rule (i.e., a rule that is not
foolproof but barely makes mistakes) and an optimal solution can be safeguarded by adjusting the
screened solution to optimality based on the KKT conditions of the inexact problem [Tibshirani et al.,
2012].
4.1
Node-wise (NW) Logistic Regression and the Pseudolikelihood (PL) Method
In NW, for each i ? V , we consider the conditional probability of Xi upon X\i , where X\i =
{Xt | t ? V \ {i}}. This is equivalent to solving p `1 -regularized logistic regression problems
separately, i.e., ?i ? V :
n
i
1 X h (k) (k)
(k)
NW
??\i
= arg min
?yi ?\i + log 1 + exp ?\i
+ ?
?\i
1 ,
(5)
?\i n
k=1
(k)
?\i
(k)
(k)
>
where
= ?\i
(2x\i ), yi = 1
(k)
unsuccessful event xi = ?1, and
(k)
represents a successful event xi
(k)
= 1, yi
= 0 represents an
>
?\i = ?i1 ?i2 ? ? ? ?i(i?1) ?i(i+1) ? ? ? ?ip ,
h
i>
(k)
(k)
(k)
(k)
(k)
(k)
x\i = xi1 xi2 ? ? ? xi(i?1) xi(i+1) ? ? ? xip
.
NW
Note that ??NW constructed from ??\i
?s is asymmetric, and ad hoc post processing techniques are used
to generate a symmetric estimation such as setting each pair of elements from ??NW in symmetric
positions to the one with a larger (or smaller) absolute value.
On the other hand, PL can be considered as solving all p `1 -regularized logistic regression problems
in (5) jointly with symmetric constraints over the parameterization [Geng et al., 2017]:
p
n
i ?
1 X X h (k) (k)
(k)
??PL = arg min
?yi ?i + log 1 + exp ?i
+ k?k1 ,
(6)
??? n
2
i=1
k=1
5
P
(k)
(k)
where ?i = j?V \{i} 2?min{i,j},max{i,j} xj .That is to say, if i < j, then ?min{i,j},max{i,j} = ?ij ;
if i > j, then ?min{i,j},max{i,j} = ?ji . Recall that ? in (6) defined in Section 2.1 represents a space
of symmetric matrices whose diagonal entries are zeros.
4.2
Regularization Parameters in NW and PL
Since the blockwise structure of a solution is given by the screening rule under a fixed regularization
parameter, the ranges of regularization parameters under which NW and PL can return nonzero
solutions need to be linked to the range [0, ?max ] in the exact problem. Theorem 3 and Theorem 4
establish such relationships for NW and PL, respectively.
Theorem 3. Let the dataset X = x(1) , x(2) , ? ? ? , x(n) be given, and let ? = ?NW
max represent the
NW
NW
?
smallest regularization parameter such that ?\i = 0 in (5), ?i ? V . Then ?max = ?max .
Theorem 4. Let the dataset X = x(1) , x(2) , ? ? ? , x(n) be given, and let ? = ?PL
max represent the
PL
PL
?
smallest regularization parameter such that ? = 0 in (6), then ?max = 2?max .
Let ? be the regularization parameter used in the exact problem. A strategy is to set the corresponding
?NW = ? when using NW and ?PL = 2? when using PL, based on the range of regularization parameters given in Theorem 3 and Theorem 4 for NW and PL. Since the magnitude of the regularization
parameter is suggestive of the magnitude of the gradient of the unregulated objective, the proposed
strategy leverages that the magnitudes of the gradients of the unregulated objectives for NW and
PL are roughly the same as, and roughly twice as large as, that of the unregulated exact objective,
respectively.
This observation has been made in the literature of binary pairwise Markov networks [H?fling and
Tibshirani, 2009, Viallon et al., 2014]. Here, by Theorem 3 and Theorem 4, we demonstrate that
this relationship is exactly true if the optimal parameterization is zero. H?fling and Tibshirani 2009
even further exploits this observation in PL for exact optimization. Their procedure can be viewed as
iteratively solving adjusted PL problems regularized by ?PL = 2? in order to obtain an exact solution
regularized by ?. The close quantitative correspondence between the derivatives of the inexact
objectives and that of the exact objective also provides insights into why combing the screening rule
with inexact methods does not lose much in practice.
4.3
Preservation for Fully Disconnectedness
While the screening rule is not foolproof when combined with NW and PL, it turns out that in terms
of identifying fully disconnected nodes, the necessary and sufficient condition in Corollary 1 can be
preserved when applying NW with caution, as shown in the following.
NW
Theorem 5. Let the dataset X = x(1) , x(2) , ? ? ? , x(n) be given. Let ??min
? ? denote a symmetric
NW
NW
?
?
matrix derived from ?
by setting each pair of elements from ?
in symmetric positions to the
one with a smaller absolute value. A sufficient condition for Xi to be fully disconnected from the
NW
remaining nodes in ??min
, where i ? V , is that ?NW ? maxj?V \{i} |EX Xi Xj |. Furthermore, when
NW
??\i
= 0, the sufficient condition is also necessary.
In practice, the utility of Theorem 5 is to provide us a lower bound for ? above which we can fully
NW
disconnect Xi (sufficiency). Moreover, if ??\i
= 0 also happens to be true, which is easily verifiable,
we can conclude that such a lower bound is tight (necessity).
5
Experiments
Experiments are conducted on both synthetic data and real world data. We will focus on efficiency in
Section 5.1 and discuss support recovery performance in Section 5.2. We consider three synthetic
networks (Table 2) with 20, 35, and 50 blocks of 20-node, 35-node, and 50-node subnetworks,
respectively. To demonstrate the estimation of networks with unbalanced-size subnetworks, we also
consider a 46-block network with power law degree distributed subnetworks of sizes ranging from 5
to 50. Within each network, the subnetwork is generated according to a power law degree distribution,
which mimics the structure of a biological network and is believed to be more challenging to recover
6
1000
50
1500
Methods
PL
NW
PL+screen
NW+screen
750
NW
PL+screen
NW+screen
1000
500
250
Runtime (s)
Runtime (s)
Runtime (s)
100
1250
Methods
PL
NW
PL+screen
NW+screen
Runtime (s)
Methods
Methods
PL
NW
PL+screen
NW+screen
1000
750
500
500
250
0
0
400
800
1200
1600
400
Sample Size
(a) Network 1
800
1200
1600
400
Sample Size
800
1200
Sample Size
(b) Network 2
(c) Network 3
1600
400
800
1200
1600
Sample Size
(d) Network 4
Figure 1: Runtime of pathwise optimization on networks in Table 2. Runtime plotted is the median
runtime over five trials. The experiments of the baseline method PL without screening can not be
fully conducted on larger networks due to high memory cost. NW: Node-wise logistic regression
without screening; NW+screen: Node-wise logistic regression with screening; PL: Pseudolikelihood
without screening; PL+screen: Pseudolikelihood with screening.
compared with other less complicated structures [Chen and Sharp, 2004, Peng et al., 2009, Danaher
et al., 2014]. Each edge of each network is associated with a weight first sampled from a standard
normal distribution, and then increased or decreased by 0.2 to further deviate from zero. For each
network, 1600 samples are generated via Gibbs sampling within each subnetwork. Experiments on
exact optimization are reported in B.2.
5.1
Pathwise Optimization
Pathwise optimization aims to compute solutions over a range of different ??s. Formally, we denote
the set of ??s used in (2) as ? = {?1 , ?2 , ? ? ? , ?? }, and without loss of generality, we assume that
?1 < ?2 < ? ? ? < ?? .
The introduction of the screening rule provides us insightful heuristics for the determination of ?.
We start by choosing a ?1 that reflects the sparse blockwise structural assumption on the data. To
achieve sparsity and avoid densely connected structures, we assume that the number of edges in the
ground truth network is O(p). This assumption coincides with networks generated according to a
power law degree distribution and hence is a faithful representation of the prior knowledge stemming
from many
biological problems. As a heuristic, we relax and apply the screening rule in (4) on each
of the p2 second empirical moments and choose ?1 such that the number of the absolute second
empirical moments that are greater than ?1 is about p log p. Given a ?1 chosen this way, one can
? 1 ) has by the screening rule. To encourage blockwise structures, we
check how many blocks ?(?
? 1 ) has more than one block. We then choose ??
magnify ?1 via ?1 ? 1.05?1 until the current ?(?
such that the number of absolute second empirical moments that are greater than ?? is about p. In
our experiments, we use an evenly spaced ? with ? = 25.
To estimate the networks in Table 2, we implement both NW and PL with and without screening
using glmnet [Friedman et al., 2010] in R as a building block for logistic regression according to
Ravikumar et al. 2010 and Geng et al. 2017. To generate a symmetric parameterization for NW, we
set each pair of elements from ?NW in symmetric positions to the element with a larger absolute value.
Given ?, we screen only at ?1 to identify various blocks. Each block is then solved separately in a
pathwise fashion under ? without further screening. The rationale of performing only one screening
is that starting from a ?1 chosen in the aforementioned way has provided us a sparse blockwise
structure that sets a significant portion of the parameterization to zeros; further screening over larger
??s hence does not necessarily offer more efficiency gain.
Figure 1 summarizes the runtime of pathwise optimization on the four synthetic networks in Table 2.
The experiments are conducted on a PowerEdge R720 server with two Intel(R) Xeon(R) E5-2620
CPUs and 128GB RAM. As many as 24 threads can be run in parallel. For robustness, each runtime
reported is the median runtime over five trials. When the sample size is less than 1600, each trial
uses a subset of samples (subsamples) that are randomly drawn from the original datasets without
replacement. As illustrated in Figure 1, the efficiency gain due to the screening rule is self-evident.
Both NW and PL benefit substantially from the application of the screening rule. The speedup is
more apparent with the increase of sample size as well as the increase of the dimension of the data. In
our experiments, we observe that even with arguably the state-of-the-art implementation [Geng et al.,
7
#blk
20
35
50
46
#nd/blk
20
35
50
5-50
TL#nd
400
1225
2500
1265
Table 2: Summary of the four synthetic networks used in the experiments. indx represents the index
of each network. #blk represents
the number of blocks each network has. #nd/blk represents the
number of nodes each block has.
TL#nd represents the total number
of nodes each network has.
1.00
Methods
900
0.75
Runtime (s)
1
2
3
4
AUC
indx
PL
NW
PL+screen
NW+screen
Mix
600
0.50
Methods
PL
NW
PL+screen
NW+screen
Mix
0.25
0.00
300
0
1
2
3
4
Network Index
(a) Edge recovery AUC
1
2
3
4
Network Index
(b) Model selection runtime
Figure 2: Model selection performance. Mix: provide PL
+screen with the regularization parameter chosen by the model
selection of NW+screen. Other legend labels are the same as
in Figure 1.
2017], PL without screening still has a significantly larger memory footprint compared with that of
NW. Therefore, the experiments for PL without screening are not fully conducted in Figure 1b,1c,
and 1d for networks with thousands of nodes. On the contrary, PL with the screening rule has a
comparable memory footprint with that of NW. Furthermore, as shown in Figure 1, after applying the
screening rule, PL also has a similar runtime with NW. This phenomenon demonstrates the utility of
the screening rule for effectively reducing the memory footprint of PL, making PL readily available
for large-scale problems.
5.2
Model Selection
Our next experiment performs model selection by choosing an appropriate ? from the regularization
parameter set ?. We leverage the Stability Approach to Regularization Selection (StARS, Liu et al.
2010) for this task. In a nutshell, StARS learns a set of various models, denoted as M, over ? using
many subsamples that are drawn randomly from the original dataset without replacement. It then
picks a ?? ? ? that strikes the best balance between network sparsity and edge selection stability
among the models in M. After the determination of ?? , it is used on the entire original dataset to
learn a model with which we compare the ground truth model and calculate its support recovery Area
Under Curve (AUC). Implementation details of model selection are provided in B.1.
In Figure 2, we summarize the experimental results of model selection, where 24 subsamples are used
for pathwise optimization in parallel to construct M. In Figure 2a, NW with and without screening
achieve the same high AUC values over all four networks, while the application of the screening
rule to NW provides roughly a 2x speedup, according to Figure 2b. The same AUC value shared by
the two variants of NW is due to the same ?? chosen by the model selection procedure. Even more
importantly, it is also because that under the same ?? , the screening rule is able to perfectly identify
the blockwise structure of the parameterization.
Due to high memory cost, the model selection for PL without screening (green bars in Figure 2)
is omitted in some networks. To control the memory footprint, the model selection for PL with
screening (golden bars in Figure 2) also needs to be carried out meticulously by avoiding small ??s
in ? that correspond to dense structures in M during estimation from subsamples. While avoiding
dense structures makes PL with screening the fastest among all (Figure 2b), it comes at the cost
of delivering the least accurate (though still reasonably effective) support recovery performance
(Figure 2a). To improve the accuracy of this approach, we also leverage the connection between
NW and PL by substituting 2??NW for the resultant regularization parameter from model selection
of PL, where ??NW is the regularization parameter selected for NW. This strategy results in better
performance in support recovery (purple bars in Figure 2a).
5.3
Real World Data
Our real world data experiment applies NW with and without screening to a real world gene mutation
dataset collected from 178 lung squamous cell carcinoma samples [Weinstein et al., 2013]. Each
sample contains 13,665 binary variables representing the mutation statuses of various genes. For ease
8
USP34
ASTN2
FBN2
CDH9
ALPK2
UNC13C
STAB2
KIAA1109
FN1
DYNC1H1
PLXNA4
ADAMTS20
MYH4
BAI3
VCAN
PDE4DIP
COL12A1
WDR17
PTPRT
HRNR
ELTD1
ZNF804A
TMEM132D
NRXN1
SYNE2
TPR
ZNF676
SCN1A
VPS13B
MAGEC1
COL6A6
RIMS2
UNC5D
ROS1
ANKRD30A
FAT1
THSD7B
CNTNAP2
MYH1
C20orf26
Figure 3: Connected components learned from lung squamous cell carcinoma mutation data. Genes in
red are (lung) cancer and other disease related genes [Uhl?n et al., 2015]. Mutation data are extracted
via the TCGA2STAT package [Wan et al., 2015] in R and the figure is rendered by Cytoscape.
of interpretation, we keep genes whose mutation rates are at least 10% across all samples, yielding
a subset of 145 genes in total. We use the model selection procedure introduced in Section 5.2 to
determine a ??NW with which we learn the gene mutation network whose connected components are
shown in Figure 3. For model selection, other than the configuration in B.1, we choose ? = 25. 384
trials are run in parallel using all 24 threads. We also choose ?1 such that about 2p log(p) absolute
second empirical moments are greater than ?1 . We choose ?? such that about 0.25p absolute second
empirical moments are greater than ?? .
In our experiment, NW with and without screening select the same ??NW , and generate the same
network. Since the dataset in question has a lower dimension and a smaller sample size compared with
the synthetic data, NW without screening is adequately efficient. Nonetheless, with screening NW is
still roughly 20% faster. This phenomenon once again indicates that in practice the screening rule
can perfectly identify the blockwise sparsity pattern in the parameterization and deliver a significant
efficiency gain. The genes in red in Figure 3 represent (lung) cancer and other disease related genes,
which are scattered across the seven subnetworks discovered by the algorithm. In our experiment, we
also notice that all the weights on the edges are positive. This is consistent with the biological belief
that associated genes tend to mutate together to cause cancer.
6
Generalization
With unary potentials, the `1 -regularized MLE for the Ising model is defined as:
?
?
p
p?1 X
p
n
X
X
X
?
1
(k)
(k) (k)
?
?? = arg min ?
?ii xi +
?ij xi xj ? + A(?) + k?k1,off ,
?
n
2
i=1
i=1 j>i
(7)
k=1
Pp Pp
where k?k1,off = i=1 j6=i |?ij |. Note that the unary potentials are not penalized, which is a
common practice [Wainwright et al., 2006, H?fling and Tibshirani, 2009, Ravikumar et al., 2010,
Viallon et al., 2014] to ensure a hierarchical parameterization. The screening rule here is to replace
(4) in Theorem 3 with:
|EX Xi Xj ? EX Xi EX Xj | ? ?.
(8)
Exhaustive justification, interpretation, and experiments are provided in Supplement C.
7
Conclusion
We have proposed a screening rule for `1 -regularized Ising model estimation. The simple closed-form
screening rule is a necessary and sufficient condition for exact blockwise structural identification.
Experimental results suggest that the proposed screening rule can provide drastic speedups for
learning when combined with various optimization algorithms. Future directions include deriving
screening rules for more general undirected graphical models [Liu et al., 2012, 2014b,a, Liu, 2014,
Liu et al., 2016], and deriving screening rules for other inexact optimization algorithms [Liu and
Page, 2013]. Further theoretical justifications regarding the conditions upon which the screening rule
can be combined with inexact algorithms to recover block structures losslessly are also desirable.
9
References
O. Banerjee, L. E. Ghaoui, and A. d?Aspremont. Model selection through sparse maximum likelihood
estimation for multivariate gaussian or binary data. Journal of Machine Learning Research, 9
(Mar):485?516, 2008.
R. F. Barber, M. Drton, et al. High-dimensional ising model selection with bayesian information
criteria. Electronic Journal of Statistics, 9(1):567?607, 2015.
H. Chen and B. M. Sharp. Content-rich biological network constructed by mining pubmed abstracts.
BMC Bioinformatics, 5(1):147, 2004.
P. Danaher, P. Wang, and D. M. Witten. The joint graphical lasso for inverse covariance estimation
across multiple classes. Journal of the Royal Statistical Society: Series B (Statistical Methodology),
76(2):373?397, 2014.
O. Fercoq, A. Gramfort, and J. Salmon. Mind the duality gap: safer rules for the lasso. In Proceedings
of The 32nd International Conference on Machine Learning, pages 333?342, 2015.
J. Friedman, T. Hastie, and R. Tibshirani. Sparse inverse covariance estimation with the graphical
lasso. Biostatistics, 9(3):432?441, 2008.
J. Friedman, T. Hastie, and R. Tibshirani. Regularization paths for generalized linear models via
coordinate descent. Journal of Statistical Software, 33(1):1, 2010.
S. Geng, Z. Kuang, and D. Page. An efficient pseudo-likelihood method for sparse binary pairwise
Markov network estimation. arXiv Preprint, 2017.
L. E. Ghaoui, V. Viallon, and T. Rabbani. Safe feature elimination for the lasso and sparse supervised
learning problems. arXiv Preprint, 2010.
T. Hastie, R. Tibshirani, and M. Wainwright. Statistical learning with sparsity: the lasso and
generalizations. CRC Press, 2015.
H. H?fling and R. Tibshirani. Estimation of sparse binary pairwise Markov networks using pseudolikelihoods. Journal of Machine Learning Research, 10(Apr):883?906, 2009.
J. Honorio and D. Samaras. Multi-task learning of gaussian graphical models. In Proceedings of the
27th International Conference on Machine Learning (ICML-10), pages 447?454, 2010.
C.-J. Hsieh, M. A. Sustik, I. S. Dhillon, P. K. Ravikumar, and R. Poldrack. Big & quic: Sparse inverse
covariance estimation for a million variables. In Advances in Neural Information Processing
Systems, pages 3165?3173, 2013.
D. Karger and N. Srebro. Learning Markov networks: Maximum bounded tree-width graphs. In
Proceedings of the Twelfth Annual ACM-SIAM Symposium on Discrete Algorithms, pages 392?401.
Society for Industrial and Applied Mathematics, 2001.
D. Koller and N. Friedman. Probabilistic graphical models: principles and techniques. MIT press,
2009.
H. Liu, K. Roeder, and L. Wasserman. Stability approach to regularization selection (stars) for high
dimensional graphical models. In Advances in Neural Information Processing Systems, pages
1432?1440, 2010.
J. Liu. Statistical Methods for Genome-wide Association Studies and Personalized Medicine. PhD
thesis, The University of Wisconsin-Madison, 2014.
J. Liu and D. Page. Structure learning of undirected graphical models with contrastive divergence.
ICML 2013 Workshop on Structured Learning: Inferring Graphs from Structured and Unstructured
Inputs, 2013.
J. Liu, P. Peissig, C. Zhang, E. Burnside, C. McCarty, and D. Page. Graphical-model based multiple
testing under dependence, with applications to genome-wide association studies. In Uncertainty in
Artificial Intelligence, volume 2012, page 511. NIH Public Access, 2012.
10
J. Liu, Z. Zhao, J. Wang, and J. Ye. Safe screening with variational inequalities and its application to
lasso. arXiv Preprint arXiv:1307.7577, 2013.
J. Liu, C. Zhang, E. Burnside, and D. Page. Learning heterogeneous hidden Markov random fields.
In Artificial Intelligence and Statistics, pages 576?584, 2014a.
J. Liu, C. Zhang, E. Burnside, and D. Page. Multiple testing under dependence via semiparametric
graphical models. In Proceedings of the 31st International Conference on Machine Learning
(ICML-14), pages 955?963, 2014b.
J. Liu, C. Zhang, D. Page, et al. Multiple testing under dependence via graphical models. The Annals
of Applied Statistics, 10(3):1699?1724, 2016.
P.-L. Loh, M. J. Wainwright, et al. Structure estimation for discrete graphical models: Generalized
covariance matrices and their inverses. In Advances in Neural Information Processing Systems,
pages 2096?2104, 2012.
P.-L. Loh, M. J. Wainwright, et al. Structure estimation for discrete graphical models: Generalized
covariance matrices and their inverses. The Annals of Statistics, 41(6):3022?3049, 2013.
S. Luo, R. Song, and D. Witten. Sure screening for gaussian graphical models. arXiv Preprint
arXiv:1407.7819, 2014.
R. Mazumder and T. Hastie. Exact covariance thresholding into connected components for large-scale
graphical lasso. Journal of Machine Learning Research, 13(Mar):781?794, 2012.
J. Pena and R. Tibshirani. Lecture notes in machine learning 10-725/statistics 36-725-convex
optimization (fall 2016), 2016.
J. Peng, P. Wang, N. Zhou, and J. Zhu. Partial correlation estimation by joint sparse regression
models. Journal of the American Statistical Association, 104(486):735?746, 2009.
P. Ravikumar, M. J. Wainwright, J. D. Lafferty, et al. High-dimensional ising model selection using
l1-regularized logistic regression. The Annals of Statistics, 38(3):1287?1319, 2010.
R. Tibshirani. Regression shrinkage and selection via the lasso. Journal of the Royal Statistical
Society. Series B (Methodological), pages 267?288, 1996.
R. Tibshirani, J. Bien, J. Friedman, T. Hastie, N. Simon, J. Taylor, and R. J. Tibshirani. Strong rules
for discarding predictors in lasso-type problems. Journal of the Royal Statistical Society: Series B
(Statistical Methodology), 74(2):245?266, 2012.
M. Uhl?n, L. Fagerberg, B. M. Hallstr?m, C. Lindskog, P. Oksvold, A. Mardinoglu, ?. Sivertsson,
C. Kampf, E. Sj?stedt, A. Asplund, et al. Tissue-based map of the human proteome. Science, 347
(6220):1260419, 2015.
V. Viallon, O. Banerjee, E. Jougla, G. Rey, and J. Coste. Empirical comparison study of approximate
methods for structure selection in binary graphical models. Biometrical Journal, 56(2):307?331,
2014.
M. Vuffray, S. Misra, A. Lokhov, and M. Chertkov. Interaction screening: Efficient and sampleoptimal learning of ising models. In Advances in Neural Information Processing Systems, pages
2595?2603, 2016.
M. J. Wainwright, J. D. Lafferty, and P. K. Ravikumar. High-dimensional graphical model selection
using l1-regularized logistic regression. In Advances in Neural Information Processing Systems,
pages 1465?1472, 2006.
Y.-W. Wan, G. I. Allen, and Z. Liu. Tcga2stat: simple tcga data access for integrated statistical
analysis in r. Bioinformatics, page btv677, 2015.
Y.-W. Wan, G. I. Allen, Y. Baker, E. Yang, P. Ravikumar, M. Anderson, and Z. Liu. Xmrf: an r
package to fit Markov networks to high-throughput genetics data. BMC Systems Biology, 10(3):69,
2016.
11
J. Wang, J. Zhou, P. Wonka, and J. Ye. Lasso screening rules via dual polytope projection. In
Advances in Neural Information Processing Systems, pages 1070?1078, 2013.
J. Wang, J. Zhou, J. Liu, P. Wonka, and J. Ye. A safe screening rule for sparse logistic regression. In
Advances in Neural Information Processing Systems, pages 1053?1061, 2014.
J. N. Weinstein, E. A. Collisson, G. B. Mills, K. R. M. Shaw, B. A. Ozenberger, K. Ellrott, I. Shmulevich, C. Sander, J. M. Stuart, C. G. A. R. Network, et al. The cancer genome atlas pan-cancer
analysis project. Nature Genetics, 45(10):1113?1120, 2013.
D. M. Witten, J. H. Friedman, and N. Simon. New insights and faster computations for the graphical
lasso. Journal of Computational and Graphical Statistics, 20(4):892?900, 2011.
Z. J. Xiang, Y. Wang, and P. J. Ramadge. Screening tests for lasso problems. IEEE Transactions on
Pattern Analysis and Machine Intelligence, PP(99):1?1, 2016. ISSN 0162-8828. doi: 10.1109/
TPAMI.2016.2568185.
S. Yang, Z. Lu, X. Shen, P. Wonka, and J. Ye. Fused multiple graphical lasso. SIAM Journal on
Optimization, 25(2):916?943, 2015.
12
| 6674 |@word trial:4 rising:1 polynomial:1 stronger:1 eliminating:1 nd:5 twelfth:1 checkable:1 prominence:1 covariance:8 hsieh:2 attainable:1 pick:1 contrastive:1 moment:10 necessity:1 liu:17 contains:1 safeguarded:1 karger:2 initial:1 configuration:1 series:3 current:1 comparing:1 luo:2 yet:1 readily:2 stemming:1 partition:10 analytic:1 atlas:1 generative:4 selected:1 intelligence:3 parameterization:12 ith:3 provides:6 node:22 zhang:4 five:2 rabbani:1 constructed:3 c2:5 symposium:1 weinstein:2 overhead:1 introduce:1 pairwise:4 peng:2 notably:1 hardness:1 roughly:4 multi:1 encouraging:1 cpu:1 increasing:1 provided:3 discover:1 notation:1 estimating:2 moreover:1 biostatistics:1 project:1 bounded:1 what:1 baker:1 substantially:2 affirmative:1 caution:1 finding:1 guarantee:1 pseudo:1 quantitative:1 golden:1 nutshell:1 runtime:13 exactly:4 demonstrates:1 control:1 arguably:1 positive:1 mistake:2 consequence:2 era:1 despite:1 mccarty:1 path:1 exert:1 twice:1 suggests:1 challenging:3 ease:2 fastest:1 ramadge:1 range:7 faithful:1 testing:3 practice:6 block:28 implement:1 footprint:4 procedure:5 peissig:1 area:1 empirical:12 evolving:1 arsenal:1 attain:1 significantly:1 projection:1 suggest:1 proteome:1 synergistic:1 selection:23 operator:1 close:1 applying:3 equivalent:1 map:1 straightforward:1 flexibly:1 independently:1 starting:1 convex:1 shen:1 identifying:2 recovery:6 unstructured:1 wasserman:1 rule:70 insight:4 estimator:1 importantly:1 fill:1 deriving:2 century:1 stability:3 exploratory:2 notion:1 justification:2 coordinate:1 annals:3 pioneered:1 exact:18 us:1 element:4 asymmetric:2 ising:24 preprint:4 wang:8 solved:3 thousand:4 calculate:1 ensures:1 connected:7 valuable:1 deeply:1 substantial:2 intuition:1 disease:2 complexity:1 instructive:1 solving:4 tight:1 deliver:2 upon:3 samara:2 efficiency:9 strikingly:1 accelerate:1 easily:3 joint:2 various:11 fast:1 effective:1 doi:1 artificial:2 choosing:2 exhaustive:1 whose:4 emerged:1 larger:5 solve:3 heuristic:2 say:1 relax:1 apparent:1 pseudolikelihoods:1 statistic:7 jointly:1 echo:1 ip:1 interplay:1 hoc:1 subsamples:4 tpami:1 interaction:2 maximal:1 combining:1 date:1 achieve:2 magnify:1 scalability:3 mutate:1 cluster:1 help:1 illustrate:1 ij:14 strong:2 p2:2 recovering:1 involves:1 come:1 direction:1 safe:3 tcga:1 owing:1 exploration:1 human:1 elimination:1 public:1 adjacency:1 crc:1 generalization:3 homotopy:1 biological:4 summation:1 adjusted:1 pl:54 considered:4 ground:2 normal:1 exp:4 nw:70 substituting:1 major:1 lokhov:1 smallest:3 omitted:1 estimation:18 realizes:1 lose:1 label:1 bridge:1 reflects:1 minimization:4 mit:1 gaussian:9 aim:1 pn:1 avoid:1 shrinkage:2 zhou:3 shmulevich:1 corollary:4 l0:7 focus:3 derived:2 iconic:1 methodological:1 likelihood:4 indicates:2 check:1 industrial:1 baseline:1 inference:1 roeder:1 unary:3 shattered:1 honorio:2 xip:1 entire:1 integrated:1 hidden:1 koller:1 i1:1 interested:2 arg:5 among:4 aforementioned:1 dual:1 denoted:2 art:1 gramfort:1 uhl:2 field:2 construct:1 once:1 beach:1 sampling:1 bmc:2 represents:8 stuart:1 biology:1 icml:3 throughput:1 geng:4 mimic:1 future:1 inherent:1 modern:1 randomly:2 fn1:1 fling:5 densely:1 divergence:1 maxj:2 replacement:2 n1:1 friedman:8 attempt:1 drton:1 screening:89 mining:1 yielding:1 amenable:1 implication:1 beforehand:1 accurate:1 edge:8 encourage:1 partial:1 necessary:8 coste:1 tree:1 taylor:1 plotted:1 theoretical:1 minimal:1 increased:1 column:1 xeon:1 assignment:1 cost:4 entry:2 subset:3 hundred:1 kuang:1 predictor:1 successful:1 conducted:4 reported:2 answer:1 synthetic:5 combined:7 st:2 fundamental:1 international:3 siam:2 probabilistic:1 xi1:1 off:2 together:1 fused:1 fat1:1 again:1 thesis:1 wan:6 choose:6 american:1 derivative:1 zhao:1 return:2 toy:1 combing:1 meticulously:1 potential:4 star:3 biometrical:1 disconnect:1 notable:1 ad:1 closed:3 linked:1 portion:4 start:1 recover:2 lung:4 parallel:4 complicated:1 red:2 defer:1 simon:2 mutation:8 contribution:1 purple:1 accuracy:1 efficiently:2 ensemble:1 correspond:2 identify:9 spaced:1 identification:7 mutated:1 bayesian:1 lu:1 biostat:1 j6:1 tissue:1 definition:2 inexact:12 vuffray:2 nonetheless:2 pp:7 steadily:1 resultant:1 proof:1 associated:2 gain:5 sampled:1 dataset:15 adjusting:1 popular:2 recall:1 knowledge:2 attained:1 indx:2 supervised:1 xxx:1 methodology:2 sufficiency:1 though:1 mar:2 generality:2 furthermore:5 anderson:1 until:2 glms:2 hand:1 correlation:1 replacing:1 banerjee:3 glance:1 logistic:12 grows:1 innate:1 facilitate:1 effect:1 ye:4 usa:1 true:2 building:1 counterpart:1 adequately:1 regularization:27 hence:5 symmetric:11 nonzero:1 iteratively:1 dhillon:1 i2:1 illustrated:2 during:1 self:1 encourages:1 auc:5 width:1 coincides:1 criterion:1 generalized:4 evident:1 demonstrate:4 performs:1 l1:2 allen:2 ranging:1 wise:5 variational:1 salmon:1 nih:1 common:1 witten:7 ji:4 blk:4 poldrack:1 exponentially:1 volume:1 million:1 association:3 interpretation:4 tpr:1 pena:1 significant:5 gibbs:1 mathematics:1 i6:1 access:2 similarity:1 multivariate:1 recent:1 burnside:3 moderate:2 discard:1 scenario:1 misra:1 server:1 inequality:1 binary:8 yi:4 additional:1 greater:4 determine:1 strike:1 preservation:1 ii:1 multiple:5 desirable:2 mix:3 danaher:4 technical:2 faster:2 determination:2 believed:1 long:1 offer:1 page3:1 post:1 mle:2 ravikumar:7 scalable:2 regression:14 variant:1 heterogeneous:1 arxiv:6 represent:6 normalization:1 cell:2 c1:5 preserved:1 background:1 semiparametric:1 separately:3 decreased:1 median:2 crucial:1 sure:1 tend:1 undirected:2 contrary:2 leveraging:1 legend:1 lafferty:2 call:1 structural:2 yang:3 leverage:4 enough:5 identically:1 sander:1 xj:17 fit:1 hastie:10 lasso:17 perfectly:2 reduce:1 regarding:1 absent:1 thread:2 utility:2 gb:1 penalty:2 song:1 loh:2 speaking:1 cause:1 rey:1 generally:1 delivering:1 verifiable:1 ten:1 generate:3 sl:1 notice:2 tibshirani:16 carcinoma:2 discrete:9 four:3 threshold:2 drawn:2 wisc:3 viallon:5 ram:1 graph:3 sum:1 screened:1 inverse:6 run:2 package:2 uncertainty:1 decide:1 electronic:1 summarizes:2 scaling:1 comparable:1 bound:2 correspondence:1 annual:1 handful:2 constraint:2 incorporation:1 x2:1 software:1 personalized:1 generates:1 speed:2 fercoq:2 extremely:1 min:9 optimality:1 performing:1 rendered:1 speedup:5 structured:2 according:7 disconnected:9 smaller:6 across:4 pan:1 making:1 happens:1 ghaoui:5 discus:2 turn:1 nonempty:1 xi2:1 needed:1 mind:1 drastic:1 end:1 subnetworks:4 sustik:1 available:2 operation:1 apply:1 observe:1 hierarchical:1 appropriate:1 lindskog:1 shaw:1 alternative:3 robustness:1 comprehensible:1 original:5 remaining:3 ensure:1 include:1 graphical:25 madison:1 medicine:1 exploit:1 k1:7 especially:3 establish:2 society:4 objective:5 question:2 quic:1 strategy:3 dependence:3 diagonal:3 losslessly:1 subnetwork:2 gradient:2 simulated:1 evenly:1 seven:1 polytope:1 barber:2 collected:1 barely:1 equip:1 issn:1 index:3 relationship:3 providing:2 balance:1 potentially:2 blockwise:28 wonka:3 negative:1 implementation:2 cl0:5 perform:1 observation:2 datasets:4 markov:6 descent:1 ever:3 discovered:1 sharp:2 david:1 introduced:1 pair:3 connection:2 learned:1 tremendous:1 nip:1 address:1 able:1 bar:3 usually:3 parallelism:1 pattern:2 sparsity:9 summarize:1 bien:1 royal:3 interpretability:1 max:17 unsuccessful:1 wainwright:7 power:3 suitable:1 event:2 memory:6 green:1 regularized:15 belief:1 squamous:2 zhu:1 representing:2 improve:1 lossless:1 deemed:1 carried:1 aspremont:1 deviate:1 prior:1 literature:4 discovery:1 xiang:2 wisconsin:2 law:3 loss:2 fully:13 lecture:1 rationale:1 srebro:2 remarkable:2 degree:3 sufficient:9 xp:1 imposes:1 consistent:1 principle:1 thresholding:1 row:1 cancer:5 cytoscape:1 penalized:2 summary:1 genetics:2 last:1 drastically:1 pseudolikelihood:7 wide:2 fall:1 absolute:9 sparse:18 distributed:2 benefit:1 curve:1 dimension:3 world:5 rich:1 genome:3 made:2 transaction:1 sj:1 approximate:2 status:1 gene:13 clique:2 keep:1 suggestive:2 kkt:1 conclude:2 discriminative:2 xi:27 thep:1 continuous:2 decade:1 why:1 table:9 nature:1 learn:4 reasonably:1 ca:1 symmetry:1 mazumder:5 e5:1 cl:12 necessarily:1 diag:1 apr:1 main:1 dense:2 big:2 x1:1 intel:1 tl:2 pubmed:1 screen:17 fashion:2 scattered:1 shrinking:1 fails:1 position:3 inferring:1 learns:1 chertkov:1 formula:2 theorem:16 discarding:2 xt:1 insightful:1 maxi:1 workshop:1 effectively:1 gained:2 supplement:2 phd:1 magnitude:3 gap:2 phenotype:1 easier:1 chen:2 mill:1 simply:1 glmnet:1 ordered:1 pathwise:7 applies:1 corresponds:1 truth:2 determines:1 extracted:1 acm:1 conditional:1 viewed:1 presentation:1 shared:1 absence:2 replace:1 hard:1 content:1 safer:1 typical:1 specifically:2 contrasted:1 craving:1 determined:1 reducing:2 principal:1 total:2 duality:1 experimental:3 formally:1 select:1 support:5 unbalanced:1 bioinformatics:2 accelerated:1 preparation:1 avoiding:2 phenomenon:2 ex:10 |
6,273 | 6,675 | Improved Dynamic Regret for Non-degenerate
Functions
Lijun Zhang? , Tianbao Yang? , Jinfeng Yi? , Rong Jin? , Zhi-Hua Zhou?
National Key Laboratory for Novel Software Technology, Nanjing University, Nanjing, China
?
Department of Computer Science, The University of Iowa, Iowa City, USA
?
AI Foundations Lab, IBM Thomas J. Watson Research Center, Yorktown Heights, NY, USA
?
Alibaba Group, Seattle, USA
[email protected], [email protected], [email protected]
[email protected], [email protected]
?
Abstract
Recently, there has been a growing research interest in the analysis of dynamic
regret, which measures the performance of an online learner against a sequence
of local minimizers. By exploiting the strong convexity, previous studies have
shown that the dynamic regret can be upper bounded by the path-length of the
comparator sequence. In this paper, we illustrate that the dynamic regret can be
further improved by allowing the learner to query the gradient of the function
multiple times, and meanwhile the strong convexity can be weakened to other
non-degenerate conditions. Specifically, we introduce the squared path-length,
which could be much smaller than the path-length, as a new regularity of the
comparator sequence. When multiple gradients are accessible to the learner, we
first demonstrate that the dynamic regret of strongly convex functions can be upper
bounded by the minimum of the path-length and the squared path-length. We then
extend our theoretical guarantee to functions that are semi-strongly convex or selfconcordant. To the best of our knowledge, this is the first time that semi-strong
convexity and self-concordance are utilized to tighten the dynamic regret.
1
Introduction
Online convex optimization is a fundamental tool for solving a wide variety of machine learning
problems [Shalev-Shwartz, 2011]. It can be formulated as a repeated game between a learner and
an adversary. On the t-th round of the game, the learner selects a point xt from a convex set X
and the adversary chooses a convex function ft : X 7? R. Then, the function is revealed to the
learner, who incurs loss ft (xt ). The standard performance measure is the regret, defined as the
difference between the learner?s cumulative loss and the cumulative loss of the optimal fixed vector
in hindsight:
T
T
X
X
ft (x).
(1)
ft (xt ) ? min
t=1
x?X
t=1
Over the past decades, various online algorithms, such as the online gradient descent [Zinkevich,
2003], have been proposed to yield sub-linear regret under different scenarios [Hazan et al., 2007,
Shalev-Shwartz et al., 2007].
Though equipped with rich theories, the notion of regret fails to illustrate the performance of online algorithms in dynamic setting, as a static comparator is used in (1). To overcome this limitation, there has been a recent surge of interest in analyzing a more stringent metric?dynamic regret
[Hall and Willett, 2013, Besbes et al., 2015, Jadbabaie et al., 2015, Mokhtari et al., 2016, Yang et al.,
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
2016], in which the cumulative loss of the learner is compared against a sequence of local minimizers, i.e.,
RT? :=R(x?1 , . . . , x?T ) =
T
X
t=1
ft (xt ) ?
T
X
ft (x?t ) =
T
X
t=1
t=1
ft (xt ) ?
T
X
t=1
min ft (x)
(2)
x?X
where x?t ? argminx?X ft (x). A more general definition of dynamic regret is to evaluate the
difference of the cumulative loss with respect to any sequence of comparators u1 , . . . , uT ? X
[Zinkevich, 2003].
It is well-known that in the worst-case, it is impossible to achieve a sub-linear dynamic regret bound,
due to the arbitrary fluctuation in the functions. However, it is possible to upper bound the dynamic
regret in terms of certain regularity of the comparator sequence or the function sequence. A natural
regularity is the path-length of the comparator sequence, defined as
PT? := P(x?1 , . . . , x?T ) =
T
X
t=2
kx?t ? x?t?1 k
(3)
that captures the cumulative Euclidean norm of the difference between successive comparators.
? For
convex functions, the dynamic regret of online gradient descent can be upper bounded by O( T PT? )
[Zinkevich, 2003]. And when all the functions are strongly convex and smooth, the upper bound can
be improved to O(PT? ) [Mokhtari et al., 2016].
In the aforementioned results, the learner uses the gradient of each function only once, and performs
one step of gradient descent to update the intermediate solution. In this paper, we examine an
interesting question: is it possible to improve the dynamic regret when the learner is allowed to
query the gradient multiple times? Note that the answer to this question is no if one aims to promote
the static regret in (1), according to the results on the minimax regret bound [Abernethy et al., 2008a].
We however show that when coming to the dynamic regret, multiple gradients can reduce the upper
bound significantly. To this end, we introduce a new regularity?the squared path-length:
ST? := S(x?1 , . . . , x?T ) =
T
X
t=2
kx?t ? x?t?1 k2
(4)
are small. For example, when
which could be much?smaller than PT? when the local variations
?
kx?t ? x?t?1 k = ?(1/ T ) for all t ? [T ], we have PT? = ?( T ) but ST? = ?(1). We advance the
analysis of dynamic regret in the following aspects.
? When all the functions are strongly convex and smooth, we propose to apply gradient descent multiple times in each round, and demonstrate that the dynamic regret is reduced from
O(PT? ) to O(min(PT? , ST? )), provided the gradients of minimizers are small. We further
present a matching lower bound which implies our result cannot be improved in general.
? When all the functions are semi-strongly convex and smooth, we show that the standard
online gradient descent still achieves the O(PT? ) dynamic regret. And if we apply gradient descent multiple times in each round, the upper bound can also be improved to
O(min(PT? , ST? )), under the same condition as strongly convex functions.
? When all the functions are self-concordant, we establish a similar guarantee if both the
gradient and Hessian of the function can be queried multiple times. Specifically, we propose to apply the damped Newton method multiple times in each round, and prove an
O(min(PT? , ST? )) bound of the dynamic regret under appropriate conditions.1
Application to Statistical Learning Most studies of dynamic regret, including this paper do not
make stochastic assumptions on the function sequence. In the following, we discuss how to interpret our results when facing the problem of statistical learning. In this case, the learner receives a
sequence of losses ?(x? z1 , y1 ), ?(x? z2 , y2 ), . . ., where (zi , yi )?s are instance-label pairs sampled
from a unknown distribution, and ?(?, ?) measures the prediction error. To avoid the random fluctuation caused by sampling, we can set ft as the loss averaged over a mini-batch of instance-label pairs.
As a result, when the underlying distribution is stationary or drifts slowly, successive functions will
be close to each other, and thus the path-length and the squared path-length are expected to be small.
1
PT? and ST? are modified slightly when functions are semi-strongly convex or self-concordant.
2
2
Related Work
The static regret in (1) has been extensively studied in the literature [Shalev-Shwartz,
2011]. It has
?
been established that the static regret can be upper bounded by O( T ), O(log T ), and O(log T )
for convex functions, strongly convex functions, and exponentially concave functions, respectively
[Zinkevich, 2003, Hazan et al., 2007]. Furthermore, those upper bounds are proved to be minimax
optimal [Abernethy et al., 2008a, Hazan and Kale, 2011].
The notion of dynamic regret is introduced by Zinkevich [2003]. If we choose the online gradient
descent as the learner, the dynamic?regret with respect to any comparator sequence u1 , . . . , uT , i.e.,
R(u1 , . . . , uT ), is on the order of T P(u1 , . . . , uTp
). When a prior knowledge of PT? is available,
?
the dynamic regret RT can be upper bounded by O( T PT? ) [Yang et al., 2016]. If all the functions
are strongly convex and smooth, the upper bound of RT? can be improved to O(PT? ) [Mokhtari et al.,
2016]. The O(PT? ) rate is also achievable when all the functions are convex and smooth, and all the
minimizers x?t ?s lie in the interior of X [Yang et al., 2016].
Another regularity of the comparator sequence, which is similar to the path-length, is defined as
P ? (u1 , . . . , uT ) =
T
X
t=2
kut ? ?t (ut?1 )k
where ?t (?) is a dynamic model that predicts a reference point for the t-th round. The advantage
of this measure is that when the comparator sequence follows the dynamical model closely, it can
be much smaller than the path-length P(u1 , . . . , uT ). A novel algorithm named dynamic mirror
descent is?proposed to take ?t (ut?1 ) into account, and the dynamic regret R(u1 , . . . , uT ) is on the
order of T P ? (u1 , . . . , uT ) [Hall and Willett, 2013]. There are also some regularities defined in
terms of the function sequence, such as the functional variation [Besbes et al., 2015]
FT := F(f1 , . . . , fT ) =
T
X
max |ft (x) ? ft?1 (x)|
(5)
max k?ft (x) ? ?ft?1 (x)k2 .
(6)
t=2
x?X
or the gradient variation [Chiang et al., 2012]
GT := G(f1 , . . . , fT ) =
T
X
t=2
x?X
Under the condition that FT ? FT and Ft is given beforehand, a restarted online gradient descent
2/3 1/3
is developed
? by Besbes et al. [2015], and the dynamic regret is upper bounded by O(T FT ) and
O(log T T FT ) for convex functions and strongly convex functions, respectively.
The regularities mentioned above reflect different aspects of the learning problem, and are not directly comparable in general. Thus, it is appealing to develop an algorithm that adapts to the smaller
regularity of the problem. Jadbabaie et al. [2015] propose an adaptive algorithm based on the optimistic mirror descent [Rakhlin and Sridharan, 2013], such that the dynamic regret is given in terms
of all the three regularities (PT? , FT , and GT ). However, it relies on the assumption that the learner
can calculate each regularity incrementally.
In the setting of prediction with expert advice, the dynamic regret is also referred to as tracking
regret or shifting regret [Herbster and Warmuth, 1998, Cesa-bianchi et al., 2012]. The path-length
of the comparator sequence is named as shift, which is just the number of times the expert changes.
Another related performance measure is the adaptive regret, which aims to minimize the static regret
over any interval [Hazan and Seshadhri, 2007, Daniely et al., 2015]. Finally, we note that the study
of dynamic regret is similar to the competitive analysis in the sense that both of them compete
against an optimal offline policy, but with significant differences in their assumptions and techniques
[Buchbinder et al., 2012].
3
Online Learning with Multiple Gradients
In this section, we discuss how to improve the dynamic regret by allowing the learner to query the
gradient multiple times. We start with strongly convex functions, and then proceed to semi-strongly
convex functions, and finally investigate self-concordant functions.
3
Algorithm 1 Online Multiple Gradient Descent (OMGD)
Require: The number of inner iterations K and the step size ?
1: Let x1 be any point in X
2: for t = 1, . . . , T do
3:
Submit xt ? X and the receive loss ft : X 7? R
4:
z1t = xt
5:
for j = 1, . . . , K do
6:
zj+1
= ?X zjt ? ??ft (zjt )
t
7:
end for
8:
xt+1 = zK+1
t
9: end for
3.1
Strongly Convex and Smooth Functions
To be self-contained, we provide the definitions of strong convexity and smoothness.
Definition 1. A function f : X 7? R is ?-strongly convex, if
f (y) ? f (x) + h?f (x), y ? xi +
?
ky ? xk2 , ?x, y ? X .
2
Definition 2. A function f : X 7? R is L-smooth, if
L
ky ? xk2 , ?x, y ? X .
2
Example 1. The following functions are both strongly convex and smooth.
f (y) ? f (x) + h?f (x), y ? xi +
1. A quadratic form f (x) = x? Ax ? 2b? x + c where aI A bI, a > 0 and b < ?;
2. The regularized logistic loss f (x) = log(1 + exp(b? x)) + ?2 kxk2 , where ? > 0.
Following previous studies [Mokhtari et al., 2016], we make the following assumptions.
Assumption 1. Suppose the following conditions hold for each ft : X 7? R.
1. ft is ?-strongly convex and L-smooth over X ;
2. k?ft (x)k ? G, ?x ? X .
When the learner can query the gradient of each function only once, the most popular learning
algorithm is the online gradient descent:
xt+1 = ?X (xt ? ??ft (xt ))
where ?X (?) denotes the projection onto the nearest point in X . Mokhtari et al. [2016] have established an O(PT? ) bound of dynamic regret, as stated below.
Theorem 1. Suppose Assumption 1 is true. By setting ? ? 1/L in online gradient descent, we have
T
X
t=1
where ? =
q
1?
ft (xt ) ? ft (x?t ) ?
1
1
GPT? +
Gkx1 ? x?1 k
1??
1??
2?
1/?+? .
We now consider the setting that the learner can access the gradient of each function multiple times.
The algorithm is a natural extension of online gradient descent by performing gradient descent multiple times in each round. Specifically, in the t-th round, given the current solution xt , we generate
a sequence of solutions, denoted by z1t , . . . , zK+1
, where K is a constant independent from T , as
t
follows:
z1t = xt , zj+1
= ?X zjt ? ??ft (zjt ) , j = 1, . . . , K.
t
Then, we set xt+1 = zK+1
. The procedure is named as Online Multiple Gradient Descent (OMGD)
t
and is summarized in Algorithm 1.
4
By applying gradient descent multiple times, we are able to extract more information from each
function and therefore are more likely to obtain a tight bound for the dynamic regret. The following
theorem shows that the multiple accesses of the gradient indeed help improve the dynamic regret.
Theorem 2. Suppose Assumption 1 is true. By setting ? ? 1/L and K = ? 1/?+?
2? ln 4? in Algorithm 1, for any constant ? > 0, we have
?
?
?
T
?2GPT + 2Gkx1 ? x1 k,
X
P
?
T
? 2
ft (xt ) ? ft (xt ) ? min
? t=1 k?ft (xt )k + 2(L + ?)S ? + (L + ?)kx1 ? x? k2 .
t=1
T
1
2?
PT
k?ft (x?t )k2 is small, Theorem 2 can be simplified as follows.
PT
Corollary 3. Suppose t=1 k?ft (x?t )k2 = O(ST? ), from Theorem 2, we have
When
t=1
T
X
t=1
In particular, if x?t
as ? ? 0, implies
T
X
t=1
ft (xt ) ? ft (x?t ) = O (min(PT? , ST? )) .
belongs to the relative interior of X (i.e., ?ft (x?t ) = 0) for all t ? [T ], Theorem 2,
ft (xt ) ? ft (x?t ) ? min 2GPT? + 2Gkx1 ? x?1 k, 2LST? + Lkx1 ? x?1 k2 .
Compared to Theorem 1, the proposed OMGD improves the dynamic regret from O(PT? ) to
O (min (PT? , ST? )), when the gradients of minimizers are small. Recall the definitions of PT? and
ST? in (3) and (4), respectively. We can see that ST? introduces a square when measuring the difference between x?t and x?t?1 . In this way, if the local variations (kx?t ? x?t?1 k?s) are small, ST? can
be significantly smaller than PT? , as indicated below.
Example 2. Suppose kx?t ? x?t?1 k = T ?? for all t ? 1 and ? > 0, we have
ST? +1 = T 1?2? ? PT? +1 = T 1?? .
?
In particular, when ? = 1/2, we have ST? +1 = 1 ? PT? +1 = T .
ST? is also closely related to the gradient variation in (6). When all the x?t ?s belong to the relative
interior of X , we have ?ft (x?t ) = 0 for all t ? [T ] and therefore
GT ?
T
X
t=2
k?ft (x?t?1 ) ? ?ft?1 (x?t?1 )k2 =
T
X
t=2
k?ft (x?t?1 ) ? ?ft (x?t )k2 ? ?2 ST?
(7)
where the last inequality follows from the property of strongly convex functions [Nesterov, 2004].
The following corollary is an immediate consequence of Theorem 2 and the inequality in (7).
Corollary 4. Suppose Assumption 1 is true, and further assume all the x?t ?s belong to the relative
interior of X . By setting ? ? 1/L and K = ? 1/?+?
2? ln 4? in Algorithm 1, we have
T
X
t=1
ft (xt ) ?
ft (x?t )
? min
2GPT?
+ 2Gkx1 ?
x?1 k,
2LGT
? 2
+ Lkx1 ? x1 k .
?2
In Theorem 2, the number of accesses of gradients K is set to be a constant depending on the
condition number of the function. One may ask whether we can obtain a tighter bound by using a
larger K. Unfortunately, according to our analysis, even if we take K = ?, which means ft (?) is
minimized exactly, the upper bound can only be improved by a constant factor and the order remains
the same. A related question is whether we can reduce the value of K by adopting more advanced
optimization techniques, such as the accelerated gradient descent [Nesterov, 2004]. This is an open
problem to us, and will be investigated as a future work.
Finally, we prove that the O(ST? ) bound is optimal for strongly convex and smooth functions.
5
Theorem 5. For any online learning algorithm A, there always exists a sequence of strongly convex
and smooth functions f1 , . . . , fT , such that
T
X
t=1
ft (xt ) ? ft (x?t ) = ?(ST? )
where x1 , . . . , xT is the solutions generated by A.
Thus, the upper bound in Theorem 2 cannot be improved in general.
3.2
Semi-strongly Convex and Smooth Functions
During the analysis of Theorems 1 and 2, we realize that the proof is built upon the fact that ?when
the function is strongly convex and smooth, gradient descent can reduce the distance to the optimal
solution by a constant factor? [Mokhtari et al., 2016, Proposition 2]. From the recent developments
in convex optimization, we know that a similar behavior also happens when the function is semistrongly convex and smooth [Necoara et al., 2015, Theorem 5.2], which motivates the study in this
section.
We first introduce the definition of semi-strong convexity [Gong and Ye, 2014].
Definition 3. A function f : X 7? R is semi-strongly convex over X , if there exists a constant ? > 0
such that for any x ? X
?
2
(8)
f (x) ? min f (x) ? kx ? ?X ? (x)k
x?X
2
where X ? = {x ? X : f (x) ? minx?X f (x)} is the set of minimizers of f over X .
The semi-strong convexity generalizes several non-strongly convex conditions, such as the quadratic
approximation property and the error bound property [Wang and Lin, 2014, Necoara et al., 2015]. A
class of functions that satisfy the semi-strongly convexity is provided below [Gong and Ye, 2014].
Example 3. Consider the following constrained optimization problem
min
x?X ?Rd
f (x) = g(Ex) + b? x
where g(?) is strongly convex and smooth, and X is either Rd or a polyhedral set. Then, f : X 7? R
is semi-strongly convex over X with some constant ? > 0.
Based on the semi-strong convexity, we assume the functions satisfy the following conditions.
Assumption 2. Suppose the following conditions hold for each ft : X 7? R.
1. ft is semi-strongly convex over X with parameter ? > 0, and L-smooth;
2. k?ft (x)k ? G, ?x ? X .
When the function is semi-strongly convex, the optimal solution may not be unique. Thus, we need
to redefine PT? and ST? to account for this freedom. We define
PT? :=
T
X
t=2
T
2
X
?
?
max
?Xt? (x) ? ?Xt?1
max
?Xt? (x) ? ?Xt?1
(x)
(x)
, and ST? :=
x?X
t=2
x?X
where Xt? = {x ? X : ft (x) ? minx?X ft (x)} is the set of minimizers of ft over X .
In this case, we will use the standard online gradient descent when the learner can query the gradient
only once, and apply the online multiple gradient descent (OMGD) in Algorithm 1, when the learner
can access the gradient multiple times. Using similar analysis as Theorems 1 and 2, we obtain the
following dynamic regret bounds for functions that are semi-strongly convex and smooth.
Theorem 6. Suppose Assumption 2 is true. By setting ? ? 1/L in online gradient descent, we have
T
X
t=1
where ? =
q
1?
?
1/?+? ,
ft (xt ) ?
T
X
t=1
min ft (x) ?
x?X
? 1 = ?X1? (x1 ).
and x
6
?1k
Gkx1 ? x
GPT?
+
1??
1??
Thus, online gradient descent still achieves an O(PT? ) bound of the dynamic regret.
ln 4? in AlgoTheorem 7. Suppose Assumption 2 is true. By setting ? ? 1/L and K = ? 1/?+?
?
rithm 1, for any constant ? > 0, we have
?
?
?1k
T
T
?2GPT + 2Gkx1 ? x
X
X
ft (xt ) ?
min ft (x) ? min G?T
x?X
?
? 1 k2
+ 2(L + ?)ST? + (L + ?)kx1 ? x
t=1
t=1
2?
PT
? 1 = ?X1? (x1 ).
where G?T = max{x?t ?Xt? }Tt=1 t=1 k?ft (x?t )k2 , and x
Again, when the gradients of minimizers are small, in other words, G?T = O(ST? ), the proposed
OMGD improves the dynamic regret form O(PT? ) to O(min(PT? , ST? )).
3.3
Self-concordant Functions
We extend our previous results to self-concordant functions, which could be non-strongly convex
and even non-smooth. Self-concordant functions play an important role in interior-point methods
for solving convex optimization problems. We note that in the study of bandit linear optimization
[Abernethy et al., 2008b], self-concordant functions have been used as barriers for constraints. However, to the best of our knowledge, this is the first time that losses themselves are self-concordant.
The definition of self-concordant functions is given below [Nemirovski, 2004].
Definition 4. Let X be a nonempty open convex set in Rd and f be a C 3 convex function defined on
X . f is called self-concordant on X , if it possesses the following two properties:
1. f (xi ) ? ? along every sequence {xi ? X } converging, as i ? ?, to a boundary point
of X ;
2. f satisfies the differential inequality
3/2
|D3 f (x)[h, h, h]| ? 2 h? ?2 f (x)h
for all x ? X and all h ? Rd , where
?3
|t =t =t =0 f (x + t1 h1 + t2 h2 + t3 h3 ) .
?t1 ?t2 ?t3 1 2 3
Example 4. We provide some examples of self-concordant functions below
[Boyd and Vandenberghe, 2004, Nemirovski, 2004].
D3 f (x)[h1 , h2 , h3 ] =
1. The function f (x) = ? log x is self-concordant on (0, ?).
2. A convex quadratic form f (x) = x? Ax ? 2b? x + c where A ? Rd?d , b ? Rd , and c ? R,
is self-concordant on Rd .
3. If f : Rd 7? R is self-concordant, and A ? Rd?k , b ? Rd , then f (Ax + b) is selfconcordant.
Using the concept of self-concordance, we make the following assumptions.
Assumption 3. Suppose the following conditions hold for each ft : Xt 7? R.
1. ft is self-concordant on domain Xt ;
2. ft is non-degenerate on Xt , i.e., ?2 ft (x) ? 0, ?x ? Xt ;
3. ft attains its minimum on Xt , and denote x?t = argminx?Xt ft (x).
Our approach is similar to previous cases except for the updating rule of xt . Since we do not assume
functions are strongly convex, we need to take into account the second order structure when updating
the current solution xt . Thus, we assume the learner can query both the gradient and Hessian of each
function multiple times. Specifically, we apply the damped Newton method [Nemirovski, 2004] to
update xt , as follows:
i?1
h
1
j
2
)
?ft (zjt ), j = 1, . . . , K
?
f
(z
z1t = xt , zj+1
= zjt ?
t
t
t
j
1 + ?t (zt )
where
r
h
i?1
j
?t (zt ) = ?ft (zjt )? ?2 ft (zjt )
?ft (zjt ).
(9)
7
Algorithm 2 Online Multiple Newton Update (OMNU)
Require: The number of inner iterations K in each round
1: Let x1 be any point in X1
2: for t = 1, . . . , T do
3:
Submit xt ? X and the receive loss ft : X 7? R
4:
z1t = xt
5:
for j = 1, . . . , K do
6:
h
i?1
1
?ft (zjt )
?2 ft (zjt )
zj+1
= zjt ?
t
j
1 + ?t (zt )
where ?t (zjt ) is given in (9)
7:
end for
8:
xt+1 = zK+1
t
9: end for
Then, we set xt+1 = zK+1
. Since the damped Newton method needs to calculate the inverse of the
t
Hessian matrix, its complexity is higher than gradient descent. The procedure is named as Online
Multiple Newton Update (OMNU) and is summarized in Algorithm 2.
To analyze the dynamic regret of OMNU, we redefine the two regularities PT? and ST? as follows:
PT? :=
T
X
kx?t ? x?t?1 kt =
T q
X
ST? :=
T
X
kx?t ? x?t?1 k2t =
T
X
(x?t ? x?t?1 )? ?2 ft (x?t )(x?t ? x?t?1 )
p
t=2
t=2
t=2
(x?t ? x?t?1 )? ?2 ft (x?t )(x?t ? x?t?1 )
t=2
Compared to the definitions in (3) and (4), we introduce ?2 ft (x?t )
where khkt =
when measuring the distance between x?t and x?t?1 . When functions are strongly convex and smooth,
these definitions are equivalent up to constant factors. We then define a quantity to compare the
second order structure of consecutive functions:
n
?1/2 2
?1/2 o
(10)
? ft (x?t ) ?2 ft?1 (x?t?1 )
? = max ?max ?2 ft?1 (x?t?1 )
h? ?2 f
?
t (xt )h.
t=2,...,T
where ?max (?) computes the maximum eigenvalue of its argument. When all the functions are ?strongly convex and L-smooth, ? ? L/?. Then, we have the following theorem regarding the
dynamic regret of the proposed OMNU algorithm.
Theorem 8. Suppose Assumption 3 is true, and further assume
1
kx?t?1 ? x?t k2t ?
, ?t ? 2.
(11)
144
When t = 1, we choose K = O(1)(f1 (x1 ) ? f1 (x?1 ) + log log ?) in OMNU such that
kx2 ? x?1 k21 ?
1
.
144?
(12)
For t ? 2, we set K = ?log4 (16?)? in OMNU, then
T
X
1
1 ?
?
?
PT , 4ST + f1 (x1 ) ? f1 (x?1 ) + .
ft (xt ) ? ft (xt ) ? min
3
36
t=1
The above theorem again implies the dynamic regret can be upper bounded by O(min(PT? , ST? ))
when the learner can access the gradient and Hessian multiple times. From the first property of
self-concordant functions in Definition 4, we know that x?t must lie in the interior of Xt , and thus
?ft (x?t ) = 0 for all t ? [T ]. As a result, we do not need the additional assumption that the gradients
of minimizers are small, which has been used before to simplify Theorems 2 and 7.
Compared to Theorems 2 and 7, Theorem 8 introduces an additional condition in (11). This condition is required to ensure that xt lies in the feasible region of ft (?), otherwise, ft (xt ) can be infinity
8
and it is impossible to bound the dynamic regret. The multiple applications of damped Newton method can enforce xt to be close to x?t?1 . Combined with (11), we conclude that xt is also close to x?t .
Then, based on the property of the Dikin ellipsoid of self-concordant functions [Nemirovski, 2004],
we can guarantee that xt is feasible for ft (?).
4
Conclusion and Future Work
In this paper, we discuss how to reduce the dynamic regret of online learning by allowing the learner
to query the gradient/Hessian of each function multiple times. By applying gradient descent multiple
times in each round, we show that the dynamic regret can be upper bounded by the minimum of the
path-length and the squared path-length, when functions are strongly convex and smooth. We then
extend this theoretical guarantee to functions that are semi-strongly convex and smooth. We finally
demonstrate that for self-concordant functions, applying the damped Newton method multiple times
achieves a similar result.
In the current study, we upper bound the dynamic regret in terms of the path-length or the squared
path-length of the comparator sequence. As we mentioned before, there also exist some regularities
defined in terms of the function sequence, e.g., the functional variation [Besbes et al., 2015]. In the
future, we will investigate whether multiple accesses of gradient/Hessian can improve the dynamic
regret when measured by certain regularities of the function sequence. Another future work is to
extend our results to the more general dynamic regret
R(u1 , . . . , uT ) =
T
X
t=1
ft (xt ) ?
T
X
ft (ut )
t=1
where u1 , . . . , uT ? X is an arbitrary sequence of comparators [Zinkevich, 2003].
Acknowledgments
This work was partially supported by the NSFC (61603177, 61333014), JiangsuSF (BK20160658),
NSF (IIS-1545995), and the Collaborative Innovation Center of Novel Software Technology and
Industrialization. Jinfeng Yi is now at Tencent AI Lab, Bellevue, WA, USA.
References
J. Abernethy, P. L. Bartlett, A. Rakhlin, and A. Tewari. Optimal stragies and minimax lower bounds
for online convex games. In Proceedings of the 21st Annual Conference on Learning Theory,
2008a.
J. Abernethy, E. Hazan, and A. Rakhlin. Competing in the dark: An efficient algorithm for bandit
linear optimization. In Proceedings of the 21st Annual Conference on Learning, pages 263?274,
2008b.
O. Besbes, Y. Gur, and A. Zeevi. Non-stationary stochastic optimization. Operations Research, 63
(5):1227?1244, 2015.
S. Boyd and L. Vandenberghe. Convex Optimization. Cambridge University Press, 2004.
N. Buchbinder, S. Chen, J. S. Naor, and O. Shamir. Unified algorithms for online learning and
competitive analysis. In Proceedings of the 25th Annual Conference on Learning Theory, 2012.
N. Cesa-bianchi, P. Gaillard, G. Lugosi, and G. Stoltz. Mirror descent meets fixed share (and feels
no regret). In Advances in Neural Information Processing Systems 25, pages 980?988, 2012.
C.-K. Chiang, T. Yang, C.-J. Lee, M. Mahdavi, C.-J. Lu, R. Jin, and S. Zhu. Online optimization
with gradual variations. In Proceedings of the 25th Annual Conference on Learning Theory, 2012.
A. Daniely, A. Gonen, and S. Shalev-Shwartz. Strongly adaptive online learning. In Proceedings of
The 32nd International Conference on Machine Learning, 2015.
P. Gong and J. Ye. Linear convergence of variance-reduced stochastic gradient without strong convexity. ArXiv e-prints, arXiv:1406.1102, 2014.
9
E. C. Hall and R. M. Willett. Dynamical models and tracking regret in online convex programming.
In Proceedings of the 30th International Conference on Machine Learning, pages 579?587, 2013.
E. Hazan and S. Kale. Beyond the regret minimization barrier: an optimal algorithm for stochastic
strongly-convex optimization. In Proceedings of the 24th Annual Conference on Learning Theory,
pages 421?436, 2011.
E. Hazan and C. Seshadhri. Adaptive algorithms for online decision problems. Electronic Colloquium on Computational Complexity, 88, 2007.
E. Hazan, A. Agarwal, and S. Kale. Logarithmic regret algorithms for online convex optimization.
Machine Learning, 69(2-3):169?192, 2007.
M. Herbster and M. K. Warmuth. Tracking the best expert. Machine Learning, 32(2):151?178,
1998.
A. Jadbabaie, A. Rakhlin, S. Shahrampour, and K. Sridharan. Online optimization: Competing
with dynamic comparators. In Proceedings of the 18th International Conference on Artificial
Intelligence and Statistics, 2015.
A. Mokhtari, S. Shahrampour, A. Jadbabaie, and A. Ribeiro. Online optimization in dynamic environments: Improved regret rates for strongly convex problems. ArXiv e-prints, arXiv:1603.04954,
2016.
I. Necoara, Y. Nesterov, and F. Glineur. Linear convergence of first order methods for non-strongly
convex optimization. ArXiv e-prints, arXiv:1504.06298, 2015.
A. Nemirovski. Interior point polynomial time methods in convex programming. Lecture notes,
Technion ? Israel Institute of Technology, 2004.
Y. Nesterov. Introductory lectures on convex optimization: a basic course, volume 87 of Applied
optimization. Kluwer Academic Publishers, 2004.
S. Rakhlin and K. Sridharan. Optimization, learning, and games with predictable sequences. In
Advances in Neural Information Processing Systems 26, pages 3066?3074, 2013.
S. Shalev-Shwartz. Online learning and online convex optimization. Foundations and Trends in
Machine Learning, 4(2):107?194, 2011.
S. Shalev-Shwartz, Y. Singer, and N. Srebro. Pegasos: primal estimated sub-gradient solver for
SVM. In Proceedings of the 24th International Conference on Machine Learning, pages 807?
814, 2007.
P.-W. Wang and C.-J. Lin. Iteration complexity of feasible descent methods for convex optimization.
Journal of Machine Learning Research, 15:1523?1548, 2014.
T. Yang, L. Zhang, R. Jin, and J. Yi. Tracking slowly moving clairvoyant: Optimal dynamic regret of
online learning with true and noisy gradient. In Proceedings of the 33rd International Conference
on Machine Learning, 2016.
M. Zinkevich. Online convex programming and generalized infinitesimal gradient ascent. In Proceedings of the 20th International Conference on Machine Learning, pages 928?936, 2003.
10
| 6675 |@word polynomial:1 achievable:1 norm:1 nd:1 open:2 gradual:1 bellevue:1 incurs:1 past:1 current:3 com:2 z2:1 dikin:1 must:1 realize:1 update:4 stationary:2 intelligence:1 warmuth:2 chiang:2 successive:2 zhang:2 height:1 along:1 differential:1 clairvoyant:1 prove:2 naor:1 introductory:1 redefine:2 polyhedral:1 introduce:4 indeed:1 expected:1 behavior:1 themselves:1 surge:1 growing:1 examine:1 zhouzh:1 zhi:1 equipped:1 solver:1 provided:2 bounded:8 underlying:1 israel:1 developed:1 unified:1 hindsight:1 guarantee:4 every:1 concave:1 seshadhri:2 exactly:1 k2:10 t1:2 nju:2 before:2 local:4 consequence:1 analyzing:1 nsfc:1 meet:1 path:16 fluctuation:2 lugosi:1 china:1 weakened:1 studied:1 k2t:2 nemirovski:5 bi:1 averaged:1 unique:1 acknowledgment:1 regret:60 procedure:2 significantly:2 matching:1 projection:1 word:1 boyd:2 nanjing:2 cannot:2 uiowa:1 close:3 interior:7 onto:1 pegasos:1 impossible:2 applying:3 lijun:1 zinkevich:7 equivalent:1 center:2 tianbao:2 kale:3 convex:59 rule:1 vandenberghe:2 gur:1 notion:2 variation:7 feel:1 pt:36 suppose:11 play:1 shamir:1 programming:3 us:1 trend:1 utilized:1 updating:2 predicts:1 ft:91 role:1 wang:2 capture:1 worst:1 calculate:2 region:1 mentioned:2 colloquium:1 convexity:9 complexity:3 environment:1 predictable:1 nesterov:4 dynamic:49 solving:2 tight:1 upon:1 learner:21 lst:1 various:1 query:7 artificial:1 shalev:6 abernethy:5 larger:1 otherwise:1 statistic:1 noisy:1 online:36 sequence:23 advantage:1 eigenvalue:1 propose:3 coming:1 degenerate:3 achieve:1 adapts:1 kx1:2 ky:2 seattle:1 exploiting:1 regularity:13 convergence:2 help:1 illustrate:2 develop:1 depending:1 gong:3 measured:1 nearest:1 h3:2 strong:8 implies:3 closely:2 stochastic:4 stringent:1 require:2 f1:7 proposition:1 tighter:1 rong:1 extension:1 hold:3 hall:3 exp:1 zeevi:1 achieves:3 consecutive:1 xk2:2 label:2 gpt:6 gaillard:1 city:1 tool:1 minimization:1 always:1 aim:2 lamda:2 modified:1 zhou:1 avoid:1 corollary:3 ax:3 attains:1 sense:1 minimizers:9 bandit:2 selects:1 aforementioned:1 denoted:1 development:1 constrained:1 once:3 beach:1 sampling:1 comparators:4 promote:1 future:4 minimized:1 t2:2 simplify:1 national:1 argminx:2 freedom:1 interest:2 investigate:2 introduces:2 primal:1 damped:5 necoara:3 kt:1 beforehand:1 stoltz:1 euclidean:1 theoretical:2 instance:2 measuring:2 daniely:2 technion:1 answer:1 mokhtari:7 chooses:1 combined:1 st:30 fundamental:1 herbster:2 international:6 accessible:1 kut:1 lee:1 jadbabaie:4 squared:6 reflect:1 cesa:2 again:2 choose:2 slowly:2 expert:3 concordance:2 account:3 mahdavi:1 summarized:2 inc:1 satisfy:2 caused:1 h1:2 lab:2 optimistic:1 hazan:8 analyze:1 competitive:2 start:1 collaborative:1 minimize:1 square:1 variance:1 who:1 yield:1 t3:2 lu:1 definition:12 infinitesimal:1 against:3 proof:1 static:5 sampled:1 proved:1 popular:1 ask:1 recall:1 knowledge:3 ut:12 improves:2 higher:1 improved:9 though:1 strongly:39 jiangsusf:1 furthermore:1 just:1 receives:1 incrementally:1 logistic:1 indicated:1 usa:5 ye:3 concept:1 y2:1 true:7 laboratory:1 round:9 game:4 self:21 during:1 yorktown:1 generalized:1 zhanglj:1 tt:1 demonstrate:3 performs:1 novel:3 recently:1 functional:2 exponentially:1 volume:1 extend:4 belong:2 kluwer:1 willett:3 interpret:1 significant:1 cambridge:1 ai:3 queried:1 smoothness:1 rd:11 moving:1 access:6 gt:3 recent:2 belongs:1 lgt:1 scenario:1 buchbinder:2 certain:2 inequality:3 watson:1 kx2:1 yi:4 minimum:3 additional:2 semi:16 ii:1 multiple:27 smooth:22 academic:1 long:1 lin:2 prediction:2 converging:1 basic:1 metric:1 arxiv:6 iteration:3 adopting:1 agarwal:1 receive:2 interval:1 publisher:1 posse:1 ascent:1 sridharan:3 yang:7 revealed:1 besbes:5 intermediate:1 variety:1 zi:1 competing:2 reduce:4 inner:2 cn:2 regarding:1 shift:1 whether:3 bartlett:1 hessian:6 proceed:1 tewari:1 dark:1 extensively:1 industrialization:1 reduced:2 generate:1 exist:1 zj:4 nsf:1 estimated:1 group:1 key:1 d3:2 compete:1 inverse:1 named:4 electronic:1 decision:1 comparable:1 bound:22 quadratic:3 annual:5 constraint:1 infinity:1 software:2 u1:10 aspect:2 argument:1 selfconcordant:2 min:18 performing:1 department:1 according:2 jr:1 smaller:5 slightly:1 appealing:1 happens:1 ln:3 remains:1 discus:3 nonempty:1 singer:1 know:2 end:5 available:1 generalizes:1 operation:1 apply:5 appropriate:1 enforce:1 batch:1 thomas:1 denotes:1 ensure:1 newton:7 lkx1:2 establish:1 utp:1 question:3 quantity:1 print:3 rt:3 gradient:51 minx:2 distance:2 length:16 mini:1 ellipsoid:1 innovation:1 unfortunately:1 glineur:1 stated:1 motivates:1 policy:1 unknown:1 zt:3 allowing:3 upper:17 bianchi:2 jin:3 descent:27 immediate:1 y1:1 arbitrary:2 drift:1 introduced:1 pair:2 required:1 z1:1 alibaba:2 established:2 nip:1 able:1 adversary:2 beyond:1 dynamical:2 below:5 gonen:1 jinfeng:2 including:1 max:8 built:1 shifting:1 natural:2 regularized:1 advanced:1 zhu:1 minimax:3 improve:4 technology:3 extract:1 prior:1 literature:1 relative:3 loss:11 lecture:2 interesting:1 limitation:1 srebro:1 facing:1 foundation:2 iowa:2 h2:2 share:1 ibm:1 course:1 supported:1 last:1 offline:1 institute:1 wide:1 barrier:2 overcome:1 boundary:1 cumulative:5 rich:1 computes:1 adaptive:4 simplified:1 ribeiro:1 tighten:1 conclude:1 xi:4 shwartz:6 decade:1 gkx1:6 zk:5 ca:1 tencent:2 investigated:1 meanwhile:1 domain:1 submit:2 repeated:1 allowed:1 x1:12 advice:1 referred:1 rithm:1 ny:1 sub:3 fails:1 lie:3 kxk2:1 theorem:21 xt:55 k21:1 rakhlin:5 svm:1 exists:2 mirror:3 kx:9 chen:1 shahrampour:2 zjt:13 logarithmic:1 likely:1 contained:1 tracking:4 partially:1 hua:1 restarted:1 satisfies:1 relies:1 comparator:10 z1t:5 formulated:1 feasible:3 change:1 specifically:4 except:1 called:1 concordant:18 log4:1 accelerated:1 evaluate:1 ex:1 |
6,274 | 6,676 | Learning Efficient Object Detection Models with
Knowledge Distillation
Guobin Chen1,2 Wongun Choi1 Xiang Yu1 Tony Han2 Manmohan Chandraker1,3
1
2
3
NEC Labs America
University of Missouri
University of California, San Diego
Abstract
Despite significant accuracy improvement in convolutional neural networks (CNN)
based object detectors, they often require prohibitive runtimes to process an image
for real-time applications. State-of-the-art models often use very deep networks
with a large number of floating point operations. Efforts such as model compression
learn compact models with fewer number of parameters, but with much reduced
accuracy. In this work, we propose a new framework to learn compact and fast object detection networks with improved accuracy using knowledge distillation [20]
and hint learning [34]. Although knowledge distillation has demonstrated excellent
improvements for simpler classification setups, the complexity of detection poses
new challenges in the form of regression, region proposals and less voluminous labels. We address this through several innovations such as a weighted cross-entropy
loss to address class imbalance, a teacher bounded loss to handle the regression
component and adaptation layers to better learn from intermediate teacher distributions. We conduct comprehensive empirical evaluation with different distillation
configurations over multiple datasets including PASCAL, KITTI, ILSVRC and
MS-COCO. Our results show consistent improvement in accuracy-speed trade-offs
for modern multi-class detection models.
1
Introduction
Recent years have seen tremendous increase in the accuracy of object detection, relying on deep
convolutional neural networks (CNNs). This has made visual object detection an attractive possibility
for domains ranging from surveillance to autonomous driving. However, speed is a key requirement in
many applications, which fundamentally contends with demands on accuracy. Thus, while advances
in object detection have relied on increasingly deeper architectures, they are associated with an
increase in computational expense at runtime. But it is also known that deep neural networks are
over-parameterized to aid generalization. Thus, to achieve faster speeds, some prior works explore
new structures such as fully convolutional networks, or lightweight models with fewer channels and
small filters [22, 25]. While impressive speedups are obtained, they are still far from real-time, with
careful redesign and tuning necessary for further improvements.
Deeper networks tend to have better performance under proper training, since they have ample
network capacity. Tasks such as object detection for a few categories might not necessarily need
that model capacity. In that direction, several works in image classification use model compression,
whereby weights in each layer are decomposed, followed by layer-wise reconstruction or fine-tuning
to recover some of the accuracy [9, 26, 41, 42]. This results in significant speed-ups, but there is
often a gap between the accuracies of original and compressed models, which is especially large
when using compressed models for more complex problems such as object detection. On the other
hand, seminal works on knowledge distillation show that a shallow or compressed model trained
to mimic the behavior of a deeper or more complex model can recover some or all of the accuracy
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
drop [3, 20, 34]. However, those results are shown only for problems such as classification, using
simpler networks without strong regularization such as dropout.
Applying distillation techniques to multi-class object detection, in contrast to image classification, is
challenging for several reasons. First, the performance of detection models suffers more degradation
with compression, since detection labels are more expensive and thereby, usually less voluminous.
Second, knowledge distillation is proposed for classification assuming each class is equally important,
whereas that is not the case for detection where the background class is far more prevalent. Third,
detection is a more complex task that combines elements of both classification and bounding box
regression. Finally, an added challenge is that we focus on transferring knowledge within the same
domain (images of the same dataset) with no additional data or labels, as opposed other works that
might rely on data from other domains (such as high-quality and low-quality image domains, or
image and depth domains).
To address the above challenges, we propose a method to train fast models for object detection with
knowledge distillation. Our contributions are four-fold:
? We propose an end-to-end trainable framework for learning compact multi-class object detection
models through knowledge distillation (Section 3.1). To the best of our knowledge, this is the first
successful demonstration of knowledge distillation for the multi-class object detection problem.
? We propose new losses that effectively address the aforementioned challenges. In particular, we
propose a weighted cross entropy loss for classification that accounts for the imbalance in the
impact of misclassification for background class as opposed to object classes (Section 3.2), a
teacher bounded regression loss for knowledge distillation (Section 3.3) and adaptation layers for
hint learning that allows the student to better learn from the distribution of neurons in intermediate
layers of the teacher (Section 3.4).
? We perform comprehensive empirical evaluation using multiple large-scale public benchmarks.
Our study demonstrates the positive impact of each of the above novel design choices, resulting in
significant improvement in object detection accuracy using compressed fast networks, consistently
across all benchmarks (Sections 4.1 ? 4.3).
? We present insights into the behavior of our framework by relating it to the generalization and
under-fitting problems (Section 4.4).
2
Related Works
CNNs for Detection. Deformable Part Model (DPM) [14] was the dominant detection framework
before the widespread use of Convolutional Neural Networks (CNNs). Following the success of
CNNs in image classification [27], Girshick et al. proposed RCNN [24] that uses CNN features
to replace handcrafted ones. Subsequently, many CNN based object detection methods have been
proposed, such as Spatial Pyramid Pooling (SPP) [19], Fast R-CNN [13], Faster-RCNN [32] and
R-FCN [29], that unify various steps in object detection into an end-to-end multi-category framework.
Model Compression. CNNs are expensive in terms of computation and memory. Very deep networks
with many convolutional layers are preferred for accuracy, while shallower networks are also widely
used where efficiency is important. Model compression in deep networks is a viable approach to
speed up runtime while preserving accuracy. Denil et al. [9] demonstrate that neural networks are
often over-parametrized and removing redundancy is possible. Subsequently, various methods [5, 7,
10, 15, 17, 30] have been proposed to accelerate the fully connected layer. Several methods based on
low-rank decomposition of the convolutional kernel tensor [10, 23, 28] are also proposed to speed
up convolutional layers. To compress the whole network, Zhang et al. [41, 42] present an algorithm
using asymmetric decomposition and additional fine-tuning. In similar spirit, Kim et al. [26] propose
one-shot whole network compression that achieves around 1.8 times improvement in runtime without
significant drop in accuracy. We will use methods presented in [26] in our experiments. Besides, a
pruning based approach has been proposed [18] but it is challenging to achieve runtime speed-up
with a conventional GPU implementation. Additionally, both weights and input activations can be the
quantized( [18]) and binarized ( [21, 31]) to lower the computationally expensive.
Knowledge Distillation. Knowledge distillation is another approach to retain accuracy with model
compression. Bucila et al. [3] propose an algorithm to train a single neural network by mimicking
the output of an ensemble of models. Ba and Caruana [2] adopt the idea of [3] to compress deep
2
Teacher
Detection
Hint
Regression
Ground Truth
Label
Weighted
Cross Entropy
Loss
Bounded
Regression
Loss
SoftMax &
SmoothL1
Loss
Classification
Regression
Classification
L2 Loss
Back Propagation
Student
Adaptation
Hint
Soft
Label
Distillation
Detection
Guided
Ground
Truth
Figure 1: The proposed learning skeme on visual object detection task using Faster-RCNN, which mainly
consists of region proposal network (RPN) and region classification network(RCN). The two networks both
use multi-task loss to jointly learn the classifier and bounding-box regressor. We employ the final output of the
teacher?s RPN and RCN as the distillation targets, and apply the intermediate layer outputs as hint. Red arrows
indicate the backpropagation pathways.
networks into shallower but wider ones, where the compressed model mimics the ?logits?. Hinton
et al. [20] propose knowledge distillation as a more general case of [3], which applies the prediction
of the teacher model as a ?soft label?, further proposing temperature cross entropy loss instead of L2
loss. Romero et al. [34] introduce a two-stage strategy to train deep networks. In their method, the
teacher?s middle layer provides ?hint? to guide the training of the student model.
Other researchers [16, 38] explore distillation for transferring knowledge between different domains,
such as high-quality and low-quality images, or RGB and depth images. In a draft manuscript
concurrent with our work, Shen et al. [36] consider the effect of distillation and hint frameworks
in learning a compact object detection model. However, they formulate the detection problem as
a binary classification task applied to pedestrians, which might not scale well to the more general
multi-category object detection setup. Unlike theirs, our method is designed for multi-category object
detection. Further, while they use external region proposals, we demonstrate distillation and hint
learning for both the region proposal and classification components of a modern end-to-end object
detection framework [32].
3
Method
In this work, we adopt the Faster-RCNN [32] as the object detection framework. Faster-RCNN
is composed of three modules: 1) A shared feature extraction through convolutional layers, 2) a
region proposal network (RPN) that generates object proposals, and 3) a classification and regression
network (RCN) that returns the detection score as well as a spatial adjustment vector for each object
proposal. Both the RCN and RPN use the output of 1) as features, RCN also takes the result of RPN
as input. In order to achieve highly accurate object detection results, it is critical to learn strong
models for all the three components.
3.1
Overall Structure
We learn strong but efficient student object detectors by using the knowledge of a high capacity teacher
detection network for all the three components. Our overall learning framework is illustrated in Figure
1. First, we adopt the hint based learning [34] (Sec.3.4) that encourages the feature representation of
a student network is similar to that of the teacher network. Second, we learn stronger classification
modules in both RPN and RCN using the knowledge distillation framework [3, 20]. In order to handle
severe category imbalance issue in object detection, we apply weighted cross entropy loss for the
distillation framework. Finally, we transfer the teacher?s regression output as a form of upper bound,
that is, if the student?s regression output is better than that of teacher, no additional loss is applied.
3
Our overall learning objective can be written as follows:
1 X RCN
1 X RCN
Lcls + ?
L
LRCN =
N i
N j reg
LRPN =
1 X RPN
1 X RPN
Lcls + ?
L
M i
M j reg
L = LRPN + LRCN + ?LHint
(1)
where N is the batch-size for RCN and M for RPN. Here, Lcls denotes the classifier loss function
that combines the hard softmax loss using the ground truth labels and the soft knowledge distillation
loss [20] of (2). Further, Lreg is the bounding box regression loss that combines smoothed L1 loss [13]
and our newly proposed teacher bounded L2 regression loss of (4). Finally, Lhint denotes the hint
based loss function that encourages the student to mimic the teacher?s feature response, expressed as
(6). In the above, ? and ? are hyper-parameters to control the balance between different losses. We
fix them to be 1 and 0.5, respectively, throughout the experiments.
3.2
Knowledge Distillation for Classification with Imbalanced Classes
Conventional use of knowledge distillation has been proposed for training classification networks,
where predictions of a teacher network are used to guide the training of a student model. Suppose we
have dataset {xi , yi }, i = 1, 2, ..., n where xi ? I is the input image and yi ? Y is its class label.
Let t be the teacher model, with Pt = softmax( ZTt ) its prediction and Zt the final score output. Here,
T is a temperature parameter (normally set to 1). Similarly, one can define Ps = softmax( ZTs ) for the
student network s. The student s is trained to optimize the following loss function:
Lcls = ?Lhard (Ps , y) + (1 ? ?)Lsof t (Ps , Pt )
(2)
where Lhard is the hard loss using ground truth labels used by Faster-RCNN, Lsof t is the soft loss
using teacher?s prediction and ? is the parameter to balance the hard and soft losses. It is known that
a deep teacher can better fit to the training data and perform better in test scenarios. The soft labels
contain information about the relationship between different classes as discovered by teacher. By
learning from soft labels, the student network inherits such hidden information.
In [20], both hard and soft losses are the cross entropy losses. But unlike simpler classification
problems, the detection problem needs to deal with a severe imbalance across different categories, that
is, the background dominates. In image classification, the only possible errors are misclassifications
between ?foreground? categories. In object detection, however, failing to discriminate between
background and foreground can dominate the error, while the frequency of having misclassification
between foreground categories is relatively rare. To address this, we adopt class-weighted cross
entropy as the distillation loss:
X
Lsof t (Ps , Pt ) = ?
wc Pt log Ps
(3)
where we use a larger weight for the background class and a relatively small weight for other classes.
For example, we use w0 = 1.5 for the background class and wi = 1 for all the others in experiments
on the PASCAL dataset.
When Pt is very similar to the hard label, with probability for one class very close to 1 and most
others very close to 0, the temperature parameter T is introduced to soften the output. Using higher
temperature will force t to produce softer labels so that the classes with near-zero probabilities will
not be ignored by the cost function. This is especially pertinent to simpler tasks, such as classification
on small datasets like MNIST. But for harder problems where the prediction error is already high, a
larger value of T introduces more noise which is detrimental to learning. Thus, lower values of T are
used in [20] for classification on larger datasets. For even harder problems such as object detection,
we find using no temperature parameter at all (equivalent to T = 1) in the distillation loss works the
best in practice (see supplementary material for an empirical study).
3.3
Knowledge Distillation for Regression with Teacher Bounds
In addition to the classification layer, most modern CNN based object detectors [26, 29, 32, 33] also
use bounding-box regression to adjust the location and size of the input proposals. Often, learning a
4
good regression model is critical to ensure good object detection accuracy [13]. Unlike distillation
for discrete categories, the teacher?s regression outputs can provide very wrong guidance toward the
student model, since the real valued regression outputs are unbounded. In addition, the teacher may
provide regression direction that is contradictory to the ground truth direction. Thus, instead of using
the teacher?s regression output directly as a target, we exploit it as an upper bound for the student to
achieve. The student?s regression vector should be as close to the ground truth label as possible in
general, but once the quality of the student surpasses that of the teacher with a certain margin, we
do not provide additional loss for the student. We call this the teacher bounded regression loss, Lb ,
which is used to formulate the regression loss, Lreg , as follows:
2
2
2
kRs ? yk2 , if kRs ? yk2 + m > kRt ? yk2
Lb (Rs , Rt , y) =
0,
otherwise
Lreg = LsL1 (Rs , yreg ) + ?Lb (Rs , Rt , yreg ),
(4)
where m is a margin, yreg denotes the regression ground truth label, Rs is the regression output of
the student network, Rt is the prediction of teacher network and ? is a weight parameter (set as 0.5 in
our experiments). Here, LsL1 is the smooth L1 loss as in [13]. The teacher bounded regression loss
Lb only penalizes the network when the error of the student is larger than that of the teacher. Note
that although we use L2 loss inside Lb , any other regression loss such as L1 and smoothed L1 can be
combined with Lb . Our combined loss encourages the student to be close to or better than teacher in
terms of regression, but does not push the student too much once it reaches the teacher?s performance.
3.4
Hint Learning with Feature Adaptation
Distillation transfers knowledge using only the final output. In [34], Romero et al. demonstrate that
using the intermediate representation of the teacher as hint can help the training process and improve
the final performance of the student. They use the L2 distance between feature vectors V and Z:
2
LHint (V, Z) = kV ? Zk2
(5)
where Z represent the intermediate layer we selected as hint in the teacher network and V represent
the output of the guided layer in the student network. We also evaluate the L1 loss:
LHint (V, Z) = kV ? Zk1
(6)
While applying hint learning, it is required that the number of neurons (channels, width and height)
should be the same between corresponding layers in the teacher and student. In order to match the
number of channels in the hint and guided layers, we add an adaptation after the guided layer whose
output size is the same as the hint layer. The adaptation layer matches the scale of neuron to make
the norm of feature in student close to teacher?s. A fully connected layer is used as adaptation layer
when both hint and guided layers are also fully connected layers. When the hint and guided layers are
convolutional layers, we use 1 ? 1 convolutions to save memory. Interestingly, we find that having
an adaptation layer is important to achieve effective knowledge transferring even when the number
of channels in the hint and guided layers are the same (see Sec. 4.3). The adaptation layer can also
match the difference when the norms of features in hint and guided layers are different. When the
hint or guided layer is convolutional and the resolution of hint and guided layers differs (for examples,
VGG16 and AlexNet), we follow the padding trick introduced in [16] to match the number of outputs.
4
Experiments
In this section, we first introduce teacher and student CNN models and datasets that are used in the
experiments. The overall results on various datasets are shown in Sec.4.1. We apply our methods to
smaller networks and lower quality inputs in Sec.4.2. Sec.4.3 describes ablation studies for three different components, namely classification/regression, distillation and hint learning. Insights obtained
for distillation and hint learning are discussed in Sec.4.4. We refer the readers to supplementary
material for further details.
Datasets We evaluate our method on several commonly used public detection datasets, namely,
KITTI [12], PASCAL VOC 2007 [11], MS COCO [6] and ImageNet DET benchmark (ILSVRC
2014) [35]. Among them, KITTI and PASCAL are relatively small datasets that contain less object
5
Student
Model Info
Teacher PASCAL [email protected] COCO@[.5,.95]
KITTI
ILSVRC
54.7
25.4
11.8
49.3
20.6
AlexNet 57.6 (+2.9) 26.5 (+1.2)
12.3 (+0.5)
51.4 (+2.1) 23.6 (+1.3)
Tucker
11M / 47ms
VGGM 58.2 (+3.5) 26.4 (+1.1)
12.2 (+0.4)
51.4 (+2.1) 23.9 (+1.6)
VGG16 59.4 (+4.7) 28.3 (+2.9)
12.6 (+0.8)
53.7 (+4.4) 24.4 (+2.1)
57.2
32.5
15.8
55.1
27.3
AlexNet 62M / 74ms
VGGM 59.2 (+2.0) 33.4 (+0.9)
16.0 (+0.2)
56.3 (+1.2) 28.7 (+1.4)
VGG16 60.1 (+2.9) 35.8 (+3.3)
16.9 (+1.1)
58.3 (+3.2) 30.1 (+2.8)
59.8
33.6
16.1
56.7
31.1
VGGM
80M / 86ms
VGG16 63.7 (+3.9) 37.2 (+3.6)
17.3 (+1.2)
58.6 (+2.3) 34.0 (+2.9)
VGG16 138M / 283ms
70.4
45.1
24.2
59.2
35.6
Table 1: Comparison of student models associated with different teacher models across four datasets, in terms
of mean Average Precision (mAP). Rows with blank (-) teacher indicate the model is without distillation, serving
as baselines. The second column reports the number of parameters and speed (per image, on GPU).
categories and labeled images, whereas MS COCO and ILSVRC 2014 are large scale datasets.
Since KITTI and ILSVRC 2014 do not provide ground-truth annotation for test sets, we use the
training/validation split introduced by [39] and [24] for analysis. For all the datasets, we follow the
PASCAL VOC convention to evaluate various models by reporting mean average precision (mAP) at
IoU = 0.5 . For MS COCO dataset, besides the PASCAL VOC metric, we also report its own metric,
which evaluates mAP averaged for IoU ? [0.5 : 0.05 : 0.95] (denoted as mAP[.5, .95]).
Models The teacher and student models defined in our experiments are standard CNN architectures,
which consist of regular convolutional layers, fully connected layers, ReLU, dropout layers and
softmax layers. We choose several popular CNN architectures as our teacher/student models, namely,
AlexNet [27], AlexNet with Tucker Decomposition [26], VGG16 [37] and VGGM [4]. We use two
different settings for the student and teacher pairs. In the first set of experiments, we use a smaller
network (that is, less parameters) as the student and use a larger one for the teacher (for example,
AlexNet as student and VGG16 as teacher). In the second set of experiments, we use smaller input
image size for the student model and larger input image size for the teacher, while keeping the
network architecture the same.
4.1
Overall Performance
Table1 shows mAP for four student models on four object detection databases, with different
architectures for teacher guidance. For student models without teacher?s supervision, we train them to
the best numbers we could achieve. Not surprisingly, larger or deeper models with more parameters
perform better than smaller or shallower models, while smaller models run faster than larger ones.
The performance of student models improves significantly with distillation and hint learning over all
different pairs and datasets, despite architectural differences between teacher and student. With a
fixed scale (number of parameters) of a student model, training from scratch or fine-tuning on its own
is not an optimal choice. Getting aid from a better teacher yields larger improvements approaching
the teacher?s performance. A deeper model as teacher leads to better student performance, which
suggests that the knowledge transferred from better teachers is more informative. Notice that the
Tucker model trained with VGG16 achieves significantly higher accuracy than the Alexnet in the
PASCAL dataset, even though the model size is about 5 times smaller. The observation may support
the hypothesis that CNN based object detectors are highly over-parameterized. On the contrary,
when the size of dataset is much larger, it becomes much harder to outperform more complicated
models. This suggests that it is worth having even higher capacity models for such large scale datasets.
Typically, when evaluating efficiency, we get 3 times faster from VGG16 as teacher to AlexNet as
student on KITTI dataset. For more detailed runtimes, please refer to supplementary material.
Further, similar to [38], we investigate another student-teacher mode: the student and teacher share
exactly the same network structure, while the input for student is down-scaled and the input for
teacher remains high resolution. Recent works [1] report that image resolution critically affects
object detection performance. On the other hand, downsampling the input size quadratically reduces
convolutional resources and speeds up computation. In Table 2, by scaling input sizes to half in
6
High-res teacher
Low-res baseline
Low-res distilled student
mAP
Speed
mAP
Speed
mAP
Speed
AlexNet
57.2
1,205 / 74 ms
53.2
726 / 47 ms 56.7(+3.5) 726 / 47 ms
Tucker
54.7
663 / 41 ms
48.6
430 / 29 ms 53.5(+4.9) 430 / 29 ms
Table 2: Comparison of high-resolution teacher model (trained on images with 688 pixels) and low-resolution
student model (trained on 344 pixels input), on PASCAL. We report mAP and speed (both CPU and GPU)
of different models. The speed of low-resolution models are about 2 times faster than the corresponding
high-resolution models, while achieving almost the same accuracy when our distillation method is used.
FLOPS(%)
20
25
30
37.5
45
Finetune
30.3
49.3
51.4
54.7
55.2
Distillation 35.5(+5.2) 55.4(+6.1) 56.8(+5.4) 59.4(+4.7) 59.5(+4.3)
Table 3: Compressed AlexNet performance evaluated on PASCAL. We compare the model fine-tuned with the
ground truth and the model trained with our full method. We vary the compression ratio by FLOPS.
PASCAL VOC dataset for the student and using the original resolution for the teacher, we get almost
the same accuracy as the high-resolution teacher while being about two times faster1 .
4.2
Speed-Accuracy Trade off in Compressed Models
It is feasible to select CNN models from a wide range of candidates to strike a balance between
speed and accuracy. However, off-the-shelf CNN models still may not meet one?s computational
requirements. Designing new models is one option. But it often requires significant labor towards
design and training. More importantly, trained models are often designed for specific tasks, but speed
and accuracy trade-offs may change when facing a different task, whereby one may as well train a
new model for the new task. In all such situations, distillation becomes an attractive option.
To understand the speed-accuracy trade off in object detection with knowledge distillation, we vary the
compression ratio (the ranks of weight matrices) of Alexnet with Tucker decomposition. We measure
the compression ratio using FLOPS of the CNN. Experiments in Table 3 show that the accuracy drops
dramatically when the network is compressed too much, for example, when compressed size is 20%
of original, accuracy drops from 57.2% to only 30.3%. However, for the squeezed networks, our
distillation framework is able to recover large amounts of the accuracy drop. For instance, for 37.5%
compression, the original squeezed net only achieves 54.7%. In contrast, our proposed method lifts
it up to 59.4% with a deep teacher (VGG16), which is even better than the uncompressed AlexNet
model 57.2%.
4.3
Ablation Study
As shown in Table 4, we compare different strategies for distillation and hint learning to highlight the
effectiveness of our proposed novel losses. We choose VGG16 as the teacher model and Tucker as
our student model for all the experiments in this section. Other choices reflect similar trends. Recall
that proposal classification and bounding box regression are the two main tasks in the Faster-RCNN
framework. Traditionally, classification is associated with cross entropy loss, denoted as CLS in
Table 4, while bounding box regression is regularized with L2 loss, denoted as L2.
To prevent the classes with small probability being ignored by the objective function, soft label with
high temperature, also named weighted cross entropy loss, is proposed for the proposal classification
task in Sec.3.2. We compare the weighted cross entropy loss defined in (3), denoted as CLS-W in
Table 4, with the standard cross entropy loss (CLS), to achieve slightly better performance on both
PASCAL and KITTI datasets.
For bounding box regression, directly parroting to teacher?s output will suffer from labeling noise. An
improvement is proposed through (4) in Sec.3.3, where the teacher?s prediction is used as a boundary
to guide the student. Such strategy, denoted as L2-B in Table 4, improves over L2 by 1.3%. Note that
a 1% improvement in object detection task is considered very significant, especially on large-scale
datasets with voluminous number of images.
1
Ideally, the convolutional layers should be about 4 times faster. However, due to the loading overhead and
the non-proportional consumption from other layers, this speed up drops to around 2 times faster.
7
Baseline L2 L2-B CLS CLS-W Hints Hints-A L2-B+CLS-W L2-B+CLS-W+Hints-A
PASCAL 54.7 54.6 55.9 57.4 57.7
56.9
58
58.4
59.4
KITTI
49.3 48.5 50.1 50.8 51.3
50.3 52.1
51.7
53.7
Table 4: The proposed method component comparison, i.e., bounded L2 for regression (L2-B, Sec.3.3) and
weighted cross entropy for classification (CLS-W, Sec.3.2) with respect to traditional methods, namely, L2
and cross entropy (CLS). Hints learning w/o adaptation layer (Hints-A and Hints) are also compared. All
comparisons take VGG16 as the teacher and Tucker as the student, with evaluations on PASCAL and KITTI.
Baseline Distillation Hint Distillation + Hint
Trainval
79.6
78.3
80.9
83.5
PASCAL
Test
54.7
58.4
58
59.4
Train
45.3
45.4
47.1
49.6
COCO
Val
25.4
26.1
27.8
28.3
Table 5: Performance of distillation and hint learning on different datasets with Tucker and VGG16 pair.
Moreover, we find that the adaptation layer proposed in Sec.3.4 is critical for hint learning. Even
if layers from teacher and student models have the same number of neurons, they are almost never
in the same feature space. Otherwise, setting the student?s subsequent structure to be the same as
teacher?s, the student would achieve identical results as the teacher. Thus, directly matching a student
layer to a teacher layer [2, 3] is unlikely to perform well. Instead, we propose to add an adaptation
layer to transfer the student layer feature space to the corresponding teacher layer feature space.
Thereby, penalizing the student feature from the teacher feature is better-defined since they lie in
the same space, which is supported by the results in Table 4. With adaptation layer, hint learning
(Hint-A) shows a 1.1% advantage over the traditional method (Hint). Our proposed overall method
(L2-B+CLS-W+Hint-A) outperforms the one without adaptive hint learning (L2-B+CLS-W) by 1.0%,
which again suggests the significant advantage of hint learning with adaptation.
4.4
Discussion
In this section, we provide further insights into distillation and hint learning. Table 5 compares the
accuracy of Tucker model learned with VGG16 on the trainval and testing split of the PASCAL and
COCO datasets. In general, distillation mostly improves the generalization capability of student,
while hint learning helps improving both the training and testing accuracy.
Distillation improves generalization: Similarly to the image classification case discussed in [20],
there also exists structural relationship among the labels in object detection task. For example,
?Car? shares more common visual characteristics with ?Truck? than with ?Person?. Such structural
information is not available in the ground truth annotations. Thus, injecting such relational information
learned with a high capacity teacher model to a student will help generalization capability of the
detection model. The result of applying the distillation only shows consistent testing accuracy
improvement in Table 5.
Hint helps both learning and generalization: We notice that the ?under-fitting? is a common
problem in object detection even with CNN based models (see low training accuracy of the baselines).
Unlike simple classification cases, where it is easy to achieve (near) perfect training accuracy [40],
the training accuracy of the detectors is still far from being perfect. It seems the learning algorithm
is suffering from the saddle point problem [8]. On the contrary, the hint may provide an effective
guidance to avoid the problem by directly having a guidance at an intermediate layer. Thereby,
the model learned with hint learning achieves noticeable improvement in both training and testing
accuracy.
Finally, by combining both distillation and hint learning, both training and test accuracies are
improved significantly compared to the baseline. Table 5 empirically verifies consistent trends on
both the PASCAL and MS COCO datasets for object detection. We believe that our methods can also
be extended to other tasks that also face similar generalization or under-fitting problems.
5
Conclusion
We propose a novel framework for learning compact and fast CNN based object detectors with
the knowledge distillation. Highly complicated detector models are used as a teacher to guide
the learning process of efficient student models. Combining the knowledge distillation and hint
8
framework together with our newly proposed loss functions, we demonstrate consistent improvements
over various experimental setups. Notably, the compact models trained with our learning framework
execute significantly faster than the teachers with almost no accuracy compromises at PASCAL
dataset. Our empirical analysis reveals the presence of under-fitting issue in object detector learning,
which could provide good insights to further advancement in the field.
Acknowledgments This work was conducted as part of Guobin Chen?s internship at NEC Labs
America in Cupertino.
References
[1] K. Ashraf, B. Wu, F. N. Iandola, M. W. Moskewicz, and K. Keutzer. Shallow networks for high-accuracy
road object-detection. CoRR, abs/1606.01561, 2016. 6
[2] J. Ba and R. Caruana. Do deep nets really need to be deep? In Advances in neural information processing
systems, pages 2654?2662, 2014. 2, 8
[3] C. Bucilua, R. Caruana, and A. Niculescu-Mizil. Model compression. In Proceedings of the 12th ACM
SIGKDD international conference on Knowledge discovery and data mining, pages 535?541. ACM, 2006.
2, 3, 8
[4] K. Chatfield, K. Simonyan, A. Vedaldi, and A. Zisserman. Return of the devil in the details: Delving deep
into convolutional nets. arXiv preprint arXiv:1405.3531, 2014. 6
[5] W. Chen, J. T. Wilson, S. Tyree, K. Q. Weinberger, and Y. Chen. Compressing neural networks with the
hashing trick. CoRR, abs/1504.04788, 2015. 2
[6] X. Chen, H. Fang, T. Lin, R. Vedantam, S. Gupta, P. Doll?r, and C. L. Zitnick. Microsoft coco captions:
Data collection and evaluation server. arXiv:1504.00325, 2015. 5
[7] Y. Cheng, F. X. Yu, R. S. Feris, S. Kumar, A. Choudhary, and S.-F. Chang. An exploration of parameter
redundancy in deep networks with circulant projections. In Proceedings of the IEEE International
Conference on Computer Vision, pages 2857?2865, 2015. 2
[8] Y. N. Dauphin, R. Pascanu, C. Gulcehre, K. Cho, S. Ganguli, and Y. Bengio. Identifying and attacking the
saddle point problem in high-dimensional non-convex optimization. In Advances in neural information
processing systems, pages 2933?2941, 2014. 8
[9] M. Denil, B. Shakibi, L. Dinh, N. de Freitas, et al. Predicting parameters in deep learning. In Advances in
Neural Information Processing Systems, pages 2148?2156, 2013. 1, 2
[10] E. L. Denton, W. Zaremba, J. Bruna, Y. LeCun, and R. Fergus. Exploiting linear structure within
convolutional networks for efficient evaluation. In Advances in Neural Information Processing Systems,
pages 1269?1277, 2014. 2
[11] M. Everingham, L. Van Gool, C. K. Williams, J. Winn, and A. Zisserman. The pascal visual object classes
(voc) challenge. International journal of computer vision, 88(2):303?338, 2010. 5
[12] A. Geiger, P. Lenz, and R. Urtasun. Are we ready for autonomous driving? the kitti vision benchmark
suite. In Computer Vision and Pattern Recognition (CVPR), 2012 IEEE Conference on, pages 3354?3361.
IEEE, 2012. 5
[13] R. Girshick. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, pages
1440?1448, 2015. 2, 4, 5
[14] R. B. Girshick, P. F. Felzenszwalb, and D. McAllester. Discriminatively trained deformable part models,
release 5, 2012. 2
[15] Y. Gong, L. Liu, M. Yang, and L. Bourdev. Compressing deep convolutional networks using vector
quantization. arXiv preprint arXiv:1412.6115, 2014. 2
[16] S. Gupta, J. Hoffman, and J. Malik. Cross modal distillation for supervision transfer. arXiv preprint
arXiv:1507.00448, 2015. 3, 5
[17] S. Han, X. Liu, H. Mao, J. Pu, A. Pedram, M. A. Horowitz, and W. J. Dally. Eie: efficient inference engine
on compressed deep neural network. arXiv preprint arXiv:1602.01528, 2016. 2
[18] S. Han, H. Mao, and W. J. Dally. Deep compression: Compressing deep neural network with pruning,
trained quantization and huffman coding. CoRR, abs/1510.00149, 2015. 2
[19] K. He, X. Zhang, S. Ren, and J. Sun. Spatial pyramid pooling in deep convolutional networks for visual
recognition. In European Conference on Computer Vision, pages 346?361. Springer, 2014. 2
[20] G. Hinton, O. Vinyals, and J. Dean. Distilling the knowledge in a neural network. arXiv preprint
arXiv:1503.02531, 2015. 1, 2, 3, 4, 8
[21] I. Hubara, M. Courbariaux, D. Soudry, R. El-Yaniv, and Y. Bengio. Binarized neural networks. In D. D.
Lee, M. Sugiyama, U. V. Luxburg, I. Guyon, and R. Garnett, editors, Advances in Neural Information
Processing Systems 29, pages 4107?4115. Curran Associates, Inc., 2016. 2
[22] F. N. Iandola, M. W. Moskewicz, K. Ashraf, S. Han, W. J. Dally, and K. Keutzer. Squeezenet: Alexnet-level
accuracy with 50x fewer parameters and <1mb model size. CoRR, abs/1602.07360, 2016. 1
9
[23] M. Jaderberg, A. Vedaldi, and A. Zisserman. Speeding up convolutional neural networks with low rank
expansions. arXiv preprint arXiv:1405.3866, 2014. 2
[24] R. J. T. JitendraMalik. Rich feature hierarchies for accurate object detection and semantic segmentation. 2,
6
[25] K. Kim, Y. Cheon, S. Hong, B. Roh, and M. Park. PVANET: deep but lightweight neural networks for
real-time object detection. CoRR, abs/1608.08021, 2016. 1
[26] Y.-D. Kim, E. Park, S. Yoo, T. Choi, L. Yang, and D. Shin. Compression of deep convolutional neural
networks for fast and low power mobile applications. arXiv preprint arXiv:1511.06530, 2015. 1, 2, 4, 6
[27] A. Krizhevsky, I. Sutskever, and G. E. Hinton. Imagenet classification with deep convolutional neural
networks. In F. Pereira, C. J. C. Burges, L. Bottou, and K. Q. Weinberger, editors, Advances in Neural
Information Processing Systems 25, pages 1097?1105. Curran Associates, Inc., 2012. 2, 6
[28] V. Lebedev, Y. Ganin, M. Rakhuba, I. Oseledets, and V. Lempitsky. Speeding-up convolutional neural
networks using fine-tuned cp-decomposition. arXiv preprint arXiv:1412.6553, 2014. 2
[29] Y. Li, K. He, J. Sun, et al. R-fcn: Object detection via region-based fully convolutional networks. In
Advances in Neural Information Processing Systems, pages 379?387, 2016. 2, 4
[30] A. Novikov, D. Podoprikhin, A. Osokin, and D. P. Vetrov. Tensorizing neural networks. In Advances in
Neural Information Processing Systems, pages 442?450, 2015. 2
[31] M. Rastegari, V. Ordonez, J. Redmon, and A. Farhadi. Xnor-net: Imagenet classification using binary
convolutional neural networks. In European Conference on Computer Vision, pages 525?542. Springer,
2016. 2
[32] S. Ren, K. He, R. Girshick, and J. Sun. Faster r-cnn: Towards real-time object detection with region
proposal networks. In Advances in neural information processing systems, pages 91?99, 2015. 2, 3, 4
[33] S. Ren, K. He, R. Girshick, X. Zhang, and J. Sun. Object detection networks on convolutional feature
maps. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2016. 4
[34] A. Romero, N. Ballas, S. E. Kahou, A. Chassang, C. Gatta, and Y. Bengio. Fitnets: Hints for thin deep
nets. arXiv preprint arXiv:1412.6550, 2014. 1, 2, 3, 5
[35] O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, A. C. Berg, and L. Fei-Fei. ImageNet Large Scale Visual Recognition Challenge. International
Journal of Computer Vision (IJCV), 115(3):211?252, 2015. 5
[36] J. Shen, N. Vesdapunt, V. N. Boddeti, and K. M. Kitani. In teacher we trust: Learning compressed models
for pedestrian detection. arXiv preprint arXiv:1612.00478, 2016. 3
[37] K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition.
arXiv preprint arXiv:1409.1556, 2014. 6
[38] J.-C. Su and S. Maji. Cross quality distillation. arXiv preprint arXiv:1604.00433, 2016. 3, 6
[39] Y. Xiang, W. Choi, Y. Lin, and S. Savarese. Data-driven 3d voxel patterns for object category recognition.
In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 1903?1911,
2015. 6
[40] C. Zhang, S. Bengio, M. Hardt, B. Recht, and O. Vinyals. Understanding deep learning requires rethinking
generalization. arXiv preprint arXiv:1611.03530, 2016. 8
[41] X. Zhang, J. Zou, K. He, and J. Sun. Accelerating very deep convolutional networks for classification and
detection. 2015. 1, 2
[42] X. Zhang, J. Zou, X. Ming, K. He, and J. Sun. Efficient and accurate approximations of nonlinear convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,
pages 1984?1992, 2015. 1, 2
10
| 6676 |@word cnn:16 middle:1 compression:14 stronger:1 norm:2 loading:1 seems:1 everingham:1 r:4 rgb:1 decomposition:5 thereby:3 harder:3 shot:1 configuration:1 lightweight:2 score:2 trainval:2 liu:2 tuned:2 interestingly:1 outperforms:1 freitas:1 blank:1 activation:1 written:1 gpu:3 subsequent:1 romero:3 informative:1 pertinent:1 lcls:4 drop:6 designed:2 rpn:9 half:1 prohibitive:1 fewer:3 selected:1 advancement:1 intelligence:1 podoprikhin:1 feris:1 provides:1 quantized:1 draft:1 location:1 pascanu:1 simpler:4 zhang:6 unbounded:1 height:1 viable:1 consists:1 yu1:1 combine:3 fitting:4 pathway:1 inside:1 overhead:1 introduce:2 ijcv:1 notably:1 behavior:2 multi:8 relying:1 decomposed:1 voc:5 ming:1 cpu:1 farhadi:1 becomes:2 bounded:7 moreover:1 alexnet:13 proposing:1 suite:1 binarized:2 runtime:4 exactly:1 zaremba:1 demonstrates:1 classifier:2 wrong:1 control:1 normally:1 scaled:1 positive:1 before:1 soudry:1 despite:2 vetrov:1 meet:1 might:3 fitnets:1 suggests:3 challenging:2 range:1 averaged:1 acknowledgment:1 lecun:1 testing:4 practice:1 differs:1 backpropagation:1 shin:1 krt:1 empirical:4 significantly:4 vedaldi:2 matching:1 ups:1 projection:1 road:1 regular:1 get:2 close:5 applying:3 seminal:1 optimize:1 conventional:2 equivalent:1 demonstrated:1 map:10 dean:1 williams:1 convex:1 shen:2 unify:1 formulate:2 resolution:9 identifying:1 insight:4 importantly:1 dominate:1 fang:1 handle:2 autonomous:2 traditionally:1 oseledets:1 diego:1 target:2 suppose:1 pt:5 caption:1 hierarchy:1 us:1 designing:1 hypothesis:1 curran:2 trick:2 element:1 trend:2 expensive:3 recognition:7 associate:2 kahou:1 asymmetric:1 labeled:1 database:1 module:2 preprint:13 region:8 compressing:3 connected:4 sun:6 trade:4 complexity:1 ideally:1 trained:10 compromise:1 efficiency:2 accelerate:1 various:5 america:2 maji:1 train:6 fast:7 effective:2 labeling:1 lift:1 hyper:1 whose:1 widely:1 larger:10 supplementary:3 valued:1 cvpr:1 otherwise:2 compressed:11 simonyan:2 jointly:1 final:4 advantage:2 net:5 propose:10 reconstruction:1 mb:1 adaptation:14 combining:2 ablation:2 vesdapunt:1 achieve:9 deformable:2 kv:2 getting:1 exploiting:1 sutskever:1 yaniv:1 requirement:2 p:5 zts:1 produce:1 table1:1 perfect:2 object:50 kitti:10 wider:1 help:4 bourdev:1 gong:1 pose:1 ganin:1 novikov:1 noticeable:1 strong:3 indicate:2 convention:1 iou:2 direction:3 guided:10 distilling:1 cnns:5 filter:1 subsequently:2 exploration:1 mcallester:1 softer:1 public:2 material:3 require:1 fix:1 generalization:8 really:1 around:2 considered:1 ground:10 driving:2 achieves:4 adopt:4 vary:2 failing:1 lenz:1 injecting:1 label:17 hubara:1 concurrent:1 weighted:8 hoffman:1 offs:2 denil:2 avoid:1 shelf:1 surveillance:1 mobile:1 wilson:1 release:1 focus:1 inherits:1 improvement:12 consistently:1 rank:3 prevalent:1 mainly:1 contrast:2 sigkdd:1 kim:3 baseline:6 inference:1 ganguli:1 el:1 niculescu:1 typically:1 transferring:3 unlikely:1 hidden:1 voluminous:3 mimicking:1 overall:6 classification:32 aforementioned:1 pascal:19 issue:2 among:2 denoted:5 dauphin:1 pixel:2 art:1 spatial:3 softmax:5 field:1 once:2 distilled:1 extraction:1 beach:1 having:4 runtimes:2 never:1 identical:1 park:2 yu:1 uncompressed:1 denton:1 thin:1 fcn:2 mimic:3 foreground:3 report:4 others:2 fundamentally:1 hint:50 few:1 employ:1 modern:3 missouri:1 composed:1 comprehensive:2 floating:1 microsoft:1 ab:5 detection:56 mining:1 possibility:1 highly:3 investigate:1 evaluation:5 severe:2 adjust:1 introduces:1 accurate:3 necessary:1 conduct:1 savarese:1 penalizes:1 re:3 guidance:4 girshick:5 instance:1 column:1 soft:9 caruana:3 soften:1 cost:1 surpasses:1 rare:1 krizhevsky:1 successful:1 conducted:1 too:2 teacher:74 combined:2 cho:1 st:1 person:1 international:5 recht:1 retain:1 lee:1 off:3 regressor:1 together:1 lebedev:1 again:1 reflect:1 opposed:2 choose:2 huang:1 external:1 horowitz:1 return:2 li:1 account:1 de:1 student:58 sec:11 coding:1 pedestrian:2 inc:2 reg:2 lab:2 dally:3 red:1 relied:1 recover:3 complicated:2 option:2 annotation:2 capability:2 contribution:1 shakibi:1 accuracy:36 convolutional:27 characteristic:1 ensemble:1 yield:1 critically:1 ren:3 worth:1 researcher:1 eie:1 chassang:1 russakovsky:1 detector:8 reach:1 suffers:1 evaluates:1 internship:1 frequency:1 tucker:9 associated:3 newly:2 dataset:9 hardt:1 popular:1 recall:1 knowledge:29 car:1 improves:4 segmentation:1 lreg:3 back:1 manuscript:1 finetune:1 higher:3 hashing:1 follow:2 response:1 improved:2 zisserman:4 modal:1 evaluated:1 box:7 though:1 execute:1 stage:1 hand:2 trust:1 su:2 nonlinear:1 propagation:1 widespread:1 mode:1 quality:7 ordonez:1 believe:1 usa:1 effect:1 contain:2 logits:1 regularization:1 kitani:1 semantic:1 illustrated:1 deal:1 attractive:2 xnor:1 width:1 encourages:3 please:1 whereby:2 m:15 hong:1 demonstrate:4 l1:5 temperature:6 cp:1 image:20 ranging:1 wise:1 novel:3 rcn:9 common:2 empirically:1 handcrafted:1 ballas:1 discussed:2 cupertino:1 he:6 relating:1 theirs:1 distillation:51 significant:7 refer:2 dinh:1 tuning:4 similarly:2 sugiyama:1 bruna:1 han:3 impressive:1 yk2:3 supervision:2 add:2 pu:1 dominant:1 imbalanced:1 recent:2 own:2 driven:1 coco:10 scenario:1 certain:1 server:1 binary:2 success:1 yi:2 seen:1 preserving:1 additional:4 deng:1 attacking:1 strike:1 vgg16:14 multiple:2 full:1 reduces:1 smooth:1 boddeti:1 faster:14 match:4 cross:15 long:1 lin:2 equally:1 impact:2 prediction:7 regression:32 vision:10 metric:2 arxiv:27 kernel:1 represent:2 pyramid:2 proposal:11 whereas:2 background:6 fine:5 addition:2 winn:1 huffman:1 krause:1 rakhuba:1 unlike:4 pooling:2 tend:1 dpm:1 ample:1 contrary:2 spirit:1 effectiveness:1 call:1 structural:2 near:2 presence:1 yang:2 bernstein:1 intermediate:6 split:2 easy:1 bengio:4 affect:1 fit:1 relu:1 misclassifications:1 architecture:5 approaching:1 idea:1 det:1 chatfield:1 accelerating:1 padding:1 effort:1 suffer:1 deep:26 ignored:2 dramatically:1 detailed:1 karpathy:1 amount:1 category:11 reduced:1 outperform:1 notice:2 per:1 serving:1 discrete:1 key:1 four:4 redundancy:2 achieving:1 prevent:1 penalizing:1 year:1 run:1 luxburg:1 parameterized:2 named:1 reporting:1 throughout:1 reader:1 almost:4 architectural:1 wu:1 guyon:1 geiger:1 keutzer:2 scaling:1 dropout:2 layer:47 bound:3 followed:1 cheng:1 fold:1 truck:1 fei:2 generates:1 wc:1 speed:19 kumar:1 relatively:3 speedup:1 transferred:1 across:3 smaller:6 increasingly:1 describes:1 slightly:1 wi:1 shallow:2 computationally:1 resource:1 remains:1 end:6 zk2:1 gulcehre:1 available:1 operation:1 lrpn:2 doll:1 apply:3 save:1 batch:1 weinberger:2 original:4 compress:2 denotes:3 tony:1 ensure:1 exploit:1 especially:3 tensor:1 objective:2 malik:1 added:1 already:1 manmohan:1 strategy:3 rt:3 traditional:2 detrimental:1 distance:1 capacity:5 parametrized:1 rethinking:1 w0:1 consumption:1 urtasun:1 reason:1 toward:1 assuming:1 besides:2 relationship:2 ratio:3 demonstration:1 balance:3 innovation:1 downsampling:1 setup:3 mostly:1 expense:1 info:1 ba:2 design:2 implementation:1 proper:1 satheesh:1 zt:1 redesign:1 perform:4 imbalance:4 shallower:3 neuron:4 upper:2 datasets:18 zk1:1 benchmark:4 convolution:1 observation:1 tensorizing:1 flop:3 hinton:3 situation:1 relational:1 extended:1 pedram:1 discovered:1 smoothed:2 lb:6 introduced:3 namely:4 required:1 pair:3 imagenet:4 california:1 engine:1 quadratically:1 learned:3 chen1:1 tremendous:1 nip:1 address:5 able:1 usually:1 pattern:5 spp:1 roh:1 challenge:6 including:1 memory:2 gool:1 power:1 misclassification:2 critical:3 rely:1 force:1 regularized:1 predicting:1 mizil:1 improve:1 choudhary:1 ready:1 speeding:2 prior:1 understanding:1 l2:18 discovery:1 val:1 xiang:2 loss:46 fully:6 highlight:1 squeezed:2 discriminatively:1 proportional:1 facing:1 validation:1 rcnn:7 consistent:4 tyree:1 editor:2 courbariaux:1 share:2 row:1 surprisingly:1 supported:1 keeping:1 moskewicz:2 guide:4 deeper:5 understand:1 circulant:1 wide:1 burges:1 face:1 felzenszwalb:1 van:1 boundary:1 depth:2 evaluating:1 rich:1 made:1 commonly:1 san:1 adaptive:1 collection:1 osokin:1 far:3 voxel:1 transaction:1 pruning:2 compact:6 preferred:1 jaderberg:1 reveals:1 vedantam:1 xi:2 fergus:1 khosla:1 table:15 additionally:1 learn:8 channel:4 transfer:4 ca:1 delving:1 rastegari:1 improving:1 expansion:1 bottou:1 excellent:1 zou:2 necessarily:1 complex:3 cl:11 domain:6 zitnick:1 european:2 garnett:1 main:1 arrow:1 bounding:7 whole:2 ztt:1 noise:2 verifies:1 suffering:1 aid:2 precision:2 mao:2 pereira:1 candidate:1 lie:1 third:1 removing:1 down:1 choi:2 specific:1 gupta:2 dominates:1 consist:1 bucila:1 mnist:1 exists:1 quantization:2 effectively:1 kr:2 corr:5 nec:2 push:1 demand:1 margin:2 gap:1 chen:4 entropy:13 explore:2 saddle:2 visual:6 gatta:1 vinyals:2 expressed:1 adjustment:1 labor:1 iandola:2 chang:1 applies:1 springer:2 truth:10 acm:2 ma:1 lempitsky:1 careful:1 towards:2 replace:1 shared:1 feasible:1 hard:5 change:1 redmon:1 degradation:1 contradictory:1 discriminate:1 experimental:1 select:1 ilsvrc:5 berg:1 support:1 devil:1 ashraf:2 evaluate:3 yoo:1 trainable:1 scratch:1 |
6,275 | 6,677 | One-Sided Unsupervised Domain Mapping
Sagie Benaim1 and Lior Wolf1,2
1
The Blavatnik School of Computer Science , Tel Aviv University, Israel
2
Facebook AI Research
Abstract
In unsupervised domain mapping, the learner is given two unmatched datasets
A and B. The goal is to learn a mapping GAB that translates a sample in A
to the analog sample in B. Recent approaches have shown that when learning
simultaneously both GAB and the inverse mapping GBA , convincing mappings
are obtained. In this work, we present a method of learning GAB without learning
GBA . This is done by learning a mapping that maintains the distance between
a pair of samples. Moreover, good mappings are obtained, even by maintaining
the distance between different parts of the same sample before and after mapping.
We present experimental results that the new method not only allows for one
sided mapping learning, but also leads to preferable numerical results over the
existing circularity-based constraint. Our entire code is made publicly available
at https://github.com/sagiebenaim/DistanceGAN.
1
Introduction
The advent of the Generative Adversarial Network (GAN) [6] technology has allowed for the
generation of realistic images that mimic a given training set by accurately capturing what is inside
the given class and what is ?fake?. Out of the many tasks made possible by GANs, the task of
mapping an image in a source domain to the analog image in a target domain is of a particular interest.
The solutions proposed for this problem can be generally separated by the amount of required
supervision. On the one extreme, fully supervised methods employ pairs of matched samples, one
in each domain, in order to learn the mapping [9]. Less direct supervision was demonstrated by
employing a mapping into a semantic space and requiring that the original sample and the analog
sample in the target domain share the same semantic representation [22].
If the two domains are highly related, it was demonstrated that just by sharing weights between the
networks working on the two domains, and without any further supervision, one can map samples
between the two domains [21, 13]. For more distant domains, it was demonstrated recently that by
symmetrically leaning mappings in both directions, meaningful analogs are obtained [28, 11, 27].
This is done by requiring circularity, i.e., that mapping a sample from one domain to the other and
then back, produces the original sample.
In this work, we go a step further and show that it is possible to learn the mapping between the
source domain and the target domain in a one-sided unsupervised way, by enforcing high crossdomain correlation between the matching pairwise distances computed in each domain. The new
constraint allows one-sided mapping and also provides, in our experiments, better numerical results
than circularity. Combining both of these constraints together often leads to further improvements.
Learning the new constraint requires comparing pairs of samples. While there is no real practical
reason not to do so, since training batches contain multiple samples, we demonstrate that similar
constraints can even be applied per image by computing the distance between, e.g., the top part of the
image and the bottom part.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
1.1
Related work
Style transfer These methods [5, 23, 10] typically receive as input a style image and a content image
and create a new image that has the style of the first and the content of the second. The problem of
image translation between domains differs since when mapping between domains, part of the content
is replaced with new content that matches the target domain and not just the style. However, the
distinction is not sharp, and many of the cross-domain mapping examples in the literature can almost
be viewed as style transfers. For example, while a zebra is not a horse in another style, the horse to
zebra mapping, performed in [28] seems to change horse skin to zebra skin. This is evident from the
stripped Putin example obtained when mapping the image of shirtless Putin riding a horse.
Generative Adversarial Networks GAN [6] methods train a generator network G that synthesizes
samples from a target distribution, given noise vectors, by jointly training a second network D. The
specific generative architecture we and others employ is based on the architecture of [18]. In image
mapping, the created image is based on an input image and not on random noise [11, 28, 27, 13, 22, 9].
Unsupervised Mapping The work that is most related to ours, employs no supervision except for
sample images from the two domains. This was done very recently [11, 28, 27] in image to image
translation and slightly earlier for translating between natural languages [24]. Note that [11] proposes
the ?GAN with reconstruction loss? method, which applies the cycle constraint in one side and trains
only one GAN. However, unlike our method, this method requires the recovery of both mappings and
is outperformed by the full two-way method.
The CoGAN method [13], learns a mapping from a random input vector to matching samples from
the two domains. It was shown in [13, 28] that the method can be modified in order to perform
domain translation. In CoGAN, the two domains are assumed to be similar and their generators (and
GAN discriminators) share many of the layers weights, similar to [21]. As was demonstrated in [28],
the method is not competitive in the field of image to image translation.
Weakly Supervised Mapping In [22], the matching between the source domain and the target
domain is performed by incorporating a fixed pre-trained feature map f and requiring f -constancy,
i.e, that the activations of f are the same for the input samples and for mapped samples.
Supervised Mapping When provided with matching pairs of (input image, output image) the
supervision can be performed directly. An example of such method that also uses GANs is [9], where
the discriminator D receives a pair of images where one image is the source image and the other is
either the matching target image (?real? pair) or a generated image (?fake? pair); The linking between
the source and the target image is further strengthened by employing the U-net architecture [19].
Domain Adaptation In this setting, we typically are given two domains, one having supervision in
the form of matching labels, while the second has little or no supervision. The goal is to learn to
label samples from the second domain. In [3], what is common to both domains and what is distinct
is separated thus improving on existing models. In [2], a transformation is learned, on the pixel
level, from one domain to another, using GANs. In [7], an unsupervised adversarial approach to
semantic segmentation, which uses both global and category specific domain adaptation techniques,
is proposed.
2
Preliminaries
In the problem of unsupervised mapping, the learning algorithm is provided with unlabeled datasets
from two domains, A and B. The first dataset includes i.i.d samples from the distribution pA and the
second dataset includes i.i.d samples from the distribution pB . Formally, given
i.i.d
{xi }m
i=1 such that xi ? pA
and
i.i.d
{xj }nj=1 such that xj ? pB ,
our goal is to learn a function GAB , which maps samples in domain A to analog samples in domain
B, see examples below. In previous work [11, 28, 27], it is necessary to simultaneously recover a
second function GBA , which similarly maps samples in domain B to analog samples in domain A.
Justification In order to allow unsupervised learning of one directional mapping, we introduce
the constraint that pairs of inputs x, x0 , which are at a certain distance from each other, are mapped
to pairs of outputs GAB (x), GAB (x0 ) with a similar distance, i.e., that the distances kx ? x0 k and
2
Figure 1: Each triplet shows the source handbag image, the target shoe as produced by CycleGAN?s [28] mapper GAB and the results of approximating GAB by a fixed nonnegative linear
transformation T , which obtains each output pixel as a linear combination of input pixels. The linear
transformation captures the essence of GAB showing that much of the mapping is achieved by a fixed
spatial transformation.
kGAB (x) ? GAB (x0 )k are highly correlated. As we show below, it is reasonable to assume that this
constraint approximately holds in many of the scenarios demonstrated by previous work on domain
translation. Although approximate, it is sufficient, since as was shown in [21], mapping between
domains requires only little supervision on top of requiring that the output distribution of the mapper
matches that of the target distribution.
Consider, for example, the case of mapping shoes to edges, as presented in Fig. 4. In this case, the
edge points are simply a subset of the image coordinates, selected by local image criterion. If image
x is visually similar to image x0 , it is likely that their edge maps are similar. In fact, this similarity
underlies the usage of gradient information in the classical computer vision literature. Therefore,
while the distances are expected to differ in the two domains, one can expect a high correlation.
Next, consider the case of handbag to shoe mapping (Fig. 4). Analogs tend to have the same
distribution of image colors in different image formations. Assuming that the spatial pixel locations
of handbags follow a tight distribution (i.e., the set of handbag images share the same shapes) and the
same holds for shoes, then there exists a set of canonical displacement fields that transform a handbag
to a shoe. If there was one displacement, which would happen to be a fixed permutation of pixel
locations, distances would be preserved. In practice, the image transformations are more complex.
To study whether the image displacement model is a valid approximation, we learned a nonnegative
2
?642
linear transformation T ? R64
that maps, one channel at a time, handbag images of size
+
64 ? 64 ? 3 to the output shoe images of the same size given by the CycleGAN method. T ?s columns
can be interpreted as weights that determine the spread of mass in the output image for each pixel
location in the input image. It was estimated by minimizing the squared error of mapping every
channel (R, G, or B) of a handbag image to the same channel in the matching shoe. Optimization
was done by gradient descent with a projection to the space of nonnegative matrices, i.e., zeroing the
negative elements of T at each iteration.
Sample mappings by the matrix T are shown in Fig. 1. As can be seen, the nonnegative linear
transformation approximates CycleGAN?s multilayer CNN GAB to some degree. Examining the
elements of T , they share some properties with permutations: the mean sum of the rows is 1.06 (SD
0.08) and 99.5% of the elements are below 0.01.
In the case of adding glasses or changing gender or hair color (Fig 3), a relatively minor image
modification, which does not significantly change the majority of the image information, suffices
in order to create the desired visual effect. Such a change is likely to largely maintain the pairwise
image distance before and after the transformation.
In the case of computer generated heads at different angles vs. rotated cars, presented in [11],
distances are highly correlated partly because the area that is captured by the foreground object is
a good indicator of the object?s yaw. When mapping between horses to zebras [28], the texture
of a horse?s skin is transformed to that of the zebra. In this case, most of the image information
is untouched and the part that is changed is modified by a uniform texture, again approximately
maintaining pairwise distances. In Fig 2(a), we compare the L1 distance in RGB space of pairs
of horse images to the distance of the samples after mapping by the CycleGAN Network [28] is
performed, using the public implementation. It is evident that the cross-domain correlation between
pairwise distances is high. We also looked at Cityscapes image and ground truth label pairs in
Fig 2(c), and found that there is high correlation between the distances. This is the also the case in
many other literature-based mappings between datasets we have tested and ground truth pairs.
While there is little downside to working with pairs of training images in comparison to working with
single images, in order to further study the amount of information needed for successful alignment,
we also consider distances between the two halves of the same image. We compare the L1 distance
3
(a)
(b)
(c)
(d)
Figure 2: Justifying the high correlation between distances in different domains. (a) Using the
CycleGAN model [28], we map horses to zebras and vice versa. Green circles are used for the
distance between two random horse images and the two corresponding translated zebra images.
Blue crosses are for the reverse direction translating zebra to horse images. The Pearson correlation
for horse to zebra translation is 0.77 (p-value 1.7e?113) and for zebra to horse it is 0.73 (p-value
8.0e?96). (b) As in (a) but using the distance between two halves of the same image that is either a
horse image translated to a zebra or vice-versa. The Pearson correlation for horse to zebra translation
is 0.91 (p-value 9.5e?23) and for zebra to horse it is 0.87 (p-value 9.7e?19). (c) Cityscapes images
and associated labels. Green circles are used for distance between two cityscapes images and the
two corresponding ground truth images The Pearson correlation is 0.65 (p-value 6.0e?16). (d) As in
(c) but using the distance between two halves of the same image. The Pearson correlation is 0.65
(p-value 1.4e?12).
between the left and right halves as computed on the input image to that which is obtained on the
generated image or the corresponding ground truth image. Fig. 1(b) and Fig. 1(d) presents the results
for horses to zebras translation and for Cityscapes image and label pairs, respectively. As can be seen,
the correlation is also very significant in this case.
From Correlations to Sum of Absolute Differences We have provided justification and empirical
evidence that for many semantic mappings, there is a high degree of correlations between the
pairwise distances in the two domains. In other words, let dk be a vector of centered and unit-variance
normalized pairwise distances in one domain and let d0k be the vector of normalized distances obtained
P
in the other domain by translating each image out of each pair between the domains, then dk d0k
should be high. When training the mapper GAB , the mean and variance used for normalization in
each domain are precomputed based on the training samples in each domain, which assumes that the
post mapping distribution of samples is similar to the training distribution.
P
The pairwise distances in the source domain dk are fixed and maximizing dk d0k causes pairwise
distances dk with large absolute value
the optimization. Instead, we propose to minimize
P to dominate
0
the sum of absolute
differences
|d
?
d
|,
which
spreads the error in distances uniformly. The
k
k
k
P
P
two losses ? dk d0k and k |dk ? d0k | are highly related and the negative correlation between them
was explicitly computed for simple distributions and shown to be very strong [1].
4
3
Unsupervised Constraints on the Learned Mapping
There are a few types of constraints suggested in the literature, which do not require paired samples.
First, one can enforce the distribution of GAB (x) : x ? pA , which we denote as GAB (pA ), to
be indistinguishable from that of pB . In addition, one can require that mapping from A to B and
back would lead to an identity mapping. Another constraint suggested, is that for every x ? B
GAB (x) = x. We review these constraints and then present the new constraints we propose.
Adversarial constraints Our training sets are viewed as two discrete distributions p?A and p?B that
are sampled from the source and target domain distributions pA and pB , respectively. For the learned
network GAB , the similarity between the distributions GAB (pA ) and pB is modeled by a GAN. This
involves the training of a discriminator network DB : B ? {0, 1}. The loss is given by:
LGAN (GAB , DB , p?A , p?B ) =ExB ?p?B [log DB (xB )] + ExA ?p?A [log(1 ? DB (GAB (xA ))]
This loss is minimized over GAB and maximized over DB . When both GAB and GBA are learned
simultaneously, there is an analog expression LGAN (GBA , DA , p?B , p?A ), in which the domains A and
B switch roles and the two losses (and four networks) are optimized jointly.
Circularity constraints In three recent reports [11, 28, 27], circularity loss was introduced for
image translation. The rationale is that given a sample from domain A, translating it to domain B and
then back to domain A should result in the identical sample. Formally, the following loss is added:
Lcycle (GAB , GBA , p?A ) = Ex?p?A kGBA (GAB (x)) ? xk1
The L1 norm employed above was found to be mostly preferable, although L2 gives similar results.
Since the circularity loss requires the recovery of the mappings in both directions, it is usually
employed symmetrically, by considering Lcycle (GAB , GBA , p?A ) + Lcycle (GBA , GAB , p?B ).
The circularity constraint is often viewed as a definite requirement for admissible functions GAB and
GBA . However, just like distance-based constraints, it is an approximate one. To see this, consider
the zebra to horse mapping example. Mapping a zebra to a horse means losing the stripes. The
inverse mapping, therefore, cannot be expected to recover the exact input stripes.
Target Domain Identity A constraint that has been used in [22] and in some of the experiments
in [28] states that GAB applied to samples from the domain B performs the identity mapping. We
did not experiment with this constraint and it is given here for completeness:
LT-ID (GAB , p?B ) = Ex?p?B kx ? GAB (x)k2
Distance Constraints The adversarial loss ensures that samples from the distribution of A are
translated to samples in the distribution of B. However, there are many such possible mappings.
Given a mapping for n samples of A to n samples of B, one can consider any permutation of the
samples in B as a valid mapping and, therefore, the space of functions mapping from A to B is very
large. Adding the circularity constraint, enforces the mapping from B to A to be the inverse of the
permutation that occurs from A to B, which reduces the amount of admissible permutations.
To further reduce this space, we propose a distance preserving map, that is, the distance between two
samples in A should be preserved in the mapping to B. We therefore consider the following loss,
which is the expectation of the absolute differences between the distances in each domain up to scale:
Ldistance (GAB , p?A ) = Exi ,xj ?p?A |
1
1
(kxi ? xj k1 ? ?A ) ?
(kGAB (xi ) ? GAB (xj )k1 ? ?B )|
?A
?B
where ?A , ?B (?A , ?B ) are the means (standard deviations) of pairwise distances in the training sets
from A and B, respectively, and are precomputed.
In practice, we compute the loss over pairs of samples that belong to the same minibatch during
training. Even for minibatches with 64 samples, as in DiscoGAN [11], considering all pairs is feasible.
If needed, for even larger mini-batches, one can subsample the pairs.
When the two mappings are simultaneously learned, Ldistance (GBA , p?B ) is similarly defined. In both
cases, the absolute difference of the L1 distances between the pairs in the two domains is considered.
5
In comparison to circularity, the distance-based constraint does not suffer from the model collapse
problem that is described in [11]. In this phenomenon, two different samples from domain A are
mapped to the same sample in domain B. The mapping in the reverse direction then generates an
average of the two original samples, since the sample in domain B should be mapped back to both the
first and second original samples in A. Pairwise distance constraints prevents this from happening.
Self-distance Constraints Whether or not the distance constraint is more effective than the circularity constraint in recovering the alignment, the distance based constraint has the advantage of being
one sided. However, it requires that pairs of samples are transfered at once, which, while having little
implications on the training process as it is currently done, might effect the ability to perform on-line
learning. Furthermore, the official CycleGAN [28] implementation employs minibatches of size one.
We, therefore, suggest an additional constraint, which employs one sample at a time and compares
the distances between two parts of the same sample.
Let L, R : Rh?w ? Rh?w/2 be the operators that given an input image return the left or right part of
it. We define the following loss:
L
?A )
self- (GAB , p
distance
1
(kL(x) ? R(x)k1 ? ?A )
?A
1
? (kL(GAB (x)) ? R(GAB (x))k1 ? ?B )|
?B
= Ex?p?A |
(1)
where ?A and ?A are the mean and standard deviation of the pairwise distances between the two
halves of the image in the training set from domain P
A, and similarly for ?B and ?B , e.g., given the
training set {xj }nj=1 ? B, ?B is precomputed as n1 j kL(xj ) ? R(xj )k1 .
3.1
Network Architecture and Training
When training the networks GAB , GBA , DB and DA , we employ the following loss, which is
minimized over GAB and GBA and maximized over DB and DA :
?1A LGAN (GAB , DB , p?A , p?B ) + ?1B LGAN (GBA , DA , p?B , p?A ) + ?2A Lcycle (GAB , GBA , p?A )+
?2B Lcycle (GBA , GAB , p?B ) + ?3A Ldistance (GAB , p?A ) + ?3B Ldistance (GBA , p?B )+
?4A Lself-distance (GAB , p?A ) + ?4B Lself-distance (GBA , p?B )
where ?iA , ?iB are trade-off parameters. We did not test the distance constraint and the self-distance
constraint jointly, so in every experiment, either ?3A = ?3B = 0 or ?4A = ?4A = 0. When
performing one sided mapping from A to B, only ?1A and either ?3A or ?4A are non-zero.
We consider A and B to be a subset of R3?s?s of images where s is either 64, 128 or 256, depending
on the image resolution. In order to directly compare our results with previous work and to employ
the strongest baseline in each dataset, we employ the generator and discriminator architectures of
both DiscoGAN [11] and CycleGAN [28].
In DiscoGAN, the generator is build of an encoder-decoder unit. The encoder consists of convolutional layers with 4 ? 4 filters followed by Leaky ReLU activation units. The decoder consists of
deconvolutional layers with 4 ? 4 filters followed by a ReLU activation units. Sigmoid is used for
the output layer and batch normalization [8] is used before the ReLU or Leaky ReLU activations.
Between 4 to 5 convolutional/deconvolutional layers are used, depending on the domains used in
A and B (we match the published code architecture per dataset). The discriminator is similar to the
encoder, but has an additional convolutional layer as the first layer and a sigmoid output unit.
The CycleGAN architecture for the generator is based on [10]. The generators consist of two 2stride convolutional layers, between 6 to 9 residual blocks depending on the image resolution and
two fractionally strided convolutions with stride 1/2. Instance normalization is used as in [10].
The discriminator uses 70 ? 70 PatchGANs [9]. For training, CycleGAN employs two additional
techniques. The first is to replace the negative log-likelihood by a least square loss [25] and the second
is to use a history of images for the discriminators, rather then only the last image generated [20].
6
Table 2:
Normalized
Table 1: Tradeoff weights for each experiment. RMSE between the angles Table 3: MNIST clasof source and translated sification on mapped
Experiment ?1A ?1B ?2A ?2B ?3A ?3B ?4A ?4B
images.
SHVN images.
DiscoGAN
Distance ?
Distance ?
Dist+Cycle
Self Dist ?
Self Dist ?
0.5
0.5
0
0.5
0.5
0
0.5 0.5 0.5 0
0
0
0
0
0
0 0.5 0
0
0
0.5 0
0
0 0.5 0
0
0.5 0.5 0.5 0.5 0.5 0
0
0
0
0
0
0 0.5 0
0.5 0
0
0
0
0 0.5
Method
car2car car2head
DiscoGAN
Distance
Dist.+Cycle
Self Dist.
0.306
0.135
0.098
0.117
0.137
0.097
0.273
0.197
Method
CycleGAN
Distance
Dist.+Cycle
Self Dist.
Accuracy
26.1%
26.8%
18.0%
25.2%
Table 4: CelebA mapping results using the VGG face descriptor.
Male ? Female
4
Blond ? Black
Method
Cosine
Similarity
Separation
Accuracy
DiscoGAN
Distance
Distance+Cycle
Self Distance
0.23
0.32
0.35
0.24
0.87
0.88
0.87
0.86
DiscoGAN
Distance
Distance+Cycle
Self Distance
0.22
0.26
0.31
0.24
0.86
0.87
0.89
0.91
Cosine
Similarity
Separation
Accuracy
Glasses ? Without
Cosine
Similarity
0.15
0.89
0.13
0.24
0.92
0.42
0.24
0.91
0.41
0.24
0.91
0.34
???? Other direction ????
0.14
0.91
0.10
0.22
0.96
0.30
0.22
0.95
0.30
0.19
0.94
0.30
Separation
Accuracy
0.84
0.79
0.82
0.80
0.90
0.89
0.85
0.81
Experiments
We compare multiple methods: the DiscoGAN or the CycleGAN baselines; the one sided mapping
using Ldistance (A ? B or B ? A); the combination of the baseline method with Ldistance ; the
self distance method. For DiscoGAN, we use a fixed weight configuration for all experiments, as
shown in Tab. 1. For CycleGAN, there is more sensitivity to parameters and while the general pattern
is preserved, we used different weight for the distance constraint depending on the experiment, digits
or horses to zebra.
Models based on DiscoGAN Datasets that were tested by DiscoGAN are evaluated here using this
architecture. In initial tests, CycleGAN is not competitive on these out of the box. The first set of
experiments maps rotated images of cars to either cars or heads. The 3D car dataset [4] consists of
rendered images of 3D cars whose degree varies at 15? intervals. Similarly, the head dataset, [17],
consists of 3D images of rotated heads which vary from ?70? to 70? . For the car2car experiment,
the car dataset is split into two parts, one of which is used for A and one for B (It is further split
into train and test set). Since the rotation angle presents the largest source of variability, and since
the rotation operation is shared between the datasets, we expect it to be the major invariant that the
network learns, i.e., a semantic mapping would preserve angles.
A regressor was trained to calculate the angle of a given car image based on the training data. Tab. 2
shows the Root Mean Square Error (RMSE) between the angle of source image and translated image.
As can be seen, the pairwise distance based mapping results in lower error than the DiscoGAN
one, combining both further improves results, and the self distance outperforms both DiscoGAN
and pairwise distance. The original DiscoGAN implementation was used, but due to differences
in evaluation (different regressors) these numbers are not compatible with the graph shown in
DiscoGAN.
For car2head, DiscoGAN?s solution produces mirror images and combination of DiscoGAN?s
circularity constraint with the distance constraint produces a solution that is rotated by 90? . We
consider these biases as ambiguities in the mapping and not as mistakes and, therefore, remove the
mean error prior to computing the RMSE. In this experiment, distance outperforms all other methods.
The combination of both methods is less competitive than both, perhaps since each method pulls
toward a different solution. Self distance, is worse than circularity in this dataset.
7
Another set of experiments arises from considering face images with and without a certain property.
CelebA [26, 14] was annotated for multiple attributes including the person?s gender, hair color, and
the existence of glasses in the image. Following [11] we perform mapping between two values of each
of these three properties. The results are shown in the supplementary material with some examples
in Fig. 3. It is evident that the DiscoGAN method (using the unmodified authors? implementation)
presents many more failure cases than our pair based method. The self-distance method was
implemented with the top and bottom image halves, instead of left to right distances, since faces are
symmetric. This method also seems to outperform DiscoGAN.
In order to evaluate how well the face translation was performed, we use the representation layer of
VGG faces [16] on the image in A and its output in B. One can assume that two images that match
will have many similar features and so the VGG representation will be similar. The cosine similarities,
as evaluated between input images and their mapped versions, are shown in Tab. 4. In all cases, the
pair-distance produces more similar input-output faces. Self-distance performs slightly worse than
pairs, but generally better than DiscoGAN. Applying circularity together with pair-distance, provides
the best results but requires, unlike the distance, learning both sides simultaneously.
While we create images that better match in the face descriptor metric, our ability to create images
that are faithful to the second distribution is not impaired. This is demonstrated by learning a linear
classifier between the two domains based on the training samples and then applying it to a set of
test image before and after mapping. The separation accuracy between the input test image and the
mapped version is also shown in Tab. 4. As can be seen, the separation ability of our method is
similar to that of DiscoGAN (it arises from the shared GAN terms).
We additionally perform a user study to asses the quality of our results. The user is first presented
with a set of real images from the dataset. Then, 50 random pairs of images are presented to a user for
a second, one trained using DiscoGAN and one using our method. The user is asked to decide which
image looks more realistic. The test was performed on 22 users. On shoes to handbags translation,
our translation performed better on 65% of the cases. For handbags to shoes, the score was 87%.
For male to female, both methods showed a similar realness score (51% to 49% of DiscoGAN?s).
We, therefore, asked a second question: given the face of a male, which of the two generated female
variants is a better fit to the original face. Our method wins 88% of the time.
In addition, in the supplementary material we compare the losses of the GAN discriminator for the
various methods and show that these values are almost identical. We also measure the losses of the
various methods during test, even if these were not directly optimized. For example, despite this
constraints not being enforced, the distance based methods seem to present a low circularity loss,
while DiscoGAN presents a relatively higher distance losses.
Sample results of mapping shoes to handbags and edges to shoes and vice versa using the DiscoGAN
baseline architecture are shown in Fig. 3. More results are shown in the supplementary. Visually, the
results of the distance-based approach seem better then DiscoGAN while the results of self-distance
are somewhat worse. The combination of DiscoGAN and distance usually works best.
Models based on CycleGAN Using the CycleGAN architecture we map horses to zebras, see
Fig. 4 and supplementary material for examples. Note that on the zebra to horse mapping, all methods
fail albeit in different ways. Subjectively, it seems that the distance + cycle method shows the most
promise in this translation.
In order to obtain numerical results, we use the baseline CycleGAN method as well as our methods
in order to translate from Street View House Numbers (SVHN) [15] to MNIST [12]. Accuracy is
then measured in the MNIST space by using a neural net trained for this task. Results are shown in
Tab. 3 and visually in the Supplementary. While the pairwise distance based method improves upon
the baseline method, there is still a large gap between the unsupervised and semi-supervised setting
presented in [22], which achieves much higher results. This can be explained by the large amount of
irrelevant information in the SVHN images (examples are shown in the supplementary). Combining
the distance based constraint with the circularity one does not work well on this dataset.
We additionally performed a qualitative evaluation using FCN score as in [28]. The FCN metric
evaluates the interoperability images by taking a generated cityscape image and generating a label
using semantic segmentation algorithm. The generated label can then be compared to the ground
truth label. FCN results are given as three measures: per-pixel accuracy, per-class accuracy and Class
8
Input
Disco GAN
Distance
Distance
+cycle
Self distance
(a)
(b)
(c)
(d)
(e)
(f)
Figure 3: Translations using various methods on the celebA dataset: (a,b) Male to and from Female.
(c,d) Blond to and from black hair. (e,f) With eyeglasses to from without eyeglasses.
Input
Disco/
CycleGAN
Distance
Distance
+cycle
Self distance
(a)
(b)
(c)
(d)
(e)
(f)
Figure 4: (a,b) Handbags to and from shoes. (c,d) Edges to/from shoes. (e,f) Horse to/from zebra.
IOU. Our distance GAN method is preferable on all three scores (0.53 vs. 0.52, 0.19 vs. 0.17, and
0.11 vs 0.11, respectively). The paired t-test p-values are 0.29, 0.002 and 0.42 respectively. In a user
study similar to the one for DiscoGAN above, our cityscapes translation scores 71% for realness
when comparing to CycleGAN?s. When looking at similarity to the ground truth image we score
68%.
5
Conclusion
We have proposed an unsupervised distance-based loss for learning a single mapping (without its
inverse), which empirically outperforms the circularity loss. It is interesting to note that the new
loss is applied to raw RGB image values. This is in contrast to all of the work we are aware of that
computes image similarity. Clearly, image descriptors or low-layer network activations can be used.
However, by considering only RGB values, we not only show the general utility of our method, but
also further demonstrate that a minimal amount of information is needed in order to form analogies
between two related domains.
Acknowledgements
This project has received funding from the European Research Council (ERC) under the European
Union?s Horizon 2020 research and innovation programme (grant ERC CoG 725974). The authors
would like to thank Laurens van der Maaten and Ross Girshick for insightful discussions.
9
References
[1] Werner Van Belle. Correlation between the inproduct and the sum of absolute differences is
-0.8485 for uniform sampled signals on [-1:1]. Available at http: // werner. yellowcouch.
org/ Papers/ sadvssip/ index. html , 2006.
[2] Konstantinos Bousmalis, Nathan Silberman, David Dohan, Dumitru Erhan, and Dilip Krishnan.
Unsupervised pixel-level domain adaptation with generative adversarial networks. In CVPR,
2017.
[3] Konstantinos Bousmalis, George Trigeorgis, Nathan Silberman, Dilip Krishnan, and Dumitru
Erhan. Domain separation networks. In D. D. Lee, M. Sugiyama, U. V. Luxburg, I. Guyon, and
R. Garnett, editors, Advances in Neural Information Processing Systems 29, pages 343?351.
Curran Associates, Inc., 2016.
[4] Sanja Fidler, Sven Dickinson, and Raquel Urtasun. 3d object detection and viewpoint estimation
with a deformable 3d cuboid model. In NIPS, 2012.
[5] Leon A. Gatys, Alexander S. Ecker, and Matthias Bethge. Image style transfer using convolutional neural networks. In CVPR, 2016.
[6] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil
Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. In NIPS. 2014.
[7] Judy Hoffman, Dequan Wang, Fisher Yu, and Trevor Darrell. Fcns in the wild: Pixel-level
adversarial and constraint-based adaptation. 12 2016.
[8] Sergey Ioffe and Christian Szegedy. Batch normalization: Accelerating deep network training
by reducing internal covariate shift. In ICML, 2015.
[9] Phillip Isola, Jun-Yan Zhu, Tinghui Zhou, and Alexei A Efros. Image-to-image translation with
conditional adversarial networks. In CVPR, 2017.
[10] Justin Johnson, Alexandre Alahi, and Li Fei-Fei. Perceptual losses for real-time style transfer
and super-resolution. In ECCV, 2016.
[11] Taeksoo Kim, Moonsu Cha, Hyunsoo Kim, Jungkwon Lee, and Jiwon Kim. Learning to discover
cross-domain relations with generative adversarial networks. arXiv preprint arXiv:1703.05192,
2017.
[12] Yann LeCun and Corinna Cortes. MNIST handwritten digit database. 2010.
[13] Ming-Yu Liu and Oncel Tuzel. Coupled generative adversarial networks. In NIPS, pages
469?477. 2016.
[14] Ziwei Liu, Ping Luo, Xiaogang Wang, and Xiaoou Tang. Deep learning face attributes in the
wild. In Proceedings of International Conference on Computer Vision (ICCV), 2015.
[15] Yuval Netzer, Tao Wang, Adam Coates, Alessandro Bissacco, Bo Wu, and Andrew Y. Ng.
Reading digits in natural images with unsupervised feature learning. In NIPS Workshop on
Deep Learning and Unsupervised Feature Learning, 2011.
[16] O. M. Parkhi, A. Vedaldi, and A. Zisserman. Deep face recognition. In British Machine Vision
Conference, 2015.
[17] Pascal Paysan, Reinhard Knothe, Brian Amberg, Sami Romdhani, and Thomas Vetter. A 3d
face model for pose and illumination invariant face recognition. In AVSS, 2009.
[18] Alec Radford, Luke Metz, and Soumith Chintala. Unsupervised representation learning with
deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434, 2015.
[19] Olaf Ronneberger, Philipp Fischer, and Thomas Brox. U-net: Convolutional networks for
biomedical image segmentation. In MICCAI, 2015.
10
[20] Ashish Shrivastava, Tomas Pfister, Oncel Tuzel, Josh Susskind, Wenda Wang, and Russ Webb.
Learning from simulated and unsupervised images through adversarial training. arXiv preprint
arXiv:1612.07828, 2016.
[21] Ilya Sutskever, Rafal Jozefowicz, Karol Gregor, Danilo Rezende, Tim Lillicrap, and Oriol
Vinyals. Towards principled unsupervised learning. In ICLR workshop, 2016.
[22] Yaniv Taigman, Adam Polyak, and Lior Wolf. Unsupervised cross-domain image generation.
In International Conference on Learning Representations (ICLR), 2017.
[23] D. Ulyanov, V. Lebedev, A. Vedaldi, and V. Lempitsky. Texture networks: Feed-forward
synthesis of textures and stylized images. In ICML, 2016.
[24] Yingce Xia, Di He, Tao Qin, Liwei Wang, Nenghai Yu, Tie-Yan Liu, and Wei-Ying Ma. Dual
learning for machine translation. arXiv preprint arXiv:1611.00179, 2016.
[25] X.Mao, Q.Li, H.Xie, R.Y. Lau, and Z.Wang. Multi-class generative adversarial networks with
the l2 loss function. arXiv preprint arXiv:1611.04076, 2016.
[26] Shuo Yang, Ping Luo, Chen Change Loy, and Xiaoou Tang. From facial parts responses to face
detection: A deep learning approach. In ICCV, pages 3676?3684, 2015.
[27] Zili Yi, Hao Zhang, Ping Tan Gong, et al. Dualgan: Unsupervised dual learning for image-toimage translation. arXiv preprint arXiv:1704.02510, 2017.
[28] Jun-Yan Zhu, Taesung Park, Phillip Isola, and Alexei A Efros. Unpaired image-to-image
translation using cycle-consistent adversarial networkss. arXiv preprint arXiv:1703.10593,
2017.
11
| 6677 |@word cnn:1 version:2 fcns:1 seems:3 norm:1 cha:1 rgb:3 d0k:5 initial:1 configuration:1 liu:3 score:6 ours:1 hyunsoo:1 deconvolutional:2 outperforms:3 existing:2 com:1 comparing:2 luo:2 activation:5 numerical:3 realistic:2 distant:1 happen:1 shape:1 christian:1 remove:1 v:4 generative:9 selected:1 half:6 alec:1 bissacco:1 provides:2 completeness:1 location:3 philipp:1 org:1 zhang:1 direct:1 qualitative:1 consists:4 wild:2 inside:1 introduce:1 x0:5 pairwise:14 expected:2 dist:7 gatys:1 multi:1 realness:2 ming:1 little:4 soumith:1 considering:4 provided:3 project:1 moreover:1 matched:1 discover:1 mass:1 advent:1 israel:1 what:4 interpreted:1 transformation:8 nj:2 every:3 alahi:1 tie:1 preferable:3 k2:1 classifier:1 sherjil:1 unit:5 grant:1 before:4 local:1 sd:1 mistake:1 ulyanov:1 despite:1 id:1 approximately:2 might:1 black:2 luke:1 collapse:1 practical:1 faithful:1 enforces:1 lecun:1 practice:2 block:1 definite:1 differs:1 union:1 digit:3 susskind:1 displacement:3 tuzel:2 area:1 empirical:1 yan:3 significantly:1 vedaldi:2 matching:7 projection:1 pre:1 word:1 vetter:1 ronneberger:1 liwei:1 suggest:1 cannot:1 unlabeled:1 operator:1 applying:2 map:10 demonstrated:6 ecker:1 maximizing:1 go:1 resolution:3 tomas:1 recovery:2 pouget:1 dominate:1 pull:1 coordinate:1 justification:2 target:12 tan:1 user:6 exact:1 losing:1 dickinson:1 us:3 curran:1 goodfellow:1 pa:6 element:3 associate:1 recognition:2 stripe:2 database:1 bottom:2 constancy:1 role:1 preprint:7 wang:6 capture:1 calculate:1 ensures:1 cycle:10 trade:1 alessandro:1 principled:1 asked:2 warde:1 trained:4 weakly:1 tight:1 upon:1 learner:1 translated:5 discogan:28 exi:1 stylized:1 xiaoou:2 various:3 train:3 separated:2 distinct:1 sven:1 effective:1 horse:22 formation:1 pearson:4 whose:1 jean:1 larger:1 supplementary:6 cvpr:3 encoder:3 ability:3 fischer:1 jointly:3 transform:1 advantage:1 net:4 matthias:1 reconstruction:1 propose:3 adaptation:4 qin:1 combining:3 translate:1 deformable:1 sutskever:1 olaf:1 impaired:1 requirement:1 darrell:1 yaniv:1 produce:4 generating:1 bousmalis:2 adam:2 gab:41 rotated:4 object:3 depending:4 andrew:1 tim:1 gong:1 pose:1 measured:1 minor:1 school:1 received:1 strong:1 recovering:1 implemented:1 involves:1 jiwon:1 differ:1 direction:5 iou:1 laurens:1 annotated:1 attribute:2 filter:2 centered:1 translating:4 public:1 material:3 require:2 suffices:1 preliminary:1 brian:1 hold:2 considered:1 ground:6 visually:3 jungkwon:1 mapping:66 major:1 vary:1 achieves:1 efros:2 interoperability:1 estimation:1 outperformed:1 label:8 currently:1 ross:1 council:1 largest:1 vice:3 create:4 hoffman:1 clearly:1 super:1 modified:2 rather:1 zhou:1 rezende:1 improvement:1 likelihood:1 contrast:1 adversarial:15 baseline:6 dilip:2 glass:3 kim:3 entire:1 typically:2 relation:1 transformed:1 tao:2 pixel:9 dual:2 html:1 pascal:1 proposes:1 spatial:2 brox:1 field:2 once:1 aware:1 having:2 beach:1 ng:1 identical:2 park:1 look:1 unsupervised:18 yu:3 yaw:1 icml:2 fcn:3 foreground:1 mimic:1 others:1 report:1 minimized:2 mirza:1 employ:9 few:1 strided:1 yoshua:1 simultaneously:5 preserve:1 replaced:1 maintain:1 n1:1 detection:2 interest:1 highly:4 alexei:2 evaluation:2 alignment:2 male:4 extreme:1 circularity:16 farley:1 xb:1 implication:1 edge:5 lcycle:5 paysan:1 necessary:1 netzer:1 facial:1 desired:1 circle:2 girshick:1 minimal:1 instance:1 column:1 earlier:1 downside:1 unmodified:1 werner:2 deviation:2 subset:2 uniform:2 examining:1 successful:1 johnson:1 varies:1 kxi:1 trigeorgis:1 st:1 person:1 international:2 sensitivity:1 lee:2 off:1 regressor:1 together:2 transfered:1 bethge:1 gans:3 ashish:1 ilya:1 squared:1 again:1 ambiguity:1 synthesis:1 rafal:1 unmatched:1 worse:3 style:8 return:1 li:2 szegedy:1 stride:2 includes:2 inc:1 explicitly:1 eyeglass:2 performed:8 root:1 view:1 tab:5 competitive:3 recover:2 maintains:1 metz:1 belle:1 rmse:3 ass:1 minimize:1 publicly:1 square:2 convolutional:7 variance:2 largely:1 accuracy:8 maximized:2 descriptor:3 directional:1 raw:1 handwritten:1 accurately:1 produced:1 rus:1 published:1 history:1 ping:3 strongest:1 romdhani:1 sharing:1 trevor:1 facebook:1 failure:1 evaluates:1 chintala:1 associated:1 lior:2 di:1 sampled:2 dataset:11 nenghai:1 color:3 car:7 improves:2 segmentation:3 back:4 alexandre:1 feed:1 higher:2 supervised:4 follow:1 danilo:1 zisserman:1 wei:1 xie:1 response:1 done:5 cyclegan:18 evaluated:2 box:1 furthermore:1 just:3 xa:1 xk1:1 biomedical:1 correlation:14 miccai:1 working:3 receives:1 mehdi:1 minibatch:1 quality:1 perhaps:1 aviv:1 riding:1 usa:1 usage:1 lillicrap:1 requiring:4 contain:1 effect:2 normalized:3 phillip:2 fidler:1 symmetric:1 dualgan:1 semantic:6 indistinguishable:1 during:2 self:17 essence:1 cosine:4 criterion:1 evident:3 demonstrate:2 performs:2 l1:4 svhn:2 image:112 recently:2 funding:1 common:1 sigmoid:2 rotation:2 empirically:1 untouched:1 analog:8 linking:1 approximates:1 belong:1 he:1 significant:1 jozefowicz:1 versa:3 ai:1 zebra:20 similarly:4 zeroing:1 erc:2 sugiyama:1 language:1 mapper:3 sanja:1 supervision:8 similarity:8 cityscape:6 subjectively:1 recent:2 female:4 showed:1 irrelevant:1 reverse:2 scenario:1 certain:2 der:1 yi:1 seen:4 captured:1 preserving:1 additional:3 somewhat:1 george:1 employed:2 isola:2 tinghui:1 determine:1 signal:1 semi:1 multiple:3 full:1 reduces:1 taesung:1 match:5 cross:5 long:1 justifying:1 post:1 paired:2 variant:1 underlies:1 hair:3 multilayer:1 vision:3 expectation:1 metric:2 arxiv:14 iteration:1 normalization:4 sergey:1 achieved:1 receive:1 preserved:3 addition:2 interval:1 source:11 unlike:2 tend:1 db:8 seem:2 symmetrically:2 yang:1 split:2 bengio:1 sami:1 krishnan:2 switch:1 xj:8 relu:4 fit:1 architecture:10 taeksoo:1 zili:1 polyak:1 reduce:1 tradeoff:1 translates:1 vgg:3 konstantinos:2 shift:1 whether:2 expression:1 utility:1 accelerating:1 suffer:1 cause:1 deep:6 generally:2 fake:2 amount:5 reinhard:1 category:1 unpaired:1 http:2 outperform:1 canonical:1 coates:1 estimated:1 per:4 blue:1 discrete:1 promise:1 four:1 fractionally:1 pb:5 changing:1 graph:1 sum:4 enforced:1 luxburg:1 inverse:4 angle:6 taigman:1 raquel:1 almost:2 reasonable:1 decide:1 guyon:1 yann:1 separation:6 wu:1 maaten:1 capturing:1 layer:10 followed:2 courville:1 exa:1 nonnegative:4 xiaogang:1 constraint:36 fei:2 generates:1 nathan:2 toimage:1 leon:1 performing:1 rendered:1 relatively:2 combination:5 slightly:2 oncel:2 modification:1 explained:1 invariant:2 iccv:2 lau:1 sided:7 bing:1 precomputed:3 r3:1 fail:1 needed:3 available:2 operation:1 celeba:3 enforce:1 batch:4 corinna:1 existence:1 original:6 thomas:2 top:3 assumes:1 gan:10 maintaining:2 k1:5 build:1 approximating:1 classical:1 gregor:1 silberman:2 skin:3 added:1 question:1 looked:1 occurs:1 gradient:2 win:1 iclr:2 distance:88 thank:1 mapped:7 simulated:1 majority:1 decoder:2 street:1 r64:1 urtasun:1 reason:1 enforcing:1 toward:1 ozair:1 assuming:1 lgan:4 code:2 modeled:1 index:1 mini:1 convincing:1 minimizing:1 innovation:1 ying:1 loy:1 mostly:1 webb:1 hao:1 negative:3 ziwei:1 implementation:4 perform:4 cogan:2 convolution:1 av:1 datasets:5 descent:1 variability:1 head:4 looking:1 sharp:1 yingce:1 introduced:1 david:2 pair:25 required:1 kl:3 optimized:2 discriminator:8 distinction:1 learned:6 nip:5 justin:1 suggested:2 below:3 usually:2 pattern:1 reading:1 green:2 including:1 ia:1 natural:2 indicator:1 residual:1 zhu:2 gba:17 github:1 technology:1 created:1 jun:2 coupled:1 review:1 literature:4 l2:2 prior:1 acknowledgement:1 fully:1 crossdomain:1 loss:23 expect:2 permutation:5 generation:2 rationale:1 interesting:1 analogy:1 generator:6 degree:3 sufficient:1 consistent:1 editor:1 leaning:1 viewpoint:1 share:4 translation:19 row:1 eccv:1 compatible:1 changed:1 last:1 side:2 allow:1 bias:1 karol:1 stripped:1 sification:1 face:14 taking:1 absolute:6 leaky:2 van:2 xia:1 valid:2 computes:1 author:2 made:2 forward:1 regressors:1 programme:1 employing:2 erhan:2 approximate:2 obtains:1 cuboid:1 global:1 ioffe:1 assumed:1 xi:3 triplet:1 table:4 additionally:2 moonsu:1 channel:3 learn:5 transfer:4 ca:1 correlated:2 shrivastava:1 tel:1 synthesizes:1 improving:1 complex:1 european:2 domain:67 shuo:1 da:4 official:1 putin:2 did:2 spread:2 garnett:1 rh:2 noise:2 subsample:1 allowed:1 xu:1 fig:11 strengthened:1 judy:1 mao:1 house:1 perceptual:1 ib:1 learns:2 admissible:2 ian:1 tang:2 british:1 cog:1 dumitru:2 specific:2 covariate:1 showing:1 insightful:1 dk:7 abadie:1 cortes:1 evidence:1 incorporating:1 exists:1 consist:1 mnist:4 adding:2 albeit:1 workshop:2 mirror:1 texture:4 illumination:1 kx:2 horizon:1 gap:1 amberg:1 chen:1 knothe:1 lt:1 simply:1 likely:2 shoe:13 visual:1 happening:1 prevents:1 josh:1 vinyals:1 bo:1 applies:1 radford:1 gender:2 wolf:1 truth:6 minibatches:2 ma:1 conditional:1 lempitsky:1 goal:3 viewed:3 identity:3 towards:1 replace:1 shared:2 content:4 change:4 feasible:1 fisher:1 parkhi:1 except:1 uniformly:1 reducing:1 yuval:1 disco:2 lebedev:1 blond:2 pfister:1 partly:1 experimental:1 meaningful:1 aaron:1 formally:2 internal:1 handbag:11 arises:2 alexander:1 oriol:1 evaluate:1 tested:2 phenomenon:1 ex:3 |
6,276 | 6,678 | Deep Mean-Shift Priors for Image Restoration
Siavash A. Bigdeli
University of Bern
[email protected]
Meiguang Jin
University of Bern
[email protected]
Paolo Favaro
University of Bern
[email protected]
Matthias Zwicker
University of Bern, and University of Maryland, College Park
[email protected]
Abstract
In this paper we introduce a natural image prior that directly represents a Gaussiansmoothed version of the natural image distribution. We include our prior in a
formulation of image restoration as a Bayes estimator that also allows us to solve
noise-blind image restoration problems. We show that the gradient of our prior
corresponds to the mean-shift vector on the natural image distribution. In addition,
we learn the mean-shift vector field using denoising autoencoders, and use it in a
gradient descent approach to perform Bayes risk minimization. We demonstrate
competitive results for noise-blind deblurring, super-resolution, and demosaicing.
1
Introduction
Image restoration tasks, such as deblurring and denoising, are ill-posed problems, whose solution
requires effective image priors. In the last decades, several natural image priors have been proposed,
including total variation [27], gradient sparsity priors [12], models based on image patches [5], and
Gaussian mixtures of local filters [24], just to name a few of the most successful ideas. See Figure 1
for a visual comparison of some popular priors. More recently, deep learning techniques have been
used to construct generic image priors.
Here, we propose an image prior that is directly based on an estimate of the natural image probability
distribution. Although this seems like the most intuitive and straightforward idea to formulate a prior,
only few previous techniques have taken this route [20]. Instead, most priors are built on intuition or
statistics of natural images (e.g., sparse gradients). Most previous deep learning priors are derived
in the context of specific algorithms to solve the restoration problem, but it is not clear how these
priors relate to the probability distribution of natural images. In contrast, our prior directly represents
the natural image distribution smoothed with a Gaussian kernel, an approximation similar to using
a Gaussian kernel density estimate. Note that we cannot hope to use the true image probability
distribution itself as our prior, since we only have a finite set of samples from this distribution. We
show a visual comparison in Figure 1, where our prior is able to capture the structure of the underlying
image, but others tend to simplify the texture to straight lines and sharp edges.
We formulate image restoration as a Bayes estimator, and define a utility function that includes the
smoothed natural image distribution. We approximate the estimator with a bound, and show that
the gradient of the bound includes the gradient of the logarithm of our prior, that is, the Gaussian
smoothed density. In addition, the gradient of the logarithm of the smoothed density is proportional
to the mean-shift vector [8], and it has recently been shown that denoising autoencoders (DAEs) learn
such a mean-shift vector field for a given set of data samples [1, 4]. Hence we call our prior a deep
mean-shift prior, and our framework is an example of Bayesian inference using deep learning.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
Input
Our prior
BM3D [9]
EPLL [39]
FoE [26]
SF [29]
Figure 1: Visualization of image priors using the method by Shaham et al. [30]: Our deep mean-shift
prior learns complex structures with different curvatures. Other priors prefer simpler structures like
lines with small curvature or sharp corners.
We demonstrate image restoration using our prior for noise-blind deblurring, super-resolution, and
image demosaicing, where we solve Bayes estimation using a gradient descent approach. We achieve
performance that is competitive with the state of the art for these applications. In summary, the main
contributions of this paper are:
? A formulation of image restoration as a Bayes estimator that leverages the Gaussian
smoothed density of natural images as its prior. In addition, the formulation allows us
to solve noise-blind restoration problems.
? An implementation of the prior, which we call deep mean-shift prior, that builds on denoising
autoencoders (DAEs). We rely on the observation that DAEs learn a mean-shift vector field,
which is proportional to the gradient of the logarithm of the prior.
? Image restoration techniques based on gradient-descent risk minimization with competitive
results for noise-blind image deblurring, super-resolution, and demosaicing. 1
2
Related Work
Image Priors. A comprehensive review of previous image priors is outside the scope of this paper.
Instead, we refer to the overview by Shaham et al. [30], where they propose a visualization technique
to compare priors. Our approach is most related to techniques that leverage CNNs to learn image
priors. These techniques build on the observation by Venkatakrishnan et al. [31] that many algorithms
that solve image restoration via MAP estimation only need the proximal operator of the regularization
term, which can be interpreted as a MAP denoiser [22]. Venkatakrishnan et al. [31] build on the
ADMM algorithm and propose to replace the proximal operator of the regularizer with a denoiser
such as BM3D [9] or NLM [5]. Unsurprisingly, this inspired several researchers to learn the proximal
operator using CNNs [6, 38, 33, 22]. Meinhardt et al. [22] consider various proximal algorithms
including the proximal gradient method, ADMM, and the primal-dual hybrid gradient method, where
in each case the proximal operator for the regularizer can be replaced by a neural network. They
show that no single method will produce systematically better results than the others.
In the proximal techniques the relation between the proximal operator of the regularizer and the
natural image probability distribution remains unclear. In contrast, we explicitly use the Gaussiansmoothed natural image distribution as a prior, and we show that we can learn the gradient of its
logarithm using a denoising autoencoder.
Romano et al. [25] designed a prior model that is also implemented by a denoiser, but that does not
build on a proximal formulation such as ADMM. Interestingly, the gradient of their regularization
term boils down to the residual of the denoiser, that is, the difference between its input and output,
which is the same as in our approach. However, their framework does not establish the connection
between the prior and the natural image probability distribution, as we do. Finally, Bigdeli and
Zwicker [4] formulate an energy function, where they use a Denoising Autoencoder (DAE) network
for the prior, as in our approach, but they do not address the case of noise-blind restoration.
Noise- and Kernel-Blind Deconvolution. Kernel-blind deconvolution has seen the most effort
recently, while we support the fully (noise and kernel) blind setting. Noise-blind deblurring is usually
1
The source code of the proposed method is available at https://github.com/siavashbigdeli/DMSP.
2
performed by first estimating the noise level and then restoration with the estimated noise. Jin et
al. [14] proposed a Bayes risk formulation that can perform deblurring by adaptively changing the
regularization without the need of the noise variance estimate. Zhang et al. [35, 36] explored a
spatially-adaptive sparse prior and scale-space formulation to handle noise- or kernel-blind deconvolution. These methods, however, are tailored specifically to image deconvolution. Also, they only
handle the noise- or kernel-blind case, but not fully blind.
3
Bayesian Formulation
We assume a standard model for image degradation,
n ? N (0, ?n2 ),
y = k ? ? + n,
(1)
where ? is the unknown image, k is the blur kernel, n is zero-mean Gaussian noise with variance ?n2 ,
and y is the observed degraded image. We restore an estimate x of the unknown image by defining
and maximizing an objective consisting of a data term and an image likelihood,
argmax ?(x) = data(x) + prior(x).
(2)
x
Our core contribution is to construct a prior that corresponds to the logarithm of the Gaussiansmoothed probability distribution of natural images. We will optimize the objective using gradient
descent, and leverage the fact that we can learn the gradient of the prior using a denoising autoencoder
(DAE). We next describe how we define our objective by formulating a Bayes estimator in Section 3.1,
then explain how we leverage DAEs to obtain the gradient of our prior in Section 3.2, describe our
gradient descent approach in Section 3.3, and finally our image restoration applications in Section 4.
3.1
Defining the Objective via a Bayes Estimator
A typical approach to solve the restoration problem is via a maximum a posteriori (MAP) estimate,
where one considers the posterior distribution of the restored image p(x|y) ? p(y|x)p(x), derives an
objective consisting of a sum of data and prior terms by taking the logarithm of the posterior, and
maximizes it (minimizes the negative log-posterior, respectively). Instead, we will compute a Bayes
estimator x for the restoration problem by maximizing the posterior expectation of a utility function,
Z
Ex? [G(?
x, x)] = G(?
x, x)p(y|?
x)p(?
x)d?
x
(3)
where G denotes the utility function (e.g., a Gaussian), which encourages its two arguments to be
similar. This is a generalization of MAP, where the utility is a Dirac impulse.
Ideally, we would like to use the true data distribution as the prior p(?
x). But we only have data
samples, hence we cannot learn this exactly. Therefore, we introduce a smoothed data distribution
Z
0
p (x) = E? [p(x + ?)] = g? (?)p(x + ?)d?,
(4)
where ? has a Gaussian distribution with zero-mean and variance ? 2 , which is represented by the
smoothing kernel g? . The key idea here is that it is possible to estimate the smoothed distribution
p0 (x) or its gradient from sample data. In particular, we will need the gradient of its logarithm, which
we will learn using denoising autoencoders (DAEs). We now define our utility function as
G(?
x, x) = g? (?
x ? x)
p0 (x)
.
p(?
x)
(5)
where we use the same Gaussian function g? with standard deviation ? as introduced for the smoothed
distribution p0 . This penalizes the estimate x if the latent parameter x
? is far from it. In addition, the
term p0 (x)/p(?
x) penalizes the estimate if its smoothed density is lower than the true density of the
latent parameter. Unlike the utility in Jin et al. [14], this approach will allow us to express the prior
directly using the smoothed distribution p0 .
By inserting our utility function into the posterior expected utility in Equation (3) we obtain
Z
Z
Ex? [G(?
x, x)] = g? ()p(y|x + ) g? (?)p(x + ?)d?d,
3
(6)
where the true density p(?
x) canceled out, as desired, and we introduced the variable substitution
=x
? ? x.
We finally formulate our objective by taking the logarithm of the expected utility in Equation (6),
and introducing a lower bound that will allow us to split Equation (6) into a data term and an image
likelihood. By exploiting the concavity of the log function, we apply Jensen?s inequality and get our
objective ?(x) as
Z
Z
log Ex? [G(?
x, x)] = log g? ()p(y|x + ) g? (?)p(x + ?)d?d
"
#
Z
Z
?
g? () log p(y|x + ) g? (?)p(x + ?)d? d
Z
=
Z
g? () log p(y|x + )d + log
|
{z
} |
Data term data(x)
g? (?)p(x + ?)d? = ?(x).
{z
}
(7)
Image likelihood prior(x)
Image Likelihood. We denote the image likelihood as
Z
prior(x) = log g? (?)p(x + ?)d?.
(8)
The key observation here is that our prior expresses the image likelihood as the logarithm of the
Gaussian-smoothed true natural image distribution p(x), which is similar to a kernel density estimate.
Data Term. Given that the degradation noise is Gaussian, we see that [14]
Z
|y ? k ? x|2
?2
data(x) = g? () log p(y|x + )d = ?
? M 2 |k|2 ? N log ?n + const, (9)
2
2?n
2?n
where M and N denote the number of pixels in x and y respectively. This will allow us to address
noise-blind problems as we will describe in detail in Section 4.
3.2
Gradient of the Prior via Denoising Autoencoders (DAE)
A key insight of our approach is that we can effectively learn the gradients of our prior in Equation (8)
using denoising autoencoders (DAEs). A DAE r? is trained to minimize [32]
LDAE = E?,x |x ? r? (x + ?)|2 ,
(10)
where the expectation is over all images x and Gaussian noise ? with variance ? 2 , and r? indicates
that the DAE was trained with noise variance ? 2 . Alain et al. [1] show that the output r? (x) of the
optimal DAE (by assuming unlimited capacity) is related to the true data distribution p(x) as
R
g? (?)p(x ? ?)?d?
E? [p(x ? ?)?]
r? (x) = x ?
=x? R
(11)
E? [p(x ? ?)]
g? (?)p(x ? ?)d?
where the noise has a Gaussian distribution g? with standard deviation ?. This is simply a continuous
formulation of mean-shift, and g? corresponds to the smoothing kernel in our prior, Equation (8).
To obtain the relation between the DAE and the desired gradient of our prior, we first rewrite the
numerator in Equation (11) using the Gaussian derivative definition to remove ?, that is
Z
Z
Z
g? (?)p(x ? ?)?d? = ?? 2 ?g? (?)p(x ? ?)d? = ?? 2 ? g? (?)p(x ? ?)d?,
(12)
where we used the Leibniz rule to interchange the ? operator with the integral. Plugging this back
into Equation (11), we have
R
Z
? 2 ? g? (?)p(x ? ?)d?
r? (x) = x + R
= x + ? 2 ? log g? (?)p(x ? ?)d?.
(13)
g? (?)p(x ? ?)d?
One can now see that the DAE error, that is, the difference r? (x) ? x between the output of the DAE
and its input, is the gradient of the image likelihood in Equation (8). Hence, a main result of our
approach is that we can write the gradient of our prior using the DAE error,
Z
1
? prior(x) = ? log g? (?)p(x + ?)d? = 2 r? (x) ? x .
(14)
?
4
1
T
t?1
2 K (Kx
?n
? y) ? ?priorsL (xt?1 ) 2. u
? = ??
u ? ?ut
NB:
1. ut =
NA:
1. ut = ?t K T (Kxt?1 ? y) ? ?priorsL (xt?1 )
4. v t = ?t xT (K t?1 xt?1 ? y) + M ? 2 k t?1
KE:
3. xt = xt?1 + u
?
2. u
? = ??
u ? ?ut
3. xt = xt?1 + u
?
5. v? = ?k v? ? ?k v t
6. k t = k t?1 + v?
Table 1: Gradient descent steps for non-blind (NB), noise-blind (NA), and kernel-blind (KE) image
deblurring. Kernel-blind deblurring involves the steps for (NA) and (KE) to update image and kernel.
3.3
Stochastic Gradient Descent
We consider the optimization as minimization of the negative of our objective ?(x) and refer to it as
gradient descent. Similar to Bigdeli and Zwicker [4], we observed that the trained DAE is overfitted
to noisy images. Because of the large gap in dimensionality between the embedding space and the
natural image manifold, the vast majority of training inputs (noisy images) for the DAE lie at a
distance very close to ? from the natural image manifold. Hence, the DAE cannot effectively learn
mean-shift vectors for locations that are closer than ? to the natural image manifold. In other words,
our DAE does not produce meaningful results for input images that do not exhibit noise close to the
DAE training ?.
To address this issue, we reformulate our prior to perform stochastic gradient descent steps that
include noise sampling. We rewrite our prior from Equation (8) as
Z
prior(x) = log g? (?)p(x + ?)d?
(15)
Z
Z
= log g?2 (?2 ) g?1 (?1 )p(x + ?1 + ?2 )d?1 d?2
(16)
"
#
Z
Z
?
g?2 (?2 ) log
g?1 (?1 )p(x + ?1 + ?2 )d?1 d?2 = priorL (x),
(17)
where ?12 + ?22 = ? 2 , we used the fact that two Gaussian convolutions are equivalent to a single
convolution with a Gaussian whose variance is the sum of the two, and we applied Jensen?s inequality
again. This leads to a new lower bound for the prior, which we call priorL (x). Note that the bound
proposed by Jin et al. [14] corresponds to the special case where ?1 = 0 and ?2 = ?.
We address our DAE overfitting issue by using the new lower bound priorL (x) with ?1 = ?2 = ??2 .
Its gradient is
Z
2
g ??2 (?2 ) r ??2 (x + ?2 ) ? (x + ?2 ) d?2 .
?priorL (x) = 2
(18)
?
In practice, computing the integral over ?2 is not possible at runtime. Instead, we approximate the
integral with a single noise sample, which leads to the stochastic evaluation of the gradient as
2
(19)
?priorsL (x) = 2 r ??2 (x + ?2 ) ? x ,
?
where ?2 ? N (0, ?2 ). This addresses the overfitting issue, since it means we add noise each time
before we evaluate the DAE. Given the stochastically sampled gradient of the prior, we apply a
gradient descent approach with momentum that consists of the following steps:
1. ut = ?? data(xt?1 ) ? ? priorsL (xt?1 )
2. u
? = ??
u ? ?ut
3. xt = xt?1 + u
?
(20)
where ut is the update step for x at iteration t, u
? is the running step, and ? and ? are the momentum
and step-size.
4
Image Restoration using the Deep Mean-Shift Prior
We next describe the detailed gradient descent steps, including the derivatives of the data term, for
different image restoration tasks. We provide a summary in Table 1. For brevity, we omit the role of
downsampling (required for super-resolution) and masking.
5
Method
FD [18]
EPLL [39]
RTF-6 [28]*
CSF [29]
DAEP [4]
IRCNN [38]
EPLL [39] + NE
EPLL [39] + NA
TV-L2 + NA
GradNet 7S [14]
Ours
Ours + NA
?n :
2.55
30.03
32.03
32.36
29.85
32.64
30.86
31.86
32.16
31.05
31.43
29.68
32.57
Levin [19]
5.10
7.65
28.40 27.32
29.79 28.31
26.34 21.43
28.13 27.28
30.07 28.30
29.85 28.83
29.77 28.28
30.25 28.96
29.14 28.03
28.88 27.55
29.45 28.95
30.21 29.00
10.2
26.52
27.20
17.33
26.70
27.15
28.05
27.16
27.85
27.16
26.96
28.29
28.23
2.55
24.44
25.38
25.70
24.73
25.42
25.60
25.36
25.57
24.61
25.57
25.69
26.00
Berkeley [2]
5.10
7.65
23.24 22.64
23.53 22.54
23.45 19.83
23.61 22.88
23.67 22.78
24.24 23.42
23.53 22.55
23.90 22.91
23.65 22.90
24.23 23.46
24.45 23.60
24.47 23.61
10.2
22.07
21.91
16.94
22.44
22.21
22.91
21.90
22.27
22.34
22.94
22.99
22.97
Table 2: Average PSNR (dB) for non-blind deconvolution on two datasets (*trained for ?n = 2.55).
Non-Blind Deblurring (NB). The gradient descent steps for non-blind deblurring with a known
kernel and degradation noise variance are given in Table 1, top row (NB). Here K denotes the Toeplitz
matrix of the blur kernel k.
Noise-Adaptive Deblurring (NA). When the degradation noise variance ?n2 is unknown, we can
solve Equation (9) for the optimal ?n2 (since it is independent of the prior), which gives
?n2 =
1
|y ? k ? x|2 + M ? 2 |k|2 .
N
(21)
By plugging this back into the equation, we get the following data term
data(x) = ?
N
log |y ? k ? x|2 + M ? 2 |k|2 ,
2
(22)
which is independent of the degradation noise variance ?n2 . We show the gradient descent steps in
?1
Table 1, second row (NA), where ?t = N |y ? Kxt?1 |2 + M ? 2 |k|2
adaptively scales the data
term with respect to the prior.
Noise- and Kernel-Blind Deblurring (NA+KE). Gradient descent in noise-blind optimization
includes an intuitive regularization for the kernel. We can use the objective in Equation (22) to
jointly optimize for the unknown image and the unknown kernel. The gradient descent steps to
update the image remain as in Table 1, second row (NA), and we take additional steps to update
the kernel estimate, as in Table 1, third row (KE). Additionally, we project the kernel by applying
t
k t = max(k t , 0) and k t = |kkt |1 after each step.
5
Experiments and Results
Our DAE uses the neural network architecture by Zhang et al. [37]. We generated training samples
by adding Gaussian noise to images from ImageNet [10]. We experimented with different noise
levels and found ?1 = 11 to perform well for all our deblurring and super-resolution experiments.
Unless mentioned, for image restoration we always take 300 iterations with step length ? = 0.1
and momentum ? = 0.9. The runtime of our method is linear in the number of pixels, and our
implementation takes about 0.2 seconds per iteration for one megapixel on an Nvidia Titan X (Pascal).
5.1
Image Deblurring: Non-Blind and Noise-Blind
In this section we evaluate our method for image deblurring using two datasets. Table 2 reports
the average PSNR for 32 images from the Levin et al. [19] and 50 images from the Berkeley [2]
segmentation dataset, where 10 images are randomly selected and blurred with 5 kernels as in Jin et
al. [14]. We highlight the best performing PSNR in bold and underline the second best value. The
6
Ground Truth
EPLL [39]
DAEP [4]
GradNet 7S [14]
Ours
Ours + NA
Figure 2: Visual comparison of our deconvolution results.
Ground Truth
Blurred with 1% noise
Ours (blind)
SSD Error Ratio
100
% Below Error Ratio
90
80
70
60
Sun et al.
Wipf and Zhang
Levin et al.
Babacan et al.
Log?TV PD
Log?TV MM
Ours
50
40
30
1
2
3
4
Figure 3: Performance of our method for fully (noise- and kernel-) blind deblurring on Levin?s set.
upper half of the table includes non-blind methods for deblurring. EPLL [39] + NE uses a noise
estimation step followed by non-blind deblurring. Noise-blind experiments are denoted by NA for
noise adaptivity. We include our results for non-blind (Ours) and noise-blind (Ours + NA). Our noise
adaptive approach consistently performs well in all experiments and on average we achieve better
results than the state of the art. Figure 2 provides a visual comparison of our results. Our prior is able
to produce sharp textures while also preserving the natural image structure.
5.2
Image Deblurring: Noise- and Kernel-Blind
We performed fully blind deconvolution with our method using Levin et al.?s [19] dataset. In this test,
we performed 1000 gradient descent iterations. We used momentum ? = 0.7 and step size ? = 0.3
for the unknown image and momentum ?k = 0.995 and step size ?k = 0.005 for the unknown
kernel. Figure 3 shows visual results of fully blind deblurring and performance comparison to state
of the art (last column). We compare the SSD error ratio and the number of images in the dataset
that achieves error ratios less than a threshold. Results for other methods are as reported by Perrone
and Favaro [23]. Our method can reconstruct all the blurry images in the dataset with errors ratios
less than 3.5. Note that our optimization performs end-to-end estimation of the final results and we
do not use the common two stage blind deconvolution (kernel estimation, followed by non-blind
deconvolution). Additionally our method uses a noise adaptive scheme where we do not assume
knowledge of the input noise level.
5.3
Super-resolution
To demonstrate the generality of our prior, we perform an additional test with single image superresolution. We evaluate our method on the two common datasets Set5 [3] and Set14 [34] for different
upsampling scales. Since these tests do not include degradation noise (?n = 0), we perform our
optimization with a rough weight for the prior and decrease it gradually to zero. We compare our
method in Table 3. The upper half of the table represents methods that are specifically trained for
super-resolution. SRCNN [11] and TNRD [7] have separate models trained for ?2, 3, 4 scales, and
we used the model for ?4 to produce the ?5 results. VDSR [16] and DnCNN-3 [37] have a single
model trained for ?2, 3, 4 scales, which we also used to produce ?5 results. The lower half of the
table represents general priors that are not designed specifically for super-resolution. Our method
performs on par with state of the art methods over all the upsampling scales.
7
Method
Bicubic
SRCNN [11]
TNRD [7]
VDSR [16]
DnCNN-3 [37]
DAEP [4]
IRCNN [38]
Ours
scale:
?2
31.80
34.50
34.62
34.50
35.20
35.23
35.07
35.16
Set5 [3]
?3
?4
28.67 26.73
30.84 28.60
31.08 28.83
31.39 29.19
31.58 29.30
31.44 29.01
31.26 29.01
31.38 29.16
?5
25.32
26.12
26.88
25.91
26.30
27.19
27.13
27.38
?2
28.53
30.52
30.53
30.72
30.99
31.07
30.79
30.99
Set14 [34]
?3
?4
25.92 24.44
27.48 25.76
27.60 25.92
27.81 26.16
27.93 26.25
27.93 26.13
27.68 25.96
27.90 26.22
?5
23.46
24.05
24.61
24.01
24.26
24.88
24.73
25.01
Table 3: Average PSNR (dB) for super-resolution on two datasets.
Matlab [21]
33.9
RTF [15]
37.8
Gharbi et al. [13]
38.4
Gharbi et al. [13] f.t.
38.6
SEM [17]
38.8
Ours
38.7
Table 4: Average PSNR (dB) in linear RGB space for demosaicing on the Panasonic dataset [15].
5.4
Demosaicing
We finally performed a demosaicing experiment on the dataset introduced by Khashabi et al. [15].
This dataset is constructed by taking RAW images from a Panasonic camera, where the images are
downsampled to construct the ground truth data. Due to the down sampling effect, in this evaluation
we train a DAE with ?1 = 3 noise standard deviation. The test dataset consists of 100 noisy images
captured by a Panasonic camera using a Bayer color filter array (RGGB). We initialize our method
with Matlab?s demosaic function [21]. To get even better initialization, we perform our initial
optimization with a large degradation noise estimate (?n = 2.5) and then perform the optimization
with a lower estimate (?n = 1). We summarize the quantitative results in Table 4. Our method
is again on par with the state of the art. Additionally, our prior is not trained for a specific color
filter array and therefore is not limited to a specific sub-pixel order. Figure 4 shows a qualitative
comparison, where our method produces much smoother results compared to the previous state of the
art.
Ground Truth
RTF [15]
Gharbi et al. [13]
SEM [17]
Ours
Figure 4: Visual comparison for demosaicing noisy images from the Panasonic data set [15].
6
Conclusions
We proposed a Bayesian deep learning framework for image restoration with a generic image prior
that directly represents the Gaussian smoothed natural image probability distribution. We showed that
we can compute the gradient of our prior efficiently using a trained denoising autoencoder (DAE).
Our formulation allows us to learn a single prior and use it for many image restoration tasks, such as
noise-blind deblurring, super-resolution, and image demosaicing. Our results indicate that we achieve
performance that is competitive with the state of the art for these applications. In the future, we would
like to explore generalizing from Gaussian smoothing of the underlying distribution to other types of
kernels. We are also considering multi-scale optimization where one would reduce the Bayes utility
support gradually to get a tighter bound with respect to maximum a posteriori. Finally, our approach
is not limited to image restoration and could be exploited to address other inverse problems.
8
Acknowledgments. MJ and PF acknowledge support from the Swiss National Science Foundation
(SNSF) on project 200021-153324.
References
[1] Guillaume Alain and Yoshua Bengio. What regularized auto-encoders learn from the data-generating
distribution. Journal of Machine Learning Research, 15:3743?3773, 2014.
[2] Pablo Arbelaez, Michael Maire, Charless Fowlkes, and Jitendra Malik. Contour detection and hierarchical
image segmentation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 33(5):898?916,
2011.
[3] Marco Bevilacqua, Aline Roumy, Christine Guillemot, and Marie-Line Alberi-Morel. Low-complexity
single-image super-resolution based on nonnegative neighbor embedding. In British Machine Vision
Conference, BMVC 2012, Surrey, UK, September 3-7, 2012, pages 1?10, 2012.
[4] Siavash Arjomand Bigdeli and Matthias Zwicker. Image restoration using autoencoding priors. arXiv
preprint arXiv:1703.09964, 2017.
[5] Antoni Buades, Bartomeu Coll, and J-M Morel. A non-local algorithm for image denoising. In Computer
Vision and Pattern Recognition (CVPR), 2005 IEEE Conference on, volume 2, pages 60?65. IEEE, 2005.
[6] JH Chang, Chun-Liang Li, Barnabas Poczos, BVK Kumar, and Aswin C Sankaranarayanan. One network to solve them all?solving linear inverse problems using deep projection models. arXiv preprint
arXiv:1703.09912, 2017.
[7] Yunjin Chen and Thomas Pock. Trainable nonlinear reaction diffusion: A flexible framework for fast and
effective image restoration. IEEE Transactions on Pattern Analysis and Machine Intelligence, 39(6):1256?
1272, 2017.
[8] Dorin Comaniciu and Peter Meer. Mean shift: A robust approach toward feature space analysis. IEEE
Transactions on Pattern Analysis and Machine Intelligence, 24(5):603?619, 2002.
[9] Kostadin Dabov, Alessandro Foi, Vladimir Katkovnik, and Karen Egiazarian. Image denoising with
block-matching and 3d filtering. In Electronic Imaging 2006, pages 606414?606414. International Society
for Optics and Photonics, 2006.
[10] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical
image database. In Computer Vision and Pattern Recognition (CVPR), 2009 IEEE Conference on, pages
248?255. IEEE, 2009.
[11] Chao Dong, Chen Change Loy, Kaiming He, and Xiaoou Tang. Image super-resolution using deep
convolutional networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 38(2):295?307,
2016.
[12] Rob Fergus, Barun Singh, Aaron Hertzmann, Sam T Roweis, and William T Freeman. Removing camera
shake from a single photograph. In ACM Transactions on Graphics (TOG), volume 25, pages 787?794.
ACM, 2006.
[13] Micha?l Gharbi, Gaurav Chaurasia, Sylvain Paris, and Fr?do Durand. Deep joint demosaicking and
denoising. ACM Transactions on Graphics (TOG), 35(6):191, 2016.
[14] M. Jin, S. Roth, and P. Favaro. Noise-blind image deblurring. In Computer Vision and Pattern Recognition
(CVPR), 2017 IEEE Conference on. IEEE, 2017.
[15] Daniel Khashabi, Sebastian Nowozin, Jeremy Jancsary, and Andrew W Fitzgibbon. Joint demosaicing and
denoising via learned nonparametric random fields. IEEE Transactions on Image Processing, 23(12):4968?
4981, 2014.
[16] Jiwon Kim, Jung Kwon Lee, and Kyoung Mu Lee. Accurate image super-resolution using very deep
convolutional networks. In Computer Vision and Pattern Recognition (CVPR), 2016 IEEE Conference on,
pages 1646?1654. IEEE, 2016.
[17] Teresa Klatzer, Kerstin Hammernik, Patrick Knobelreiter, and Thomas Pock. Learning joint demosaicing
and denoising based on sequential energy minimization. In Computational Photography (ICCP), 2016
IEEE International Conference on, pages 1?11. IEEE, 2016.
[18] Dilip Krishnan and Rob Fergus. Fast image deconvolution using hyper-laplacian priors. In Advances in
Neural Information Processing Systems, pages 1033?1041, 2009.
9
[19] Anat Levin, Rob Fergus, Fr?do Durand, and William T Freeman. Image and depth from a conventional
camera with a coded aperture. ACM Transactions on Graphics (TOG), 26(3):70, 2007.
[20] Anat Levin and Boaz Nadler. Natural image denoising: Optimality and inherent bounds. In Computer
Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on, pages 2833?2840. IEEE, 2011.
[21] Henrique S Malvar, Li-wei He, and Ross Cutler. High-quality linear interpolation for demosaicing of bayerpatterned color images. In Acoustics, Speech, and Signal Processing, 2004. Proceedings.(ICASSP?04).
IEEE International Conference on, volume 3, pages iii?485. IEEE, 2004.
[22] Tim Meinhardt, Michael M?ller, Caner Hazirbas, and Daniel Cremers. Learning proximal operators: Using
denoising networks for regularizing inverse imaging problems. arXiv preprint arXiv:1704.03488, 2017.
[23] Daniele Perrone and Paolo Favaro. A logarithmic image prior for blind deconvolution. International
Journal of Computer Vision, 117(2):159?172, 2016.
[24] J. Portilla, V. Strela, M. J. Wainwright, and E. P. Simoncelli. Image denoising using scale mixtures of
gaussians in the wavelet domain. IEEE Transactions on Image Processing, 12(11):1338?1351, Nov 2003.
[25] Yaniv Romano, Michael Elad, and Peyman Milanfar. The little engine that could: Regularization by
denoising (red). arXiv preprint arXiv:1611.02862, 2016.
[26] Stefan Roth and Michael J Black. Fields of experts: A framework for learning image priors. In Computer
Vision and Pattern Recognition (CVPR), 2005 IEEE Conference on, volume 2, pages 860?867. IEEE, 2005.
[27] Leonid I. Rudin, Stanley Osher, and Emad Fatemi. Nonlinear total variation based noise removal algorithms.
Physica D: Nonlinear Phenomena, 60(1):259 ? 268, 1992.
[28] Uwe Schmidt, Jeremy Jancsary, Sebastian Nowozin, Stefan Roth, and Carsten Rother. Cascades of
regression tree fields for image restoration. IEEE transactions on pattern analysis and machine intelligence,
38(4):677?689, 2016.
[29] Uwe Schmidt and Stefan Roth. Shrinkage fields for effective image restoration. In Computer Vision and
Pattern Recognition (CVPR), 2014 IEEE Conference on, pages 2774?2781. IEEE, 2014.
[30] Tamar Rott Shaham and Tomer Michaeli. Visualizing image priors. In European Conference on Computer
Vision, pages 136?153. Springer, 2016.
[31] Singanallur V Venkatakrishnan, Charles A Bouman, and Brendt Wohlberg. Plug-and-play priors for model
based reconstruction. In GlobalSIP, pages 945?948. IEEE, 2013.
[32] Pascal Vincent, Hugo Larochelle, Yoshua Bengio, and Pierre-Antoine Manzagol. Extracting and composing
robust features with denoising autoencoders. In Proceedings of the 25th International Conference on
Machine Learning, pages 1096?1103. ACM, 2008.
[33] Lei Xiao, Felix Heide, Wolfgang Heidrich, Bernhard Sch?lkopf, and Michael Hirsch. Discriminative
transfer learning for general image restoration. arXiv preprint arXiv:1703.09245, 2017.
[34] Roman Zeyde, Michael Elad, and Matan Protter. On single image scale-up using sparse-representations.
In International Conference on Curves and Surfaces, pages 711?730. Springer, 2010.
[35] Haichao Zhang and David Wipf. Non-uniform camera shake removal using a spatially-adaptive sparse
penalty. In Advances in Neural Information Processing Systems, pages 1556?1564, 2013.
[36] Haichao Zhang and Jianchao Yang. Scale adaptive blind deblurring. In Advances in Neural Information
Processing Systems, pages 3005?3013, 2014.
[37] Kai Zhang, Wangmeng Zuo, Yunjin Chen, Deyu Meng, and Lei Zhang. Beyond a gaussian denoiser:
Residual learning of deep cnn for image denoising. arXiv preprint arXiv:1608.03981, 2016.
[38] Kai Zhang, Wangmeng Zuo, Shuhang Gu, and Lei Zhang. Learning deep cnn denoiser prior for image
restoration. arXiv preprint arXiv:1704.03264, 2017.
[39] Daniel Zoran and Yair Weiss. From learning models of natural image patches to whole image restoration.
In Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on, pages 479?486. IEEE,
2011.
10
| 6678 |@word cnn:2 version:1 seems:1 underline:1 rgb:1 p0:5 set5:2 initial:1 substitution:1 daniel:3 ours:11 interestingly:1 reaction:1 com:1 blur:2 remove:1 designed:2 update:4 half:3 selected:1 intelligence:5 rudin:1 kyoung:1 core:1 provides:1 location:1 simpler:1 zhang:9 favaro:5 hazirbas:1 constructed:1 qualitative:1 consists:2 megapixel:1 introduce:2 expected:2 multi:1 bm3d:2 inspired:1 freeman:2 little:1 pf:1 considering:1 project:2 estimating:1 underlying:2 buades:1 maximizes:1 superresolution:1 what:1 strela:1 interpreted:1 minimizes:1 berkeley:2 quantitative:1 runtime:2 exactly:1 uk:1 omit:1 before:1 felix:1 local:2 pock:2 meng:1 interpolation:1 black:1 initialization:1 micha:1 limited:2 demosaicking:1 acknowledgment:1 camera:5 practice:1 block:1 yunjin:2 swiss:1 fitzgibbon:1 maire:1 cascade:1 projection:1 matching:1 word:1 downsampled:1 get:4 cannot:3 close:2 operator:7 nb:4 risk:3 context:1 applying:1 optimize:2 equivalent:1 map:4 conventional:1 roth:4 maximizing:2 straightforward:1 resolution:13 formulate:4 ke:5 estimator:7 insight:1 rule:1 array:2 zuo:2 embedding:2 handle:2 meer:1 variation:2 play:1 epll:6 deblurring:23 us:3 recognition:8 database:1 observed:2 role:1 preprint:7 capture:1 sun:1 decrease:1 overfitted:1 mentioned:1 intuition:1 pd:1 alessandro:1 complexity:1 hertzmann:1 mu:1 ideally:1 dncnn:2 barnabas:1 venkatakrishnan:3 trained:9 singh:1 rewrite:2 solving:1 zoran:1 rtf:3 tog:3 gu:1 icassp:1 joint:3 xiaoou:1 various:1 represented:1 regularizer:3 train:1 fast:2 effective:3 describe:4 hyper:1 outside:1 matan:1 whose:2 posed:1 solve:8 cvpr:8 kai:3 elad:2 reconstruct:1 toeplitz:1 statistic:1 jointly:1 itself:1 noisy:4 final:1 autoencoding:1 kxt:2 matthias:2 propose:3 reconstruction:1 fr:2 inserting:1 iccp:1 achieve:3 roweis:1 intuitive:2 dirac:1 exploiting:1 yaniv:1 produce:6 generating:1 tim:1 andrew:1 wangmeng:2 implemented:1 c:1 involves:1 indicate:1 jiwon:1 larochelle:1 csf:1 filter:3 cnns:2 stochastic:3 nlm:1 caner:1 generalization:1 tighter:1 physica:1 mm:1 marco:1 ground:4 nadler:1 scope:1 achieves:1 dorin:1 shaham:3 estimation:5 emad:1 ross:1 morel:2 minimization:4 hope:1 rough:1 gaurav:1 khashabi:2 gaussian:20 always:1 super:13 snsf:1 stefan:3 shrinkage:1 derived:1 guillemot:1 consistently:1 likelihood:7 indicates:1 contrast:2 kim:1 dilip:1 posteriori:2 inference:1 relation:2 pixel:3 issue:3 canceled:1 flexible:1 dual:1 ill:1 pascal:2 denoted:1 uwe:2 art:7 smoothing:3 special:1 initialize:1 field:7 construct:3 beach:1 sampling:2 represents:5 park:1 wipf:2 future:1 others:2 report:1 simplify:1 inherent:1 roman:1 few:2 yoshua:2 richard:1 randomly:1 kwon:1 national:1 comprehensive:1 replaced:1 argmax:1 consisting:2 william:2 detection:1 fd:1 wohlberg:1 evaluation:2 photonics:1 mixture:2 cutler:1 primal:1 accurate:1 bicubic:1 edge:1 integral:3 closer:1 bayer:1 unless:1 tree:1 logarithm:9 penalizes:2 desired:2 dae:20 bouman:1 column:1 restoration:29 introducing:1 deviation:3 uniform:1 successful:1 levin:7 graphic:3 reported:1 encoders:1 proximal:10 adaptively:2 st:1 density:8 international:6 lee:2 dong:2 michael:6 na:13 again:2 corner:1 stochastically:1 expert:1 derivative:2 li:6 jeremy:2 bold:1 includes:4 blurred:2 titan:1 jitendra:1 cremers:1 explicitly:1 blind:41 performed:4 wolfgang:1 red:1 competitive:4 bayes:10 masking:1 jia:2 contribution:2 minimize:1 degraded:1 egiazarian:1 variance:9 convolutional:2 efficiently:1 lkopf:1 bayesian:3 raw:1 vincent:1 foi:1 dabov:1 globalsip:1 researcher:1 straight:1 foe:1 explain:1 sebastian:2 definition:1 energy:2 surrey:1 boil:1 sampled:1 dataset:8 popular:1 knowledge:1 ut:7 dimensionality:1 psnr:5 segmentation:2 color:3 stanley:1 back:2 wei:3 bmvc:1 formulation:9 generality:1 just:1 stage:1 autoencoders:7 nonlinear:3 quality:1 impulse:1 lei:3 name:1 usa:1 effect:1 true:6 gharbi:4 hence:4 regularization:5 spatially:2 visualizing:1 daes:6 numerator:1 encourages:1 comaniciu:1 daniele:1 demonstrate:3 performs:3 christine:1 image:114 photography:1 recently:3 charles:2 common:2 hugo:1 overview:1 volume:4 he:2 refer:2 bevilacqua:1 ssd:2 surface:1 heidrich:1 add:1 patrick:1 curvature:2 posterior:5 showed:1 inf:3 route:1 nvidia:1 tnrd:2 inequality:2 durand:2 exploited:1 seen:1 preserving:1 additional:2 captured:1 deng:1 ller:1 signal:1 smoother:1 simoncelli:1 panasonic:4 plug:1 long:1 coded:1 plugging:2 laplacian:1 regression:1 vision:11 expectation:2 arxiv:14 iteration:4 kernel:27 tailored:1 addition:4 priorsl:4 source:1 unibe:3 sch:1 umd:1 unlike:1 tend:1 db:3 call:3 extracting:1 leverage:4 yang:1 split:1 bengio:2 iii:1 krishnan:1 architecture:1 reduce:1 idea:3 tamar:1 shift:13 jianchao:1 utility:10 effort:1 penalty:1 milanfar:1 peter:1 karen:1 poczos:1 speech:1 romano:2 matlab:2 deep:15 clear:1 detailed:1 shake:2 nonparametric:1 http:1 estimated:1 per:1 write:1 paolo:2 express:2 key:3 threshold:1 changing:1 marie:1 diffusion:1 vast:1 imaging:2 zeyde:1 sum:2 inverse:3 electronic:1 patch:2 leibniz:1 prefer:1 bound:8 followed:2 nonnegative:1 optic:1 kerstin:1 fei:2 unlimited:1 argument:1 babacan:1 formulating:1 demosaicing:11 performing:1 kumar:1 optimality:1 tv:3 perrone:2 remain:1 sam:1 rob:3 osher:1 gradually:2 taken:1 equation:12 visualization:2 remains:1 end:2 available:1 gaussians:1 apply:2 hierarchical:2 generic:2 blurry:1 pierre:1 fowlkes:1 schmidt:2 yair:1 thomas:2 denotes:2 running:1 include:4 zwicker:5 top:1 const:1 build:4 establish:1 society:1 objective:9 malik:1 restored:1 antoine:1 unclear:1 exhibit:1 gradient:40 september:1 distance:1 separate:1 maryland:1 arbelaez:1 capacity:1 majority:1 upsampling:2 manifold:3 considers:1 toward:1 denoiser:6 rother:1 assuming:1 code:1 length:1 fatemi:1 reformulate:1 ratio:5 downsampling:1 vladimir:1 loy:1 liang:1 manzagol:1 relate:1 negative:2 implementation:2 unknown:7 perform:8 upper:2 observation:3 convolution:2 datasets:4 finite:1 acknowledge:1 jin:7 descent:16 meinhardt:2 defining:2 portilla:1 smoothed:12 sharp:3 tomer:1 introduced:3 pablo:1 david:1 required:1 paris:1 trainable:1 connection:1 imagenet:2 teresa:1 acoustic:1 engine:1 learned:1 nip:1 address:6 able:2 beyond:1 usually:1 below:1 pattern:13 aswin:1 sparsity:1 summarize:1 built:1 including:3 max:1 wainwright:1 natural:22 rely:1 hybrid:1 restore:1 regularized:1 residual:2 scheme:1 github:1 roumy:1 ne:2 autoencoder:4 auto:1 set14:2 chao:1 prior:77 review:1 l2:1 removal:2 haichao:2 protter:1 unsurprisingly:1 fully:5 par:2 highlight:1 adaptivity:1 proportional:2 filtering:1 deyu:1 foundation:1 xiao:1 systematically:1 nowozin:2 row:4 summary:2 jung:1 last:2 bern:4 alain:2 allow:3 jh:1 katkovnik:1 neighbor:1 taking:3 sparse:4 curve:1 depth:1 contour:1 concavity:1 interchange:1 adaptive:6 coll:1 far:1 transaction:10 approximate:2 nov:1 boaz:1 aperture:1 michaeli:1 bernhard:1 hirsch:1 overfitting:2 kkt:1 fergus:3 discriminative:1 continuous:1 latent:2 decade:1 table:15 additionally:3 learn:13 mj:1 robust:2 ca:1 composing:1 transfer:1 sem:2 complex:1 european:1 domain:1 main:2 whole:1 noise:50 n2:6 srcnn:2 sub:1 momentum:5 sf:1 lie:1 third:1 anat:2 learns:1 wavelet:1 tang:1 down:2 british:1 antoni:1 removing:1 specific:3 xt:12 jensen:2 explored:1 experimented:1 chun:1 deconvolution:11 derives:1 sankaranarayanan:1 socher:1 adding:1 effectively:2 sequential:1 texture:2 heide:1 kx:1 gap:1 chen:3 generalizing:1 bvk:1 photograph:1 simply:1 explore:1 logarithmic:1 visual:6 kaiming:1 chang:1 springer:2 ch:3 corresponds:4 truth:4 acm:5 peyman:1 carsten:1 replace:1 admm:3 leonid:1 change:1 jancsary:2 specifically:3 typical:1 sylvain:1 denoising:22 degradation:7 total:2 meaningful:1 aaron:1 college:1 guillaume:1 support:3 brevity:1 evaluate:3 regularizing:1 phenomenon:1 ex:3 |
6,277 | 6,679 | Greedy Algorithms for Cone Constrained
Optimization with Convergence Guarantees
Francesco Locatello
MPI for Intelligent Systems - ETH Zurich
Michael Tschannen
ETH Zurich
[email protected]
[email protected]
Gunnar R?tsch
ETH Zurich
Martin Jaggi
EPFL
[email protected]
[email protected]
Abstract
Greedy optimization methods such as Matching Pursuit (MP) and Frank-Wolfe
(FW) algorithms regained popularity in recent years due to their simplicity, effectiveness and theoretical guarantees. MP and FW address optimization over the
linear span and the convex hull of a set of atoms, respectively. In this paper, we
consider the intermediate case of optimization over the convex cone, parametrized
as the conic hull of a generic atom set, leading to the first principled definitions
of non-negative MP algorithms for which we give explicit convergence rates and
demonstrate excellent empirical performance. In particular, we derive sublinear
(O(1/t)) convergence on general smooth and convex objectives, and linear convergence (O(e?t )) on strongly convex objectives, in both cases for general sets
of atoms. Furthermore, we establish a clear correspondence of our algorithms
to known algorithms from the MP and FW literature. Our novel algorithms and
analyses target general atom sets and general objective functions, and hence are
directly applicable to a large variety of learning settings.
1
Introduction
In recent years, greedy optimization algorithms have attracted significant interest in the domains
of signal processing and machine learning thanks to their ability to process very large data sets.
Arguably two of the most popular representatives are Frank-Wolfe (FW) [12, 21] and Matching
Pursuit (MP) algorithms [34], in particular Orthogonal MP (OMP) [9, 49]. While the former targets
minimization of a convex function over bounded convex sets, the latter apply to minimization over a
linear subspace. In both cases, the domain is commonly parametrized by a set of atoms or dictionary
elements, and in each iteration, both algorithms rely on querying a so-called linear minimization
oracle (LMO) to find the direction of steepest descent in the set of atoms. The iterate is then updated
as a linear or convex combination, respectively, of previous iterates and the newly obtained atom
from the LMO. The particular choice of the atom set allows to encode structure such as sparsity and
non-negativity (of the atoms) into the solution. This enables control of the trade-off between the
amount of structure in the solution and approximation quality via the number of iterations, which
was found useful in a large variety of use cases including structured matrix and tensor factorizations
[50, 53, 54, 18].
In this paper, we target an important ?intermediate case? between the two domain parameterizations
given by the linear span and the convex hull of an atom set, namely the parameterization of the
optimization domain as the conic hull of a possibly infinite atom set. In this case, the solution
can be represented as a non-negative linear combination of the atoms, which is desirable in many
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
applications, e.g., due to the physics underlying the problem at hand, or for the sake of interpretability.
Concrete examples include unmixing problems [11, 16, 3], model selection [33], and matrix and
tensor factorizations [4, 24]. However, existing convergence analyses do not apply to the currently
used greedy algorithms. In particular, all existing MP variants for the conic hull case [5, 38, 52] are
not guaranteed to converge and may get stuck far away from the optimum (this can be observed in
the experiments in Section 6). From a theoretical perspective, this intermediate case is of paramount
interest in the context of MP and FW algorithms. Indeed, the atom set is not guaranteed to contain
an atom aligned with a descent direction for all possible suboptimal iterates, as is the case when the
optimization domain is the linear span or the convex hull of the atom set [39, 32]. Hence, while conic
constraints have been widely studied in the context of a manifold of different applications, none of
the existing greedy algorithms enjoys explicit convergence rates.
We propose and analyze new MP algorithms tailored for the minimization of smooth convex functions
over the conic hull of an atom set. Specifically, our key contributions are:
? We propose the first (non-orthogonal) MP algorithm for optimization over conic hulls
guaranteed to converge, and prove a corresponding sublinear convergence rate with explicit constants. Surprisingly, convergence is achieved without increasing computational
complexity compared to ordinary MP.
? We propose new away-step, pairwise, and fully corrective MP variants, inspired by variants
of FW [28] and generalized MP [32], respectively, that allow for different degrees of weight
corrections for previously selected atoms. We derive corresponding sublinear and linear (for
strongly convex objectives) convergence rates that solely depend on the geometry of the
atom set.
? All our algorithms apply to general smooth convex functions. This is in contrast to all prior
work on non-negative MP, which targets quadratic objectives [5, 38, 52]. Furthermore, if
the conic hull of the atom set equals its linear span, we recover both algorithms and rates
derived in [32] for generalized MP variants.
? We make no assumptions on the atom set which is simply a subset of a Hilbert space, in
particular we do not assume the atom set to be finite.
Before presenting our algorithms (Section 3) along with the corresponding convergence guarantees
(Section 4), we briefly review generalized MP variants. A detailed discussion of related work can
be found in Section 5 followed by illustrative experiments on a least squares problem on synthetic
data, and non-negative matrix factorization as well as non-negative garrote logistic regression as
applications examples on real data (numerical evaluations of more applications and the dependency
between constants in the rate and empirical convergence can be found in the supplementary material).
Notation. Given a non-empty subset A of some Hilbert space, let conv(A) be the convex hull
of A, and let lin(A) denote its linear span. Given a closed set A, we call its diameter diam(A) =
maxz1 ,z2 ?A kz1 ? z2 k and its radius radius(A) = maxz?A kzk. kxkA := inf{c > 0 : x ?
c ? conv(A)} is the atomic norm of x over a set A (also known as the gauge function of conv(A)).
We call a subset A of a Hilbert space symmetric if it is closed under negation.
2
Review of Matching Pursuit Variants
Let H be a Hilbert space with associated inner product hx, yi, ? x, y ? H. The inner product induces
the norm kxk2 := hx, xi, ? x ? H. Let A ? H be a compact set (the ?set of atoms? or dictionary)
and let f : H ? R be convex and L-smooth (L-Lipschitz gradient in the finite dimensional case).
If H is an infinite-dimensional Hilbert space, then f is assumed to be Fr?chet differentiable. The
generalized MP algorithm studied in [32], presented in Algorithm 1, solves the following optimization
problem:
min f (x).
(1)
x?lin(A)
In each iteration, MP queries a linear minimization oracle (LMO) solving the following linear
problem:
LMOA (y) := arg min hy, zi
(2)
z?A
for a given query y ? H. The MP update step minimizes a quadratic upper bound gxt (x) =
f (xt ) + h?f (xt ), x ? xt i + L2 kx ? xt k2 of f at xt , where L is an upper bound on the smoothness
2
constant of f with respect to a chosen norm k ? k. Optimizing this norm problem instead of f
directly allows for substantial efficiency gains in the case of complicated f . For symmetric A and for
f (x) = 12 ky ? xk2 , y ? H, Algorithm 1 recovers MP (Variant 0) [34] and OMP (Variant 1) [9, 49],
see [32] for details.
Approximate linear oracles. Solving the
Algorithm 1 Norm-Corrective Generalized Match- LMO defined in (2) exactly is often hard in
practice, in particular when applied to matrix
ing Pursuit
(or tensor) factorization problems, while ap1: init x0 ? lin(A), and S := {x0 }
proximate versions can be much more efficient.
2: for t = 0 . . . T
Algorithm 1 allows for an approximate LMO.
3:
Find zt := (Approx-)LMOA (?f (xt ))
For given quality parameter ? ? (0, 1] and
4:
S := S ? {zt }
given direction d ? H, the approximate LMO
5:
Let b := xt ? L1 ?f (xt )
? ? A such
for Algorithm 1 returns a vector z
6:
Variant 0:
that
2
Update xt+1 := arg min kz ? bk
?i ? ?hd, zi,
hd, z
(3)
z:=xt +?zt
??R
relative to z = LMOA (d) being an exact solu7:
Variant 1:
tion.
2
Update xt+1 := arg min kz ? bk
z?lin(S)
8:
Optional: Correction of some/all atoms z0...t Discussion and limitations of MP. The analysis of the convergence of Algorithm 1 in [32]
9: end for
critically relies on the assumption that the origin is in the relative interior of conv(A) with
respect to its linear span. This assumption originates from the fact that the convergence of MP- and
FW-type algorithms fundamentally depends on an alignment assumption of the search direction
returned by the LMO (i.e., zt in Algorithm 1) and the gradient of the objective at the current iteration
(see third premise in [39]). Specifically, for Algorithm 1, the LMO is assumed to select a descent
direction, i.e., h?f (xt ), zt i < 0, so that the resulting weight (i.e., ? for Variant 0) is always positive.
In this spirit, Algorithm 1 is a natural candidate to minimize f over the conic hull of A. However,
if the optimization domain is a cone, the alignment assumption does not hold as there may be
non-stationary points x in the conic hull of A for which minz?A h?f (x), zi = 0. Algorithm 1 is
therefore not guaranteed to converge when applied to conic problems. The same issue arises for
essentially all existing non-negative variants of MP, see, e.g., Alg. 2 in [38] and in Alg. 2 in [52]. We
now present modifications corroborating this issue along with the resulting MP-type algorithms for
conic problems and corresponding convergence guarantees.
3
Greedy Algorithms on Conic Hulls
The cone cone(A ? y) tangent to the convex set conv(A) at a point y is formed by the half-lines
emanating from y and intersecting conv(A) in at least one point distinct from y. Without loss of
generality we consider 0 ? A and assume the set cone(A) (i.e., y = 0) to be closed. If A is finite
P|A|
the cone constraint can be written as cone(A) := {x : x = i=1 ?i ai s.t. ai ? A, ?i ? 0 ?i}. We
consider conic optimization problems of the form:
min
f (x).
x?cone(A)
(4)
Note that if the set A is symmetric or if the origin is in the relative interior of conv(A) w.r.t. its linear
span then cone(A) = lin(A). We will show later how our results recover known MP rates when the
origin is in the relative interior of conv(A).
As a first algorithm to solve problems of the form (4), we present the Non-Negative Generalized
Matching Pursuit (NNMP) in Algorithm 2 which is an extension of MP to general f and non-negative
weights.
Discussion: Algorithm 2 differs from Algorithm 1 (Variant 0) in line 4, adding the iterationdependent atom ? kxxttkA to the set of possible search directions1 . We use the atomic norm for the
This additional direction makes sense only if xt 6= 0. Therefore, we set ? kxxttkA = 0 if xt = 0, i.e., no
direction is added.
1
3
Algorithm 2 Non-Negative Matching Pursuit
1: init x0 = 0 ? A
2: for t = 0 . . . T
?t := (Approx-)LMOA (?f (xt ))
3:
Find z
4:
zt = arg min n ?xt o h?f (xt ), zi
?t , kx
z? z
t kA
h??f (xt ),zt i
Lkzt k2
5:
? :=
6:
Update xt+1 := xt + ?zt
7: end for
Figure 1: Two dimensional example for TA (xt ) where
A = {a1 , a2 }, for three different iterates x0 , x1 and
x2 . The shaded area corresponds to TA (xt ) and the
white area to lin(A) \ TA (xt ).
normalization because it yields the best constant in the convergence rate. In practice, one can replace
it with the Euclidean norm, which is often much less expensive to compute. This iteration-dependent
additional search direction allows to reduce the weights of the atoms that were previously selected,
thus admitting the algorithm to ?move back? towards the origin while maintaining the cone constraint.
This idea is informally explained here and formally studied in Section 4.1.
Recall the alignment assumption of the search direction and the gradient of the objective at the current
iterate discussed in Section 2 (see also [39]). Algorithm 2 obeys this assumption. The intuition
behind this is the following. Whenever xt is not a minimizer of (4) and minz?A h?f (xt ), zi = 0,
the vector ? kxxttkA is aligned with ?f (xt ) (i.e., h?f (xt ), ? kxxttkA i < 0), preventing the algorithm
from stopping at a suboptimal iterate. To make this intuition more formal, let us define the set of
feasible descent directions of Algorithm 2 at a point x ? cone(A) as:
n
x o
TA (x) := d ? H : ?z ? A ? ?
s.t. hd, zi < 0 .
(5)
kxkA
If at some iteration t = 0, 1, . . . the gradient ?f (xt ) is not in TA (xt ) Algorithm 2 terminates as
minz?A hd, zi = 0 and hd, ?xt i ? 0 (which yields zt = 0). Even though, in general, not every
direction in H is a feasible descent direction, ?f (xt ) ?
/ TA only occurs if xt is a constrained
minimum of Equation 4:
? ? cone(A) and ?f (?
? is a solution to minx?cone(A) f (x).
Lemma 1. If x
x) 6? TA then x
Initializing Algorithm 2 with x0 = 0 guarantees that the iterates xt always remain inside cone(A)
even though this is not enforced explicitly (by convexity of f , see proof of Theorem 2 in Appendix D
for details).
Limitations of Algorithm 2: Let us call active the atoms which have nonzero weights in the
Pt?1
representation of xt = i=0 ?i zi computed by Algorithm 2. Formally, the set of active atoms is
defined as S := {zi : ?i > 0, i = 0, 1, . . . , t ? 1}. The main drawback of Algorithm 2 is that when
the direction ? kxxttkA is selected, the weight of all active atoms is reduced. This can lead to the
algorithm alternately selecting ? kxxttkA and an atom from A, thereby slowing down convergence in a
similar manner as the zig-zagging phenomenon well-known in the Frank-Wolfe framework [28]. In
order to achieve faster convergence we introduce the corrective variants of Algorithm 2.
3.1
Corrective Variants
To achieve faster (linear) convergence (see Section 4.2) we introduce variants of Algorithm 2, termed
Away-steps MP (AMP) and Pairwise MP (PWMP), presented in Algorithm 3. Here, inspired by the
away-steps and pairwise variants of FW [12, 28], instead of reducing the weights of the active atoms
uniformly as in Algorithm 2, the LMO is queried a second time on the active set S to identify the
direction of steepest ascent in S. This allows, at each iteration, to reduce the weight of a previously
selected atom (AMP) or swap weight between atoms (PWMP). This selective ?reduction? or ?swap
of weight? helps to avoid the zig-zagging phenomenon which prevent Algorithm 2 from converging
linearly.
At each iteration, Algorithm 3 updates the weights of zt and vt as ?zt = ?zt + ? and ?vt = ?vt ? ?,
respectively. To ensure that xt+1 ? cone(A), ? has to be clipped according to the weight which is
currently on vt , i.e., ?max = ?vt . If ? = ?max , we set ?vt = 0 and remove vt from S as the atom vt
is no longer active. If dt ? A (i.e., we take a regular MP step and not an away step), the line search
is unconstrained (i.e., ?max = ?).
4
For both algorithm variants, the second LMO query increases the computational complexity. Note
that an exact search on S is feasible in practice as |S| has at most t elements at iteration t.
Taking an additional computational burden allows to update the weights of all active atoms in the
spirit of OMP. This approach is implemented in the Fully Corrective MP (FCMP), Algorithm 4.
Algorithm 3 Away-steps (AMP) and Pairwise Algorithm 4 Fully Corrective Non-Negative
(PWMP) Non-Negative Matching Pursuit
Matching Pursuit (FCMP)
1: init x0 = 0 ? A, and S := {x0 }
2: for t = 0 . . . T
3:
Find zt := (Approx-)LMOA (?f (xt ))
4:
Find vt := (Approx-)LMOS (??f (xt ))
5:
S = S ? zt
6:
AMP: dt = arg mind?{zt ,?vt }h?f (xt ), di
7:
PWMP: dnt = zt ? vt
o
h??f (xt ),dt i
, ?max
Lkdt k2
? := min
(?max see text)
9:
Update ?zt , ?vt and S according to ?
(? see text)
10:
Update xt+1 := xt + ?dt
11: end for
8:
1: init x0 = 0 ? A, S = {x0 }
2: for t = 0 . . . T
3:
Find zt := (Approx-)LMOA (?f (xt ))
4:
S := S ? {zt }
5:
Variant 0:
xt+1 = arg min kx?(xt ? L1 ?f (xt ))k2
x?cone(S)
Variant 1:
xt+1 = arg minx?cone(S) f (x)
7:
Remove atoms with zero weights from S
8: end for
6:
At each iteration, Algorithm 4 maintains the set of active atoms S by adding zt and removing atoms
with zero weights after the update. In Variant 0, the algorithm minimizes the quadratic upper bound
gxt (x) on f at xt (see Section 2) imitating a gradient descent step with projection onto a ?varying?
target, i.e., cone(S). In Variant 1, the original objective f is minimized over cone(S) at each iteration,
which is in general more efficient than minimizing f over cone(A) using a generic solver for cone
constrained problems. For f (x) = 12 ky ? xk2 , y ? H, Variant 1 recovers Algorithm 1 in [52] and
the OMP variant in [5] which both only apply to this specific objective f .
3.2
Computational Complexity
cost per iteration convergence
k(t)
algorithm
We briefly discuss the computaNNMP
C
+
O(d)
O(1/t)
tional complexity of the algorithms
t
PWMP
C + O(d + td)
O e??k(t)
we introduced. For H = Rd , sums
?
3|A|!+1
and inner products have cost O(d).
AMP
C + O(d + td)
O e? 2 k(t)
t/2
Let us assume that each call of the
t
??k(t)
FCMP v. 0 C + O(d) + h0
O e
3|A|!+1
LMO has cost C on the set A and
??k(t)
FCMP v. 1 C + O(d) + h1
O e
t
O(td) on S. The variants 0 and 1
of FCMP solve a cone problem at
each iteration with cost h0 and h1 , Table 1: Computational complexity versus convergence rate (see Section 4) for strongly convex objectives
respectively. In general, h0 can be
much smaller than h1 . In Table 1
we report the cost per iteration for every algorithm along with the asymptotic convergence rates
derived in Section 4.
4
Convergence Rates
In this section, we present convergence guarantees for Algorithms 2, 3, and 4. All proofs are deferred
to the Appendix in the supplementary material. We write x? ? arg minx?cone(A) f (x) for an optimal
solution. Our rates will depend on the atomic norm of the solution and the iterates of the respective
algorithm variant:
? = max {kx? kA , kx0 kA . . . , kxT kA } .
(6)
If the optimum is not unique, we consider x? to be one of largest atomic norm. A more intuitive
and looser notion is to simply upper-bound ? by the diameter of the level set of the initial iterate
x0 measured by the atomic norm. Then, boundedness follows since the presented method is a
descent method (due to Lemma 1 and line search on the quadratic upper bound, each iteration strictly
5
decreases the objective and our method stops only at the optimum). This justifies the statement
f (xt ) ? f (x0 ). Hence, ? must be bounded for any sequence of iterates produced by the algorithm,
and the convergence rates presented in this section are valid as T goes to infinity. A similar notion to
measure the convergence of MP was established in [32]. All of our algorithms and rates can be made
affine invariant. We defer this discussion to Appendix B.
4.1
Sublinear Convergence
We now present the convergence results for the non-negative and Fully-Corrective Matching Pursuit
algorithms. Sublinear convergence of Algorithm 3 is addressed in Theorem 3.
Theorem 2. Let A ? H be a bounded set with 0 ? A, ? := max {kx? kA , kx0 kA , . . . , kxT kA , }
and f be L-smooth over ? conv(A ? ?A). Then, Algorithms 2 and 4 converge for t ? 0 as
4 2? L?2 radius(A)2 + ?0
?
f (xt ) ? f (x ) ?
,
?t + 4
where ? ? (0, 1] is the relative accuracy parameter of the employed approximate LMO (see Equation (3)).
Relation to FW rates. By rescaling A by a large enough factor ? > 0, FW with ? A as atom
set could in principle be used to solve (4). In fact, for large enough ? , only the constraints of (4)
become active when minimizing f over conv(? A). The sublinear convergence rate obtained with
this approach is up to constants identical to that in Theorem 2 for our MP variants, see [21]. However,
as the correct scaling is unknown, one has to either take the risk of choosing ? too small and hence
failing to recover an optimal solution of (4), or to rely on too large ? which can result in slow
convergence. In contrast, knowledge of ? is not required to run our MP variants.
Relation to MP rates. If A is symmetric, we have that lin(A) = cone(A) and it is easy to show
that the additional direction ? kxxtt k in Algorithm 2 is never selected. Therefore, Algorithm 2 becomes
equivalent to Variant 0 of Algorithm 1, while Variant 1 of Algorithm 1 is equivalent to Variant 0 of
Algorithm 4. The rate specified in Theorem 2 hence generalizes the sublinear rate in [32, Theorem 2]
for symmetric A.
4.2
Linear Convergence
We start by recalling some of the geometric complexity quantities that were introduced in the context
of FW and are adapted here to the optimization problem we aim to solve (minimization over cone(A)
instead of conv(A)).
Directional Width. The directional width of a set A w.r.t. a direction r ? H is defined as:
r
dirW (A, r) := max krk
,s ? v
s,v?A
(7)
Pyramidal Directional Width [28]. The Pyramidal Directional Width of a set A with respect to a
direction r and a reference point x ? conv(A) is defined as:
P dirW (A, r, x) := min dirW (S ? {s(A, r)}, r),
S?Sx
(8)
where Sx := {S | S ? A and x is a proper convex combination of all the elements in S} and
r
s(A, r) := maxs?A h krk
, si.
Inspired by the notion of pyramidal width in [28], which is the minimal pyramidal directional width
computed over the set of feasible directions, we now define the cone width of a set A where only
the generating faces (g-faces) of cone(A) (instead of the faces of conv(A)) are considered. Before
doing so we introduce the notions of face, generating face, and feasible direction.
Face of a convex set. Let us consider a set K with a k?dimensional affine hull along with a
point x ? K. Then, K is a k?dimensional face of conv(A) if K = conv(A) ? {y : hr, y ? xi =
0} for some normal vector r and conv(A) is contained in the half-space determined by r, i.e.,
hr, y ? xi ? 0, ? y ? conv(A). Intuitively, given a set conv(A) one can think of conv(A) being a
dim(conv(A))?dimensional face of itself, an edge on the border of the set a 1-dimensional face and
a vertex a 0-dimensional face.
6
Face of a cone and g-faces. Similarly, a k?dimensional face of a cone is an open and unbounded
set cone(A) ? {y : hr, y ? xi = 0} for some normal vector r and cone(A) is contained in the half
space determined by r. We can define the generating faces of a cone as:
g-faces(cone(A)) := {B ? conv(A) : B ? faces(cone(A))} .
Note that g-faces(cone(A)) ? faces(conv(A)) and conv(A) ? g-faces(cone(A)). Furthermore,
for each K ? g-faces(cone(A)), cone(K) is a k?dimensional face of cone(A).
We now introduce the notion of feasible directions. A direction d is feasible from x ? cone(A) if it
points inwards cone(A), i.e., if ?? > 0 s.t. x + ?d ? cone(A). Since a face of the cone is itself a
cone, if a direction is feasible from x ? cone(K) \ 0, it is feasible from every positive rescaling of x.
We therefore can consider only the feasible directions on the generating faces (which are closed and
bounded sets). Finally, we define the cone width of A.
Cone Width.
CWidth(A) :=
min
K?g-faces(cone(A))
x?K
r?cone(K?x)\{0}
P dirW (K ? A, r, x)
(9)
We are now ready to show the linear convergence of Algorithms 3 and 4.
Theorem 3. Let A ? H be a bounded set with 0 ? A and let the objective function f : H ? R be
both L-smooth and ?-strongly convex over ? conv(A ? ?A). Then, the suboptimality of the iterates
of Algorithms 3 and 4 decreases geometrically at each step in which ? < ?vt (henceforth referred to
as ?good steps?) as:
?t+1 ? (1 ? ?) ?t ,
(10)
2
?
where ? := ? 2 ?LCWidth(A)
diam(A)2 ? (0, 1], ?t := f (xt )?f (x ) is the suboptimality at step t and ? ? (0, 1]
is the relative accuracy parameter of the employed approximate LMO (3). For AMP (Algorithm 3),
? AMP = ?/2. If ? = 0 Algorithm 3 converges with rate O(1/k(t)) where k(t) is the number of
?good steps? up to iteration t.
Discussion. To obtain a linear convergence rate, one needs to upper-bound the number of ?bad
steps? t ? k(t) (i.e., steps with ? ? ?vt ). We have that k(t) = t for Variant 1 of FCMP (Algorithm 4),
k(t) ? t/2 for AMP (Algorithm 3) and k(t) ? t/(3|A|! + 1) for PWMP (Algorithm 3) and Variant 0
of FCMP (Algorithm 4). This yields a global linear convergence rate of ?t ? ?0 exp (??k(t)). The
bound for PWMP is very loose and only meaningful for finite sets A. However, it can be observed
in the experiments in the supplementary material (Appendix A) that only a very small fraction of
iterations result in bad PWMP steps in practice. Further note that Variant 1 of FCMP (Algorithm 4)
does not produce bad steps. Also note that the bounds on the number of good steps given above are
the same as for the corresponding FW variants and are obtained using the same (purely combinatorial)
arguments as in [28].
Relation to previous MP rates. The linear convergence of the generalized (not non-negative) MP
variants studied in [32] crucially depends on the geometry of the set which is characterized by the
Minimal Directional Width mDW(A):
mDW(A) :=
min maxh
d?lin(A) z?A
d6=0
d
, zi .
kdk
(11)
The following Lemma relates the Cone Width with the minimal directional width.
Lemma 4. If the origin is in the relative interior of conv(A) with respect to its linear span, then
cone(A) = lin(A) and CWidth(A) = mDW(A).
Now, if the set A is symmetric or, more generally, if cone(A) spans the linear space lin(A) (which
implies that the origin is in the relative interior of conv(A)), there are no bad steps. Hence, by
Lemma 4, the linear rate obtained in Theorem 3 for non-negative MP variants generalizes the one
presented in [32, Theorem 7] for generalized MP variants.
7
Relation to FW rates. Optimization over conic hulls with non-negative MP is more similar to FW
than to MP itself in the following sense. For MP, every direction in lin(A) allows for unconstrained
steps, from any iterate xt . In contrast, for our non-negative MPs, while some directions allow for
unconstrained steps from some iterate xt , others are constrained, thereby leading to the dependence
of the linear convergence rate on the cone width, a geometric constant which is very similar in spirit
to the Pyramidal Width appearing in the linear convergence bound in [28] for FW. Furthermore, as
for Algorithm 3, the linear rate of Away-steps and Pairwise FW holds only for good steps. We finally
relate the cone width with the Pyramidal Width [28]. The Pyramidal Width is defined as
PWidth(A) :=
min
K?faces(conv(A))
x?K
r?cone(K?x)\{0}
P dirW (K ? A, r, x).
We have CWidth(A) ? PWidth(A) as the minimization in the definition (9) of CWidth(A) is only
over the subset g-faces(cone(A)) of faces(conv(A)). As a consequence, the decrease per iteration
characterized in Theorem 3 is larger than what one could obtain with FW on the rescaled convex set
? A (see Section 4.1 for details about the rescaling). Furthermore, the decrease characterized in [28]
scales as 1/? 2 due to the dependence on 1/ diam(conv(A))2 .
5
Related Work
The line of recent works by [44, 46, 47, 48, 37, 32] targets the generalization of MP from the
least-squares objective to general smooth objectives and derives corresponding convergence rates
(see [32] for a more in-depth discussion). However, only little prior work targets MP variants with
non-negativity constraint [5, 38, 52]. In particular, the least-squares objective was addressed and
no rigorous convergence analysis was carried out. [5, 52] proposed an algorithm equivalent to our
Algorithm 4 for the least-squares case. More specifically, [52] then developed an acceleration heuristic,
whereas [5] derived a coherence-based recovery guarantee for sparse linear combinations of atoms.
Apart from MP-type algorithms, there is a large variety of non-negative least-squares algorithms,
e.g., [30], in particular also for matrix and tensor spaces. The gold standard in factorization problems
is projected gradient descent with alternating minimization, see [43, 4, 45, 23]. Other related works
are [40], which is concerned with the feasibility problem on symmetric cones, and [19], which
introduces a norm-regularized variant of problem (4) and solves it using FW on a rescaled convex
set. To the best of our knowledge, in the context of MP-type algorithms, we are the first to combine
general convex objectives with conic constraints and to derive corresponding convergence guarantees.
Boosting: In an earlier line of work, a flavor of the generalized MP became popular in the context
of boosting, see [35]. The literature on boosting is vast, we refer to [42, 35, 7] for a general overview.
Taking the optimization perspective given in [42], boosting is an iterative greedy algorithm minimizing
a (strongly) convex objective over the linear span of a possibly infinite set called hypothesis class.
The convergence analysis crucially relies on the assumption of the origin being in the relative interior
of the hypothesis class, see Theorem 1 in [17]. Indeed, Algorithm 5.2 of [35] might not converge
if the [39] alignment assumption is violated. Here, we managed to relax this assumption while
preserving essentially the same asymptotic rates in [35, 17]. Our work is therefore also relevant in
the context of (non-negative) boosting.
6
Illustrative Experiments
We illustrate the performance of the presented algorithms on three different exemplary tasks, showing
that our algorithms are competitive with established baselines across a wide range of objective functions, domains, and data sets while not being specifically tailored to any of these tasks (see Section 3.2
for a discussion of the computational complexity of the algorithms). Additional experiments targeting
KL divergence NMF, non-negative tensor factorization, and hyperspectral image unmixing can be
found in the appendix.
Synthetic data. We consider minimizing the least squares objective on the conic hull
of 100 unit-norm vectors sampled at random in the first orthant of R50 . We compare
the convergence of Algorithms 2, 3, and 4 with the Fast Non-Negative MP (FNNOMP)
of [52], and Variant 3 (line-search) of the FW algorithm in [32] on the atom set rescaled
by ? = 10kyk (see Section 4.1), observing linear convergence for our corrective variants.
8
Suboptimality
104
10
2
10
0
Synthetic data
PWMP (Alg. 3)
NNMP (Alg. 2)
FCMP (Alg. 4)
FW
AMP (Alg. 3)
FNNOMP
Figure 2 shows the suboptimality ?t , averaged over 20
realizations of A and y, as a function of the iteration t. As
expected, FCMP achieves fastest convergence followed
by PWMP, AMP and NNMP. The FNNOMP gets stuck
instead. Indeed, [52] only show that the algorithm terminates and not its convergence.
10-2
Non-negative matrix factorization. The second task
consists of decomposing a given matrix into the product
10
0
10
20
30
40
50
of two non-negative matrices as in Equation (1) of [20].
Iteration
We consider the intersection of the positive semidefinite
Figure 2: Synthetic data experiment.
cone and the positive orthant. We parametrize the set A as
the set of matrices obtained as an outer product of vectors
from A1 = {z ? Rk : zi ? 0 ? i} and A2 = {z ? Rd : zi ? 0 ? i}. The LMO is approximated
using a truncated power method [55], and we perform atom correction with greedy coordinate descent
see, e.g., [29, 18], to obtain a better objective value while maintaining the same (small) number of
atoms. We consider three different datasets: The Reuters Corpus2 , the CBCL face dataset3 and the
KNIX dataset4 . The subsample of the Reuters corpus we used is a term frequency matrix of 7,769
documents and 26,001 words. The CBCL face dataset is composed of 2,492 images of 361 pixels
each, arranged into a matrix. The KNIX dataset contains 24 MRI slices of a knee, arranged in a
matrix of size 262, 144 ? 24. Pixels are divided by their overall mean intensity. For interpretability
reasons, there is interest to decompose MRI data into non-negative factorizations [25]. We compare
PWMP and FCMP against the multiplicative (mult) and the alternating (als) algorithm of [4], and the
greedy coordinate descent (GCD) of [20]. Since the Reuters corpus is much larger than the CBCL
and the KNIX dataset we only used the GCD for which a fast implementation in C is available. We
report the objective value for fixed values of the rank in Table 2, showing that FCMP outperform all
the baselines across all the datasets. PWMP achieves smallest error on the Reuters corpus.
-4
Non-negative garrote. We consider the non-negative garrote which is a common approach to
model order selection [6]. We evaluate NNMP, PWMP, and FCMP in the experiment described
in [33], where the non-negative garrote is used to perform model order selection for logistic regression
(i.e., for a non-quadratic objective function). We evaluated training and test accuracy on 100 random
splits of the sonar dataset from the UCI machine learning repository. In Table 3 we compare the
median classification accuracy of our algorithms with that of the cyclic coordinate descent algorithm
(NNG) from [33].
Reuters
CBCL
CBCL
KNIX
K = 10 K = 10 K = 50 K = 10
mult
2.4241e3 1.1405e3 2.4471e03
als
2.73e3
3.84e3 2.7292e03
GCD
5.9799e5 2.2372e3
806
2.2372e03
PWMP 5.9591e5 2.2494e3 789.901 2.2494e03
FCMP
5.9762e5 2.2364e3 786.15 2.2364e03
Table 2: Objective value for least-squares non-negative
matrix factorization with rank K.
algorithm
7
training accuracy
test accuracy
NNMP 0.8345 ? 0.0242 0.7419 ? 0.0389
PWMP 0.8379 ? 0.0240 0.7419 ? 0.0392
FCMP 0.8345 ? 0.0238 0.7419 ? 0.0403
NNG
0.8069 ? 0.0518 0.7258 ? 0.0602
Table 3: Logistic Regression with non-negative
Garrote, median ? std. dev.
Conclusion
In this paper, we considered greedy algorithms for optimization over the convex cone, parametrized
as the conic hull of a generic atom set. We presented a novel formulation of NNMP along with a
comprehensive convergence analysis. Furthermore, we introduced corrective variants with linear
convergence guarantees, and verified this convergence rate in numerical applications. We believe that
the generality of our novel analysis will be useful to design new, fast algorithms with convergence
guarantees, and to study convergence of existing heuristics, in particular in the context of non-negative
matrix and tensor factorization.
2
http://www.nltk.org/book/ch02.html
http://cbcl.mit.edu/software-datasets/FaceData2.html
4
http://www.osirix-viewer.com/resources/dicom-image-library/
3
9
References
[1] Animashree Anandkumar, Rong Ge, Daniel J Hsu, Sham M Kakade, and Matus Telgarsky.
Tensor decompositions for learning latent variable models. Journal of Machine Learning
Research, 15(1):2773?2832, 2014.
[2] M?rio C?sar Ugulino Ara?jo, Teresa Cristina Bezerra Saldanha, Roberto Kawakami Harrop
Galvao, Takashi Yoneyama, Henrique Caldas Chame, and Valeria Visani. The successive projections algorithm for variable selection in spectroscopic multicomponent analysis. Chemometrics
and Intelligent Laboratory Systems, 57(2):65?73, 2001.
[3] Jonas Behr, Andr? Kahles, Yi Zhong, Vipin T Sreedharan, Philipp Drewe, and Gunnar R?tsch.
Mitie: Simultaneous rna-seq-based transcript identification and quantification in multiple
samples. Bioinformatics, 29(20):2529?2538, 2013.
[4] Michael W Berry, Murray Browne, Amy N Langville, V Paul Pauca, and Robert J Plemmons.
Algorithms and applications for approximate nonnegative matrix factorization. Computational
statistics & data analysis, 52(1):155?173, 2007.
[5] Alfred M Bruckstein, Michael Elad, and Michael Zibulevsky. On the uniqueness of nonnegative
sparse solutions to underdetermined systems of equations. IEEE Transactions on Information
Theory, 54(11):4813?4820, 2008.
[6] P B?hlmann and B Yu. Boosting, model selection, lasso and nonnegative garrote. Technical
Report 127, Seminar f?r Statistik ETH Z?rich, 2005.
[7] Peter B?hlmann and Bin Yu. Boosting. Wiley Interdisciplinary Reviews: Computational
Statistics, 2(1):69?74, 2010.
[8] Martin Burger. Infinite-dimensional optimization and optimal design. 2003.
[9] Sheng Chen, Stephen A Billings, and Wan Luo. Orthogonal least squares methods and their
application to non-linear system identification. International Journal of control, 50(5):1873?
1896, 1989.
[10] Andrzej Cichocki and PHAN Anh-Huy. Fast local algorithms for large scale nonnegative matrix
and tensor factorizations. IEICE transactions on fundamentals of electronics, communications
and computer sciences, 92(3):708?721, 2009.
[11] Ernie Esser, Yifei Lou, and Jack Xin. A method for finding structured sparse solutions to
nonnegative least squares problems with applications. SIAM Journal on Imaging Sciences,
6(4):2010?2046, 2013.
[12] M Frank and P Wolfe. An algorithm for quadratic programming. Naval research logistics
quarterly, 1956.
[13] Nicolas Gillis. Successive nonnegative projection algorithm for robust nonnegative blind source
separation. SIAM Journal on Imaging Sciences, 7(2):1420?1450, 2014.
[14] Nicolas Gillis and Fran?ois Glineur. Accelerated multiplicative updates and hierarchical als
algorithms for nonnegative matrix factorization. Neural Computation, 24(4):1085?1105, 2012.
[15] Nicolas Gillis, Da Kuang, and Haesun Park. Hierarchical clustering of hyperspectral images
using rank-two nonnegative matrix factorization. IEEE Transactions on Geoscience and Remote
Sensing, 53(4):2066?2078, 2015.
[16] Nicolas Gillis and Robert Luce. A fast gradient method for nonnegative sparse regression with
self dictionary. arXiv preprint arXiv:1610.01349, 2016.
[17] Alexander Grubb and J Andrew Bagnell. Generalized boosting algorithms for convex optimization. arXiv preprint arXiv:1105.2054, 2011.
[18] Xiawei Guo, Quanming Yao, and James T Kwok. Efficient sparse low-rank tensor completion
using the Frank-Wolfe algorithm. In AAAI Conference on Artificial Intelligence, 2017.
[19] Zaid Harchaoui, Anatoli Juditsky, and Arkadi Nemirovski. Conditional gradient algorithms for
norm-regularized smooth convex optimization. Mathematical Programming, 152(1-2):75?112,
2015.
[20] Cho-Jui Hsieh and Inderjit S Dhillon. Fast coordinate descent methods with variable selection
for non-negative matrix factorization. In Proceedings of the 17th ACM SIGKDD international
conference on Knowledge discovery and data mining, pages 1064?1072. ACM, 2011.
[21] Martin Jaggi. Revisiting Frank-Wolfe: Projection-Free Sparse Convex Optimization. In ICML
2013 - Proceedings of the 30th International Conference on Machine Learning, 2013.
10
[22] Hyunsoo Kim, Haesun Park, and Lars Elden. Non-negative tensor factorization based on
alternating large-scale non-negativity-constrained least squares. In Bioinformatics and Bioengineering, 2007. BIBE 2007. Proceedings of the 7th IEEE International Conference on, pages
1147?1151. IEEE, 2007.
[23] Jingu Kim, Yunlong He, and Haesun Park. Algorithms for nonnegative matrix and tensor
factorizations: A unified view based on block coordinate descent framework. Journal of Global
Optimization, 58(2):285?319, 2014.
[24] Jingu Kim and Haesun Park. Fast nonnegative tensor factorization with an active-set-like
method. In High-Performance Scientific Computing, pages 311?326. Springer, 2012.
[25] Ivica Kopriva and Andrzej Cichocki. Nonlinear band expansion and 3d nonnegative tensor
factorization for blind decomposition of magnetic resonance image of the brain. In International
Conference on Latent Variable Analysis and Signal Separation, pages 490?497. Springer, 2010.
[26] Abhishek Kumar, Vikas Sindhwani, and Prabhanjan Kambadur. Fast conical hull algorithms for
near-separable non-negative matrix factorization. In ICML (1), pages 231?239, 2013.
[27] Simon Lacoste-Julien and Martin Jaggi. An Affine Invariant Linear Convergence Analysis for
Frank-Wolfe Algorithms. In NIPS 2013 Workshop on Greedy Algorithms, Frank-Wolfe and
Friends, December 2013.
[28] Simon Lacoste-Julien and Martin Jaggi. On the Global Linear Convergence of Frank-Wolfe
Optimization Variants. In NIPS 2015, pages 496?504, 2015.
[29] S?ren Laue. A Hybrid Algorithm for Convex Semidefinite Optimization. In ICML, 2012.
[30] Charles L Lawson and Richard J Hanson. Solving least squares problems, volume 15. SIAM,
1995.
[31] Daniel D Lee and H Sebastian Seung. Algorithms for non-negative matrix factorization. In
Advances in neural information processing systems, pages 556?562, 2001.
[32] Francesco Locatello, Rajiv Khanna, Michael Tschannen, and Martin Jaggi. A unified optimization view on generalized matching pursuit and frank-wolfe. In Proc. International Conference
on Artificial Intelligence and Statistics (AISTATS), 2017.
[33] Enes Makalic and Daniel F Schmidt. Logistic regression with the nonnegative garrote. In
Australasian Joint Conference on Artificial Intelligence, pages 82?91. Springer, 2011.
[34] St?phane Mallat and Zhifeng Zhang. Matching pursuits with time-frequency dictionaries. IEEE
Transactions on Signal Processing, 41(12):3397?3415, 1993.
[35] Ron Meir and Gunnar R?tsch. An introduction to boosting and leveraging. In Advanced lectures
on machine learning, pages 118?183. Springer, 2003.
[36] Jos? MP Nascimento and Jos? MB Dias. Vertex component analysis: A fast algorithm to unmix
hyperspectral data. IEEE transactions on Geoscience and Remote Sensing, 43(4):898?910,
2005.
[37] Hao Nguyen and Guergana Petrova. Greedy strategies for convex optimization. Calcolo, pages
1?18, 2014.
[38] Robert Peharz, Michael Stark, and Franz Pernkopf. Sparse nonnegative matrix factorization
using l0-constraints. In IEEE, editor, Proceedings of MLSP, pages 83 ? 88, Aug 2010.
[39] Javier Pena and Daniel Rodriguez. Polytope conditioning and linear convergence of the frankwolfe algorithm. arXiv preprint arXiv:1512.06142, 2015.
[40] Javier Pena and Negar Soheili. Solving conic systems via projection and rescaling. Mathematical
Programming, pages 1?25, 2016.
[41] Aleksei Pogorelov. Extrinsic geometry of convex surfaces, volume 35. American Mathematical
Soc., 1973.
[42] Gunnar R?tsch, Sebastian Mika, Manfred K Warmuth, et al. On the convergence of leveraging.
In NIPS, pages 487?494, 2001.
[43] F Sha, LK Saul, and Daniel D Lee. Multiplicative updates for nonnegative quadratic programming in support vector machines. Advances in Neural Information Processing Systems, 15,
2002.
[44] Shai Shalev-Shwartz, Nathan Srebro, and Tong Zhang. Trading Accuracy for Sparsity in
Optimization Problems with Sparsity Constraints. SIAM Journal on Optimization, 20:2807?
2832, 2010.
11
[45] Amnon Shashua and Tamir Hazan. Non-negative tensor factorization with applications to
statistics and computer vision. In Proceedings of the 22nd international conference on Machine
learning, pages 792?799. ACM, 2005.
[46] Vladimir Temlyakov. Chebushev Greedy Algorithm in convex optimization. arXiv.org, December 2013.
[47] Vladimir Temlyakov. Greedy algorithms in convex optimization on Banach spaces. In 48th
Asilomar Conference on Signals, Systems and Computers, pages 1331?1335. IEEE, 2014.
[48] VN Temlyakov. Greedy approximation in convex optimization. Constructive Approximation,
41(2):269?296, 2015.
[49] Joel A Tropp. Greed is good: algorithmic results for sparse approximation. IEEE Transactions
on Information Theory, 50(10):2231?2242, 2004.
[50] Zheng Wang, Ming jun Lai, Zhaosong Lu, Wei Fan, Hasan Davulcu, and Jieping Ye. Rank-one
matrix pursuit for matrix completion. In ICML, pages 91?99, 2014.
[51] Max Welling and Markus Weber. Positive tensor factorization. Pattern Recognition Letters,
22(12):1255?1261, 2001.
[52] Mehrdad Yaghoobi, Di Wu, and Mike E Davies. Fast non-negative orthogonal matching pursuit.
IEEE Signal Processing Letters, 22(9):1229?1233, 2015.
[53] Yuning Yang, Siamak Mehrkanoon, and Johan A K Suykens. Higher order Matching Pursuit
for Low Rank Tensor Learning. arXiv.org, March 2015.
[54] Quanming Yao and James T Kwok. Greedy learning of generalized low-rank models. In IJCAI,
2016.
[55] Xiao-Tong Yuan and Tong Zhang. Truncated power method for sparse eigenvalue problems. J.
Mach. Learn. Res., 14(1):899?925, April 2013.
12
| 6679 |@word repository:1 version:1 mri:2 briefly:2 norm:13 nd:1 open:1 crucially:2 hsieh:1 decomposition:2 kz1:1 thereby:2 boundedness:1 reduction:1 initial:1 cyclic:1 cristina:1 electronics:1 selecting:1 contains:1 ap1:1 kahles:1 daniel:5 document:1 amp:10 hyunsoo:1 frankwolfe:1 existing:5 kx0:2 ka:7 current:2 z2:2 luo:1 com:1 si:1 must:1 attracted:1 written:1 numerical:2 enables:1 zaid:1 siamak:1 remove:2 update:11 juditsky:1 stationary:1 greedy:16 selected:5 intelligence:3 half:3 lkdt:1 slowing:1 parameterization:1 kyk:1 warmuth:1 steepest:2 manfred:1 iterates:7 parameterizations:1 boosting:9 ron:1 philipp:1 successive:2 org:3 zhang:3 unbounded:1 mathematical:3 along:5 become:1 dicom:1 jonas:1 yuan:1 prove:1 consists:1 combine:1 inside:1 manner:1 introduce:4 x0:11 pairwise:5 expected:1 indeed:3 plemmons:1 brain:1 ara:1 inspired:3 ming:1 td:3 little:1 solver:1 increasing:1 becomes:1 burger:1 conv:29 notation:1 bounded:5 underlying:1 anh:1 what:1 minimizes:2 tschannen:2 developed:1 unified:2 finding:1 guarantee:10 every:4 exactly:1 k2:4 control:2 originates:1 unit:1 arguably:1 before:2 positive:5 local:1 consequence:1 mach:1 solely:1 mika:1 might:1 studied:4 shaded:1 fastest:1 factorization:24 nemirovski:1 range:1 obeys:1 averaged:1 unique:1 atomic:5 practice:4 block:1 differs:1 area:2 empirical:2 eth:4 mult:2 matching:12 projection:5 davy:1 word:1 regular:1 jui:1 get:2 onto:1 interior:6 selection:6 targeting:1 context:7 risk:1 www:2 equivalent:3 maxz:1 tamir:1 jieping:1 go:1 rajiv:1 convex:33 simplicity:1 knee:1 recovery:1 amy:1 hd:5 notion:5 coordinate:5 sar:1 updated:1 target:7 pt:1 mallat:1 exact:2 programming:4 hypothesis:2 origin:7 element:3 wolfe:10 recognition:1 expensive:1 approximated:1 std:1 observed:2 mike:1 preprint:3 wang:1 initializing:1 revisiting:1 remote:2 trade:1 rescaled:3 zibulevsky:1 decrease:4 zig:2 principled:1 substantial:1 intuition:2 convexity:1 complexity:7 tsch:4 seung:1 chet:1 depend:2 solving:4 purely:1 efficiency:1 swap:2 joint:1 represented:1 corrective:9 distinct:1 fast:10 query:3 emanating:1 artificial:3 choosing:1 shalev:1 h0:3 heuristic:2 elad:1 larger:2 supplementary:3 solve:4 relax:1 widely:1 ability:1 statistic:4 think:1 itself:3 sequence:1 eigenvalue:1 kxt:2 differentiable:1 exemplary:1 propose:3 mb:1 product:5 fr:1 aligned:2 relevant:1 lmoa:6 uci:1 realization:1 achieve:2 gold:1 intuitive:1 ky:2 chemometrics:1 ijcai:1 convergence:54 empty:1 optimum:3 unmixing:2 generating:4 produce:1 converges:1 phane:1 telgarsky:1 help:1 derive:3 andrew:1 illustrate:1 completion:2 friend:1 measured:1 pauca:1 transcript:1 aug:1 solves:2 soc:1 implemented:1 ois:1 tion:1 implies:1 trading:1 direction:25 nng:2 radius:3 drawback:1 correct:1 hull:18 lars:1 material:3 bin:1 premise:1 hx:2 generalization:1 decompose:1 spectroscopic:1 underdetermined:1 strictly:1 viewer:1 rong:1 hold:2 correction:3 extension:1 considered:2 normal:2 exp:1 cbcl:6 algorithmic:1 matus:1 dictionary:4 achieves:2 a2:2 smallest:1 xk2:2 uniqueness:1 failing:1 proc:1 applicable:1 combinatorial:1 currently:2 largest:1 gauge:1 minimization:8 mit:1 rna:1 always:2 aim:1 avoid:1 zhong:1 varying:1 encode:1 l0:1 derived:3 takashi:1 naval:1 rank:7 contrast:3 rigorous:1 sigkdd:1 kim:3 sense:2 baseline:2 dim:1 tional:1 rio:1 dependent:1 stopping:1 epfl:2 relation:4 selective:1 pixel:2 arg:8 overall:1 classification:1 issue:2 html:2 resonance:1 constrained:5 equal:1 never:1 calcolo:1 beach:1 atom:43 identical:1 park:4 yu:2 icml:4 minimized:1 report:3 others:1 fundamentally:1 richard:1 intelligent:2 composed:1 divergence:1 comprehensive:1 geometry:3 negation:1 recalling:1 interest:3 mining:1 zheng:1 evaluation:1 zhaosong:1 alignment:4 joel:1 introduces:1 deferred:1 admitting:1 semidefinite:2 behind:1 bioengineering:1 edge:1 dataset3:1 respective:1 orthogonal:4 euclidean:1 re:1 h1:3 theoretical:2 minimal:3 earlier:1 dev:1 hlmann:2 ordinary:1 cost:5 vertex:2 subset:4 kuang:1 too:2 dependency:1 synthetic:4 cho:1 st:2 thanks:1 international:7 siam:4 fundamental:1 interdisciplinary:1 lee:2 physic:1 off:1 jos:2 michael:6 concrete:1 yao:2 intersecting:1 jo:1 aaai:1 wan:1 possibly:2 henceforth:1 book:1 american:1 leading:2 return:1 stark:1 rescaling:4 unmix:1 australasian:1 mlsp:1 dnt:1 explicitly:1 nari:1 blind:2 depends:2 mp:49 multiplicative:3 grubb:1 view:2 closed:4 valeria:1 hazan:1 analyze:1 doing:1 start:1 shashua:1 competitive:1 recover:3 complicated:1 shai:1 observing:1 defer:1 arkadi:1 langville:1 maintains:1 simon:2 contribution:1 minimize:1 square:11 formed:1 accuracy:7 became:1 yield:3 identify:1 directional:7 identification:2 produced:1 critically:1 none:1 ren:1 lu:1 simultaneous:1 whenever:1 sebastian:2 definition:2 against:1 frequency:2 james:2 proof:2 di:2 recovers:2 associated:1 stop:1 sampled:1 hsu:1 newly:1 popular:2 animashree:1 gain:1 recall:1 dataset:4 knowledge:3 hilbert:5 peharz:1 javier:2 back:1 ta:7 dt:4 higher:1 wei:1 april:1 arranged:2 evaluated:1 formulation:1 strongly:5 generality:2 furthermore:6 though:2 osirix:1 sheng:1 hand:1 tropp:1 nonlinear:1 rodriguez:1 kawakami:1 logistic:4 khanna:1 quality:2 scientific:1 believe:1 ieice:1 usa:1 ye:1 contain:1 managed:1 former:1 hence:6 alternating:3 symmetric:7 dhillon:1 nonzero:1 laboratory:1 white:1 self:1 width:17 illustrative:2 vipin:1 mpi:1 suboptimality:4 generalized:12 presenting:1 demonstrate:1 l1:2 weber:1 image:5 jack:1 novel:3 charles:1 common:1 overview:1 conditioning:1 gcd:3 volume:2 banach:1 discussed:1 he:1 pena:2 raetsch:1 significant:1 refer:1 queried:1 ai:2 smoothness:1 approx:5 unconstrained:3 rd:2 similarly:1 yifei:1 esser:1 longer:1 maxh:1 surface:1 jaggi:6 recent:3 perspective:2 optimizing:1 inf:2 apart:1 termed:1 vt:14 yi:2 preserving:1 minimum:1 regained:1 additional:5 omp:4 employed:2 converge:5 signal:5 stephen:1 relates:1 multiple:1 desirable:1 sham:1 harchaoui:1 ing:1 technical:1 match:1 characterized:3 smooth:8 faster:2 long:1 lin:11 divided:1 lai:1 proximate:1 a1:2 feasibility:1 converging:1 variant:44 regression:5 essentially:2 vision:1 arxiv:8 iteration:20 normalization:1 tailored:2 achieved:1 suykens:1 whereas:1 addressed:2 pyramidal:7 source:1 median:2 hasan:1 ascent:1 december:2 leveraging:2 spirit:3 effectiveness:1 anandkumar:1 call:4 ee:1 near:1 yang:1 intermediate:3 split:1 enough:2 easy:1 gillis:4 iterate:6 concerned:1 browne:1 zi:12 variety:3 lasso:1 suboptimal:2 inner:3 reduce:2 idea:1 billing:1 luce:1 amnon:1 greed:1 peter:1 returned:1 e3:7 useful:2 generally:1 detailed:1 clear:1 informally:1 amount:1 facedata2:1 band:1 induces:1 maxz1:1 diameter:2 reduced:1 http:3 outperform:1 meir:1 andr:1 extrinsic:1 popularity:1 per:3 alfred:1 write:1 gunnar:4 key:1 prevent:1 verified:1 corpus2:1 lacoste:2 imaging:2 vast:1 geometrically:1 fraction:1 cone:60 year:2 enforced:1 run:1 sum:1 letter:2 clipped:1 wu:1 vn:1 seq:1 fran:1 looser:1 separation:2 garrote:7 coherence:1 scaling:1 appendix:5 bound:9 guaranteed:4 conical:1 followed:2 correspondence:1 fan:1 quadratic:7 paramount:1 nonnegative:16 oracle:3 adapted:1 constraint:8 infinity:1 x2:1 software:1 hy:1 markus:1 sake:1 statistik:1 nathan:1 argument:1 min:12 span:10 kumar:1 separable:1 martin:7 structured:2 according:2 march:1 combination:4 remain:1 across:2 smaller:1 terminates:2 kakade:1 modification:1 explained:1 intuitively:1 invariant:2 imitating:1 ene:1 asilomar:1 equation:4 resource:1 r50:1 previously:3 zurich:3 loose:1 discus:1 mind:1 ge:1 end:4 dia:1 pursuit:14 parametrize:1 generalizes:2 decomposing:1 available:1 apply:4 quarterly:1 hierarchical:2 away:7 generic:3 kwok:2 magnetic:1 appearing:1 schmidt:1 vikas:1 original:1 andrzej:2 clustering:1 ensure:1 include:1 maintaining:2 anatoli:1 murray:1 establish:1 tensor:16 move:1 objective:23 added:1 quantity:1 occurs:1 laue:1 strategy:1 sha:1 dependence:2 mehrdad:1 bagnell:1 minx:3 gradient:8 subspace:1 lou:1 d6:1 outer:1 parametrized:3 manifold:1 polytope:1 dataset4:1 reason:1 prabhanjan:1 kambadur:1 minimizing:4 vladimir:2 inwards:1 robert:3 statement:1 relate:1 glineur:1 hao:1 frank:10 negative:35 design:2 implementation:1 proper:1 zt:20 unknown:1 perform:2 upper:6 francesco:2 datasets:3 finite:4 zagging:2 descent:13 orthant:2 optional:1 truncated:2 logistics:1 communication:1 pernkopf:1 intensity:1 nmf:1 bk:2 introduced:3 namely:1 required:1 specified:1 kl:1 hanson:1 teresa:1 established:2 alternately:1 nip:4 address:1 pattern:1 sparsity:3 max:10 including:1 interpretability:2 power:2 natural:1 rely:2 quantification:1 regularized:2 hybrid:1 hr:3 advanced:1 library:1 julien:2 conic:18 lk:1 ready:1 carried:1 negativity:3 jun:1 cichocki:2 roberto:1 text:2 prior:2 literature:2 geometric:2 berry:1 review:3 discovery:1 l2:1 relative:9 tangent:1 nascimento:1 fully:4 negar:1 lecture:1 asymptotic:2 loss:1 sublinear:7 limitation:2 srebro:1 querying:1 versus:1 degree:1 affine:3 xiao:1 principle:1 editor:1 surprisingly:1 free:1 enjoys:1 formal:1 allow:2 saul:1 wide:1 face:29 taking:2 sparse:9 slice:1 kzk:1 depth:1 valid:1 rich:1 kz:2 preventing:1 kdk:1 commonly:1 made:1 stuck:2 franz:1 projected:1 nguyen:1 far:1 welling:1 transaction:6 temlyakov:3 approximate:6 compact:1 global:3 active:10 bruckstein:1 corpus:3 corroborating:1 assumed:2 kxka:2 xi:4 shwartz:1 abhishek:1 search:8 iterative:1 latent:2 behr:1 sonar:1 table:6 learn:1 johan:1 robust:1 ca:1 nicolas:4 init:4 alg:6 e5:3 expansion:1 excellent:1 domain:7 krk:2 da:1 aistats:1 main:1 linearly:1 reuters:5 border:1 subsample:1 huy:1 paul:1 x1:1 representative:1 gxt:2 referred:1 slow:1 tong:3 wiley:1 seminar:1 lmo:14 explicit:3 haesun:4 candidate:1 kxk2:1 lawson:1 minz:3 third:1 zhifeng:1 theorem:11 removing:1 rk:1 bad:4 specific:1 nltk:1 xt:52 showing:2 z0:1 down:1 sensing:2 derives:1 burden:1 workshop:1 adding:2 hyperspectral:3 justifies:1 sx:2 kx:5 chen:1 flavor:1 phan:1 intersection:1 simply:2 contained:2 geoscience:2 inderjit:1 sindhwani:1 springer:4 ch:4 corresponds:1 minimizer:1 relies:2 acm:3 conditional:1 diam:3 acceleration:1 towards:1 lipschitz:1 replace:1 feasible:10 fw:20 hard:1 infinite:4 specifically:4 reducing:1 determined:2 uniformly:1 lemma:5 called:2 xin:1 meaningful:1 select:1 formally:2 support:1 guo:1 latter:1 arises:1 alexander:1 bioinformatics:2 ethz:3 accelerated:1 violated:1 constructive:1 evaluate:1 phenomenon:2 later:1 dirw:5 |
6,278 | 668 | Single-iteration Threshold Hamming
Networks
Eytan Ruppin
Isaac Meilijson
Moshe Sipper
School of Mathematical Sciences
Raymond and Beverly Sackler Faculty of Exact Sciences
Tel Aviv University, 69978 Tel Aviv, Israel
Abstract
We analyze in detail the performance of a Hamming network classifying inputs that are distorted versions of one of its m stored
memory patterns. The activation function of the memory neurons
in the original Hamming network is replaced by a simple threshold
function. The resulting Threshold Hamming Network (THN) correctly classifies the input pattern, with probability approaching 1,
using only O(mln m) connections, in a single iteration . The THN
drastically reduces the time and space complexity of Hamming Network classifiers.
1
Introduction
Originally presented in (Steinbuch 1961, Taylor 1964) the Hamming network (HN)
has received renewed attention in recent years (Lippmann et. al. 1987, Baum et.
al. 1988). The HN calculates the Hamming distance between the input pattern
and each memory pattern, and selects the memory with the smallest distance. It
is composed of two subnets: The similarity subnet, consisting of an n-neuron input
layer connected with an m-neuron memory layer, calculates the number of equal bits
between the input and each memory pattern. The winner-take-all (WTA) subnet,
consisting of a fully connected m-neuron topology, selects the memory neuron that
best matches the input pattern.
564
Single-iteration Threshold Hamming Networks
The similarity subnet uses mn connections and performs a single iteration. The
WTA sub net has m 2 connections. With randomly generated input and memory
patterns, it converges in 8(m In(mn)) iterations (Floreen 1991). Since m is exponential in n, the space and time complexity of the network is primarily due to
the WTA subnet (Domany & Orland 1987). We analyze the performance of the
HN in the practical scenario where the input pattern is a distorted version of some
stored memory vector. We show that it is possible to replace the original activation function of thf' neurons in the memory layer by a simple threshold function,
and completely discard the WTA subnet. If the threshold is properly tuned, only
the neuron standing for the 'correct' memory is likely to be activated. The resulting Threshold Hamming Network (THN) will perform correctly (with probability
approaching 1) in a single iteration, using only O(m In m) connections instead of
the O( m 2 ) connections in the original HN. We identify the optimal threshold, and
measure its performance relative to the original HN.
2
The Threshold Hamming Network
e",
We examine a HN storing m + 1 memory patterns
1 ~ jJ ~ m + 1, each
being an n-dimensional vector of ?1. The input pattern x is generated by selecting
some memory pattern ~I-' (w.l.g., ~m+l), and letting each bit Xi be either ~f or
-~f with probabilities a and (I - a) respectively, where a > 0.5. To analyze this
HN, we use some tight. approximations to the binomial distribution. Due to space
considerations, their proofs are omitted.
Lemlna 1.
Let X,...., Bin(n,p) . If Xn are integers such that
P(X
> xn) ~
-
(1 - ~
limn-+oo~
= /3 E (p, 1), then
exp{ -n[/3ln /3 + (1 _ /3) In 1 - /3]}
)v1211'n/3(1
- /3)
1p
p
in the sense that the ratio between LHS and RHS converges to 1 as n
the special case p = ~, let G{/3) = In 2 + /31n/3 + (1- /3) In{1 - /3), then
~ 00.
P(X> x ),...., -:--_e7'xp,-,{=--;:-=n=G::::(/3:::::)==}=.
n ,...., (2 - ~ )V211'n/3(1 - 13)
Lenllna 2.
Let Xi ,...., Bin(n,~) be independent, 'Y E (0,1), and let
m
(1)
p
Xn
For
(2)
be as in Lemma 1. If
= (2 - ~)\h7rn/3(I- 13) (ln~) enG ({3),
(3)
then
(4)
Lenllna 3.
Let y,...., Bin(n,Q') with a >~, let (Xi) and 'Y be as in Lemma 2, and let
Let
Xn
be the integer closest to
T}
E (0,1).
nf3, where
/3=a_/a(l-a)z
n
1)
V
_~
2n
(5)
565
566
Meilijson, Ruppin, and Sipper
and zTj is the
T] -
quantile of the standard normal distribution, i.e.,
T]
= -1-
vf2;
jZ'I e-
x
2
/2dx
(6)
-00
Then, if Y and (Xd are independent
P (max(X 1 , X 2 , " ' , Xm) < Y) 2:: P(max(X1 , X 2 , " ' , Xm) <
as n
-+-
00,
~
Xn
Y) =>
,T]
(7)
for m as in (3).
Based on the above binomial probability approximations, we can now propose and
analyze a n-neuron Threshold Hamming Network (THN) that classifies the input
patterns with probability of error not exceeding f, when the input vector is generated
with an initial bit-similarity a: Let Xi be the similarity between the input vector
and the j'th memory pattern (1 ~ j < m), and let Y be the similarity with
the 'correct' memory pattern ~m+l. Choose, and T] so that ,T] 2:: 1 - f, e.g.,
T] = VI="f; determine f3 by (5) and m by (3). Discard the WTA subnet, and
simply replace the neurons of the memory layer by m neurons having a threshold
Xn , the integer closest to nf3. If any memory neuron with similarity at least Xn
is declared 'the winner', then, by Lemma 3, the probability of error is at most f,
where 'error' may be due to the existence of no winner, wrong winner, or multiple
wmners.
,=
3
The Hamming Network and an Optimal Threshold
Hamming Network
We now calculate the choice of the threshold Xn that maximizes the storage capacity m = men, f, a). Let </J (eI? denote the standard normal density (cumulative
distribution function), and let r = </J/(l- eI? denote the corresponding failure rate
function. Then,
Lenuna 4.
The opt.imal proportion between the two error probabilities is
1- ,
---1 - T]
r(zTj)
vna(1 - a) In
(8)
6 '
~ -r==~==~--~
which we will denote by 8.
Proof:
Let AI = max(X 1 ,X2 ,""Xm ), and let Y denote the similarity with the
'correct' memory pattern, as before.
We have seen that P(M < x) ~
exp{ -m J exp{ -nG({3)} 1 } . Since G'(f3) = In (1~ ~)' then by Taylor expansion
21!"n{3(1-{3)(2-~)
P(AI<x)=P(M<xo+x-xo)~exp{-m
I-'
exp{-n[G(f3 + X-XO)]}
n
1
V211' n f3(1 - (3)(2 - fj)
exp{ -nG(f3) - (x - xo) In (1~{3)}
exp{ -m
V211' n f3(1 _ (3)(2 _ ~)
}
=,
}~
(-..L )",0-"
l-fj
(9)
Single-iteration Threshold Hamming Networks
(in accordance with Gnedenko extreme-value distribution of type 1 (Leadbetter et.
al. 1983)). Similarly,
P(Y
< x) = exp{ln P(Y < Xo + x - xo)}
~(z)
x - Xo
P(Y < x o)ex P {?1>$(z) Jna(l- a)}
~
x- Xo
= (1-7])exp{r(z) Jna(1- a)}
(10)
where ~ is the standard normal density function, ?1> is the standard normal cumulative distribution function, ?1>$ = 1 - ?1> and r =
is the corresponding failure
rate function. The probability of correct recognition using a threshold x can now
be expressed as
-J.
P(M < x)P(Y ~ x)
= ,(6)"'0-"'(1- (1-7])exp{r(z)
x - Xo })
Jna(l- a)
(11)
We differentiate expression (11) with respect to Xo - x, and equate the derivative
at Xo = x to zero, to obtain the relation between , and 7] that yields the optimal
threshold, i.e., that which maximizes the probability of correct recognition. This
yields
1'(Z)
1 - 7]
exp{--}
(12)
Jna(l-a)ln~ 7]
,=
We now approximate
1- ,
~
- In ,
~
r(z)
Jna(l- a)ln
4
( 1 - 7] )
(13)
and thus the optimal proportion between the two error probabilities is
1 -;
-~
1 - 7]
r(z)
jna(1 - a) In ~
= {yo
(14)
o
Based on Lemma 4, if the desired probability of error is (, we choose
{Jf.
,=I-I+{Y'
_ 17] (1
t
+ {y)
(15)
We start with, = 7] = ..;r=f, obtain {3 from (5) and {y from (8), and recompute 7]
and, from (15). The limiting values of j3 and, in this iterative process give the
maximal capacity m and threshold x n .
We now compute the error probability t( m, n, a) of the original HN (with the WTA
subnet) for arbitrary tn, n and a, and compare it with (.
Lemma 5.
For arbitrary n, a and t, let m, {3", 7] and {y be as calculated above. Then, the
probability of error ((m, n, a) of the HN satisfies
((m,n,a)~r(I-{Y)
1- e- 61n 6
-.L
{yIn
1-{3
{y6
({y)lH(1+6
1+
(16)
567
568
Meiiijson, Ruppin, and Sipper
where
(17)
is the Gamma function.
Proof:
P(Y ~ M) = LP(Y ~ x)P(M = x) =
x
LP(Y ~ x)[P(M < x+ 1) - P(M < x)] ~
x
L P(Y ~ xo)e- (xo-x)ln 6
6
x
(18)
We now approximate this sum by the integral of the summand: let b
=~
and
c = 6ln ~. We have seen that the probability of incorrect performance of the
WTA subnet is equal to
L
P(Y :S M) ~
P(Y ~ xo)e-c(xo-x)[(P(M < xo))b(ro-r-l) - (P(M < xo))b(ro-r)] ~
x
Now we transform variables t
= bY In ~ to get the integral in the form
This is the convergent difference between two divergent Gamma function integrals.
~e perform inte~rat~on by parts to obtain a representation as an integr~l wi~h rK2
mstead of t-(1+ 2) m the mtegrand. For 0 ~ K2 < 1, the correspondmg mtegral
converges. The final result is then
(1 - 7])
1 - eC
C
c
In b
1
'Y
c
r(l - -)(1n -)1iib
(21 )
Hence, we have
1
1 - e -61n -1L
l-{J
P(Y ~ M) ~ (1-7])
-L r(l- 6)(ln _)6 ~
6ln l-f3
'Y
1 - e- 6ln 6
r(l - 6)
-L
6In 1 _/3
(1
(f6)6
6)1+6 f
+
(22)
Single-iteration Threshold Hamming Networks
% error -+
threshold , m
predicted
THN
predicted
HN
experimental
THN
experimental"
HN
2.46
0.144
2.552
0.103
!
133 , 145
= 1.0
= 1.552)
(1 - 'Y = 1.03
1 - T/
1.46)
(1 - 'Y
1 - T/
=
134 ? 346
3.4
=
0.272
135 , 825
4.714
0.494
(1 - 'Y = 1.776
1 - 11 = 2.991)
136 , 1970
6.346
=
4.152
0.485
(1 - 'Y = 1.606
1 - 11 = 2.576)
0.857
(1 - 'Y = 2.274
1 - T/
4.167)
0.253
3.468
(1 - 'Y = 1.373
1 - T/ = 2.168)
(1 - 'Y
1.37
1-T/=2.1l)
6.447
0.863
(1 - 'Y = 2.335
1 - T/
4.162)
=
Table 1: The performance of a HN and optimal THN: A comparison between calculated and experimental results (a = 0.7,n = 210).
as claimed. Expression (22) is presented as K(f, 8, (3)f, where K(f, 8, (3) is the factor
(:::; 1) by which the probability of error f of the THN should be multiplied in order
to get the probability of error of the original HN with the WTA subnet. For small
8, K is close to 1, however, as will be seen in the next section, K is typically larger.
4
Numerical results
The experimental results presented in table 1 testify to the accuracy of the HN and
THN calculations. Figure 1 presents the calculated error probabilities for various
values of input similarity a and memory capacity m, as a function of the input size
n. As is evident, the performance of the THN is worse than that of the HN, but due
to the exponential growth of m, it requires only a minor increment in n to obtain
a THN that performs as well as the original HN.
To examine the sensitivity of the THN network to threshold variation, we have fixed
a = 0.7, n = 210, m = 825, and let the threshold vary between 132 and 138. As we
can see in figure 2, the threshold 135 is indeed optimal, but the performance with
threshold values of 134 and 136 is practically identical. The magnitude of the two
error types varies considerably with the threshold value, but this variation has no
effect on the overall performance near the optimum. These two error probabilities
might as well be taken equal to each other.
Conclusion In this paper we analyzed in detail the performance of a Hamming
Network and a Threshold Hamming Network. Given a desired storage capacity and
performance, we described how to compute the corresponding minimal network size
required. The THN drastically reduces the time and connectivity requirements of
Hamming Network classifiers.
569
570
Meilijson, Ruppin, and Sipper
alpha=0.6,m=10 3
0.0001
0.0003
0.000
0.002
epsilon
(error
probability)
THN~
HN
0.007
-+-
0.14
0.37
800
1000
1200
1400
1600
n (network size)
1800
2000
2200
alpha=0 .7,m=10 6
0.0001
0.0003
0.000
epsilon
(error
probability)
THN~
HN
-+-
0.14
0.37
300 320 340 360 380 400 420 440 460 480 500 520 540 560 580 600
n (network size)
a1pha=0.8,m=10 9
0.0001....,.-----------r--------..,...---,
0.0003
0.000
THN~
HN
epsilon
(error
probability)
-+-
0.37
160
180
200
220
240
260
n (network size)
280
300
320
Figure 1: Probability of error as a function of network size: three networks are
depicted , displaying the performance at various values of (}' and m . For graphical
convenience, we have plotted log ~ versus n.
Single-iteration Threshold Hamming Networks
THN performance
10
epsilon ~
1 - gamma +1 - eta -e-
9
8
7
% error
6
5
4
3
2
1
0
132
133
134
135
threshold
Figure 2: Threshold sensitivity of the THN (a
136
137
138
= 0.7, n = 210, m = 825).
References
[1] K. Steinbuch. Dei lernmatrix. /(ybernetic, 1:36-45, 1961.
[2] \iV.K. Taylor. Cortico-thalamic organization and memory. Proc. of the Royal
Society of London B, 159:466-478, 1964.
[3] R.P. Lippmann, B. Gold, and M.L. Malpass. A comparison of Hamming and
Hopfield neural nets for pattern classification. Technical Report TR-769, MIT
Lincoln Laboratory, 1987.
[4] E.E. Baum, J. Moody, and F. Wilczek. Internal representations for associative
memory. Biological Cybernetics, 59:217-228, 1987.
[5] P. Floreen. The convergence of hamming memory networks. IEEE Trans. on
Neural Networks, 2(4):449-457, 1991.
[6] E. Domany and H. Orland. A maximum overlap neural network for pattern
recognition. Physics Letters A, 125:32-34,1987.
[7] M.R. Leadbetter, G. Lindgren, and H. Rootzen. Extremes and related properties of random sequences and processes. Springer-Verlag, Berlin-HeidelbergNew York , 1983.
571
| 668 |@word effect:1 predicted:2 version:2 faculty:1 rk2:1 proportion:2 hence:1 society:1 moshe:1 correct:5 laboratory:1 eng:1 tr:1 bin:3 subnet:9 distance:2 berlin:1 rat:1 initial:1 capacity:4 selecting:1 opt:1 tuned:1 renewed:1 biological:1 evident:1 tn:1 performs:2 fj:2 practically:1 activation:2 normal:4 dx:1 exp:11 ruppin:4 consideration:1 ratio:1 inte:1 numerical:1 vary:1 smallest:1 omitted:1 winner:4 proc:1 perform:2 mln:1 neuron:11 ai:2 v211:3 similarly:1 royal:1 recompute:1 mit:1 e7:1 arbitrary:2 similarity:8 mathematical:1 lindgren:1 closest:2 incorrect:1 recent:1 yo:1 required:1 properly:1 connection:5 discard:2 scenario:1 claimed:1 verlag:1 indeed:1 sense:1 examine:2 trans:1 seen:3 pattern:17 typically:1 xm:3 relation:1 determine:1 selects:2 classifies:2 max:3 multiple:1 maximizes:2 overall:1 classification:1 reduces:2 israel:1 memory:22 technical:1 match:1 overlap:1 calculation:1 special:1 imal:1 orland:2 mn:2 equal:3 f3:7 having:1 ng:2 dei:1 identical:1 y6:1 calculates:2 j3:1 jna:6 xd:1 growth:1 thf:1 ro:2 k2:1 classifier:2 wrong:1 report:1 summand:1 primarily:1 iteration:9 randomly:1 raymond:1 composed:1 gamma:3 before:1 relative:1 accordance:1 fully:1 replaced:1 consisting:2 limn:1 men:1 testify:1 versus:1 organization:1 might:1 xp:1 ztj:2 analyzed:1 extreme:2 integer:3 near:1 displaying:1 activated:1 classifying:1 storing:1 practical:1 integral:3 approaching:2 topology:1 drastically:2 lh:2 domany:2 cortico:1 school:1 iv:1 taylor:3 desired:2 plotted:1 expression:2 calculated:3 minimal:1 xn:8 cumulative:2 eta:1 aviv:2 get:2 convenience:1 close:1 york:1 jj:1 ec:1 storage:2 approximate:2 lippmann:2 alpha:2 baum:2 attention:1 stored:2 a_:1 varies:1 xi:4 vna:1 considerably:1 iterative:1 density:2 sensitivity:2 nf3:2 correctly:2 table:2 standing:1 jz:1 physic:1 variation:2 increment:1 tel:2 vf2:1 limiting:1 moody:1 thn:18 connectivity:1 expansion:1 exact:1 gnedenko:1 threshold:28 us:1 hn:19 choose:2 worse:1 recognition:3 iib:1 rh:1 derivative:1 lenllna:2 steinbuch:2 f6:1 year:1 sum:1 x1:1 letter:1 distorted:2 calculate:1 connected:2 vi:1 sub:1 exceeding:1 exponential:2 analyze:4 meilijson:3 bit:3 complexity:2 start:1 thalamic:1 layer:4 sackler:1 convergent:1 tight:1 accuracy:1 beverly:1 equate:1 x2:1 yield:2 completely:1 identify:1 divergent:1 declared:1 hopfield:1 various:2 magnitude:1 cybernetics:1 london:1 depicted:1 yin:1 simply:1 likely:1 larger:1 failure:2 wi:1 lp:2 wta:8 expressed:1 isaac:1 proof:3 hamming:21 transform:1 xo:17 springer:1 final:1 associative:1 taken:1 differentiate:1 sequence:1 ln:10 satisfies:1 net:2 propose:1 integr:1 maximal:1 letting:1 replace:2 jf:1 originally:1 multiplied:1 lincoln:1 gold:1 lemma:5 eytan:1 experimental:4 ex:1 convergence:1 optimum:1 requirement:1 existence:1 ei:2 wilczek:1 original:7 converges:3 binomial:2 epsilon:4 lernmatrix:1 oo:1 graphical:1 internal:1 subnets:1 minor:1 received:1 quantile:1 |
6,279 | 6,680 | A New Theory for Matrix Completion
Guangcan Liu?
Qingshan Liu?
Xiao-Tong Yuan?
School of Information & Control, Nanjing University of Information Science & Technology
NO 219 Ningliu Road, Nanjing, Jiangsu, China, 210044
{gcliu,qsliu,xtyuan}@nuist.edu.cn
Abstract
Prevalent matrix completion theories reply on an assumption that the locations of
the missing data are distributed uniformly and randomly (i.e., uniform sampling).
Nevertheless, the reason for observations being missing often depends on the unseen
observations themselves, and thus the missing data in practice usually occurs in a
nonuniform and deterministic fashion rather than randomly. To break through the
limits of random sampling, this paper introduces a new hypothesis called isomeric
condition, which is provably weaker than the assumption of uniform sampling and
arguably holds even when the missing data is placed irregularly. Equipped with
this new tool, we prove a series of theorems for missing data recovery and matrix
completion. In particular, we prove that the exact solutions that identify the target
matrix are included as critical points by the commonly used nonconvex programs.
Unlike the existing theories for nonconvex matrix completion, which are built
upon the same condition as convex programs, our theory shows that nonconvex
programs have the potential to work with a much weaker condition. Comparing to
the existing studies on nonuniform sampling, our setup is more general.
1
Introduction
Missing data is a common occurrence in modern applications such as computer vision and image
processing, reducing significantly the representativeness of data samples and therefore distorting
seriously the inferences about data. Given this pressing situation, it is crucial to study the problem
of recovering the unseen data from a sampling of observations. Since the data in reality is often
organized in matrix form, it is of considerable practical significance to study the well-known problem
of matrix completion [1] which is to fill in the missing entries of a partially observed matrix.
Problem 1.1 (Matrix Completion). Denote the (i, j)th entry of a matrix as [?]ij . Let L0 ? Rm?n be
an unknown matrix of interest. In particular, the rank of L0 is unknown either. Given a sampling of
the entries in L0 and a 2D index set ? ? {1, 2, ? ? ? , m} ? {1, 2, ? ? ? , n} consisting of the locations
of the observed entries, i.e., given
{[L0 ]ij |(i, j) ? ?}
and
?,
can we restore the missing entries whose indices are not included in ?, in an exact and scalable
fashion? If so, under which conditions?
Due to its unique role in a broad range of applications, e.g., structure from motion and magnetic
resonance imaging, matrix completion has received extensive attentions in the literatures, e.g., [2?13].
?
The work of Guangcan Liu is supported in part by national Natural Science Foundation of China (NSFC)
under Grant 61622305 and Grant 61502238, in part by Natural Science Foundation of Jiangsu Province of China
(NSFJPC) under Grant BK20160040.
?
The work of Qingshan Liu is supported by NSFC under Grant 61532009.
?
The work of Xiao-Tong Yuan is supported in part by NSFC under Grant 61402232 and Grant 61522308, in
part by NSFJPC under Grant BK20141003.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
Figure 1: Left and Middle: Typical configurations for the locations of the observed entries. Right: A
real example from the Oxford motion database. The black areas correspond to the missing entries.
In general, given no presumption about the nature of matrix entries, it is virtually impossible to
restore L0 as the missing entries can be of arbitrary values. That is, some assumptions are necessary
for solving Problem 1.1. Based on the high-dimensional and massive essence of today?s data-driven
community, it is arguable that the target matrix L0 we wish to recover is often low rank [23]. Hence,
one may perform matrix completion by seeking a matrix with the lowest rank that also satisfies the
constraints given by the observed entries:
min rank (L) ,
L
s.t.
[L]ij = [L0 ]ij , ?(i, j) ? ?.
(1)
Unfortunately, this idea is of little practical because the problem above is NP-hard and cannot be
solved in polynomial time [15]. To achieve practical matrix completion, Cand?s and Recht [4]
suggested to consider an alternative that minimizes instead the nuclear norm which is a convex
envelope of the rank function [12]. Namely,
min kLk? ,
L
s.t.
[L]ij = [L0 ]ij , ?(i, j) ? ?,
(2)
where k ? k? denotes the nuclear norm, i.e., the sum of the singular values of a matrix. Rather
surprisingly, it is proved in [4] that the missing entries, with high probability, can be exactly restored
by the convex program (2), as long as the target matrix L0 is low rank and incoherent and the set ? of
locations corresponding to the observed entries is a set sampled uniformly at random. This pioneering
work provides people several useful tools to investigate matrix completion and many other related
problems. Those assumptions, including low-rankness, incoherence and uniform sampling, are now
standard and widely used in the literatures, e.g., [14, 17, 22, 24, 28, 33, 34, 36]. In particular, the
analyses in [17, 33, 36] show that, in terms of theoretical completeness, many nonconvex optimization
based methods are as powerful as the convex program (2). Unfortunately, these theories still depend
on the assumption of uniform sampling, and thus they cannot explain why there are many nonconvex
methods which often do better than the convex program (2) in practice.
The missing data in practice, however, often occurs in a nonuniform and deterministic fashion instead
of randomly. This is because the reason for an observation being missing usually depends on the
unseen observations themselves. For example, in structure from motion and magnetic resonance
imaging, typically the locations of the observed entries are concentrated around the main diagonal of
a matrix4 , as shown in Figure 1. Moreover, as pointed out by [19, 21, 23], the incoherence condition
is indeed not so consistent with the mixture structure of multiple subspaces, which is also a ubiquitous
phenomenon in practice. There has been sparse research in the direction of nonuniform sampling,
e.g., [18, 25?27, 31]. In particular, Negahban and Wainwright [26] studied the case of weighted
entrywise sampling, which is more general than the setup of uniform sampling but still a special
form of random sampling. Kir?ly et al. [18] considered deterministic sampling and is most related to
this work. However, they had only established conditions to decide whether a particular entry of the
matrix can be restored. In other words, the setup of [18] may not handle well the dependence among
the missing entries. In summary, matrix completion still starves for practical theories and methods,
although has attained considerable improvements in these years.
To break through the limits of the setup of random sampling, in this paper we introduce a new
hypothesis called isomeric condition, which is a mixed concept that combines together the rank and
coherence of L0 with the locations and amount of the observed entries. In general, isomerism (noun
4
This statement means that the observed entries are concentrated around the main diagonal after a permutation
of the sampling pattern ?.
2
of isomeric) is a very mild hypothesis and only a little bit more strict than the well-known oracle
assumption; that is, the number of observed entries in each row and column of L0 is not smaller than
the rank of L0 . It is arguable that the isomeric condition can hold even when the missing entries have
irregular locations. In particular, it is provable that the widely used assumption of uniform sampling
is sufficient to ensure isomerism, not necessary. Equipped with this new tool, isomerism, we prove a
set of theorems pertaining to missing data recovery [35] and matrix completion. For example, we
prove that, under the condition of isomerism, the exact solutions that identify the target matrix are
included as critical points by the commonly used bilinear programs. This result helps to explain the
widely observed phenomenon that there are many nonconvex methods performing better than the
convex program (2) on real-world matrix completion tasks. In summary, the contributions of this
paper mainly include:
We invent a new hypothesis called isomeric condition, which provably holds given the
standard assumptions of uniform sampling, low-rankness and incoherence. In addition,
we also exemplify that the isomeric condition can hold even if the target matrix L0 is not
incoherent and the missing entries are placed irregularly. Comparing to the existing studies
about nonuniform sampling, our setup is more general.
Equipped with the isomeric condition, we prove that the exact solutions that identify L0
are included as critical points by the commonly used bilinear programs. Comparing to the
existing theories for nonconvex matrix completion, our theory is built upon a much weaker
assumption and can therefore partially reveal the superiorities of nonconvex programs over
the convex methods based on (2).
We prove that the isomeric condition is sufficient and necessary for the column and row
projectors of L0 to be invertible given the sampling pattern ?. This result implies that
the isomeric condition is necessary for ensuring that the minimal rank solution to (1) can
identify the target L0 .
The rest of this paper is organized as follows. Section 2 summarizes the mathematical notations used
in the paper. Section 3 introduces the proposed isomeric condition, along with some theorems for
matrix completion. Section 4 shows some empirical results and Section 5 concludes this paper. The
detailed proofs to all the proposed theorems are presented in the Supplementary Materials.
2
Notations
Capital and lowercase letters are used to represent matrices and vectors, respectively, except that the
lowercase letters, i, j, k, m, n, l, p, q, r, s and t, are used to denote some integers, e.g., the location of
an observation, the rank of a matrix, etc. For a matrix M , [M ]ij is its (i, j)th entry, [M ]i,: is its ith row
and [M ]:,j is its jth column. Let ?1 and ?2 be two 1D index sets; namely, ?1 = {i1 , i2 , ? ? ? , ik } and
?2 = {j1 , j2 , ? ? ? , js }. Then [M ]?1 ,: denotes the submatrix of M obtained by selecting the rows with
indices i1 , i2 , ? ? ? , ik , [M ]:,?2 is the submatrix constructed by choosing the columns j1 , j2 , ? ? ? , js ,
and similarly for [M ]?1 ,?2 . For a 2D index set ? ? {1, 2, ? ? ? , m} ? {1, 2, ? ? ? , n}, we imagine it
as a sparse matrix and, accordingly, define its ?rows?, ?columns? and ?transpose? as follows: The
ith row ?i = {j1 |(i1 , j1 ) ? ?, i1 = i}, the jth column ?j = {i1 |(i1 , j1 ) ? ?, j1 = j} and the
transpose ?T = {(j1 , i1 )|(i1 , j1 ) ? ?}.
The special symbol (?)+ is reserved to denote the Moore-Penrose pseudo-inverse of a matrix. More
T
precisely, for a matrix M with Singular Value Decomposition (SVD)5 M = UM ?M VM
, its pseudo?1 T
+
inverse is given by M = VM ?M UM . For convenience, we adopt the conventions of using
span{M } to denote the linear space spanned by the columns of a matrix M , using y ? span{M } to
denote that a vector y belongs to the space span{M }, and using Y ? span{M } to denote that all the
column vectors of a matrix Y belong to span{M }.
Capital letters U , V , ? and their variants (complements, subscripts, etc.) are reserved for left singular
vectors, right singular vectors and index set, respectively. For convenience, we shall abuse the
notation U (resp. V ) to denote the linear space spanned by the columns of U (resp. V ), i.e., the
column space (resp. row space). The orthogonal projection onto the column space U , is denoted by
PU and given by PU (M ) = U U T M , and similarly for the row space PV (M ) = M V V T . The same
In this paper, SVD always refers to skinny SVD. For a rank-r matrix M ? Rm?n , its SVD is of the form
T
UM ?M VM
, where UM ? Rm?r , ?M ? Rr?r and VM ? Rn?r .
5
3
notation is also used to represent a subspace of matrices (i.e., the image of an operator), e.g., we say
that M ? PU for any matrix M which satisfies PU (M ) = M . We shall also abuse the notation ?
to denote the linear space of matrices supported on ?. Then the symbol P? denotes the orthogonal
projection onto ?, namely,
[M ]ij , if (i, j) ? ?,
[P? (M )]ij =
0,
otherwise.
Similarly, the symbol P?? denotes the orthogonal projection onto the complement space of ?. That
is, P? + P?? = I, where I is the identity operator.
Three types of matrix norms are used in this paper, and they are all functions of the singular values:
1) The operator norm or 2-norm (i.e., largest singular value) denoted by kM k, 2) the Frobenius norm
(i.e., square root of the sum of squared singular values) denoted by kM kF and 3) the nuclear norm
or trace norm (i.e., sum of singular values) denoted by kM k? . The only used vector norm is the `2
norm, which is denoted by k ? k2 . The symbol | ? | is reserved for the cardinality of an index set.
3
Isomeric Condition and Matrix Completion
This section introduces the proposed isomeric condition and a set of theorems for matrix completion.
But most of the detailed proofs are deferred until the Supplementary Materials.
3.1 Isomeric Condition
In general cases, as aforementioned, matrix completion is an ill-posed problem. Thus, some assumptions are necessary for studying Problem 1.1. To eliminate the disadvantages of the setup of random
sampling, we define and investigate a so-called isomeric condition.
3.1.1
Definitions
For the ease of understanding, we shall begin with a concept called k-isomerism (or k-isomeric in
adjective form), which could be regarded as an extension of low-rankness.
Definition 3.1 (k-isomeric). A matrix M ? Rm?l is called k-isomeric if and only if any k rows of
M can linearly represent all rows in M . That is,
rank ([M ]?,: ) = rank (M ) , ?? ? {1, 2, ? ? ? , m}, |?| = k,
where | ? | is the cardinality of an index set.
In general, k-isomerism is somewhat similar to Spark [37] which defines the smallest linearly
dependent subset of the rows of a matrix. For a matrix M to be k-isomeric, it is necessary that
rank (M ) ? k, not sufficient. In fact, k-isomerism is also somehow related to the concept of
coherence [4, 21]. When the coherence of a matrix M ? Rm?l is not too high, the rows of M will
sufficiently spread, and thus M could be k-isomeric with a small k, e.g., k = rank (M ). Whenever
the coherence of M is very high, one may need a large k to satisfy the k-isomeric property. For
example, consider an extreme case where M is a rank-1 matrix with one row being 1 and everywhere
else being 0. In this case, we need k = m to ensure that M is k-isomeric.
While Definition 3.1 involves all 1D index sets of cardinality k, we often need the isomeric property
to be associated with a certain 2D index set ?. To this end, we define below a concept called
?-isomerism (or ?-isomeric in adjective form).
Definition 3.2 (?-isomeric). Let M ? Rm?l and ? ? {1, 2, ? ? ? , m} ? {1, 2, ? ? ? , n}. Suppose
that ?j 6= ? (empty set), ?1 ? j ? n. Then the matrix M is called ?-isomeric if and only if
rank [M ]?j ,: = rank (M ) , ?j = 1, 2, ? ? ? , n.
Note here that only the number of rows in M is required to coincide with the row indices included in
?, and thereby l 6= n is allowable.
Generally, ?-isomerism is less strict than k-isomerism. Provided that |?j | ? k, ?1 ? j ? n, a matrix
M is k-isomeric ensures that M is ?-isomeric as well, but not vice versa. For the extreme example
where M is nonzero at only one row, interestingly, M can be ?-isomeric as long as the locations of
the nonzero elements are included in ?.
With the notation of ?T = {(j1 , i1 )|(i1 , j1 ) ? ?}, the isomeric property could be also defined on
the column vectors of a matrix, as shown in the following definition.
4
Definition 3.3 (?/?T -isomeric). Let M ? Rm?n and ? ? {1, 2, ? ? ? , m}?{1, 2, ? ? ? , n}. Suppose
?i 6= ? and ?j 6= ?, ?i = 1, ? ? ? , m, j = 1, ? ? ? , n. Then the matrix M is called ?/?T -isomeric if
and only if M is ?-isomeric and M T is ?T -isomeric as well.
To solve Problem 1.1 without the imperfect assumption of missing at random, as will be shown later,
we need to assume that L0 is ?/?T -isomeric. This condition has excluded the unidentifiable cases
where any rows or columns of L0 are wholly missing. In fact, whenever L0 is ?/?T -isomeric, the
number of observed entries in each row and column of L0 has to be greater than or equal to the rank
of L0 ; this is consistent with the results in [20]. Moreover, ?/?T -isomerism has actually well treated
the cases where L0 is of high coherence. For example, consider an extreme case where L0 is 1 at only
one element and 0 everywhere else. In this case, L0 cannot be ?/?T -isomeric unless the nonzero
element is observed. So, generally, it is possible to restore the missing entries of a highly coherent
matrix, as long as the ?/?T -isomeric condition is obeyed.
3.1.2
Basic Properties
While its definitions are associated with a certain matrix, the isomeric condition is actually characterizing some properties of a space, as shown in the lemma below.
Lemma 3.1. Let L0 ? Rm?n and ? ? {1, 2, ? ? ? , m} ? {1, 2, ? ? ? , n}. Denote the SVD of L0 as
U0 ?0 V0T . Then we have:
1. L0 is ?-isomeric if and only if U0 is ?-isomeric.
2. LT0 is ?T -isomeric if and only if V0 is ?T -isomeric.
Proof. It could be manipulated that
[L0 ]?j ,: = ([U0 ]?j ,: )?0 V0T , ?j = 1, ? ? ? , n.
Since ?0 V0T is row-wisely full rank, we have
rank [L0 ]?j ,: = rank [U0 ]?j ,: , ?j = 1, ? ? ? , n.
As a result, L0 is ?-isomeric is equivalent to that U0 is ?-isomeric. In a similar way, the second
claim is proved.
It is easy to see that the above lemma is still valid even when the condition of ?-isomerism is replaced
by k-isomerism. Thus, hereafter, we may say that a space is isomeric (k-isomeric, ?-isomeric or
?T -isomeric) as long as its basis matrix is isomeric. In addition, the isomeric property is subspace
successive, as shown in the next lemma.
Lemma 3.2. Let ? ? {1, 2, ? ? ? , m} ? {1, 2, ? ? ? , n} and U0 ? Rm?r be the basis matrix of a
Euclidean subspace embedded in Rm . Suppose that U is a subspace of U0 , i.e., U = U0 U0T U . If U0
is ?-isomeric then U is ?-isomeric as well.
Proof. By U = U0 U0T U and U0 is ?-isomeric,
rank [U ]?j ,: = rank ([U0 ]?j ,: )U0T U = rank U0T U
= rank U0 U0T U = rank (U ) , ?1 ? j ? n.
The above lemma states that, in one word, the subspace of an isomeric space is isomeric as well.
3.1.3
Important Properties
As aforementioned, the isometric condition is actually necessary for ensuring that the minimal rank
solution to (1) can identify L0 . To see why, let?s assume that U0 ? ?? 6= {0}, where we denote by
U0 ?0 V0T the SVD of L0 . Then one could construct a nonzero perturbation, denoted as ? ? U0 ? ?? ,
? 0 = L0 + ? to the problem in (1). Since ? ? U0 , we
and accordingly, obtain a feasible solution L
?
? 0 ) < rank (L0 ). Such
have rank(L0 ) ? rank (L0 ). Even more, it is entirely possible that rank(L
a case is unidentifiable in nature, as the global optimum to problem (1) cannot identify L0 . Thus,
5
to ensure that the global minimum to (1) can identify L0 , it is essentially necessary to show that
U0 ? ?? = {0} (resp. V0 ? ?? = {0}), which is equivalent to that the operator PU0 P? PU0 (resp.
PV0 P? PV0 ) is invertible (see Lemma 6.8 of the Supplementary Materials). Interestingly, the isomeric
condition is indeed a sufficient and necessary condition for the operators PU0 P? PU0 and PV0 P? PV0
to be invertible, as shown in the following theorem.
Theorem 3.1. Let L0 ? Rm?n and ? ? {1, 2, ? ? ? , m} ? {1, 2, ? ? ? , n}. Let the SVD of L0 be
U0 ?0 V0T . Denote PU0 (?) = U0 U0T (?) and PV0 (?) = (?)V0 V0T . Then we have the following:
1. The linear operator PU0 P? PU0 is invertible if and only if U0 is ?-isomeric.
2. The linear operator PV0 P? PV0 is invertible if and only if V0 is ?T -isomeric.
The necessity stated above implies that the isomeric condition is actually a very mild hypothesis. In
general, there are numerous reasons for the target matrix L0 to be isomeric. Particularly, the widely
used assumptions of low-rankness, incoherence and uniform sampling are indeed sufficient (but not
necessary) to ensure isomerism, as shown in the following theorem.
Theorem 3.2. Let L0 ? Rm?n and ? ? {1, 2, ? ? ? , m} ? {1, 2, ? ? ? , n}. Denote n1 = max(m, n)
and n2 = min(m, n). Suppose that L0 is incoherent and ? is a 2D index set sampled uniformly
at random, namely Pr((i, j) ? ?) = ?0 and Pr((i, j) ?
/ ?) = 1 ? ?0 . For any ? > 0, if ?0 > ?
is obeyed and rank (L0 ) < ?n2 /(c log n1 ) holds for some numerical constant c then, with high
probability at least 1 ? n?10
, L0 is ?/?T -isomeric.
1
It is worth noting that the isomeric condition can be obeyed in numerous circumstances other than
the case of uniform sampling plus incoherence. For example,
"
#
1 0 0
? = {(1, 1), (1, 2), (1, 3), (2, 1), (3, 1)} and L0 = 0 0 0 ,
0 0 0
where L0 is a 3?3 matrix with 1 at (1, 1) and 0 everywhere else. In this example, L0 is not incoherent
and the sampling is not uniform either, but it could be verified that L0 is ?/?T -isomeric.
3.2 Results
In this subsection, we shall show how the isomeric condition can take effect in the context of
nonuniform sampling, establishing some theorems pertaining to missing data recovery [35] as well
as matrix completion.
3.2.1
Missing Data Recovery
Before exploring the matrix completion problem, for the ease of understanding, we would like
to consider a missing data recovery problem studied by Zhang [35], which could be described as
follows: Let y0 ? Rm be a data vector drawn form some low-dimensional subspace, denoted as
y0 ? S0 ? Rm . Suppose that y0 contains some available observations in yb ? Rk and some missing
entries in yu ? Rm?k . Namely, after a permutation,
yb
y0 =
, yb ? Rk , yu ? Rm?k .
(3)
yu
Given the observations in yb , we seek to restore the unseen entries in yu . To do this, we consider the
prevalent idea that represents a data vector as a linear combination of the bases in a given dictionary:
y0 = Ax0 ,
(4)
where A ? Rm?p is a dictionary constructed in advance and x0 ? Rp is the representation of y0 .
Utilizing the same permutation used in (3), we can partition the rows of A into two parts according to
the indices of the observed and missing entries, respectively:
Ab
A=
, Ab ? Rk?p , Au ? R(m?k)?p .
(5)
Au
In this way, the equation in (4) gives that
yb = Ab x0
and
6
y u = Au x 0 .
As we now can see, the unseen data yu could be restored, as long as the representation x0 is retrieved
by only accessing the available observations in yb . In general cases, there are infinitely many
representations that satisfy y0 = Ax0 , e.g., x0 = A+ y0 , where (?)+ is the pseudo-inverse of a matrix.
Since A+ y0 is the representation of minimal `2 norm, we revisit the traditional `2 program:
min
x
1
2
kxk2 ,
2
s.t.
yb = Ab x,
(6)
where k ? k2 is the `2 norm of a vector. Under some verifiable conditions, the above `2 program
is indeed consistently successful in a sense as in the following: For any y0 ? S0 with an arbitrary
partition y0 = [yb ; yu ] (i.e., arbitrarily missing), the desired representation x0 = A+ y0 is the unique
minimizer to the problem in (6). That is, the unseen data yu is exactly recovered by firstly computing
the minimizer x? to problem (6) and then calculating yu = Au x? .
Theorem 3.3. Let y0 = [yb ; yu ] ? Rm be an authentic sample drawn from some low-dimensional
subspace S0 embedded in Rm , A ? Rm?p be a given dictionary and k be the number of available
observations in yb . Then the convex program (6) is consistently successful, provided that S0 ?
span{A} and the dictionary A is k-isomeric.
Unlike the theory in [35], the condition of which is unverifiable, our k-isomeric condition could be
verified in finite time. Notice, that the problem of missing data recovery is closely related to matrix
completion, which is actually to restore the missing entries in multiple data vectors simultaneously.
Hence, Theorem 3.3 can be naturally generalized to the case of matrix completion, as will be shown
in the next subsection.
3.2.2
Matrix Completion
The spirits of the `2 program (6) can be easily transferred to the case of matrix completion. Following (6), one may consider Frobenius norm minimization for matrix completion:
min
X
1
2
kXkF , s.t. P? (AX ? L0 ) = 0,
2
(7)
where A ? Rm?p is a dictionary assumed to be given. As one can see, the problem in (7) is equivalent
to (6) if L0 is consisting of only one column vector. The same as (6), the convex program (7) can
also exactly recover the desired representation matrix A+ L0 , as shown in the theorem below. The
difference is that we here require ?-isomerism instead of k-isomerism.
Theorem 3.4. Let L0 ? Rm?n and ? ? {1, 2, ? ? ? , m} ? {1, 2, ? ? ? , n}. Suppose that A ? Rm?p
is a given dictionary. Provided that L0 ? span{A} and A is ?-isomeric, the desired representation
X0 = A+ L0 is the unique minimizer to the problem in (7).
Theorem 3.4 tells us that, in general, even when the locations of the missing entries are interrelated
and nonuniformly distributed, the target matrix L0 can be restored as long as we have found a proper
dictionary A. This motivates us to consider the commonly used bilinear program that seeks both A
and X simultaneously:
min
A,X
1
1
2
2
kAkF + kXkF , s.t. P? (AX ? L0 ) = 0,
2
2
(8)
where A ? Rm?p and X ? Rp?n . The problem above is bilinear and therefore nonconvex. So, it
would be hard to obtain a strong performance guarantee as done in the convex programs, e.g., [4, 21].
Interestingly, under a very mild condition, the problem in (8) is proved to include the exact solutions
that identify the target matrix L0 as the critical points.
Theorem 3.5. Let L0 ? Rm?n and ? ? {1, 2, ? ? ? , m} ? {1, 2, ? ? ? , n}. Denote the rank and SVD
of L0 as r0 and U0 ?0 V0T , respectively. If L0 is ?/?T -isomeric then the exact solution, denoted by
(A0 , X0 ) and given by
1
1
A0 = U0 ?02 QT , X0 = Q?02 V0T , ?Q ? Rp?r0 , QT Q = I,
is a critical point to the problem in (8).
To exhibit the power of program (8), however, the parameter p, which indicates the number of
columns in the dictionary matrix A, must be close to the true rank of the target matrix L0 . This is
7
observed entries (%)
95
convex (nonuniform) nonconvex (nonuniform)
95
95
convex (uniform)
nonconvex (uniform)
95
75
75
75
75
55
55
55
55
35
35
35
35
15
15
15
15
1
1
1 15 35 55 75 95
rank(L0)
1
1 15 35 55 75 95
rank(L0)
1
1 15 35 55 75 95
rank(L0)
1 15 35 55 75 95
rank(L0)
Figure 2: Comparing the bilinear program (9) (p = m) with the convex method (2). The numbers
plotted on the above figures are the success rates within 20 random trials. The white and black points
mean ?succeed? and ?fail?, respectively. Here the success is in a sense that PSNR ? 40dB, where
PSNR standing for peak signal-to-noise ratio.
impractical in the cases where the rank of L0 is unknown. Notice, that the ?-isomeric condition
imposed on A requires
rank (A) ? |?j |, ?j = 1, 2, ? ? ? , n.
This, together with the condition of L0 ? span{A}, essentially need us to solve a low rank matrix
recovery problem [14]. Hence, we suggest to combine the formulation (7) with the popular idea of
nuclear norm minimization, resulting in a bilinear program that jointly estimates both the dictionary
matrix A and the representation matrix X by
min kAk? +
A,X
1
2
kXkF , s.t. P? (AX ? L0 ) = 0,
2
(9)
which, by coincidence, has been mentioned in a paper about optimization [32]. Similar to (8), the
program in (9) has the following theorem to guarantee its performance.
Theorem 3.6. Let L0 ? Rm?n and ? ? {1, 2, ? ? ? , m} ? {1, 2, ? ? ? , n}. Denote the rank and SVD
of L0 as r0 and U0 ?0 V0T , respectively. If L0 is ?/?T -isomeric then the exact solution, denoted by
(A0 , X0 ) and given by
2
1
A0 = U0 ?03 QT , X0 = Q?03 V0T , ?Q ? Rp?r0 , QT Q = I,
is a critical point to the problem in (9).
Unlike (8), which possesses superior performance only if p is close to rank (L0 ) and the initial
solution is chosen carefully, the bilinear program in (9) can work well by simply choosing p = m
and using A = I as the initial solution. To see why, one essentially needs to figure out the conditions
under which a specific optimization procedure can produce an optimal solution that meets an exact
solution. This requires extensive justifications and we leave it as future work.
4
Simulations
To verify the superiorities of the nonconvex matrix completion methods over the convex program (2),
we would like to experiment with randomly generated matrices. We generate a collection of m ? n
(m = n = 100) target matrices according to the model of L0 = BC, where B ? Rm?r0 and
C ? Rr0 ?n are N (0, 1) matrices. The rank of L0 , i.e., r0 , is configured as r0 = 1, 5, 10, ? ? ? , 90, 95.
Regarding the index set ? consisting of the locations of the observed entries, we consider two settings: One is to create ? by using a Bernoulli model to randomly sample a subset from
{1, ? ? ? , m} ? {1, ? ? ? , n} (referred to as ?uniform?), the other is as in Figure 1 that makes the
locations of the observed entries be concentrated around the main diagonal of a matrix (referred to as
?nonuniform?). The observation fraction is set to be |?|/(mn) = 0.01, 0.05, ? ? ? , 0.9, 0.95. For each
pair of (r0 , |?|/(mn)), we run 20 trials, resulting in 8000 simulations in total.
When p = m and the identity matrix is used to initialize the dictionary A, we have empirically found
that program (8) has the same performance as (2). This is not strange, because it has been proven
in [16] that kLk? = minA,X 21 (kAk2F + kXk2F ), s.t. L = AX. Figure 2 compares the bilinear
8
program (9) to the convex method (2). It can be seen that (9) works distinctly better than (2). Namely,
while handling the nonuniformly missing data, the number of matrices successfully restored by the
bilinear program (9) is 102% more than the convex program (2). Even for dealing with the missing
entries chosen uniformly at random, in terms of the number of successfully restored matrices, the
bilinear program (9) can still outperform the convex method (2) by 44%. These results illustrate that,
even in the cases where the rank of L0 is unknown, the bilinear program (9) can do much better than
the convex optimization based method (2).
5
Conclusion and Future Work
This work studied the problem of matrix completion with nonuniform sampling, a significant setting
not extensively studied before. To figure out the conditions under which exact recovery is possible,
we proposed a so-called isomeric condition, which provably holds when the standard assumptions
of low-rankness, incoherence and uniform sampling arise. In addition, we also exemplified that
the isomeric condition can be obeyed in the other cases beyond the setting of uniform sampling.
Even more, our theory implies that the isomeric condition is indeed necessary for making sure
that the minimal rank completion can identify the target matrix L0 . Equipped with the isomeric
condition, finally, we mathematically proved that the widely used bilinear programs can include the
exact solutions that recover the target matrix L0 as the critical points; this guarantees the recovery
performance of bilinear programs to some extend.
However, there still remain several problems for future work. In particular, it is unknown under which
conditions a specific optimization procedure for (9) can produce an optimal solution that exactly
restores the target matrix L0 . To do this, one needs to analyze the convergence property as well as
the recovery performance. Moreover, it is unknown either whether the isomeric condition suffices
for ensuring that the minimal rank completion can identify the target L0 . These require extensive
justifications and we leave them as future work.
Acknowledgment
We would like to thanks the anonymous reviewers and meta-reviewers for providing us many valuable
comments to refine this paper.
References
[1] Emmanuel Cand?s and Terence Tao. The power of convex relaxation: Near-optimal matrix completion.
IEEE Transactions on Information Theory, 56(5):2053?2080, 2009.
[2] Emmanuel Cand?s and Yaniv Plan. Matrix completion with noise. In IEEE Proceeding, volume 98, pages
925?936, 2010.
[3] William E. Bishop and Byron M. Yu. Deterministic symmetric positive semidefinite matrix completion.
In Neural Information Processing Systems, pages 2762?2770, 2014.
[4] Emmanuel Cand?s and Benjamin Recht. Exact matrix completion via convex optimization. Foundations
of Computational Mathematics, 9(6):717?772, 2009.
[5] Eyal Heiman, Gideon Schechtman, and Adi Shraibman. Deterministic algorithms for matrix completion.
Random Structures and Algorithms, 45(2):306?317, 2014.
[6] Raghunandan H. Keshavan, Andrea Montanari, and Sewoong Oh. Matrix completion from a few entries.
IEEE Transactions on Information Theory, 56(6):2980?2998, 2010.
[7] Raghunandan H. Keshavan, Andrea Montanari, and Sewoong Oh. Matrix completion from noisy entries.
Journal of Machine Learning Research, 11:2057?2078, 2010.
[8] Akshay Krishnamurthy and Aarti Singh. Low-rank matrix and tensor completion via adaptive sampling.
In Neural Information Processing Systems, pages 836?844, 2013.
[9] Troy Lee and Adi Shraibman. Matrix completion from any given set of observations. In Neural Information
Processing Systems, pages 1781?1787, 2013.
[10] Rahul Mazumder, Trevor Hastie, and Robert Tibshirani. Spectral regularization algorithms for learning
large incomplete matrices. Journal of Machine Learning Research, 11:2287?2322, 2010.
[11] Karthik Mohan and Maryam Fazel. New restricted isometry results for noisy low-rank recovery. In IEEE
International Symposium on Information Theory, pages 1573?1577, 2010.
9
[12] B. Recht, W. Xu, and B. Hassibi. Necessary and sufficient conditions for success of the nuclear norm
heuristic for rank minimization. Technical report, CalTech, 2008.
[13] Markus Weimer, Alexandros Karatzoglou, Quoc V. Le, and Alex J. Smola. Cofi rank - maximum margin
matrix factorization for collaborative ranking. In Neural Information Processing Systems, 2007.
[14] Emmanuel J. Cand?s, Xiaodong Li, Yi Ma, and John Wright. Robust principal component analysis?
Journal of the ACM, 58(3):1?37, 2011.
[15] Alexander L. Chistov and Dima Grigoriev. Complexity of quantifier elimination in the theory of algebraically closed fields. In Proceedings of the Mathematical Foundations of Computer Science, pages
17?31, 1984.
[16] Maryam Fazel, Haitham Hindi, and Stephen P. Boyd. A rank minimization heuristic with application to
minimum order system approximation. In American Control Conference, pages 4734?4739, 2001.
[17] Rong Ge, Jason D. Lee, and Tengyu Ma. Matrix completion has no spurious local minimum. In Neural
Information Processing Systems, pages 2973?2981, 2016.
[18] Franz J. Kir?ly, Louis Theran, and Ryota Tomioka. The algebraic combinatorial approach for low-rank
matrix completion. J. Mach. Learn. Res., 16(1):1391?1436, January 2015.
[19] Guangcan Liu and Ping Li. Recovery of coherent data via low-rank dictionary pursuit. In Neural
Information Processing Systems, pages 1206?1214, 2014.
[20] Daniel L. Pimentel-Alarc?n and Robert D. Nowak. The Information-theoretic requirements of subspace
clustering with missing data. In International Conference on Machine Learning, 48:802?810, 2016.
[21] Guangcan Liu and Ping Li. Low-rank matrix completion in the presence of high coherence. IEEE
Transactions on Signal Processing, 64(21):5623?5633, 2016.
[22] Guangcan Liu, Zhouchen Lin, Shuicheng Yan, Ju Sun, Yong Yu, and Yi Ma. Robust recovery of subspace
structures by low-rank representation. IEEE Transactions on Pattern Recognition and Machine Intelligence,
35(1):171?184, 2013.
[23] Guangcan Liu, Qingshan Liu, and Ping Li. Blessing of dimensionality: Recovering mixture data via
dictionary pursuit. IEEE Transactions on Pattern Recognition and Machine Intelligence, 39(1):47?60,
2017.
[24] Guangcan Liu, Huan Xu, Jinhui Tang, Qingshan Liu, and Shuicheng Yan. A deterministic analysis for
LRR. IEEE Transactions on Pattern Recognition and Machine Intelligence, 38(3):417?430, 2016.
[25] Raghu Meka, Prateek Jain, and Inderjit S. Dhillon. Matrix completion from power-law distributed samples.
In Neural Information Processing Systems, pages 1258?1266, 2009.
[26] Sahand Negahban and Martin J. Wainwright. Restricted strong convexity and weighted matrix completion:
Optimal bounds with noise. Journal of Machine Learning Research, 13:1665?1697, 2012.
[27] Yudong Chen, Srinadh Bhojanapalli, Sujay Sanghavi, and Rachel Ward. Completing any low-rank matrix,
provably. Journal of Machine Learning Research, 16: 2999-3034, 2015.
[28] Praneeth Netrapalli, U. N. Niranjan, Sujay Sanghavi, Animashree Anandkumar, and Prateek Jain. Nonconvex robust PCA. In Neural Information Processing Systems, pages 1107?1115, 2014.
[29] Yuzhao Ni, Ju Sun, Xiaotong Yuan, Shuicheng Yan, and Loong-Fah Cheong. Robust low-rank subspace
segmentation with semidefinite guarantees. In International Conference on Data Mining Workshops, pages
1179?1188, 2013.
[30] R. Rockafellar. Convex Analysis. Princeton University Press, Princeton, NJ, USA, 1970.
[31] Ruslan Salakhutdinov and Nathan Srebro. Collaborative filtering in a non-uniform world: Learning with
the weighted trace norm. In Neural Information Processing Systems, pages 2056?2064, 2010.
[32] Fanhua Shang, Yuanyuan Liu, and James Cheng. Scalable algorithms for tractable schatten quasi-norm
minimization. In AAAI Conference on Artificial Intelligence, pages 2016?2022, 2016.
[33] Ruoyu Sun and Zhi-Quan Luo. Guaranteed matrix completion via non-convex factorization. IEEE
Transactions on Information Theory, 62(11):6535?6579, 2016.
[34] Huan Xu, Constantine Caramanis, and Sujay Sanghavi. Robust PCA via outlier pursuit. IEEE Transactions
on Information Theory, 58(5):3047?3064, 2012.
[35] Yin Zhang. When is missing data recoverable? CAAM Technical Report TR06-15, 2006.
[36] Tuo Zhao, Zhaoran Wang, and Han Liu. A nonconvex optimization framework for low rank matrix
estimation. In Neural Information Processing Systems, pages 559?567, 2015.
[37] David L. Donoho and Michael Elad. Optimally sparse representation in general (nonorthogonal) dictionaries via `1 minimization. Proceedings of the National Academy of Sciences, 100(5): 2197-2202,
2003.
10
| 6680 |@word mild:3 trial:2 middle:1 polynomial:1 norm:17 km:3 shuicheng:3 seek:2 simulation:2 theran:1 decomposition:1 thereby:1 klk:2 necessity:1 liu:13 series:1 configuration:1 selecting:1 hereafter:1 contains:1 seriously:1 initial:2 interestingly:3 bc:1 daniel:1 existing:4 recovered:1 comparing:4 luo:1 must:1 john:1 numerical:1 partition:2 j1:10 intelligence:4 accordingly:2 ith:2 alexandros:1 provides:1 completeness:1 location:12 successive:1 firstly:1 zhang:2 mathematical:2 along:1 constructed:2 symposium:1 ik:2 yuan:3 prove:6 combine:2 introduce:1 x0:10 indeed:5 qingshan:4 andrea:2 themselves:2 cand:5 salakhutdinov:1 zhi:1 little:2 equipped:4 cardinality:3 xtyuan:1 begin:1 moreover:3 notation:6 provided:3 lowest:1 prateek:2 bhojanapalli:1 minimizes:1 shraibman:2 impractical:1 nj:1 guarantee:4 pseudo:3 exactly:4 um:4 rm:27 k2:2 dima:1 control:2 grant:7 ly:2 superiority:2 louis:1 arguably:1 before:2 positive:1 local:1 limit:2 bilinear:13 mach:1 nsfc:3 oxford:1 establishing:1 subscript:1 incoherence:6 meet:1 abuse:2 black:2 plus:1 au:4 china:3 studied:4 kxk2f:1 ease:2 factorization:2 range:1 lrr:1 fazel:2 practical:4 unique:3 acknowledgment:1 fah:1 practice:4 procedure:2 wholly:1 area:1 empirical:1 yan:3 significantly:1 projection:3 boyd:1 word:2 road:1 refers:1 suggest:1 nanjing:2 cannot:4 convenience:2 onto:3 operator:7 close:2 context:1 impossible:1 equivalent:3 deterministic:6 projector:1 missing:33 imposed:1 reviewer:2 attention:1 fanhua:1 convex:22 spark:1 recovery:13 utilizing:1 regarded:1 fill:1 nuclear:5 spanned:2 oh:2 handle:1 justification:2 krishnamurthy:1 resp:5 target:15 today:1 imagine:1 massive:1 exact:11 suppose:6 hypothesis:5 element:3 recognition:3 particularly:1 database:1 observed:16 role:1 coincidence:1 solved:1 wang:1 ensures:1 sun:3 valuable:1 mentioned:1 accessing:1 benjamin:1 convexity:1 complexity:1 depend:1 solving:1 singh:1 upon:2 basis:2 easily:1 caramanis:1 jain:2 pertaining:2 artificial:1 tell:1 choosing:2 whose:1 heuristic:2 widely:5 supplementary:3 posed:1 say:2 solve:2 otherwise:1 elad:1 ward:1 unseen:6 jointly:1 noisy:2 jinhui:1 pressing:1 rr:1 maryam:2 j2:2 achieve:1 academy:1 yuanyuan:1 frobenius:2 convergence:1 nuist:1 empty:1 optimum:1 yaniv:1 produce:2 requirement:1 leave:2 help:1 illustrate:1 completion:44 ij:9 qt:4 school:1 received:1 strong:2 netrapalli:1 recovering:2 involves:1 implies:3 convention:1 direction:1 closely:1 karatzoglou:1 material:3 elimination:1 require:2 suffices:1 anonymous:1 mathematically:1 extension:1 exploring:1 rong:1 hold:6 around:3 considered:1 wright:1 sufficiently:1 nonorthogonal:1 claim:1 dictionary:13 adopt:1 smallest:1 aarti:1 ruslan:1 estimation:1 combinatorial:1 cheong:1 largest:1 vice:1 create:1 successfully:2 tool:3 weighted:3 minimization:6 always:1 rather:2 l0:80 ax:4 improvement:1 consistently:2 prevalent:2 rank:60 mainly:1 indicates:1 bernoulli:1 sense:2 inference:1 dependent:1 lowercase:2 typically:1 eliminate:1 a0:4 spurious:1 quasi:1 i1:10 tao:1 provably:4 among:1 aforementioned:2 ill:1 denoted:9 resonance:2 noun:1 special:2 initialize:1 restores:1 plan:1 equal:1 construct:1 field:1 beach:1 sampling:28 represents:1 broad:1 yu:11 future:4 sanghavi:3 np:1 report:2 few:1 modern:1 randomly:5 manipulated:1 simultaneously:2 national:2 replaced:1 consisting:3 skinny:1 raghunandan:2 n1:2 karthik:1 ab:4 william:1 interest:1 mining:1 investigate:2 highly:1 deferred:1 introduces:3 mixture:2 extreme:3 semidefinite:2 nowak:1 necessary:12 huan:2 orthogonal:3 unless:1 incomplete:1 euclidean:1 desired:3 plotted:1 re:1 theoretical:1 minimal:5 column:16 disadvantage:1 bk20141003:1 ax0:2 kxkf:3 entry:34 subset:2 jiangsu:2 uniform:16 successful:2 too:1 optimally:1 obeyed:4 st:1 recht:3 peak:1 negahban:2 thanks:1 international:3 ju:2 standing:1 lee:2 vm:4 invertible:5 terence:1 together:2 michael:1 squared:1 aaai:1 american:1 zhao:1 li:4 potential:1 zhaoran:1 representativeness:1 rockafellar:1 configured:1 satisfy:2 ranking:1 depends:2 later:1 break:2 root:1 closed:1 eyal:1 analyze:1 jason:1 recover:3 guangcan:7 contribution:1 collaborative:2 square:1 ni:1 reserved:3 correspond:1 identify:10 worth:1 explain:2 ping:3 whenever:2 trevor:1 definition:7 james:1 naturally:1 proof:4 associated:2 sampled:2 proved:4 popular:1 animashree:1 exemplify:1 subsection:2 dimensionality:1 ubiquitous:1 organized:2 psnr:2 segmentation:1 carefully:1 actually:5 attained:1 isometric:1 rahul:1 entrywise:1 yb:10 unidentifiable:2 done:1 formulation:1 smola:1 reply:1 until:1 keshavan:2 somehow:1 defines:1 reveal:1 usa:2 effect:1 xiaodong:1 concept:4 true:1 verify:1 hence:3 regularization:1 excluded:1 symmetric:1 moore:1 nonzero:4 dhillon:1 i2:2 white:1 essence:1 kak:1 generalized:1 allowable:1 mina:1 theoretic:1 motion:3 image:2 common:1 superior:1 empirically:1 volume:1 belong:1 extend:1 significant:1 versa:1 rr0:1 meka:1 sujay:3 mathematics:1 similarly:3 pointed:1 zhouchen:1 had:1 han:1 v0:4 etc:2 pu:4 base:1 j:2 isometry:1 retrieved:1 constantine:1 belongs:1 driven:1 certain:2 nonconvex:14 meta:1 arbitrarily:1 success:3 yi:2 caltech:1 seen:1 minimum:3 greater:1 somewhat:1 ruoyu:1 r0:8 algebraically:1 signal:2 u0:25 stephen:1 multiple:2 full:1 recoverable:1 technical:2 long:7 lin:1 grigoriev:1 niranjan:1 ensuring:3 scalable:2 variant:1 basic:1 vision:1 invent:1 essentially:3 circumstance:1 represent:3 irregular:1 addition:3 else:3 singular:8 crucial:1 envelope:1 rest:1 unlike:3 posse:1 strict:2 sure:1 comment:1 virtually:1 pv0:7 db:1 byron:1 quan:1 spirit:1 integer:1 anandkumar:1 near:1 noting:1 presence:1 easy:1 hastie:1 imperfect:1 idea:3 cn:1 regarding:1 praneeth:1 whether:2 distorting:1 pca:2 sahand:1 algebraic:1 useful:1 generally:2 detailed:2 amount:1 verifiable:1 extensively:1 concentrated:3 generate:1 outperform:1 arguable:2 wisely:1 revisit:1 notice:2 tibshirani:1 shall:4 nevertheless:1 authentic:1 drawn:2 capital:2 verified:2 imaging:2 relaxation:1 fraction:1 sum:3 year:1 run:1 inverse:3 letter:3 powerful:1 everywhere:3 rachel:1 decide:1 strange:1 coherence:6 summarizes:1 bit:1 submatrix:2 entirely:1 bound:1 completing:1 guaranteed:1 cheng:1 refine:1 oracle:1 constraint:1 precisely:1 alex:1 unverifiable:1 yong:1 markus:1 nathan:1 min:7 span:8 xiaotong:1 performing:1 cofi:1 tengyu:1 martin:1 transferred:1 according:2 combination:1 smaller:1 remain:1 gcliu:1 y0:13 making:1 quoc:1 caam:1 outlier:1 restricted:2 pr:2 quantifier:1 equation:1 fail:1 irregularly:2 ge:1 tractable:1 end:1 raghu:1 studying:1 available:3 pursuit:3 spectral:1 magnetic:2 occurrence:1 alternative:1 rp:4 denotes:4 clustering:1 ensure:4 include:3 calculating:1 emmanuel:4 seeking:1 tensor:1 occurs:2 restored:6 dependence:1 nonuniformly:2 diagonal:3 traditional:1 exhibit:1 subspace:11 schatten:1 kak2f:1 reason:3 provable:1 index:14 ratio:1 providing:1 setup:6 unfortunately:2 matrix4:1 robert:2 statement:1 ryota:1 troy:1 trace:2 stated:1 kir:2 proper:1 motivates:1 unknown:6 perform:1 observation:12 finite:1 january:1 situation:1 rn:1 nonuniform:10 perturbation:1 arbitrary:2 community:1 tuo:1 david:1 complement:2 namely:6 required:1 pair:1 extensive:3 coherent:2 established:1 nip:1 beyond:1 suggested:1 usually:2 pattern:5 below:3 exemplified:1 gideon:1 program:31 pioneering:1 built:2 including:1 adjective:2 max:1 wainwright:2 power:3 critical:7 natural:2 treated:1 restore:5 hindi:1 mn:2 technology:1 numerous:2 concludes:1 incoherent:4 literature:2 understanding:2 kf:1 law:1 embedded:2 permutation:3 presumption:1 mixed:1 kakf:1 filtering:1 proven:1 srebro:1 foundation:4 sufficient:6 consistent:2 s0:4 xiao:2 sewoong:2 haitham:1 row:20 summary:2 placed:2 supported:4 surprisingly:1 transpose:2 jth:2 weaker:3 u0t:6 characterizing:1 akshay:1 v0t:10 sparse:3 distinctly:1 distributed:3 yudong:1 world:2 valid:1 commonly:4 collection:1 coincide:1 adaptive:1 franz:1 transaction:8 dealing:1 global:2 assumed:1 why:3 reality:1 nature:2 learn:1 robust:5 ca:1 mazumder:1 adi:2 significance:1 main:3 spread:1 linearly:2 montanari:2 weimer:1 noise:3 arise:1 n2:2 xu:3 referred:2 fashion:3 heiman:1 tong:2 hassibi:1 tomioka:1 pv:1 wish:1 kxk2:1 srinadh:1 tang:1 theorem:18 rk:3 specific:2 bishop:1 symbol:4 workshop:1 province:1 mohan:1 margin:1 rankness:5 chen:1 yin:1 interrelated:1 simply:1 infinitely:1 penrose:1 partially:2 inderjit:1 tr06:1 minimizer:3 satisfies:2 acm:1 ma:3 succeed:1 identity:2 donoho:1 considerable:2 hard:2 feasible:1 included:6 typical:1 except:1 uniformly:4 reducing:1 shang:1 lemma:7 principal:1 called:10 total:1 blessing:1 svd:9 schechtman:1 people:1 pu0:7 alexander:1 princeton:2 phenomenon:2 handling:1 |
6,280 | 6,681 | Robust Hypothesis Test for Nonlinear Effect
with Gaussian Processes
Jeremiah Zhe Liu, Brent Coull
Department of Biostatistics
Harvard University
Cambridge, MA 02138
{zhl112@mail, bcoull@hsph}.harvard.edu
Abstract
This work constructs a hypothesis test for detecting whether an data-generating
function h : Rp ? R belongs to a specific reproducing kernel Hilbert space
H0 , where the structure of H0 is only partially known. Utilizing the theory of
reproducing kernels, we reduce this hypothesis to a simple one-sided score test
for a scalar parameter, develop a testing procedure that is robust against the misspecification of kernel functions, and also propose an ensemble-based estimator
for the null model to guarantee test performance in small samples. To demonstrate
the utility of the proposed method, we apply our test to the problem of detecting
nonlinear interaction between groups of continuous features. We evaluate the
finite-sample performance of our test under different data-generating functions and
estimation strategies for the null model. Our results reveal interesting connections
between notions in machine learning (model underfit/overfit) and those in statistical
inference (i.e. Type I error/power of hypothesis test), and also highlight unexpected
consequences of common model estimating strategies (e.g. estimating kernel
hyperparameters using maximum likelihood estimation) on model inference.
1
Introduction
We study the problem of constructing a hypothesis test for an unknown data-generating function h :
Rp ? R, when h is estimated with a black-box algorithm (specifically, Gaussian Process regression)
from n observations {yi , xi }ni=1 . Specifically, we are interested in testing for the hypothesis:
H0 : h ? H0
v.s.
Ha : h ? Ha
where H0 , Ha are the function spaces for h under the null and the alternative hypothesis. We assume
only partial knowledge about H0 . For example, we may assume H0 = {h|h(xi ) = h(xi,1 )} is the
space of functions depend only on x1 , while claiming no knowledge about other properties (linearity,
smoothness, etc) about h. We pay special attention to the setting where the sample size n is small.
This type of tests carries concrete significance in scientific studies. In areas such as genetics,
drug trials and environmental health, a hypothesis test for feature effect plays a critical role in
answering scientific questions of interest. For example, assuming for simplicity x2?1 = [x1 , x2 ]T ,
an investigator might inquire the effect of drug dosage x1 on patient?s biometric measurement y
(corresponds to the null hypothesis H0 = {h(x) = h(x2 )}), or whether the adverse health effect
of air pollutants x1 is modified by patients? nutrient intake x2 (corresponds to the null hypothesis
H0 = {h(x) = h1 (x1 ) + h2 (x2 )}). In these studies, h usually represents some complex, nonlinear
biological process whose exact mathematical properties are not known. Sample size in these studies
are often small (n ? 100 ? 200), due to the high monetary and time cost in subject recruitment and
the lab analysis of biological samples.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
There exist two challenges in designing such a test. The first challenge arises from the low interpretability of the blackbox model. It is difficult to formulate a hypothesis about feature effect in
? implicitly using a collection of basis functions
these models, since the blackbox models represents h
constructed from the entire feature vector x, rather than a set of model parameters that can be
interpreted in the context of some effect of interest. For example, consider testing for the interaction
effect between x1 and x2 . With linear model h(xi ) = xi1 ?1 + xi2 ?2 + xi1 xi2 ?3 , we can simply
represent the interaction effect using a single parameter
Pn ?3 , and test for H0 : ?3 = 0. On the other
hand, Gaussian process (GP) [16] models h(xi ) = j=1 k(xi , xj )?j using basis functions defined
by the kernel function k. Since k is an implicit function that takes the entire feature vector as input, it
is not clear how to represent the interaction effect in GP models. We address this challenge assuming
h belongs to a reproducing kernel Hilbert space (RKHS) governed by the kernel function k? , such
that H = H0 when ? = 0, and H = Ha otherwise. In this way, ? encode exactly the feature effect
of interest, and the null hypothesis h ? H0 can be equivalently stated as H0 : ? = 0. To test for the
hypothesis, we re-formulate the GP estimates as the variance components of a linear mixed model
(LMM) [13], and derive a variance component score test which requires only model estimates under
the null hypothesis.
Clearly, performance of the hypothesis test depends on the quality of the model estimate under the
null hypothesis, which give rise to the second challenge: estimating the null model when only having
partial knowledge about H0 . In the case of Gaussian process, this translates to only having partial
knowledge about the kernel function k0 . The performance of Gaussian process is sensitive to the
choices of the kernel function k(z, z0 ). In principle, the RKHS H generated by a proper kernel
function k(z, z0 ) should be rich enough so it contains the data-generating function h, yet restrictive
? does not overfit in small samples. Choosing a kernel function that is too restrictive
enough such that h
or too flexible will lead to either model underfit or overfit, rendering the subsequent hypothesis
tests not valid. We address this challenge by proposing an ensemble-based estimator for h we term
Cross-validated Kernel Ensemble (CVEK). Using a library of base kernels, CVEK learns a proper H
from data by directly minimizing the ensemble model?s cross-validation error, therefore guaranteeing
robust test performance for a wide range of data-generating functions.
The rest of the paper is structured as follows. After a brief review of Gaussian process and its
connection with linear mixed model in Section 2, we introduce the test procedure for general
hypothesis h ? H0 in Section 3, and its companion estimation procedure CVEK in Section 4. To
demonstrate the utility of the proposed test, in section 5, we adapt our test to the problem of detecting
nonlinear interaction between groups of continuous features, and in section 6 we conduct simulation
studies to evaluate the finite-sample performance of the interaction test, under different kernel
estimation strategies, and under a range of data-generating functions with different mathematical
properties. Our simulation study reveals interesting connection between notions in machine learning
and those in statistical inference, by elucidating the consequence of model estimation (underfit /
overfit) on the Type I error and power of the subsequent hypothesis test. It also cautions against the
use of some common estimation strategies (most notably, selecting kernel hyperparameters using
maximum likelihood estimation) when conducting hypothesis test in small samples, by highlighting
inflated Type I errors from hypothesis tests based on the resulting estimates. We note that the methods
and conclusions from this work is extendable beyond the Gaussian Process models, due to GP?s
connection to other blackbox models such as random forest [5] and deep neural network [19].
2
Background on Gaussian Process
Assume we observe data from n independent subjects. For the ith subject, let yi be a continuous
response, xi be the set of p continuous features that has nonlinear effect on yi . We assume that the
outcome yi depends on features xi through below data-generating model:
iid
yi |h = ? + h(xi ) + i
where i ? N (0, ?)
(1)
and h : Rp ? R follows the Gaussian process prior GP(0, k) governed by the positive definite kernel
function k, such that the function evaluated at the observed record follows the multivariate normal
(MVN) distribution:
h = [h(x1 ), . . . , h(xn )] ? M V N (0, K)
2
with covariance matrix Ki,j = k(xi , xj ). Under above construction, the predictive distribution of h
evaluated at the samples is also multivariate normal:
h|{yi , xi }ni=1 ? M V N (h? , K? )
h? = K(K + ?I)?1 (y ? u)
K? = K ? K(K + ?I)?1 K
To understand the impact of ? and k on h? , recall that Gaussian process can be understood as the
Bayesian version of the kernel machine regression, where h? equivalently arise from the below
optimization problem:
h? = argmin ||y ? ? ? h(x)||2 + ?||h||2H
h?Hk
where Hk is the RKHS generated by kernel function k. From this perspective, h? is the element
in a spherical ball in Hk that best approximates the observed data y. The norm of h? , ||h||2H , is
constrained by the tuning parameter ?, and the mathematical properties (e.g. smoothness, spectral
density, etc) of h? are governed by the kernel function k. It should be noticed that although h? may
arise from Hk , the probability of the Gaussian Process h ? Hk is 0 [14].
Gaussian Process as Linear Mixed Model
[13] argued that if define ? =
?2
? ,
y =?+h+
h? can arise exactly from a linear mixed model (LMM):
where
h ? N (0, ? K)
? N (0, ? 2 I)
(2)
Therefore ? can be treated as part of the LMM?s variance components parameters. If K is correctly specified, then the variance components parameters (?, ? 2 ) can be estimated unbiasedly by
maximizing the Restricted Maximum Likelihood (REML)[12]:
LREML (?, ?, ? 2 |K) = ?log|V| ? log|1T V?1 1| ? (y ? ?)T V?1 (y ? ?)
(3)
where V = ? K + ? 2 I, and 1 a n ? 1 vector whose all elements are 1. However, it is worth noting
2
that REML is a model-based procedure. Therefore improper estimates for ? = ?? may arise when
the family of kernel functions are mis-specified.
3
A recipe for general hypothesis h ? H0
The GP-LMM connection introduced in Section 2 opens up the arsenal of statistical tools from Linear
Mixed Model for inference tasks in Gaussian Process. Here, we use the classical variance component
test [12] to construct a testing procedure for the hypothesis about Gaussian process function:
H0 : h ? H0 .
(4)
We first translate above hypothesis into a hypothesis in terms of model parameters. The key of our
approach is to assume that h lies in a RKHS generated by a garrote kernel function k? (z, z0 ) [15],
which is constructed by including an extra garrote parameter
? to a given kernel function. When
? = 0, the garrote kernel function k0 (x, x0 ) = k? (x, x0 )
generates exactly H0 , the space of
?=0
functions under the null hypothesis. In order to adapt this general hypothesisio to their hypothesis of
interest, practitioners need only to specify the form of the garrote kernel so that H0 corresponds to
the null hypothesis. For example, If k? (x) = k(? ? x1 , x2 , . . . , xp ), ? = 0 corresponds to the null
hypothesis H0 : h(x) = h(x2 , . . . , xp ), i.e. the function h(x) does not depend on x1 . (As we?ll see
in section 5, identifying such k0 is not always straightforward). As a result, the general hypothesis
(4) is equivalent to
H0 : ? = 0.
(5)
We now construct a test statistic T?0 for (5) by noticing that the garrote parameter ? can be treated as a
variance component parameter in the linear mixed model. This is because the Gaussian process under
garrote kernel can be formulated into below LMM:
y =?+h+
where
h ? N (0, ? K? )
3
? N (0, ? 2 I)
where K? is the kernel matrix generated by k? (z, z0 ). Consequently, we can derive a variance
component test for H0 by calculating the square derivative of LREML with respect ? under H0 [12]:
h
i
? T V0?1 ?K0 V0?1 (y ? ?)
?
T?0 = ?? ? (y ? ?)
(6)
where V0 = ?
? 2 I + ??K0 . In this expression, K0 = K? , and ?K0 is the null derivative kernel
?=0
?
matrix whose (i, j)th entry is ??
k? (x, x0 ) .
?=0
As discussed previously, misspecifying the null kernel function k0 negatively impacts the performance
of the resulting hypothesis test. To better understand the mechanism at play, we express the test
?
? ? h:
statistic T?0 from (6) in terms of the model residual ? = y ? ?
h
i
??
T
?,
?
?K
(7)
?
T?0 =
0
?
?4
? = (?
where we have used the fact V0?1 (y ? ?)
? 2 )?1 (?) [10]. As shown, the test statistic T?0 is a
scaled quadratic-form statistic that is a function of the model residual. If k0 is too restrictive, model
estimates will underfit the data even under the null hypothesis, introducing extraneous correlation
among the ?i ?s, therefore leading to overestimated T?0 and eventually underestimated p-value under
the null. In this case, the test procedure will frequently reject the null hypothesis (i.e. suggest the
existence of nonlinear interaction) even when there is in fact no interaction, yielding an invalid test
due to inflated Type I error. On the other hand, if k0 is too flexible, model estimates will likely
overfit the data in small samples, producing underestimated residuals, an underestimated test statistic,
and overestimated p-values. In this case, the test procedure will too frequently fail to reject the
null hypothesis (i.e. suggesting there is no interaction) when there is in fact interaction, yielding an
insensitive test with diminished power.
The null distribution of T? can be approximated using a scaled chi-square distribution ??2? using
Satterthwaite method [20] by matching the first two moments of T :
? ? ? = E(T ),
with solution (see Appendix for derivation):
h
i
?
? = ?I?? / ?? ? tr V0?1 ?K0
2 ? ?2 ? ? = V ar(T )
h
i2
?? = ?? ? tr V0?1 ?K0
/(2 ? ?I?? )
where ?I?? and ?I?? are the submatrices of the REML information matrix. Numerically more accurate,
but computationally less efficient approximation methods are also available [2].
Finally, the p-value of this test is calculated by examining the tail probability of ?
? ?2?? :
p = P (?
??2?? > T?) = P (?2?? > T?/?
?)
A complete summary of the proposed testing procedure is available in Algorithm 1.
Algorithm 1 Variance Component Test for h ? H0
1: procedure VCT FOR I NTERACTION
Input: Null Kernel Matrix K0 , Derivative Kernel Matrix ?K0 , Data (y, X)
Output: Hypothesis Test p-value p
# Step 1: Estimate Null Model using REML
? ??, ?
2:
(?,
? 2 ) = argmin LREML (?, ?, ? 2 |K0 ) as in (3)
# Step 2: Compute Test Statistic and Null Distribution Parameters
? T V?1 ?K0 V?1 (y ? X?)
?
3:
T?0 = ?? ? (y ? X?)
0 i
0 h
h
i2
4:
?
? = ?I?? / ?? ? tr V?1 ?K0 , ?? = ?? ? tr V?1 ?K0
/(2 ? ?I?? )
0
0
# Step 3: Compute p-value and reach conclusion
5:
p = P (?
??2?? > T?) = P (?2?? > T?/?
?)
6: end procedure
4
In light of the discussion about model misspecification in Introduction section, we highlight the fact
that our proposed test (6) is robust against model misspecification under the alternative [12], since
the calculation of test statistics do not require detailed parametric assumption about k? . However, the
test is NOT robust against model misspecification under the null, since the expression of both test
statistic T?0 and the null distribution parameters (?
?, ??) still involve the kernel matrices generated by
k0 (see Algorithm 1). In the next section, we address this problem by proposing a robust estimation
procedure for the kernel matrices under the null.
4
Estimating Null Kernel Matrix using Cross-validated Kernel Ensemble
Observation in (7) motivates the need for a kernel estimation strategy that is flexible so that it does
not underfit under the null, yet stable so that it does not overfit under the alternative. To this end, we
propose estimating h using the ensemble of a library of fixed base kernels {kd }D
d=1 :
?
h(x)
=
D
X
? d (x)
ud h
u ? ? = {u|u ? 0, ||u||22 = 1},
(8)
d=1
? d is the kernel predictor generated by dth base kernel kd . In order to maximize model stability,
where h
? We term
the ensemble weights u are estimated to minimize the overall cross-validation error of h.
this method the Cross-Validated Kernel Ensemble (CVEK). Our proposed method belongs to the
well-studied family of algorithms known as ensembles of kernel predictors (EKP) [7, 8, 3, 4], but with
specialized focus in maximizing the algorithm?s cross-validation stability. Furthermore, in addition to
? CVEK will also produce the ensemble estimate of the kernel matrix
producing ensemble estimates h,
? 0 that is required by Algorithm 1. The exact algorithm proceeds in three stages as follows:
K
?
? ?1 y,
Stage 1: For each basis kernel in the library {kd }D
d=1 , we first estimate hd = Kd (Kd + ?d I)
th
? d is selected by minimizing the
the prediction based on d kernel, where the tunning parameter ?
leave-one-out cross validation (LOOCV) error [6]:
? d,? ) where Ad,? = Kd (Kd + ?I)?1 .
LOOCV(?|kd ) = (I ? diag(Ad,? ))?1 (y ? h
(9)
? d |kd .
We denote estimate the final LOOCV error for dth kernel ?d = LOOCV ?
D
Stage 2: Using the estimated LOOCV errors {?
d }D
d=1 , estimate the ensemble weights u = {ud }d=1
such that it minimizes the overall LOOCV error:
D
X
?
u = argmin ||
ud ?d ||2
where ? = {u|u ? 0, ||u||22 = 1},
u??
d=1
and produce the final ensemble prediction:
?=
h
D
X
d=1
u
?d hd =
D
X
?
u
?d Ad,?? d y = Ay,
d=1
? = PD u
where A
? d is the ensemble hat matrix.
d=1 ?d Ad,?
? estimate the ensemble kernel matrix K
? by solving:
Stage 3: Using the ensemble hat matrix A,
? K
? + ?I)?1 = A.
?
K(
? then K
? adopts
Specifically, if we denote UA and {?A,k }nk=1 the eigenvector and eigenvalues of A,
the form (see Appendix for derivation):
?
A,k
? = UA diag
K
UTA
1 ? ?A,k
5
Application: Testing for Nonlinear Interaction
Recall in Section 3, we assume that we are given a k? that generates exactly H0 . However, depending
on the exact hypothesis of interest, identifying such k0 is not always straightforward. In this section,
5
we revisit the example about interaction testing discussed in challenge 1 at the Introduction section,
and consider how to build a k0 for below hypothesis of interest
H0 : h(x) = h1 (x1 ) + h2 (x2 )
Ha : h(x) = h1 (x1 ) + h2 (x2 ) + h12 (x1 , x2 )
where h12 is the "pure interaction" function that is orthogonal to main effect functions h1 and h2 .
Recall as discussed previously, this hypothesis is difficult to formulate with Gaussian process models,
since the kernel functions k(x, x0 ) in general do not explicitly separate the main and the interaction
effect. Therefore rather than directly define k0 , we need to first construct H0 and Ha that corresponds
to the null and alternative hypothesis, and then identify the garrote kernel function k? such it generates
exactly H0 when ? = 0 and Ha when ? > 0.
We build H0 using the tensor-product construction of RKHS on the product domain (x1,i , x2,i ) ?
Rp1 ? Rp2 [9], due to this approach?s unique ability in explicitly characterizing the space of "pure
interaction" functions. Let 1 = {f |f ? 1} be the RKHS of constant functions, and H1 , H2
be the RKHS of centered functions for x1 x2 , respectively. We can then define the full space as
H = ?2m=1 (1 ? Hm ). H describes the space of functions that depends jointly on {x1 , x2 }, and
adopts below orthogonal decomposition:
H = (1 ? H1 ) ? (1 ? H2 )
n
o n
o
?
= 1 ? H1 ? H2 ? H1 ? H2 = 1 ? H12
? H12
?
where we have denoted H12
= H1 ? H2 and H12 = H1 ? H2 , respectively. We see that H12 is
indeed the space of?pure interaction" functions , since H12 contains functions on the product domain
?
Rp1 ? Rp2 , but is orthogonal to the space of additive main effect functions H12
. To summarize, we
have identified two function spaces H0 and Ha that has the desired interpretation:
?
H0 = H12
?
Ha = H12
? H12
We are now ready to identify the garrote kernel k? (x, x0 ). To this end, we notice that both H0 and
H12 are composite spaces built from basis RKHSs using direct sum and tensor product. If denote
km (xm , x0m ) the reproducing kernel associated with Hm , we can construct kernel functions for
composite spaces H0 and H12 as [1]:
k0 (x, x0 ) = k1 (x1 , x01 ) + k2 (x2 , x02 )
k12 (x, x0 ) = k1 (x1 , x01 ) k2 (x2 , x02 )
and consequently, the garrote kernel function for Ha :
k? (x, x0 ) = k0 (x, x0 ) + ? ? k12 (x, x0 ).
(10)
Finally, using the chosen form of the garrote kernel function, the (i, j)th element of the null derivative
?
kernel matrix K0 is ??
k? (x, x0 ) = k12 (x, x0 ), i.e. the null derivative kernel matrix ?K0 is simply
the kernel matrix K12 that corresponds to the interaction space. Therefore the score test statistic T?0
in (6) simplifies to:
? T V?1 K12 V?1 (y ? X?)
?
T?0 = ?? ? (y ? X?)
0
0
(11)
where V0 = ?
? 2 I + ??K0 .
6
Simulation Experiment
We evaluated the finite-sample performance of the proposed interaction test in a simulation study
that is analogous to a real nutrition-environment interaction study. We generate two groups of input
features (xi,1 , xi,2 ) ? Rp1 ? Rp2 independently from standard Gaussian distribution, representing
normalized data representing subject?s level of exposure to p1 environmental pollutants and the levels
of a subject?s intake of p2 nutrients during the study. Throughout the simulation scenarios, we keep
n = 100, and p1 = p2 = 5. We generate the outcome yi as:
yi = h1 (xi,1 ) + h2 (xi,2 ) + ? ? h12 (xi,1 , xi,2 ) + i
6
(12)
where h1 , h2 , h12 are sampled from RKHSs H1 , H2 and H1 ? H2 , generated using a ground-truth
kernel ktrue . We standardize all sampled functions to have unit norm, so that ? represents the strength
of interaction relative to the main effect.
For each simulation scenario, we first generated data using ? and ktrue as above, then selected a
kmodel to estimate the null model and obtain p-value using Algorithm 1. We repeated each scenario
1000 times, and evaluate the test performance using the empirical probability P? (p ? 0.05). Under
null hypothesis H0 : ? = 0, P? (p ? 0.05) estimates the test?s Type I error, and should be smaller or
equal to the significance level 0.05. Under alternative hypothesis Ha : ? > 0, P? (p ? 0.05) estimates
the test?s power, and should ideally approach 1 quickly as the strength of interaction ? increases.
In this study, we varied ktrue to produce data-generating functions h? (xi,1 , xi,2 ) with different
smoothness and complexity properties, and varied kmodel to reflect different common modeling
strategies for the null model in addition to using CVEK. We then evaluated how these two aspects
impact the hypothesis test?s Type I error and power.
Data-generating Functions
We sampled the data-generate function by using ktrue from Mat?rn kernel family [16]:
? ?
21?? ?
k(r|?, ?) =
2??||r|| K?
2??||r|| ,
where
r = x ? x0 ,
?(?)
with two non-negative hyperparameters (?, ?). For a function h sampled using Mat?rn kernel, ?
determines the function?s smoothness, since h is k-times mean square differentiable if and only
if ? > k. In the case of ? ? ?, Mat?rn kernel reduces to the Gaussian RBF kernel which is
infinitely differentiable. ? determines the function?s complexity, this is because in Bochner?s spectral
decomposition[16]
Z
T
k(r|?, ?) = e2?is r dS(s|?, ?),
(13)
? determines how much weight the spectral density S(s) puts on the slow-varying, low-frequency
basis functions. In this work, we vary ? ? { 32 , 52 , ?} to generate once-, twice, and infinitelydifferentiable functions, and vary ? ? {0.5, 1, 1.5} to generate functions with varying degree of
complexity.
Modeling Strategies
Polynomial Kernels is a family of simple parametric kernels that is equivalent to the polynomial
ridge regression model favored by statisticians due to its model interpretability. In this work, we
use the linear kernel klinear (x, x0 |p) = xT x0 and quadratic kernel kquad (x, x0 |p) = (1 + xT x0 )2
which are common choices from this family.
Gaussian RBF Kernels kRBF (x, x0 |?) = exp(??||x ? x0 ||2 ) is a general-purpose kernel family that
generates nonlinear, but infinitely differentiable (therefore very smooth) functions. Under this kernel,
we consider two hyperparameter selection strategies common in machine learning applications: RBFMedian where we set ? to the sample median of {||xi ? xj ||}i6=j , and RBF-MLE who estimates ?
by maximizing the model likelihood.
Mat?rn and Neural Network Kernels are two flexible kernels favored by machine learning practitioners for their expressiveness. Mat?rn kernels generates functions that are more flexible compared
to that of Gaussian RBF due to the relaxed smoothness constraint [17]. In order to investigate the
consequences of added flexibility relative to the true model, we use Matern 1/2, Matern 3/2 and
Matern 5/2, corresponding to Mat?rn
kernels with ? = 1/2,
3/2, and 5/2. Neural network kernels
?0
2??
xT x
[16] knn (x, x0 |?) = ?2 ? sin?1 ?
, on the other hand, represent a 1-layer
T
0T 0
? )(1+2??
(1+2??
x x
x
? )
x
Bayesian neural network with infinite hidden unit and probit link function, with ? being the prior
variance on hidden weights. Therefore knn is flexible in the sense that it is an universal approximator
for arbitrary continuous functions on the compact domain [11]. In this work, we denote NN 0.1, NN
1 and NN 10 to represent Bayesian networks with different prior constraints ? ? {0.1, 1, 10}.
Result
The simulation results are presented graphically in Figure 1 and documented in detail in the Appendix.
We first observe that for reasonably specified kmodel , the proposed hypothesis test always has the
7
1.0
0.8
0.6
0.0
0.2
0.4
Linear
Quadratic
RBF_MLE
RBF_Median
Matern 1/2
Matern 3/2
Matern 5/2
(b) ktrue = Mat?rn 3/2, ? = 1
(c) ktrue = Mat?rn 3/2, ? = 1.5
0.6
0.8
1.0
(a) ktrue = Mat?rn 3/2, ? = 0.5
0.0
0.2
0.4
NN 0.1
NN 1
NN 10
CVKE_RBF
CVKE_NN
(e) ktrue = Mat?rn 5/2, ? = 1
(f) ktrue = Mat?rn 5/2, ? = 1.5
0.0
0.2
0.4
0.6
0.8
1.0
(d) ktrue = Mat?rn 5/2, ? = 0.5
0.0
0.2
0.4
0.6
0.8
1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0
0.2
0.4
0.6
0.8
1.0
(g) ktrue = Gaussian RBF, ? = 0.5 (h) ktrue = Gaussian RBF, ? = 1 (i) ktrue = Gaussian RBF, ? = 1.5
Figure 1: Estimated P? (p < 0.05) (y-axis) as a function of Interaction Strength ? ? [0, 1] (x-axis).
Skype Blue: Linear (Solid) and Quadratic (Dashed) Kernels, Black: RBF-Median (Solid) and RBFMLE (Dashed), Dark Blue: Mat?rn Kernels with ? = 12 , 32 , 25 , Purple: Neural Network Kernels
with ? = 0.1, 1, 10, Red: CVEK based on RBF (Solid) and Neural Networks (Dashed).
Horizontal line marks the test?s significance level (0.05). When ? = 0, P? should be below this line.
correct Type I error and reasonable power. We also observe that the complexity of the data-generating
function h? (12) plays a role in test performance, in the sense that the power of the hypothesis
tests increases as the Mat?rn ktrue ?s complexity parameter ? becomes larger, which corresponds to
functions that put more weight on the complex, fast-varying eigenfunctions in (13).
We observed clear differences in test performance from different estimation strategies. In general,
polynomial models (linear and quadratic) are too restrictive and appear to underfit the data under
both the null and the alternative, producing inflated Type I error and diminished power. On the other
hand, lower-order Mat?rn kernels (Mat?rn 1/2 and Mat?rn 3/2, dark blue lines) appear to be too
flexible. Whenever data are generated from smoother ktrue , Mat?rn 1/2 and 3/2 overfit the data and
produce deflated Type I error and severely diminished power, even if the hyperparameter ? are fixed
at true value. Therefore unless there?s strong evidence that h exhibits behavior consistent with that
described by these kernels, we recommend avoid the use of either polynomial or lower-order Mat?rn
kernels for hypothesis testing. Comparatively, Gaussian RBF works well for a wider range of ktrue ?s,
but only if the hyperparameter ? is selected carefully. Specifically, RBF-Median (black dashed line)
works generally well, despite being slightly conservative (i.e. lower power) when the data-generation
function is smooth and of low complexity. RBF-MLE (black solid line), on the other hand, tends to
underfit the data under the null and exhibits inflated Type I error, possibly because of the fact that ? is
not strongly identified when the sample size is too small [18]. The situation becomes more severe as
h? becomes rougher and more complex, in the moderately extreme case of non-differentiable h with
? = 1.5, the Type I error is inflated to as high as 0.238. Neural Network kernels also perform well
for a wide range of ktrue , and its Type I error is more robust to the specification of hyperparameters.
Finally, the two ensemble estimators CVEK-RBF (based on kRBF ?s with log(?) ? {?2, ?1, 0, 1, 2})
and CVEK-NN (based on kNN ?s with ? ? {0.1, 1, 10, 50}) perform as well or better than the nonensemble approaches for all ktrue ?s, despite being slightly conservative under the null. As compared
to CVEK-NN, CVEK-RBF appears to be slightly more powerful.
8
References
[1] N. Aronszajn. Theory of reproducing kernels. Transactions of the American Mathematical
Society, 68(3):337?404, 1950.
[2] D. A. Bodenham and N. M. Adams. A comparison of efficient approximations for a weighted
sum of chi-squared random variables. Statistics and Computing, 26(4):917?928, July 2016.
[3] C. Cortes, M. Mohri, and A. Rostamizadeh. Two-Stage Learning Kernel Algorithms. 2010.
[4] C. Cortes, M. Mohri, and A. Rostamizadeh. Ensembles of Kernel Predictors. arXiv:1202.3712
[cs, stat], Feb. 2012. arXiv: 1202.3712.
[5] A. Davies and Z. Ghahramani. The Random Forest Kernel and other kernels for big data from
random partitions. arXiv:1402.4293 [cs, stat], Feb. 2014. arXiv: 1402.4293.
[6] A. Elisseeff and M. Pontil. Leave-one-out Error and Stability of Learning Algorithms with
Applications. In J. Suykens, G. Horvath, S. Basu, C. Micchelli, and J. Vandewalle, editors,
Learning Theory and Practice. IOS Press, 2002.
[7] T. Evgeniou, L. Perez-Breva, M. Pontil, and T. Poggio. Bounds on the Generalization Performance of Kernel Machine Ensembles. In Proceedings of the Seventeenth International
Conference on Machine Learning, ICML ?00, pages 271?278, San Francisco, CA, USA, 2000.
Morgan Kaufmann Publishers Inc.
[8] T. Evgeniou, M. Pontil, and A. Elisseeff. Leave One Out Error, Stability, and Generalization of
Voting Combinations of Classifiers. Machine Learning, 55(1):71?97, Apr. 2004.
[9] C. Gu. Smoothing Spline ANOVA Models. Springer Science & Business Media, Jan. 2013.
Google-Books-ID: 5VxGAAAAQBAJ.
[10] D. A. Harville. Maximum Likelihood Approaches to Variance Component Estimation and to
Related Problems. Journal of the American Statistical Association, 72(358):320?338, 1977.
[11] K. Hornik. Approximation capabilities of multilayer feedforward networks. Neural Networks,
4(2):251?257, 1991.
[12] X. Lin. Variance component testing in generalised linear models with random effects.
Biometrika, 84(2):309?326, June 1997.
[13] D. Liu, X. Lin, and D. Ghosh. Semiparametric Regression of Multidimensional Genetic Pathway
Data: Least-Squares Kernel Machines and Linear Mixed Models. Biometrics, 63(4):1079?1088,
Dec. 2007.
[14] M. N. Luki?c and J. H. Beder. Stochastic Processes with Sample Paths in Reproducing Kernel
Hilbert Spaces. Transactions of the American Mathematical Society, 353(10):3945?3969, 2001.
[15] A. Maity and X. Lin. Powerful tests for detecting a gene effect in the presence of possible
gene-gene interactions using garrote kernel machines. Biometrics, 67(4):1271?1284, Dec. 2011.
[16] C. E. Rasmussen and C. K. I. Williams. Gaussian Processes for Machine Learning. University
Press Group Limited, Jan. 2006. Google-Books-ID: vWtwQgAACAAJ.
[17] J. Snoek, H. Larochelle, and R. P. Adams. Practical Bayesian Optimization of Machine Learning
Algorithms. arXiv:1206.2944 [cs, stat], June 2012. arXiv: 1206.2944.
[18] G. Wahba. Spline Models for Observational Data. SIAM, Sept. 1990. Google-Books-ID:
ScRQJEETs0EC.
[19] A. G. Wilson, Z. Hu, R. Salakhutdinov, and E. P. Xing. Deep Kernel Learning. arXiv:1511.02222
[cs, stat], Nov. 2015. arXiv: 1511.02222.
[20] D. Zhang and X. Lin. Hypothesis testing in semiparametric additive mixed models. Biostatistics
(Oxford, England), 4(1):57?74, Jan. 2003.
9
| 6681 |@word trial:1 version:1 polynomial:4 norm:2 open:1 km:1 hu:1 simulation:7 covariance:1 decomposition:2 elisseeff:2 tr:4 solid:4 carry:1 moment:1 liu:2 contains:2 score:3 selecting:1 genetic:1 rkhs:7 intake:2 yet:2 subsequent:2 additive:2 partition:1 selected:3 rp1:3 ith:1 record:1 detecting:4 zhang:1 unbiasedly:1 mathematical:5 constructed:2 direct:1 bodenham:1 pathway:1 introduce:1 x0:20 snoek:1 notably:1 indeed:1 behavior:1 p1:2 frequently:2 blackbox:3 chi:2 salakhutdinov:1 spherical:1 ua:2 becomes:3 estimating:5 linearity:1 biostatistics:2 medium:1 null:36 argmin:3 interpreted:1 minimizes:1 eigenvector:1 proposing:2 caution:1 ghosh:1 guarantee:1 multidimensional:1 voting:1 exactly:5 biometrika:1 scaled:2 k2:2 x0m:1 classifier:1 unit:2 appear:2 producing:3 positive:1 generalised:1 understood:1 tends:1 consequence:3 severely:1 despite:2 io:1 id:3 oxford:1 path:1 black:4 might:1 twice:1 studied:1 limited:1 range:4 seventeenth:1 unique:1 practical:1 testing:10 practice:1 definite:1 procedure:11 pontil:3 jan:3 area:1 empirical:1 drug:2 arsenal:1 reject:2 submatrices:1 matching:1 composite:2 universal:1 davy:1 suggest:1 selection:1 put:2 context:1 equivalent:2 maximizing:3 straightforward:2 attention:1 exposure:1 independently:1 satterthwaite:1 graphically:1 formulate:3 williams:1 simplicity:1 identifying:2 pure:3 estimator:3 utilizing:1 tunning:1 hd:2 stability:4 notion:2 analogous:1 construction:2 play:3 exact:3 designing:1 hypothesis:46 harvard:2 element:3 standardize:1 approximated:1 observed:3 role:2 inquire:1 improper:1 pd:1 environment:1 complexity:6 moderately:1 ideally:1 depend:2 solving:1 predictive:1 negatively:1 basis:5 gu:1 k0:27 derivation:2 fast:1 choosing:1 h0:35 outcome:2 whose:3 larger:1 otherwise:1 ability:1 statistic:10 knn:3 gp:6 jointly:1 final:2 eigenvalue:1 differentiable:4 propose:2 interaction:23 product:4 monetary:1 translate:1 flexibility:1 recipe:1 produce:4 generating:10 guaranteeing:1 leave:3 adam:2 wider:1 derive:2 develop:1 depending:1 stat:4 strong:1 p2:2 c:4 larochelle:1 inflated:5 correct:1 stochastic:1 nutrient:2 centered:1 observational:1 argued:1 require:1 generalization:2 biological:2 ground:1 normal:2 exp:1 vary:2 purpose:1 estimation:11 loocv:6 sensitive:1 tool:1 weighted:1 clearly:1 gaussian:25 always:3 modified:1 rather:2 pn:1 avoid:1 varying:3 wilson:1 encode:1 validated:3 focus:1 june:2 likelihood:5 hk:5 rostamizadeh:2 sense:2 inference:4 nn:8 entire:2 hidden:2 interested:1 biometric:1 overall:2 among:1 flexible:7 denoted:1 extraneous:1 favored:2 constrained:1 special:1 smoothing:1 equal:1 construct:5 evgeniou:2 having:2 beach:1 once:1 represents:3 icml:1 recommend:1 spline:2 dosage:1 uta:1 statistician:1 interest:6 investigate:1 elucidating:1 severe:1 extreme:1 yielding:2 light:1 perez:1 accurate:1 partial:3 poggio:1 orthogonal:3 unless:1 conduct:1 biometrics:2 desired:1 re:1 hsph:1 modeling:2 ar:1 cost:1 introducing:1 entry:1 predictor:3 examining:1 vandewalle:1 too:8 extendable:1 st:1 density:2 international:1 siam:1 overestimated:2 xi1:2 quickly:1 concrete:1 squared:1 reflect:1 possibly:1 brent:1 american:3 derivative:5 leading:1 book:3 suggesting:1 inc:1 explicitly:2 depends:3 ad:4 h1:14 matern:6 lab:1 red:1 xing:1 capability:1 minimize:1 air:1 ni:2 square:4 purple:1 variance:11 conducting:1 who:1 ensemble:19 kaufmann:1 identify:2 bayesian:4 iid:1 reml:4 worth:1 reach:1 whenever:1 against:4 frequency:1 e2:1 associated:1 mi:1 sampled:4 recall:3 knowledge:4 hilbert:3 carefully:1 appears:1 response:1 specify:1 evaluated:4 box:1 strongly:1 furthermore:1 implicit:1 stage:5 correlation:1 overfit:7 hand:5 d:1 horizontal:1 aronszajn:1 nonlinear:8 google:3 quality:1 reveal:1 scientific:2 usa:2 effect:17 normalized:1 true:2 i2:2 ll:1 during:1 sin:1 ay:1 complete:1 demonstrate:2 ridge:1 ktrue:18 common:5 specialized:1 insensitive:1 discussed:3 tail:1 approximates:1 interpretation:1 numerically:1 association:1 measurement:1 cambridge:1 smoothness:5 tuning:1 i6:1 stable:1 specification:1 v0:7 etc:2 base:3 feb:2 multivariate:2 perspective:1 belongs:3 krbf:2 scenario:3 yi:8 morgan:1 relaxed:1 bochner:1 lmm:5 ud:3 maximize:1 x02:2 dashed:4 smoother:1 full:1 july:1 reduces:1 smooth:2 adapt:2 calculation:1 cross:7 long:1 lin:4 england:1 mle:2 impact:3 prediction:2 regression:4 multilayer:1 patient:2 arxiv:8 kernel:89 represent:4 suykens:1 dec:2 background:1 addition:2 semiparametric:2 underestimated:3 median:3 publisher:1 extra:1 rest:1 eigenfunctions:1 subject:5 practitioner:2 noting:1 presence:1 feedforward:1 enough:2 rendering:1 xj:3 identified:2 wahba:1 reduce:1 simplifies:1 translates:1 whether:2 expression:2 utility:2 deep:2 generally:1 clear:2 detailed:1 involve:1 dark:2 documented:1 generate:5 exist:1 revisit:1 notice:1 estimated:5 correctly:1 blue:3 hyperparameter:3 mat:19 express:1 group:4 key:1 harville:1 anova:1 sum:2 recruitment:1 noticing:1 powerful:2 family:6 throughout:1 reasonable:1 breva:1 h12:16 garrote:11 appendix:3 ki:1 layer:1 pay:1 bound:1 quadratic:5 strength:3 constraint:2 x2:16 rp2:3 generates:5 aspect:1 department:1 structured:1 ball:1 combination:1 kd:9 describes:1 smaller:1 slightly:3 restricted:1 sided:1 computationally:1 previously:2 eventually:1 mechanism:1 xi2:2 fail:1 end:3 available:2 apply:1 observe:3 spectral:3 alternative:6 rkhss:2 rp:3 existence:1 hat:2 calculating:1 restrictive:4 k1:2 build:2 ghahramani:1 classical:1 comparatively:1 society:2 tensor:2 micchelli:1 noticed:1 question:1 added:1 strategy:9 parametric:2 exhibit:2 separate:1 link:1 mail:1 nutrition:1 assuming:2 horvath:1 minimizing:2 equivalently:2 difficult:2 claiming:1 stated:1 rise:1 negative:1 proper:2 motivates:1 unknown:1 perform:2 observation:2 finite:3 situation:1 misspecification:4 rn:19 varied:2 reproducing:6 arbitrary:1 expressiveness:1 introduced:1 required:1 specified:3 connection:5 rougher:1 nip:1 address:3 beyond:1 dth:2 proceeds:1 usually:1 below:6 xm:1 challenge:6 summarize:1 built:1 interpretability:2 including:1 power:10 critical:1 treated:2 business:1 residual:3 representing:2 brief:1 library:3 axis:2 ready:1 hm:2 health:2 sept:1 mvn:1 review:1 prior:3 relative:2 probit:1 highlight:2 mixed:8 interesting:2 generation:1 approximator:1 validation:4 h2:14 ekp:1 x01:2 degree:1 xp:2 consistent:1 principle:1 editor:1 genetics:1 summary:1 mohri:2 rasmussen:1 pollutant:2 understand:2 wide:2 basu:1 characterizing:1 k12:5 calculated:1 xn:1 valid:1 rich:1 adopts:2 collection:1 san:1 transaction:2 nov:1 compact:1 implicitly:1 keep:1 gene:3 reveals:1 francisco:1 xi:20 zhe:1 continuous:5 reasonably:1 robust:7 ca:2 hornik:1 forest:2 complex:3 constructing:1 domain:3 diag:2 significance:3 main:4 apr:1 underfit:7 big:1 hyperparameters:4 arise:4 misspecifying:1 repeated:1 x1:17 slow:1 lie:1 governed:3 answering:1 learns:1 z0:4 companion:1 specific:1 xt:3 deflated:1 cortes:2 evidence:1 nk:1 simply:2 likely:1 infinitely:2 highlighting:1 unexpected:1 partially:1 scalar:1 springer:1 corresponds:7 truth:1 environmental:2 determines:3 ma:1 formulated:1 consequently:2 invalid:1 rbf:14 adverse:1 diminished:3 specifically:4 infinite:1 conservative:2 mark:1 vct:1 arises:1 investigator:1 evaluate:3 |
6,281 | 6,682 | Lower bounds on the robustness to adversarial
perturbations
Jonathan Peck1,2 , Joris Roels2,3 , Bart Goossens3 , and Yvan Saeys1,2
1
Department of Applied Mathematics, Computer Science and Statistics, Ghent University, Ghent, 9000, Belgium
2
Data Mining and Modeling for Biomedicine, VIB Inflammation Research Center, Ghent, 9052, Belgium
3
Department of Telecommunications and Information Processing, Ghent University, Ghent, 9000, Belgium
Abstract
The input-output mappings learned by state-of-the-art neural networks are significantly discontinuous. It is possible to cause a neural network used for image
recognition to misclassify its input by applying very specific, hardly perceptible
perturbations to the input, called adversarial perturbations. Many hypotheses have
been proposed to explain the existence of these peculiar samples as well as several
methods to mitigate them, but a proven explanation remains elusive. In this work,
we take steps towards a formal characterization of adversarial perturbations by
deriving lower bounds on the magnitudes of perturbations necessary to change the
classification of neural networks. The proposed bounds can be computed efficiently,
requiring time at most linear in the number of parameters and hyperparameters
of the model for any given sample. This makes them suitable for use in model
selection, when one wishes to find out which of several proposed classifiers is
most robust to adversarial perturbations. They may also be used as a basis for
developing techniques to increase the robustness of classifiers, since they enjoy the
theoretical guarantee that no adversarial perturbation could possibly be any smaller
than the quantities provided by the bounds. We experimentally verify the bounds
on the MNIST and CIFAR-10 data sets and find no violations. Additionally, the
experimental results suggest that very small adversarial perturbations may occur
with non-zero probability on natural samples.
1
Introduction
Despite their big successes in various AI tasks, neural networks are basically black boxes: there is no
clear fundamental explanation how they are able to outperform the more classical approaches. This
has led to the identification of several unexpected and counter-intuitive properties of neural networks.
In particular, Szegedy et al. [2014] discovered that the input-output mappings learned by state-of-theart neural networks are significantly discontinuous. It is possible to cause a neural network used for
image recognition to misclassify its input by applying a very specific, hardly perceptible perturbation
to the input. Szegedy et al. [2014] call these perturbations adversarial perturbations, and the inputs
resulting from applying them to natural samples are called adversarial examples.
In this paper, we hope to shed more light on the nature and cause of adversarial examples by
deriving lower bounds on the magnitudes of perturbations necessary to change the classification of
neural network classifiers. Such lower bounds are indispensable for developing rigorous methods
that increase the robustness of classifiers without sacrificing accuracy. Since the bounds enjoy the
theoretical guarantee that no adversarial perturbation could ever be any smaller, a method which
increases these lower bounds potentially makes the classifier more robust. They may also aid model
selection: if the bounds can be computed efficiently, then one can use them to compare different
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
models with respect to their robustness to adversarial perturbations and select the model that scores
the highest in this regard without the need for extensive empirical tests.
The rest of the paper is organized as follows. Section 2 discusses related work that has been done
on the phenomenon of adversarial perturbations; Section 3 details the theoretical framework used
to prove the lower bounds; Section 4 proves lower bounds on the robustness of different families of
classifiers to adversarial perturbations; Section 5 empirically verifies that the bounds are not violated;
Section 6 concludes the paper and provides avenues for future work.
2
Related work
Since the puzzling discovery of adversarial perturbations, several hypotheses have been proposed to
explain why they exist, as well as a number of methods to make classifiers more robust to them.
2.1
Hypotheses
The leading hypothesis explaining the cause of adversarial perturbations is the linearity hypothesis by
Goodfellow et al. [2015]. According this view, neural network classifiers tend to act very linearly on
their input data despite the presence of non-linear transformations within their layers. Since the input
data on which modern classifiers operate is often very high in dimensionality, such linear behavior
can cause minute perturbations to the input to have a large impact on the output. In this vein, Lou
et al. [2016] propose a variant of the linearity hypothesis which claims that neural network classifiers
operate highly linearly on certain regions of their inputs, but non-linearly in other regions. Rozsa et al.
[2016] conjecture that adversarial examples exist because of evolutionary stalling: during training,
the gradients of samples that are classified correctly diminish, so the learning algorithm ?stalls? and
does not create significantly flat regions around the training samples. As such, most of the training
samples will lie close to some decision boundary, and only a small perturbation is required to push
them into a different class.
2.2
Proposed solutions
Gu and Rigazio [2014] propose the Deep Contractive Network, which includes a smoothness penalty
in the training procedure inspired by the Contractive Autoencoder. This penalty encourages the
Jacobian of the network to have small components, thus making the network robust to small changes
in the input. Based on their linearity hypothesis, Goodfellow et al. [2015] propose the fast gradient
sign method for efficiently generating adversarial examples. They then use this method as a regularizer
during training in an attempt to make networks more robust. Lou et al. [2016] use their ?local linearity
hypothesis? as the basis for training neural network classifiers using foveations, i.e. a transformation
which selects certain regions from the input and discards all other information. Rozsa et al. [2016]
introduce Batch-Adjusted Network Gradients (BANG) based on their idea of evolutionary stalling.
BANG normalizes the gradients on a per-minibatch basis so that even correctly classified samples
retain significant gradients and the learning algorithm does not stall.
The solutions proposed above provide attractive intuitive explanations for the cause of adversarial
examples, and empirical results seem to suggest that they are effective at eliminating them. However,
none of the hypotheses on which these methods are based have been formally proven. Hence, even
with the protections discussed above, it may still be possible to generate adversarial examples for
classifiers using techniques which defy the proposed hypotheses. As such, there is a need to formally
characterize the nature of adversarial examples. Fawzi et al. [2016] take a step in this direction by
deriving precise bounds on the norms of adversarial perturbations of arbitrary classifiers in terms of
the curvature of the decision boundary. Their analysis encourages to impose geometric constraints
on this curvature in order to improve robustness. However, it is not obvious how such constraints
relate to the parameters of the models and hence how one would best implement such constraints
in practice. In this work, we derive lower bounds on the robustness of neural networks directly in
terms of their model parameters. We consider only feedforward networks comprised of convolutional
layers, pooling layers, fully-connected layers and softmax layers.
2
3
Theoretical framework
The theoretical framework used in this paper draws heavily from Fawzi et al. [2016] and Papernot
et al. [2016]. In the following, k?k denotes the Euclidean norm and k?kF denotes the Frobenius norm.
We assume we want to train a classifier f : Rd ? {1, . . . , C} to correctly assign one of C different
classes to input vectors x from a d-dimensional Euclidean space. Let ? denote the probability measure
on Rd and let f ? be an oracle that always returns the correct label for any input. The distribution ?
is assumed to be of bounded support, i.e. Px?? (x ? X ) = 1 with X = {x ? Rd | kxk ? M } for
some M > 0.
Formally, adversarial perturbations are defined relative to a classifier f and an input x. A perturbation
? is called an adversarial perturbation of x for f if f (x + ?) 6= f (x) while f ? (x + ?) = f ? (x).
An adversarial perturbation ? is called minimal if no other adversarial perturbation ? for x and f
satisfies k?k < k?k. In this work, we will focus on minimal adversarial perturbations.
The robustness of a classifier f is defined as the expected norm of the smallest perturbation necessary
to change the classification of an arbitrary input x sampled from ?:
?adv (f ) = Ex?? [?adv (x; f )],
where
?adv (x; f ) = min {k?k | f (x + ?) 6= f (x)}.
??Rd
A multi-index is a tuple of non-negative integers, generally denoted by Greek letters such as ? and ?.
For a multi-index ? = (?1 , . . . , ?m ) and a function f we define
??f =
|?| = ?1 + ? ? ? + ?n ,
? |?| f
.
n
. . . ?x?
n
1
?x?
1
The Jacobian matrix of a function f : Rn ? Rm : x 7? [f1 (x), . . . , fm (x)]T is defined as
? ?f1
?f1 ?
. . . ?x
?x1
n
?
?
?
..
..
f = ? ...
?.
.
.
?x
?fm
?fm
. . . ?xn
?x1
3.1
Families of classifiers
The derivation of the lower bounds will be built up incrementally. We will start with the family
of linear classifiers, which are among the simplest. Then, we extend the analysis to Multi-Layer
Perceptrons, which are the oldest neural network architectures. Finally, we analyze Convolutional
Neural Networks. In this section, we introduce each of these families of classifiers in turn.
A linear classifier is a classifier f of the form
f (x) = arg max wi ? x + bi .
i=1,...,C
The vectors wi are called weights and the scalars bi are called biases.
A Multi-Layer Perceptron (MLP) is a classifier given by
f (x) = arg max softmax(hL (x))i ,
i=1,...,C
hL (x) = gL (VL hL?1 (x) + bL ),
..
.
h1 (x) = g1 (V1 x + b1 ).
An MLP is nothing more than a series of linear transformations Vl hl?1 (x) + bl followed by nonlinear activation functions gl (e.g. a ReLU [Glorot et al., 2011]). Here, softmax is the softmax
function:
exp(wi ? y + bi )
softmax(y)i = P
.
j exp(wj ? y + bj )
3
This function is a popular choice as the final layer for an MLP used for classification, but it is by no
means the only possibility. Note that having a softmax as the final layer essentially turns the network
into a linear classifier of the output of its penultimate layer, hL (x).
A Convolutional Neural Network (CNN) is a neural network that uses at least one convolution
operation. For an input tensor X ? Rc?d?d and a kernel tensor W ? Rk?c?q?q , the discrete
convolution of X and W is given by
q X
q
c X
X
(X ? W)ijk =
wi,n,m,l xn,m+s(q?1),l+s(q?1) .
n=1 m=1 l=1
Here, s is the stride of the convolution. The output of such a layer is a 3D tensor of size k ? t ? t
k
where t = d?q
s + 1. After the convolution operation, usually a bias b ? R is added to each of the
feature maps. The different components (W ? X)i constitute the feature maps of this convolutional
layer. In a slight abuse of notation, we will write W ? X + b to signify the tensor W ? X where each
of the k feature maps has its respective bias added in:
(W ? X + b)ijk = (W ? X)ijk + bi .
CNNs also often employ pooling layers, which perform a sort of dimensionality reduction. If we
write the output of a pooling layer as Z(X), then we have
zijk (X) = p({xi,n+s(j?1),m+s(k?1) | 1 ? n, m ? q}).
Here, p is the pooling operation, s is the stride and q is a parameter. The output tensor Z(X) has
dimensions c ? t ? t. For ease of notation, we assume each pooling operation has an associated
function I such that
zijk (X) = p({xinm | (n, m) ? I(j, k)}).
In the literature, the set I(j, k) is referred to as the receptive field of the pooling layer. Each receptive
field corresponds to some q ? q region in the input X. Common pooling operations include taking
the maximum of all inputs, averaging the inputs and taking an Lp norm of the inputs.
4
Lower bounds on classifier robustness
Comparing the architectures of several practical CNNs such as LeNet [Lecun et al., 1998], AlexNet
[Krizhevsky et al., 2012], VGGNet [Simonyan and Zisserman, 2015], GoogLeNet [Szegedy et al.,
2015] and ResNet [He et al., 2016], it would seem the only useful approach is a ?modular? one. If we
succeed in lower-bounding the robustness of some layer given the robustness of the next layer, we can
work our way backwards through the network, starting at the output layer and going backwards until
we reach the input layer. That way, our approach can be applied to any feedforward neural network
as long as the robustness bounds of the different layer types have been established. To be precise, if a
given layer computes a function h of its input y and if the following layer has a robustness bound of
? in the sense that any adversarial perturbation to this layer has a Euclidean norm of at least ?, then
we want to find a perturbation r such that
kh(y + r)k = kh(y)k + ?.
This is clearly a necessary condition for any adversarial perturbation to the given layer. Hence, any
adversarial perturbation q to this layer will satisfy kqk ? krk. Of course, the output layer of the
network will require special treatment. For softmax output layers, ? is the norm of the smallest
perturbation necessary to change the maximal component of the classification vector.
The obvious downside of this idea is that we most likely introduce cumulative approximation errors
which increase as the number of layers of the network increases. In turn, however, we get a flexible
and efficient framework which can handle any feedforward architecture composed of known layer
types.
4.1
Softmax output layers
We now want to find the smallest perturbation r to the input x of a softmax layer such that f (x+r) 6=
f (x). It can be proven (Theorem A.3) that any such perturbation satisfies
|(wc0 ? wc ) ? x + bc0 ? bc |
krk ? min
,
c0 6=c
kwc0 ? wc k
where f (x) = c. Moreover, there exist classifiers for which this bound is tight (Theorem A.4).
4
4.2
Fully-connected layers
To analyze the robustness of fully-connected layers to adversarial perturbations, we assume the next
layer has a robustness of ? (this will usually be the softmax output layer, however there exist CNNs
which employ fully-connected layers in other locations than just at the end [Lin et al., 2014]). We
then want to find a perturbation r such that
khL (x + r)k = khL (x)k + ?.
We find
Theorem 4.1. Let hL : Rd ? Rn be twice differentiable with second-order derivatives bounded by
M . Then for any x ? Rd ,
q
?
2
kJ (x)k + 2M n? ? kJ (x)k
?
krk ?
,
(1)
M n
where J (x) is the Jacobian matrix of hL at x.
The proof can be found in Appendix A. In Theorem A.5 it is proved that the assumptions on hL
are usually satisfied in practice. The proof of this theorem also yields an efficient algorithm for
approximating M , a task which otherwise might involve a prohibitively expensive optimization
problem.
4.3
Convolutional layers
The next layer of the network is assumed to have a robustness bound of ?, in the sense that any
adversarial perturbation Q to X must satisfy kQkF ? ?. We can now attempt to bound the norm of a
perturbation R to X such that
kReLU(W ? (X + R) + b)kF = kReLU(W ? X + b)kF + ?.
We find
Theorem 4.2. Consider a convolutional layer with filter tensor W ? Rk?c?q?q and stride s whose
input consists of a 3D tensor X ? Rc?d?d . Suppose the next layer has a robustness bound of ?, then
any adversarial perturbation to the input of this layer must satisfy
?
kRkF ?
.
(2)
kWkF
The proof of Theorem 4.2 can be found in Appendix A.
4.4
Pooling layers
To facilitate the analysis of the pooling layers, we make the following assumption which is satisfied
by the most common pooling operations (see Appendix B):
Assumption 4.3. The pooling operation satisfies
zijk (X + R) ? zijk (X) + zijk (R).
We have
Theorem 4.4. Consider a pooling layer whose operation satisfies Assumption 4.3. Let the input be of
size c ? d ? d and the receptive field of size q ? q. Let the output be of size c ? t ? t. If the robustness
bound of the next layer is ?, then the following bounds hold for any adversarial perturbation R:
? MAX or average pooling:
?
.
t
(3)
?
.
tq 2/p
(4)
kRkF ?
? Lp pooling:
kRkF ?
Proof can be found in Appendix A.
5
Figure 1: Illustration of LeNet architecture. Image taken from Lecun et al. [1998].
Table 1: Normalized summary of norms of adversarial perturbations found by FGS on MNIST and
CIFAR-10 test sets
Data set
Mean
Median
Std
Min
Max
MNIST
CIFAR-10
5
0.933448
0.0218984
0.884287
0.0091399
0.4655439
0.06103627
0.000023
0.0000012
3.306903
1.6975207
Experimental results
We tested the theoretical bounds on the MNIST and CIFAR-10 test sets using the Caffe [Jia et al.,
2014] implementation of LeNet [Lecun et al., 1998]. The MNIST data set [LeCun et al., 1998]
consists of 70,000 28 ? 28 images of handwritten digits; the CIFAR-10 data set [Krizhevsky and
Hinton, 2009] consists of 60,000 32 ? 32 RGB images of various natural scenes, each belonging to
one of ten possible classes. The architecture of LeNet is depicted in Figure 1. The kernels of the two
convolutional layers will be written as W1 and W2 , respectively. The output sizes of the two pooling
layers will be written as t1 and t2 . The function computed by the first fully-connected layer will be
denoted by h with Jacobian J . The last fully-connected layer has a weight matrix V and bias vector
b. For an input sample x, the theoretical lower bound on the adversarial robustness of the network
with respect to x is given by ?1 , where
q
?
2
kJ (x)k + 2M 500?6 ? kJ (x)k
0
0
|(vc ? vc ) ? x + bc ? bc |
?
,
?5 =
?6 = min
,
c0 6=c
kvc0 ? vc k
M 500
?5
?4
,
?4 =
,
?3 =
t2
kW2 kF
?3
?2
?2 =
.
,
?1 =
t1
kW1 kF
Because our method only computes norms and does not provide a way to generate actual adversarial
perturbations, we used the fast gradient sign method (FGS) [Goodfellow et al., 2015] to adversarially
perturb each sample in the test sets in order to assess the tightness of our theoretical bounds. FGS
linearizes the cost function of the network to obtain an estimated perturbation
? = ?sign?x L(x, ?).
Here, ? > 0 is a parameter of the algorithm, L is the loss function and ? is the set of parameters of
the network. The magnitudes of the perturbations found by FGS depend on the choice of ?, so we
had to minimize this value in order to obtain the smallest perturbations the FGS method could supply.
This was accomplished using a simple binary search for the smallest value of ? which still resulted in
misclassification. As the MNIST and CIFAR-10 samples have pixel values within the range [0, 255],
we upper-bounded ? by 100.
No violations of the bounds were detected in our experiments. Figure 2 shows histograms of the
norms of adversarial perturbations found by FGS and Table 1 summarizes their statistics. Histograms
of the theoretical bounds of all samples in the test set are shown in Figure 3; their statistics are
summarized in Table 2. Note that the statistics of Tables 1 and 2 have been normalized by dividing
them by the dimensionality of their respective data sets (i.e. 28 ? 28 for MNIST and 3 ? 32 ? 32
for CIFAR-10) to allow for a meaningful comparison between the two networks. Figure 4 provides
histograms of the per-sample log-ratio between the norms of the adversarial perturbations and their
corresponding theoretical lower bounds.
6
(a) MNIST
(b) CIFAR-10
Figure 2: Histograms of norms of adversarial perturbations found by FGS on MNIST and CIFAR-10
test sets
(a) MNIST
(b) CIFAR-10
Figure 3: Histograms of theoretical bounds on MNIST and CIFAR-10 test sets
Although the theoretical bounds on average deviate considerably from the perturbations found by
FGS, one has to take into consideration that the theoretical bounds were constructed to provide a
worst-case estimate for the norms of adversarial perturbations. These estimates may not hold for
all (or even most) input samples. Furthermore, the smallest perturbations we were able to generate
on the two data sets have norms that are much closer to the theoretical bound than their averages
(0.0179 for MNIST and 0.0000012 for CIFAR-10). This indicates that the theoretical bound is
not necessarily very loose, but rather that very small adversarial perturbations occur with non-zero
probability on natural samples. Note also that the FGS method does not necessarily generate minimal
perturbations even with the smallest choice of ?: the method depends on the linearity hypothesis and
uses a first-order Taylor approximation of the loss function. Higher-order methods may find much
smaller perturbations by exploiting non-linearities in the network, but these are generally much less
efficient than FGS.
There is a striking difference in magnitude between MNIST and CIFAR-10 of both the empirical
and theoretical perturbations: the perturbations on MNIST are much larger than the ones found for
Table 2: Normalized summary of theoretical bounds on MNIST and CIFAR-10 test sets
Data set
Mean
Median
Std
Min
Max
MNIST
CIFAR-10
7.274e?8
4.812e?13
6.547e?8
4.445e?13
4.229566e?8
2.605381e?13
7
4.073e?10
7.563e?15
2.932e?7
2.098e?12
(a) MNIST
(b) CIFAR-10
Figure 4: Histograms of the per-sample log-ratio between adversarial perturbation and lower bound
for MNIST and CIFAR-10 test sets. A higher ratio indicates a bigger deviation of the theoretical
bound from the empirical norm.
CIFAR-10. This result can be explained by the linearity hypothesis of Goodfellow et al. [2015].
The input samples of CIFAR-10 are much larger in dimensionality than MNIST samples, so the
linearity hypothesis correctly predicts that networks trained on CIFAR-10 are more susceptible to
adversarial perturbations due to the highly linear behavior these classifiers are conjectured to exhibit.
However, these differences may also be related to the fact that LeNet achieves much lower accuracy
on the CIFAR-10 data set than it does on MNIST (over 99% on MNIST compared to about 60% on
CIFAR-10).
6
Conclusion and future work
Despite attracting a significant amount of research interest, a precise characterization of adversarial
examples remains elusive. In this paper, we derived lower bounds on the norms of adversarial
perturbations in terms of the model parameters of feedforward neural network classifiers consisting
of convolutional layers, pooling layers, fully-connected layers and softmax layers. The bounds can be
computed efficiently and thus may serve as an aid in model selection or the development of methods
to increase the robustness of classifiers. They enable one to assess the robustness of a classifier
without running extensive tests, so they can be used to compare different models and quickly select
the one with highest robustness. Furthermore, the bounds enjoy a theoretical guarantee that no
adversarial perturbation could ever be smaller, so methods which increase these bounds may make
classifiers more robust. We tested the validity of our bounds on MNIST and CIFAR-10 and found no
violations. Comparisons with adversarial perturbations generated using the fast gradient sign method
suggest that these bounds can be close to the actual norms in the worst case.
We have only derived lower bounds for feedforward networks consisting of fully-connected layers,
convolutional layers and pooling layers. Extending this analysis to recurrent networks and other types
of layers such as Batch Normalization [Ioffe and Szegedy, 2015] and Local Response Normalization
[Krizhevsky et al., 2012] is an obvious avenue for future work.
It would also be interesting to quantify just how tight the above bounds really are. In the absence
of a precise characterization of adversarial examples, the only way to do this would be to generate
adversarial perturbations using optimization techniques that make no assumptions on their underlying
cause. Szegedy et al. [2014] use a box-constrained L-BFGS approach to generate adversarial examples
without any assumptions, so using this method for comparison could provide a more accurate picture
of how tight the theoretical bounds are. It is much less efficient than the FGS method, however.
The analysis presented here is a ?modular? one: we consider each layer in isolation, and derive bounds
on their robustness in terms of the robustness of the next layer. However, it may also be insightful to
study the relationship between the number of layers, the breadth of each layer and the robustness of
the network. Providing estimates on the approximation errors incurred by this layer-wise approach
could also be useful.
8
Finally, there is currently no known precise characterization of the trade-off between classifier robustness and accuracy. Intuitively, one might expect that as the robustness of the classifier increases, its
accuracy will also increase up to a point since it is becoming more robust to adversarial perturbations.
Once the robustness exceeds a certain threshold, however, we expect the accuracy to drop because the
decision surfaces are becoming too flat and the classifier becomes too insensitive to changes. Having
a precise characterization of this relationship between robustness and accuracy may aid methods
designed to protect classifiers against adversarial examples while also maintaining state-of-the-art
accuracy.
References
A. Fawzi, S.-M. Moosavi-Dezfooli, and P. Frossard. Robustness of classifiers: from adversarial to
random noise. In Proceedings of Advances in Neural Information Processing Systems 29, pages
1632?1640. Curran Associates, Inc., 2016.
X. Glorot, A. Bordes, and Y. Bengio. Deep sparse rectifier neural networks. In Proceedings of
the Fourteenth International Conference on Artificial Intelligence and Statistics, volume 15 of
Proceedings of Machine Learning Research, pages 315?323, Fort Lauderdale, FL, USA, 11?13
Apr 2011. PMLR.
I. Goodfellow, J. Shlens, and C. Szegedy. Explaining and harnessing adversarial examples. In
Proceedings of the Third International Conference on Learning Representations, volume 3 of
Proceedings of the International Conference on Learning Representations, San Diego, CA, USA,
7?9 May 2015. ICLR.
S. Gu and L. Rigazio. Towards deep neural network architectures robust to adversarial examples.
NIPS Workshop on Deep Learning and Representation Learning, 2014.
K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. In Proceedings
of the IEEE Conference on Computer Vision and Pattern Recognition, pages 770?778, Las Vegas,
NV, USA, 26 Jun ? 1 Jul 2016. CVPR.
S. Ioffe and C. Szegedy. Batch normalization: Accelerating deep network training by reducing
internal covariate shift. In Proceedings of the 32nd International Conference on Machine Learning,
volume 37 of Proceedings of the International Conference on Machine Learning, pages 448?456,
Lille, 6?11 Jul 2015. JMLR.
Y. Jia, E. Shelhamer, J. Donahue, S. Karayev, J. Long, R. Girshick, S. Guadarrama, and T. Darrell.
Caffe: Convolutional architecture for fast feature embedding. In Proceedings of the 22nd ACM
International Conference on Multimedia, pages 675?678. ACM, 2014.
A. Krizhevsky and G. Hinton. Learning multiple layers of features from tiny images. Technical
report, University of Toronto, 2009.
A. Krizhevsky, I. Sutskever, and G. E. Hinton. Imagenet classification with deep convolutional neural
networks. In Proceedings of the 25th International Conference on Advances in Neural Information
Processing Systems, volume 25 of Advances in Neural Information Processing Systems, pages
1097?1105, Lake Tahoe, USA, 3?8 Dec 2012. NIPS.
Y. Lecun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-based learning applied to document
recognition. Proceedings of the IEEE, 86(11):2278?2324, Nov 1998.
Y. LeCun, C. Cortes, and C. J. Burges. The MNIST database of handwritten digits. http://yann.
lecun.com/exdb/mnist/, 1998. Accessed 2017-04-17.
M. Lin, Q. Chen, and S. Yan. Network in network. Proceedings of International Conference on
Learning Representations, 2014.
Y. Lou, X. Boix, G. Roig, T. Poggio, and Q. Zhao. Foveation-based mechanisms alleviate adversarial
examples. arXiv preprint arXiv:1511.06292, 2016.
N. Papernot, P. McDaniel, X. Wu, S. Jha, and A. Swami. Distillation as a defense to adversarial
perturbations against deep neural networks. In 2016 IEEE Symposium on Security and Privacy
(SP), pages 582?597, May 2016.
9
A. Rozsa, M. Gunther, and T. E. Boult. Towards robust deep neural networks with BANG. arXiv
preprint arXiv:1612.00138, 2016.
K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition.
In Proceedings of the Third International Conference on Learning Representations, volume 3 of
Proceedings of the International Conference on Learning Representations, San Diego, CA, USA,
7?9 May 2015. ICLR.
C. Szegedy, W. Zaremba, and I. Sutskever. Intriguing properties of neural networks. In Proceedings
of the Second International Conference on Learning Representations, volume 2 of Proceedings
of the International Conference on Learning Representations, Banff, Canada, 14?16 Apr 2014.
ICLR.
C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and
A. Rabinovich. Going deeper with convolutions. In Proceedings of the IEEE Conference on
Computer Vision and Pattern Recognition, Boston, MA, USA, 7-12 Jun 2015. CVPR.
10
| 6682 |@word moosavi:1 cnn:1 eliminating:1 norm:18 nd:2 c0:2 rgb:1 reduction:1 liu:1 series:1 score:1 bc:3 document:1 guadarrama:1 comparing:1 com:1 protection:1 activation:1 intriguing:1 must:2 written:2 drop:1 designed:1 bart:1 intelligence:1 oldest:1 krkf:3 characterization:5 provides:2 location:1 toronto:1 tahoe:1 banff:1 zhang:1 accessed:1 rc:2 constructed:1 supply:1 symposium:1 prove:1 consists:3 privacy:1 introduce:3 expected:1 frossard:1 behavior:2 multi:4 inspired:1 actual:2 becomes:1 provided:1 linearity:8 bounded:3 notation:2 moreover:1 alexnet:1 underlying:1 transformation:3 guarantee:3 mitigate:1 act:1 shed:1 zaremba:1 prohibitively:1 classifier:35 rm:1 enjoy:3 t1:2 local:2 despite:3 becoming:2 abuse:1 black:1 might:2 twice:1 ease:1 contractive:2 bi:4 kw2:1 range:1 practical:1 lecun:7 practice:2 implement:1 digit:2 procedure:1 empirical:4 yan:1 significantly:3 suggest:3 get:1 close:2 selection:3 applying:3 map:3 center:1 elusive:2 starting:1 deriving:3 shlens:1 embedding:1 handle:1 roig:1 diego:2 suppose:1 heavily:1 us:2 curran:1 hypothesis:13 goodfellow:5 associate:1 recognition:7 expensive:1 std:2 predicts:1 database:1 vein:1 preprint:2 worst:2 region:5 wj:1 connected:8 adv:3 sun:1 counter:1 highest:2 trade:1 trained:1 depend:1 tight:3 swami:1 serve:1 basis:3 gu:2 various:2 regularizer:1 derivation:1 train:1 fast:4 effective:1 krelu:2 detected:1 artificial:1 caffe:2 harnessing:1 whose:2 modular:2 larger:2 cvpr:2 tightness:1 otherwise:1 statistic:5 simonyan:2 g1:1 final:2 karayev:1 differentiable:1 propose:3 maximal:1 dezfooli:1 intuitive:2 frobenius:1 kh:2 exploiting:1 sutskever:2 darrell:1 extending:1 generating:1 resnet:1 derive:2 recurrent:1 dividing:1 quantify:1 direction:1 greek:1 discontinuous:2 correct:1 cnns:3 filter:1 vc:3 enable:1 require:1 assign:1 f1:3 really:1 alleviate:1 adjusted:1 hold:2 around:1 diminish:1 exp:2 mapping:2 bj:1 claim:1 achieves:1 smallest:7 belgium:3 label:1 currently:1 create:1 hope:1 clearly:1 always:1 rather:1 derived:2 focus:1 indicates:2 adversarial:57 rigorous:1 sense:2 vl:2 stalling:2 going:2 selects:1 pixel:1 arg:2 classification:6 among:1 flexible:1 denoted:2 development:1 art:2 softmax:11 special:1 constrained:1 field:3 once:1 having:2 beach:1 adversarially:1 lille:1 theart:1 future:3 t2:2 report:1 employ:2 modern:1 composed:1 resulted:1 consisting:2 tq:1 attempt:2 misclassify:2 mlp:3 interest:1 mining:1 highly:2 possibility:1 violation:3 light:1 accurate:1 peculiar:1 tuple:1 closer:1 necessary:5 poggio:1 respective:2 euclidean:3 taylor:1 sacrificing:1 girshick:1 theoretical:20 fawzi:3 minimal:3 modeling:1 downside:1 inflammation:1 rabinovich:1 cost:1 deviation:1 comprised:1 krizhevsky:5 too:2 characterize:1 considerably:1 st:1 fundamental:1 wc0:1 international:12 retain:1 off:1 lauderdale:1 quickly:1 w1:1 satisfied:2 possibly:1 khl:2 derivative:1 leading:1 return:1 zhao:1 szegedy:9 bfgs:1 stride:3 summarized:1 includes:1 jha:1 inc:1 satisfy:3 depends:1 view:1 h1:1 analyze:2 start:1 sort:1 jul:2 jia:3 ass:2 minimize:1 accuracy:7 convolutional:12 efficiently:4 yield:1 identification:1 handwritten:2 basically:1 none:1 ren:1 biomedicine:1 classified:2 explain:2 reach:1 papernot:2 against:2 obvious:3 associated:1 proof:4 sampled:1 proved:1 treatment:1 popular:1 dimensionality:4 organized:1 higher:2 zisserman:2 response:1 done:1 box:2 furthermore:2 just:2 until:1 nonlinear:1 incrementally:1 minibatch:1 zijk:5 usa:7 facilitate:1 validity:1 requiring:1 verify:1 normalized:3 hence:3 lenet:5 attractive:1 during:2 encourages:2 exdb:1 image:8 wise:1 consideration:1 vega:1 common:2 empirically:1 insensitive:1 volume:6 discussed:1 extend:1 slight:1 googlenet:1 he:2 significant:2 distillation:1 anguelov:1 ai:1 smoothness:1 rd:6 mathematics:1 had:1 kw1:1 surface:1 attracting:1 curvature:2 fgs:11 conjectured:1 discard:1 indispensable:1 certain:3 binary:1 success:1 accomplished:1 impose:1 multiple:1 exceeds:1 technical:1 long:3 cifar:23 lin:2 bigger:1 impact:1 variant:1 essentially:1 vision:2 arxiv:4 histogram:6 kernel:2 normalization:3 dec:1 want:4 signify:1 median:2 w2:1 rest:1 operate:2 rigazio:2 pooling:17 tend:1 nv:1 kwkf:1 seem:2 call:1 integer:1 linearizes:1 presence:1 backwards:2 feedforward:5 bengio:2 relu:1 isolation:1 architecture:7 fm:3 stall:2 idea:2 avenue:2 haffner:1 shift:1 defense:1 accelerating:1 penalty:2 cause:7 hardly:2 constitute:1 deep:10 generally:2 useful:2 clear:1 involve:1 kqkf:1 amount:1 ten:1 mcdaniel:1 simplest:1 generate:6 http:1 outperform:1 exist:4 sign:4 estimated:1 correctly:4 per:3 discrete:1 write:2 threshold:1 gunther:1 breadth:1 kqk:1 v1:1 fourteenth:1 letter:1 telecommunication:1 striking:1 family:4 yann:1 wu:1 lake:1 draw:1 decision:3 appendix:4 summarizes:1 bound:49 layer:63 fl:1 followed:1 oracle:1 occur:2 constraint:3 scene:1 flat:2 wc:2 min:5 px:1 conjecture:1 department:2 developing:2 according:1 belonging:1 smaller:4 wi:4 lp:2 perceptible:2 making:1 hl:8 explained:1 intuitively:1 taken:1 remains:2 discus:1 turn:3 loose:1 mechanism:1 end:1 operation:8 pmlr:1 batch:3 robustness:30 existence:1 denotes:2 running:1 include:1 maintaining:1 joris:1 perturb:1 prof:1 approximating:1 classical:1 bl:2 tensor:7 added:2 quantity:1 receptive:3 evolutionary:2 gradient:8 exhibit:1 iclr:3 lou:3 penultimate:1 index:2 relationship:2 illustration:1 vib:1 ratio:3 providing:1 sermanet:1 reed:1 susceptible:1 potentially:1 relate:1 negative:1 implementation:1 perform:1 upper:1 convolution:5 hinton:3 ever:2 precise:6 discovered:1 perturbation:65 rn:2 arbitrary:2 canada:1 fort:1 required:1 extensive:2 imagenet:1 security:1 learned:2 established:1 protect:1 nip:3 able:2 usually:3 pattern:2 built:1 max:5 explanation:3 suitable:1 misclassification:1 natural:4 residual:1 improve:1 picture:1 vggnet:1 concludes:1 jun:2 autoencoder:1 kj:4 deviate:1 geometric:1 discovery:1 literature:1 kf:5 relative:1 fully:8 loss:2 expect:2 boult:1 interesting:1 proven:3 shelhamer:1 incurred:1 vanhoucke:1 tiny:1 bordes:1 normalizes:1 course:1 summary:2 gl:2 last:1 formal:1 bias:4 allow:1 perceptron:1 burges:1 explaining:2 deeper:1 taking:2 sparse:1 regard:1 boundary:2 dimension:1 xn:2 cumulative:1 computes:2 san:2 erhan:1 nov:1 ioffe:2 b1:1 assumed:2 xi:1 search:1 why:1 table:5 additionally:1 nature:2 robust:9 ca:3 defy:1 bottou:1 necessarily:2 krk:3 sp:1 apr:2 linearly:3 big:1 bounding:1 hyperparameters:1 noise:1 nothing:1 verifies:1 x1:2 referred:1 boix:1 aid:3 wish:1 lie:1 jmlr:1 jacobian:4 third:2 donahue:1 minute:1 rk:2 theorem:8 specific:2 rectifier:1 covariate:1 insightful:1 cortes:1 glorot:2 workshop:1 mnist:24 magnitude:4 push:1 chen:1 boston:1 depicted:1 led:1 likely:1 unexpected:1 kxk:1 scalar:1 corresponds:1 satisfies:4 acm:2 ma:1 succeed:1 bang:3 towards:3 absence:1 change:6 experimentally:1 foveation:1 reducing:1 averaging:1 ghent:5 called:6 multimedia:1 experimental:2 la:1 ijk:3 meaningful:1 perceptrons:1 select:2 formally:3 puzzling:1 internal:1 support:1 jonathan:1 violated:1 tested:2 phenomenon:1 ex:1 |
6,282 | 6,683 | Minimizing a Submodular Function from Samples
Eric Balkanski
Harvard University
[email protected]
Yaron Singer
Harvard University
[email protected]
Abstract
In this paper we consider the problem of minimizing a submodular function from
training data. Submodular functions can be efficiently minimized and are consequently heavily applied in machine learning. There are many cases, however, in
which we do not know the function we aim to optimize, but rather have access
to training data that is used to learn it. In this paper we consider the question of
whether submodular functions can be minimized when given access to its training
data. We show that even learnable submodular functions cannot be minimized
within any non-trivial approximation when given access to polynomially-many samples. Specifically, we show that there is a class of submodular functions with range
in [0, 1] such that, despite being PAC-learnable and minimizable in polynomial-time,
no algorithm can obtain an approximation strictly better than 1/2 o(1) using
polynomially-many samples drawn from any distribution. Furthermore, we show
that this bound is tight via a trivial algorithm that obtains an approximation of 1/2.
1
Introduction
For well over a decade now, submodular minimization has been heavily studied in machine learning
(e.g. [SK10, JB11, JLB11, NB12, EN15, DTK16]). This focus can be largely attributed to the fact
that if a set function f : 2N ! R is submodular, meaning it has the following property of diminishing
returns: f (S [ {a}) f (S)
f (T [ {a}) f (T ) for all S ? T ? N and a 62 T , then it can
be optimized efficiently: its minimizer can be found in time that is polynomial in the size of the
ground set N [GLS81, IFF01]. In many cases, however, we do not know the submodular function,
and instead learn it from data (e.g. [BH11, IJB13, FKV13, FK14, Bal15, BVW16]). The question
we address in this paper is whether submodular functions can be (approximately) minimized when
the function is not known but can be learned from training data.
An intuitive approach for optimization from training data is to learn a surrogate function from training
data that predicts the behavior of the submodular function well, and then find the minimizer of the
surrogate learned and use that as a proxy for the true minimizer we seek. The problem however, is that
this approach does not generally guarantee that the resulting solution is close to the true minimum of
the function. One pitfall is that the surrogate may be non-submodular, and despite approximating the
true submodular function arbitrarily well, the surrogate can be intractable to minimize. Alternatively,
it may be that the surrogate is submodular, but its minimum is arbitrarily far from the minimum of
the true function we aim to optimize (see examples in Appendix A).
Since optimizing a surrogate function learned from data may generally result in poor approximations,
one may seek learning algorithms that are guaranteed to produce surrogates whose optima wellapproximate the true optima and are tractable to compute. More generally, however, it is possible that
there is some other approach for optimizing the function from the training samples, without learning
a model. Therefore, at a high level, the question is whether a reasonable number of training samples
suffices to minimize a submodular function. We can formalize this as optimization from samples.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
Optimization from samples. We will say that a class of functions F = {f : 2N ! [0, 1]} is
?-optimizable from samples over distribution D if for every f 2 F and 2 (0, 1), when given
poly(|N |) i.i.d. samples {(Si , f (Si ))}m
over the
i=1 where Si ? D, with probability at least 1
samples one can construct an algorithm that returns a solution S ? N s.t,
f (S)
min f (T ) ? ?.
T ?N
This framework was recently introduced in [BRS17] for the problem of submodular maximization
where the standard notion of approximation is multiplicative. For submodular minimization, since
the optimum may have zero value, the suitable measure is that of additive approximations for [0, 1]bounded functions, and the goal is to obtain a solution which is a o(1) additive approximation to the
minimum (see e.g. [CLSW16, EN15, SK10]). The question is then:
Can submodular functions be minimized from samples?
Since submodular functions can be minimized in polynomial-time, it is tempting to conjecture that
when the function is learnable it also has desirable approximation guarantees from samples, especially
in light of positive results in related settings of submodular maximization:
? Constrained maximization. For functions that can be maximized in polynomial time under
a cardinality constraint, like modular and unit-demand functions, there are polynomial
time algorithms that obtain an arbitrarily good approximation using polynomially-many
samples [BRS16, BRS17]. For general monotone submodular functions which are NPhard to maximize under cardinality constraints, there is no algorithm that can obtain a
reasonable approximation from polynomially-many samples [BRS17]. For the problem of
unconstrained minimization, submodular functions can be optimized in polynomial time;
? Unconstrained maximization. For unconstrained maximization of general submodular
functions, the problem is NP-hard to maximize (e.g. MAX-CUT) and one seeks constant
factor approximations. For this problem, there is an extremely simple algorithm that uses
no queries and obtains a good approximation: choose elements uniformly at random with
probability 1/2 each. This algorithm achieves a constant factor approximation of 1/4 for
general submodular functions. For symmetric submodular functions (i.e. f (S) = f (N \ S)),
this algorithm is a 1/2-approximation which is optimal, since no algorithm can obtain an
approximation ratio strictly better than 1/2 using polynomially-many value queries, even
for symmetric submodular functions [FMV11]. For unconstrained symmetric submodular
minimization, there is an appealing analogue: the empty set and the ground set N are
guaranteed to be minimizers of the function (see Section 2). This algorithm, of course, uses
no queries either. The parallel between these two problems seems quite intuitive, and it
is tempting to conjecture that like for unconstrained submodular maximization, there are
optimization from samples algorithms for general unconstrained submodular minimization
with good approximation guarantees.
Main result. Somewhat counter-intuitively, we show that despite being computationally tractable
to optimize, submodular functions cannot be minimized from samples to within a desirable guarantee,
even when these functions are learnable. In particular, we show that there is no algorithm for
minimizing a submodular function from polynomially-many samples drawn from any distribution
that obtains an additive approximation of 1/2 o(1), even when the function is PAC-learnable.
Furthermore, we show that this bound is tight: the algorithm which returns the empty set or ground
set each with probability 1/2 achieves at least a 1/2 approximation. Notice that this also implies that
in general, there is no learning algorithm that can produce a surrogate whose minima is close to the
minima of the function we aim to optimize, as otherwise this would contradict our main result.
Technical overview. At a high level, hardness results in optimization from samples are shown
by constructing a family of functions, where the values of functions in the family are likely to be
indistinguishable for the samples, while having very different optimizers. The main technical difficulty
is to construct a family of functions that concurrently satisfy these two properties (indistinguishability
and different optimizers), and that are also PAC-learnable. En route to our main construction, we
first construct a family of functions that are completely indistinguishable given samples drawn from
the uniform distribution, in which case we obtain a 1/2 o(1) impossibility result (Section 2). The
2
general result that holds for any distribution requires heavier machinery to argue about more general
families of functions where some subset of functions can be distinguished from others given samples.
Instead of satisfying the two desired properties for all functions in a fixed family, we show that these
properties hold for all functions in a randomized subfamily (Section 3.2). We then develop an efficient
learning algorithm for the family of functions constructed for the main hardness result (Section 3.3).
This algorithm builds multiple linear regression predictors and a classifier to direct a fresh set to the
appropriate linear predictor. The learning of the classifier and the linear predictors relies on multiple
observations about the specific structure of this class of functions.
1.1
Related work
The problem of optimization from samples was introduced in the context of constrained submodular
maximization [BRS17, BRS16]. In general, for maximizing a submodular function under a cardinality
constraint, no algorithm can obtain a constant factor approximation guarantee from any samples. As
discussed above, for special classes of submodular functions that can be optimized in polynomial time
under a cardinality constraint, and for unconstrained maximization, there are desirable optimization
from samples guarantees. It is thus somewhat surprising that submodular minimization, which is an
unconstrained optimization problem that is optimizable in polynomial time in the value query model,
is hard to optimize from samples. From a technical perspective the constructions are quite different.
In maximization, the functions constructed in [BRS17, BRS16] are monotone so the ground set
would be an optimal solution if the problem was unconstrained. Instead, we need to construct novel
non-monotone functions. In convex optimization, recent work shows a tight 1/2-inapproximability
for convex minimization from samples [BS17]. Although there is a conceptual connection between
that paper and this one, from a technical perspective these papers are orthogonal. The discrete
analogue of the family of convex functions constructed in that paper is not (even approximately) a
family of submodular functions, and the constructions are significantly different.
2
Warm up: the Uniform Distribution
As a warm up to our main impossibility result, we sketch a tight lower bound for the special case in
which the samples are drawn from the uniform distribution. At a high level, the idea is to construct a
function which considers some special subset of ?good? elements that make its value drops when a
set contains all such ?good? elements. When samples are drawn from the uniform distribution and
?good? elements are sufficiently rare, there is a relatively simple construction that obfuscates which
elements the function considers ?good?, which then leads to the inapproximability.
2.1
Hardness for uniform distribution
p
We construct a family of functions F where fi 2 F is defined in terms of a set Gi ? N of size n.
For each such function we call Gi the set of good elements, and Bi = N \ Gi its bad elements. We
denote the number of good and bad elements in a set S by gS and bS , dropping the subscripts (S and
i) when clear from context, so g = |Gi \ S| and b = |Bi \ S|. The function fi is defined as follows:
8
p
1
1
>
< +
? (g + b) if g < n
fi (S) := 2 2n
p
>
: 1 ?b
if g = n
2n
It is easy to verify that these functions are submodular with range in [0, 1] (see illustration in
Figure 1a). Given samples drawn uniformly at random (u.a.r.),
pit is impossible to distinguish good
and bad elements since with high probability (w.h.p.) g < n for all samples. Informally, this
implies that a good learner for F over the uniform distribution D is f 0 (S) = 1/2 + |S|/(2n).
Intuitively, F is not 1/2 o(1) optimizable from samples because if an algorithm cannot learn the
set of good elements Gi , then it cannot find S such fi (S) < 1/2 o(1) whereas the optimal solution
Si? = Gi has value fi (Gi ) = 0.
Theorem 1. Submodular functions f : 2N ! [0, 1] are not 1/2 o(1) optimizable from samples
drawn from the uniform distribution for the problem of submodular minimization.
3
Proof. The details for the derivation of concentration bounds are in Appendix B. Consider fk drawn
u.a.r. from F and let f ? = fk and G? = Gk . Since the samples are all drawn from the uniform
distribution, by standard application of the Chernoff bound we have that every set Si in the sample
respects |Si | ? 3n/4, w.p. 1 e ?(n) . For sets S1 , . . . , Sm , all of size at most 3n/4, when fj is
p
1/2
drawn u.a.r. from F we get that |Si \ Gj | < n, w.p. 1 e ?(n ) for all i 2 [m], again by
1/2
Chernoff, and since m = poly(n). Notice that this implies that w.p. 1 e ?(n ) for all i 2 [m]:
1 |Si |
fj (Si ) = +
2
2n
Now, let F 0 be the collection of all functions fj for which fj (Si ) = 1/2 + |Si |/(2n) on all sets
1/2
0
{Si }m
e ?(n ) )|F|. Thus, since f ? is drawn u.a.r.
i=1 . The argument above implies that |F | = (1
1/2
from F we have that f ? 2 F 0 w.p. 1 e ?(n ) , and we condition on this event.
Let S be the (possibly randomized) solution returned by the algorithm. Observe that S is independent
of f ? 2 F 0 . In other words, the algorithm cannot learn any information about which function in F 0
1/6
generates the samples. By Chernoff, if we fix S and choose f u.a.r. from F, then, w.p. 1 e ?(n ) :
1
f (S)
o(1).
2
0
1/2
1/6
|
Since |F
e ?(n ) , it is also the case that f ? (S) 1/2 o(1) w.p. 1 e ?(n ) over the
|F | = 1
choice of f ? 2 F 0 . By the probabilistic method and since all the events we conditioned on occur
with exponentially high probability, there exists f ? 2 F s.t. the value of the set S returned by the
algorithm is 1/2 o(1) whereas the optimal solution is f ? (G? ) = 0.
2.2
A tight upper bound
We now show that the result above is tight. In particular, by randomizing between the empty set and
the ground set we get a solution whose value is at most 1/2. In the case of symmetric functions, i.e.
f (S) = f (N \ S) for all S ? N , ; and N are minima since f (N ) + f (;) ? f (S) + f (N \ S) for
all S ? N as shown below.1 Notice, that this does not require any samples.
Proposition 2. The algorithm which returns the empty set ; or the ground N with probability 1/2
each is a 1/2 additive approximation for the problem of unconstrained submodular minimization.
Proof. Let S ? N , observe that
f (N \ S) f (;) f ((N \ S) [ S) f (S) = f (N )
where the inequality is by submodularity. Thus, we obtain
1
1
1
(f (N ) + f (;)) ? f (S) + f (N \ S) ? f (S) +
2
2
2
In particular, this holds for S 2 argminT ?N f (T ).
3
f (S)
1
.
2
General Distribution
In this section, we show our main result, namely that there exists a family of submodular functions
such that, despite being PAC-learnable for all distributions, no algorithm can obtain an approximation
better than 1/2 o(1) for the problem of unconstrained minimization.
The functions in this section build upon the previous construction, though are inevitably more involved
in order to achieve learnability and inapproximability on any distribution. The functions constructed
for the uniform distribution do not yield inapproximability for general distributions due to the fact that
the indistinguishability between two functions no longer holds when sets S of large size are sampled
with non-negligible probability. Intuitively, in the previous construction, once a set is sufficiently
large the good elements of the function can be distinguished from the bad ones. The main idea to get
around this issue is to introduce masking elements M . We construct functions such that, for sets S of
large size, good and bad elements are indistinguishable if S contains at least one masking element.
1
Although ; and N are trivial minima if f is symmetric, the problem of minimizing a symmetric submodular
function over proper nonempty subsets is non-trivial (see [Que98]).
4
1
f(S)
f(S)
1
0.5
0.5
m = |S|
b = |S|
g = |S|
b = |S|-1, m=1
g = |S|-1, m=1
b = |S|
g = |S|
0
0 ?n - 1 ?n
0
n ? ?n
|S|
(a) Uniform distribution
0
1 n1/4
?n
|S|
n/2
(b) General distribution
Figure 1: An illustration of the value of a set S of good (blue), bad (red), and masking (green)
elements as a function of |S| for the functions constructed. For the general distribution case, we
also illustrate the value of a set S of good (dark blue) and bad (dark red) elements when S also
contains at least one masking element.
The construction. Each function fi 2 F is defined in terms of a partition Pi of the ground set
into good, bad, andpmasking elements. The
p partitions we consider are Pi = (Gi , Bi , Mi ) with
|Gi | = n/2, |Bi | = n, and |Mi | = n/2
n. Again, when clear from context, we drop indices of
i and S and the number of good, bad, and masking elements in a set S are denoted by g, b, and m.
For such a given partition Pi , the function fi is defined as follows (see illustration in Figure 1b):
8
1
>
>
p ? (b + g)
Region X : if m = 0 and g < n1/4
>
>2 n
>
>
?
? 1 ?
?
1 < 1
fi (S) = +
p ? b + n1/4
? g n1/4
Region Y : if m = 0 and g n1/4
2 >
n
2
n
>
>
>
>
1
>
:1
? (b + g)
Region Z : otherwise
2 n
3.1
Submodularity
In the appendix, we prove that the functions fi constructed pas above are indeed submodular
(Lemma 10). By rescaling fi with an additive term of n1/4 /(2 n) = 1/(2n1/4 ), it can be easily
verified that its range is in [0, 1]. We use the non-normalized definition as above for ease of notation.
3.2
Inapproximability
We now show that F cannot be minimized within a 1/2 o(1) approximation given samples from
any distribution. We first define F M , which is a randomized subfamily of F. We then give a general
lemma that shows that if two conditions of indistinguishability and gap are satisfied then we obtain
inapproximability. We then show that these two conditions are satisfied for the subfamily F M .
A randomization over masking elements. Instead of considering a function f drawn u.a.r. from
F as in the uniform case, we consider functions f in a randomized subfamily of functions F M ? F
to obtain the indistinguishability and gap
the family of functions F, let M be a
p conditions. Given
uniformly random subset of size n/2
n and define F M ? F:
F M := {fi 2 F : (Gi , Bi , M )}.
Since masking elements are distinguishable from good and bad elements, they need to be the same
set of elements for each function in family F M to obtain indistinguishability of functions in F M .
The inapproximability lemma. In addition to this randomized subfamily of functions, another
main conceptual departure of the following inapproximability lemma from the uniform case is that
no assumption can be made about the samples, such as their size, since the distribution is arbitrary.
We denote by U (A) the uniform distribution over the set A.
5
Lemma 3. Let F be a family of functions and F 0 = {f1 , . . . , fp } ? F be a subfamily of functions
drawn from some distribution. Assume the following two conditions hold:
1/4
1. Indistinguishability. For all S ? N , w.p. 1 e ?(n
fi (S) = fj (S);
)
over F 0 : for every fi , fj 2 F 0 ,
2. ?-gap. Let Si? be a minimizer of fi , then w.p. 1 over F 0 : for all S ? N ,
E 0 [fi (S) fi (Si? )] ?;
fi ?U (F )
Then, F is not ?-minimizable from strictly less than e?(n
1/4
)
samples over any distribution D.
Note that the ordering of the quantifiers is crucial. The proof is deferred to the appendix, but the main
ideas are summarized as follows. We use a probabilistic argument to switch from the randomization
over F 0 to the randomization over S ? D and show that there exists a deterministic F ? F such that
fi (S) = fj (S) for all fi , fj 2 F w.h.p. over S ? D. By a union bound this holds for all samples
S. Thus, for such a family of functions F = {f1 , . . . , fp }, the choices of an algorithm that is given
samples from fi for i 2 [p] are independent of i. By the ?-gap condition, this implies that there exists
fi 2 F for which a solution S returned by the algorithm is at least ? away from fi (Si? ).
Indistinguishability and gap of F. We now show the indistinguishability and gap conditions, with
? = 1/2 o(1), which immediately imply a 1/2 o(1) inapproximability by Lemma 3. For the
indistinguishability, it suffices to show that good and bad elements are indistinguishable since the
masking elements are identical for all functions in F M . Good and bad elements are indistinguishable
since, w.h.p., a set S is not in region Y, which is the only region distinguishing good and bad elements.
Lemma 4. For all S ? N s.t. |S| < n1/4 : For all fi 2 F M ,
(
1
p
? (b + g)
if m = 0 (Region X )
1
fi (S) = + 21 n 1
2
otherwise (Region Z)
2
n ? (b + g)
1/4
and for all S ? N such that |S|
n1/4 , with probability 1 e ?(n ) over F M : For all fi 2 F M ,
1
fi (S) = 1
? (b + g)
(Region Z).
n
Proof. Let S ? N . If |S| < n1/4 , then the proof follows immediately from the definition of fi . If
|S| n1/4 , then, the number of masking elements m in S is m = |M \ S| for all fi 2 F M . We
1/4
then get m 1, for all fi 2 F M , with probability 1 e ?(n ) over F M by Chernoff bound. The
proof then follows again immediately from the definition of fi .
Next, we show the gap. The gap is since the good elements can be any subset of N \ M .
Lemma 5. Let Si? be a minimizer of fi . With probability 1 over F M , for all S ? N ,
1
E
[fi (S)]
o(1).
M
2
fi ?U (F )
Proof. Let S ? N and fi ? U (F M ). Note that the order of the quantifiers in the statement of the
lemma implies that S can be dependent on M , but that it is independent of i. There are three cases. If
m 1, then S is in region Z and fi (S) 1/2. If m = 0 and |S| ? n7/8 , then S is in region X or Y
and fi (S) 1/2 n7/8 /n = 12 o(1). Otherwise, m = 0 and |S| n7/8 . Since S is independent
of i, by Chernoff bound, we get
p
p
n/2 + n
n/2 + n
p
(1 o(1)) ? |S| ?
? b and
? g ? (1 + o(1)) ? |S|
n/2
n
with probability 1
1
fi (S)
+ (1
2
?(n1/4 )
. Thus S is in region Y and
p
1
n
1
n/2
p ? |S| (1 + o(1)) ?
p ? |S|
o(1)) p ?
n n/2 + n
2 n n/2 + n
e
Thus, we obtain Efi ?U (F M ) [fi (S)]
1
2
o(1).
6
1
2
o(1).
Combining the above three lemmas, we obtain the inapproximability result.
Lemma 6. The problem of submodular minimization cannot be approximated with a 1/2
additive approximation given poly(n) samples from any distribution D.
o(1)
Proof. For any set S ? N , observe that the number g + b of elements in S that are either good or
bad is the same for any two functions fi , fj 2 F M and for any F M . Thus, by Lemma 4, we obtain
the indistinguishability condition. Next, the optimal solution Si? = Gi of fi has value fi (Gi ) = o(1),
so by Lemma 5, we obtain the ?-gap condition with ? = 1/2 o(1). Thus F is not 1/2 o(1)
minimizable from samples from any distribution D by Lemma 3. The class of functions F is a class
of submodular functions by Lemma 10 (in Appendix C).
3.3
Learnability of F
We now show that every function in F is efficiently learnable from samples drawn from any distribution D. Specifically, we show that for any ?, 2 (0, 1) the functions are (?, ) PAC learnable
with the absolute loss function (or any Lipschitz loss function) using poly(1/?, 1/ , n) samples and
running time. At a high level, since each function fi is piecewise-linear over three different regions
Xi , Yi , and Zi , the main idea is to exploit this structure by first training a classifier to distinguish
between regions and then apply linear regression in different regions.
The learning algorithm. Since every function f 2 F is piecewise linear over three different
regions, there are three different linear functions fX , fY , fZ s.t. for every S ? N its value f (S)
can be expressed as fR (S) for some region R 2 {X , Y, Z}. The learning algorithm produces a
predictor f? by using a multi-label classifier and a set of linear predictors {fX? , fY? } [ {[i2M? fZ?i }.
? [ {[ ? Z?i },
The multi-label classifier creates a mapping from sets to regions, g : 2N ! {X? , Y}
i2M
?
?
?
s.t. X , Y, Z are approximated by X , Y, [i2M? Zi . Given a sample S ? D, using the algorithm then
retuns f?(S) = fg(S) (S). We give a formal description below (detailed description is in Appendix D).
Algorithm 1 A learning algorithm for f 2 F which combines classification and linear regression.
Input: samples S = {(Sj , f (Sj ))}j2[m]
? M
?)
(Z,
(;, ;)
for i = 1 to n do
?
Z?i
{S : ai 2 S, S 62 Z}
reg
fZ?i
ERM ({(Sj , f (Sj )) : Sj 2 Z?i })
linear regression
P
if (Sj ,f (Sj )):Sj 2Z?i |fZ?i (Sj ) f (Sj )| = 0 then
?
? [ {ai }
Z?
Z? [ Z?i , M
M
cla
? j ? m/2})
C
ERM ({(Sj , f (Sj )) : Sj 62 Z,
train a classifier for regions X , Y
?
? C(S) = 1}, {S : S 62 Z,
? C(S) = 1})
(X? , Y)
({S : S 62 Z,
8
p
if S2X?
<|S|/(2 n)
reg
?
?
return f
S 7! fY? (S) = ERM ({(Sj , f (Sj )) : Sj 2 Y, j > m/2}) if S2Y?
:
? })
?
fZ?i (S) : i = min({i0 : ai0 2 S \ M
if S2Z
Overview of analysis of the learning algorithm. There are two main challenges in training the
algorithm. The first is that the region X , Y, or Z that a sample (Sj , f (Sj )) belongs to is not known.
? Z? using the samples,
Thus, even before being able to train a classifier which learns the regions X? , Y,
we need to learn the region a sample Sj belongs to using f (Sj ). The second is that the samples SR
used for training a linear regression predictor fR over region R need to be carefully selected so that
SR is a collection of i.i.d. samples from the distribution S ? D conditioned on S 2 R (Lemma 20).
We first discuss the challenge of labeling samples with the region they belong to. Observe that for a
fixed masking element ai 2 M , f 2 F is linear over all sets S containing ai since these sets are all in
region Z. Thus, there must exist a linear regression predictor fZ?i = ERMreg (?) with zero empirical
loss over all samples Sj containing ai if ai 2 M (and thus Sj 2 Z). ERMreg (?) minimizes the
empirical loss on the input samples over the class of linear regression predictors with bounded norm
7
Figure 2: An illustration of the regions. The dots represent the samples, the corresponding full circles represent
? Z? learned by the
the regions X (red), Y (blue), and Z (green). The ellipsoids represent the regions X? , Y,
?
classifier. Notice that Z has no false negatives.
(Lemma 19). If fZ?i has zero empirical loss, we directly classify any set S containing ai as being in
? Next, for a sample (Sj , f (Sj )) not in Z,
? we can label these samples since Sj 2 X if and only if
Z.
p
f (Sj ) = |Sj |/(2 n). With these labeled samples S 0 , we train a binary classifier C = ERMcla (S 0 )
? ERMcla (S 0 ) minimizes the empirical loss on
that indicates if S s.t. S 62 Z? is in region X? or Y.
0
labeled samples S over the class of halfspaces w 2 Rn (Lemma 23).
Regarding the second challenge, we cannot use all samples Sj s.t. Sj 2 Y? to train a linear predictor
? so they are not a collection of i.i.d.
fY? for region Y? since these same samples were used to define Y,
?
samples from the distribution S ? D conditioned on S 2 Y. To get around this issue, we partition
the samples into two distinct collections, one to train the classifier C and one to train fY? (Lemma 24).
? we predict f ? (T ) where i is s.t. ai 2 T \ M
? (breaking ties lexicographically)
Next, given T 2 Z,
Zi
?
? (Lemma 22). Since we break
which performs well since fZ?i has zero empirical error for ai 2 M
ties lexicographically, f?Z?i must be trained over samples Sj such that ai 2 Sj and ai0 62 Sj for i0 s.t.
? to obtain i.i.d. samples from the same distribution as T ? D conditioned on T
i0 < i and ai0 2 M
being directed to f?Z?i (Lemma 21).
The analysis of the learning algorithm leads to the following main learning result.
Lemma 7. Let f? be the predictor returned by Algorithm 1, then w.p. 1
over m 2 O(n3 +
2
2
n (log(2n/ ))/? ) samples S drawn i.i.d. from any distribution D, ES?D [|f?(S) f (S)|] ? ?.
3.4
Main Result
We conclude this section with our main result which combines Lemmas 6 and 7.
Theorem 8. There exists a family of [0, 1]-bounded submodular functions F that is efficiently PAClearnable and that cannot be optimized from polynomially many samples drawn from any distribution
D within a 1/2 o(1) additive approximation for unconstrained submodular minimization.
4
Discussion
In this paper, we studied the problem of submodular minimization from samples. Our main result
is an impossibility, showing that even for learnable submodular functions it is impossible to find
a non-trivial approximation to the minimizer with polynomially-many samples, drawn from any
distribution. In particular, this implies that minimizing a general submodular function learned from
data cannot yield desirable guarantees. In general, it seems that the intersection between learning and
optimization is elusive, and a great deal still remains to be explored.
8
References
[Bal15] Maria-Florina Balcan. Learning submodular functions with applications to multi-agent systems. In
Proceedings of the 2015 International Conference on Autonomous Agents and Multiagent Systems,
AAMAS 2015, Istanbul, Turkey, May 4-8, 2015, page 3, 2015.
[BH11] Maria-Florina Balcan and Nicholas JA Harvey. Learning submodular functions. In Proceedings of
the forty-third annual ACM symposium on Theory of computing, pages 793?802. ACM, 2011.
[BRS16] Eric Balkanski, Aviad Rubinstein, and Yaron Singer. The power of optimization from samples. In
Advances in Neural Information Processing Systems, pages 4017?4025, 2016.
[BRS17] Eric Balkanski, Aviad Rubinstein, and Yaron Singer. The limitations of optimization from samples.
Proceedings of the Forty-Ninth Annual ACM on Symposium on Theory of Computing, 2017.
[BS17] Eric Balkanski and Yaron Singer. The sample complexity of optimizing a convex function. In
COLT, 2017.
[BVW16] Maria-Florina Balcan, Ellen Vitercik, and Colin White. Learning combinatorial functions from
pairwise comparisons. In Proceedings of the 29th Conference on Learning Theory, COLT 2016,
New York, USA, June 23-26, 2016, pages 310?335, 2016.
[CLSW16] Deeparnab Chakrabarty, Yin Tat Lee, Aaron Sidford, and Sam Chiu-wai Wong. Subquadratic
submodular function minimization. arXiv preprint arXiv:1610.09800, 2016.
[DTK16] Josip Djolonga, Sebastian Tschiatschek, and Andreas Krause. Variational inference in mixed
probabilistic submodular models. In Advances in Neural Information Processing Systems 29: Annual
Conference on Neural Information Processing Systems 2016, December 5-10, 2016, Barcelona,
Spain, pages 1759?1767, 2016.
[EN15] Alina Ene and Huy L. Nguyen. Random coordinate descent methods for minimizing decomposable
submodular functions. In Proceedings of the 32nd International Conference on Machine Learning,
ICML 2015, Lille, France, 6-11 July 2015, pages 787?795, 2015.
[FK14] Vitaly Feldman and Pravesh Kothari. Learning coverage functions and private release of marginals.
In COLT, pages 679?702, 2014.
[FKV13] Vitaly Feldman, Pravesh Kothari, and Jan Vondr?k. Representation, approximation and learning of
submodular functions using low-rank decision trees. In COLT, pages 711?740, 2013.
[FMV11] Uriel Feige, Vahab S. Mirrokni, and Jan Vondr?k. Maximizing non-monotone submodular functions.
SIAM J. Comput., 40(4):1133?1153, 2011.
[GLS81] Martin Grotschel, Laszlo Lovasz, and Alexander Schrijver. The ellipsoid method and its consequences in combinatorial optimization. Combinatorica, 1(2):169?197, 1981.
[IFF01] Satoru Iwata, Lisa Fleischer, and Satoru Fujishige. A combinatorial strongly polynomial algorithm
for minimizing submodular functions. J. ACM, 48(4):761?777, 2001.
[IJB13] Rishabh K. Iyer, Stefanie Jegelka, and Jeff A. Bilmes. Curvature and optimal algorithms for learning
and minimizing submodular functions. In Advances in Neural Information Processing Systems 26:
27th Annual Conference on Neural Information Processing Systems 2013. Proceedings of a meeting
held December 5-8, 2013, Lake Tahoe, Nevada, United States., pages 2742?2750, 2013.
[JB11] Stefanie Jegelka and Jeff Bilmes. Submodularity beyond submodular energies: coupling edges in
graph cuts. In Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on, pages
1897?1904. IEEE, 2011.
[JLB11] Stefanie Jegelka, Hui Lin, and Jeff A Bilmes. On fast approximate submodular minimization. In
Advances in Neural Information Processing Systems, pages 460?468, 2011.
[NB12] Mukund Narasimhan and Jeff A Bilmes. A submodular-supermodular procedure with applications
to discriminative structure learning. arXiv preprint arXiv:1207.1404, 2012.
[Que98] Maurice Queyranne. Minimizing symmetric submodular functions. Mathematical Programming,
82(1-2):3?12, 1998.
[SK10] Peter Stobbe and Andreas Krause. Efficient minimization of decomposable submodular functions.
In Advances in Neural Information Processing Systems, pages 2208?2216, 2010.
[SSBD14] Shai Shalev-Shwartz and Shai Ben-David. Understanding machine learning: From theory to
algorithms. Cambridge university press, 2014.
9
| 6683 |@word private:1 polynomial:9 seems:2 norm:1 nd:1 seek:3 tat:1 cla:1 contains:3 united:1 surprising:1 si:17 must:2 additive:7 partition:4 drop:2 selected:1 tahoe:1 mathematical:1 constructed:6 direct:1 symposium:2 prove:1 combine:2 introduce:1 pairwise:1 indeed:1 hardness:3 behavior:1 multi:3 pitfall:1 paclearnable:1 cardinality:4 considering:1 spain:1 grotschel:1 bounded:3 notation:1 minimizable:3 minimizes:2 narasimhan:1 guarantee:7 every:6 tie:2 classifier:10 indistinguishability:10 unit:1 positive:1 negligible:1 before:1 consequence:1 despite:4 subscript:1 approximately:2 studied:2 pit:1 ease:1 tschiatschek:1 range:3 bi:5 directed:1 union:1 optimizers:2 procedure:1 jan:2 empirical:5 significantly:1 word:1 get:6 cannot:10 close:2 satoru:2 context:3 impossible:2 wong:1 optimize:5 deterministic:1 maximizing:2 elusive:1 convex:4 decomposable:2 immediately:3 ellen:1 notion:1 fx:2 autonomous:1 coordinate:1 construction:7 heavily:2 programming:1 us:2 distinguishing:1 harvard:4 element:31 pa:1 satisfying:1 approximated:2 recognition:1 cut:2 predicts:1 labeled:2 preprint:2 region:29 ordering:1 counter:1 halfspaces:1 complexity:1 trained:1 tight:6 upon:1 creates:1 eric:4 learner:1 completely:1 easily:1 derivation:1 train:6 distinct:1 fast:1 fk14:2 query:4 labeling:1 rubinstein:2 shalev:1 whose:3 modular:1 quite:2 cvpr:1 say:1 otherwise:4 gi:12 nevada:1 ai0:3 fr:2 j2:1 combining:1 achieve:1 intuitive:2 description:2 empty:4 optimum:3 sea:1 produce:3 ben:1 illustrate:1 develop:1 coupling:1 coverage:1 implies:7 submodularity:3 brs16:4 require:1 ja:1 suffices:2 fix:1 f1:2 randomization:3 proposition:1 strictly:3 hold:6 sufficiently:2 around:2 ground:7 s2x:1 great:1 mapping:1 predict:1 achieves:2 pravesh:2 label:3 combinatorial:3 minimization:16 lovasz:1 concurrently:1 aim:3 rather:1 release:1 focus:1 june:1 maria:3 rank:1 indicates:1 impossibility:3 obfuscates:1 inference:1 dependent:1 minimizers:1 i0:3 istanbul:1 diminishing:1 france:1 issue:2 classification:1 colt:4 denoted:1 constrained:2 special:3 construct:7 once:1 having:1 beach:1 chernoff:5 identical:1 lille:1 icml:1 djolonga:1 minimized:8 np:1 others:1 piecewise:2 subquadratic:1 n1:12 deferred:1 rishabh:1 light:1 held:1 laszlo:1 edge:1 machinery:1 orthogonal:1 tree:1 desired:1 circle:1 josip:1 vahab:1 classify:1 sidford:1 maximization:9 subset:5 rare:1 uniform:13 predictor:10 learnability:2 randomizing:1 st:1 international:2 randomized:5 siam:1 probabilistic:3 lee:1 again:3 satisfied:2 balkanski:4 containing:3 choose:2 possibly:1 maurice:1 return:5 rescaling:1 summarized:1 satisfy:1 brs17:6 multiplicative:1 break:1 red:3 parallel:1 yaron:5 masking:10 shai:2 minimize:2 largely:1 efficiently:4 maximized:1 yield:2 bilmes:4 sebastian:1 wai:1 stobbe:1 definition:3 energy:1 involved:1 chakrabarty:1 proof:8 attributed:1 mi:2 sampled:1 formalize:1 carefully:1 supermodular:1 though:1 strongly:1 furthermore:2 uriel:1 sketch:1 usa:2 verify:1 true:5 bs17:2 normalized:1 symmetric:7 deal:1 white:1 indistinguishable:5 performs:1 fj:9 balcan:3 meaning:1 variational:1 novel:1 recently:1 fi:42 overview:2 exponentially:1 discussed:1 belong:1 marginals:1 cambridge:1 feldman:2 ai:10 unconstrained:12 fk:2 submodular:64 dot:1 access:3 longer:1 gj:1 curvature:1 recent:1 perspective:2 optimizing:3 belongs:2 route:1 harvey:1 inequality:1 binary:1 arbitrarily:3 meeting:1 yi:1 minimum:8 somewhat:2 forty:2 maximize:2 colin:1 tempting:2 july:1 multiple:2 desirable:4 full:1 turkey:1 technical:4 lexicographically:2 long:1 lin:1 regression:7 florina:3 vision:1 arxiv:4 represent:3 whereas:2 addition:1 krause:2 crucial:1 sr:2 fujishige:1 december:2 vitaly:2 n7:3 call:1 easy:1 switch:1 zi:3 andreas:2 idea:4 regarding:1 aviad:2 fleischer:1 whether:3 heavier:1 queyranne:1 peter:1 returned:4 york:1 generally:3 clear:2 informally:1 detailed:1 dark:2 fz:8 exist:1 notice:4 blue:3 discrete:1 dropping:1 drawn:17 alina:1 verified:1 graph:1 monotone:4 family:16 reasonable:2 lake:1 decision:1 appendix:6 bound:9 guaranteed:2 distinguish:2 g:1 annual:4 occur:1 constraint:4 n3:1 generates:1 argument:2 min:2 extremely:1 relatively:1 conjecture:2 martin:1 poor:1 feige:1 sam:1 appealing:1 b:1 s1:1 intuitively:3 quantifier:2 erm:3 ene:1 computationally:1 remains:1 discus:1 nonempty:1 singer:4 know:2 tractable:2 optimizable:4 efi:1 apply:1 observe:4 away:1 appropriate:1 nicholas:1 distinguished:2 running:1 exploit:1 especially:1 build:2 approximating:1 question:4 concentration:1 mirrokni:1 surrogate:8 argue:1 considers:2 fy:5 trivial:5 vitercik:1 fresh:1 index:1 illustration:4 ratio:1 minimizing:9 ellipsoid:2 statement:1 gk:1 negative:1 proper:1 upper:1 observation:1 kothari:2 sm:1 descent:1 inevitably:1 i2m:3 rn:1 ninth:1 arbitrary:1 introduced:2 david:1 namely:1 optimized:4 connection:1 learned:5 barcelona:1 nip:1 address:1 able:1 beyond:1 below:2 pattern:1 departure:1 fp:2 challenge:3 max:1 green:2 analogue:2 power:1 suitable:1 event:2 difficulty:1 warm:2 imply:1 stefanie:3 understanding:1 loss:6 multiagent:1 mixed:1 limitation:1 agent:2 jegelka:3 proxy:1 pi:3 course:1 formal:1 lisa:1 absolute:1 fg:1 collection:4 made:1 nguyen:1 far:1 polynomially:8 sj:32 approximate:1 obtains:3 contradict:1 argmint:1 vondr:2 conceptual:2 conclude:1 xi:1 discriminative:1 alternatively:1 shwartz:1 decade:1 learn:6 ca:1 poly:4 constructing:1 main:16 huy:1 aamas:1 en:1 nphard:1 comput:1 breaking:1 third:1 learns:1 theorem:2 bad:14 specific:1 pac:5 showing:1 learnable:10 explored:1 mukund:1 intractable:1 exists:5 false:1 hui:1 iyer:1 conditioned:4 demand:1 gap:9 intersection:1 yin:1 distinguishable:1 likely:1 subfamily:6 expressed:1 deeparnab:1 inapproximability:10 minimizer:6 iwata:1 relies:1 acm:4 goal:1 consequently:1 ericbalkanski:1 jeff:4 lipschitz:1 hard:2 specifically:2 uniformly:3 lemma:23 e:1 schrijver:1 aaron:1 chiu:1 combinatorica:1 alexander:1 reg:2 |
6,283 | 6,684 | Introspective Classification with Convolutional Nets
Long Jin
UC San Diego
[email protected]
Justin Lazarow
UC San Diego
[email protected]
Zhuowen Tu
UC San Diego
[email protected]
Abstract
We propose introspective convolutional networks (ICN) that emphasize the importance of having convolutional neural networks empowered with generative
capabilities. We employ a reclassification-by-synthesis algorithm to perform training using a formulation stemmed from the Bayes theory. Our ICN tries to iteratively:
(1) synthesize pseudo-negative samples; and (2) enhance itself by improving the
classification. The single CNN classifier learned is at the same time generative
? being able to directly synthesize new samples within its own discriminative
model. We conduct experiments on benchmark datasets including MNIST, CIFAR10, and SVHN using state-of-the-art CNN architectures, and observe improved
classification results.
1
Introduction
Great success has been achieved in obtaining powerful discriminative classifiers via supervised
training, such as decision trees [34], support vector machines [42], neural networks [23], boosting
[7], and random forests [2]. However, recent studies reveal that even modern classifiers like deep
convolutional neural networks [20] still make mistakes that look absurd to humans [11]. A common
way to improve the classification performance is by using more data, in particular ?hard examples?,
to train the classifier. Different types of approaches have been proposed in the past including
bootstrapping [31], active learning [37], semi-supervised learning [51], and data augmentation [20].
However, the approaches above utilize data samples that are either already present in the given
training set, or additionally created by humans or separate algorithms.
In this paper, we focus on improving convolutional neural networks by endowing them with synthesis
capabilities to make them internally generative. In the past, attempts have been made to build
connections between generative models and discriminative classifiers [8, 27, 41, 15]. In [44], a
self supervised boosting algorithm was proposed to train a boosting algorithm by sequentially
learning weak classifiers using the given data and self-generated negative samples; the generative via
discriminative learning work in [40] generalizes the concept that unsupervised generative modeling
can be accomplished by learning a sequence of discriminative classifiers via self-generated pseudonegatives. Inspired by [44, 40] in which self-generated samples are utilized, as well as recent success
in deep learning [20, 9], we propose here an introspective convolutional network (ICN) classifier and
study how its internal generative aspect can benefit CNN?s discriminative classification task. There is
a recent line of work using a discriminator to help with an external generator, generative adversarial
networks (GAN) [10], which is different from our objective here. We aim at building a single CNN
model that is simultaneously discriminative and generative.
The introspective convolutional networks (ICN) being introduced here have a number of properties. (1)
We introduce introspection to convolutional neural networks and show its significance in supervised
classification. (2) A reclassification-by-synthesis algorithm is devised to train ICN by iteratively
augmenting the negative samples and updating the classifier. (3) A stochastic gradient descent
sampling process is adopted to perform efficient synthesis for ICN. (4) We propose a supervised
formulation to directly train a multi-class ICN classifier. We show consistent improvement over
state-of-the-art CNN classifiers (ResNet [12]) on benchmark datasets in the experiments.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
2
Related work
Our ICN method is directly related to the generative via discriminative learning framework [40].
It also has connection to the self-supervised learning method [44], which is focused on density
estimation by combining weak classifiers. Previous algorithms connecting generative modeling
with discriminative classification [8, 27, 41, 15] fall in the category of hybrid models that are direct
combinations of the two. Some existing works on introspective learning [22, 3, 38] have a different
scope to the problem being tackled here. Other generative modeling schemes such as MiniMax
entropy [50], inducing features [6], auto-encoder [1], and recent CNN-based generative modeling
approaches [48, 47] are not for discriminative classification and they do not have a single model that
is both generative and discriminative. Below we discuss the two methods most related to ICN, namely
generative via discriminative learning (GDL) [40] and generative adversarial networks (GAN) [10].
Relationship with generative via discriminative learning (GDL) [40]
ICN is largely inspired by GDL and it follows a similar pipeline developed in [40]. However, there is
also a large improvement of ICN to GDL, which is summarized below.
? CNN vs. Boosting. ICN builds on top of convolutional neural networks (CNN) by explicitly
revealing the introspectiveness of CNN whereas GDL adopts the boosting algorithm [7].
? Supervised classification vs. unsupervised modeling. ICN focuses on the supervised classification task with competitive results on benchmark datasets whereas GDL was originally applied
to generative modeling and its power for the classification task itself was not addressed.
? SGD sampling vs. Gibbs sampling. ICN carries efficient SGD sampling for synthesis through
backpropagation which is much more efficient than the Gibbs sampling strategy used in GDL.
? Single CNN vs. Cascade of classifiers. ICN maintains a single CNN classifier whereas GDL
consists of a sequence of boosting classifiers.
? Automatic feature learning vs. manually specified features. ICN has greater representational
power due to the end-to-end training of CNN whereas GDL relies on manually designed features.
Comparison with Generative Adversarial Networks (GANs) [10]
Recent efforts in adversarial learning [10] are also very interesting and worth comparing with.
? Introspective vs. adversarial. ICN emphasizes being introspective by synthesizing samples
from its own classifier while GAN focuses on adversarial ? using a distinct discriminator to
guide the generator.
? Supervised classification vs. unsupervised modeling. The main focus of ICN is to develop a
classifier with introspection to improve the supervised classification task whereas GAN is mostly
for building high-quality generative models under unsupervised learning.
? Single model vs. two separate models. ICN retains a CNN discriminator that is itself a generator
whereas GAN maintains two models, a generator and a discriminator, with the discriminator in
GAN trained to classify between ?real? (given) and ?fake? (generated by the generator) samples.
? Reclassification-by-synthesis vs. minimax. ICN engages an iterative procedure, reclassificationby-synthesis, stemmed from the Bayes theory whereas GAN has a minimax objective function to
optimize. Training an ICN classifier is the same as that for the standard CNN.
? Multi-class formulation. In a GAN-family work [36], a semi-supervised learning task is devised
by adding an additional ?not-real? class to the standard k classes in multi-class classification; this
results in a different setting to the standard multi-class classification with additional model parameters. ICN instead, aims directly at the supervised multi-class classification task by maintaining
the same parameter setting within the softmax function without additional model parameters.
Later developments alongside GAN [35, 36, 49, 3] share some similar aspects to GAN, which
also do not achieve the same goal as ICN does. Since the discriminator in GAN is not meant to
perform the generic two-class/multi-class classification task, some special settings for semi-supervised
learning [10, 35, 49, 3, 36] were created. ICN instead has a single model that is both generative and
discriminative, and thus, an improvement to ICN?s generator leads to a direct means to ameliorate its
discriminator. Other work like [11] was motivated from an observation that adding small perturbations
to an image leads to classification errors that are absurd to humans; their approach is however taken
by augmenting positive samples from existing input whereas ICN is able to synthesize new samples
from scratch. A recent work proposed in [21] is in the same family of ICN, but [21] focuses on
unsupervised image modeling using a cascade of CNNs.
2
3
Method
The pipeline of ICN is shown in Figure 1, which has an immediate improvement over GDL [40]
in several aspects that have been described in the previous section. One particular gain of ICN is
its representation power and efficient sampling process through backpropagation as a variational
sampling strategy.
3.1 Formulation
We start the discussion by introducing the basic formulation and borrow the notation from [40].
Let x be a data sample (vector) and y ? {?1, +1} be its label, indicating either a negative or
a positive sample (in multi-class classification y ? {1, ..., K}). We study binary classification
first. A discriminative classifier computes p(y|x), the probability of x being positive or negative.
p(y = ?1|x) + p(y = +1|x) = 1. A generative model instead models p(y, x) = p(x|y)p(y), which
captures the underlying generation process of x for class y. In binary classification, positive samples
are of primary interest. Under the Bayes rule:
p(y = +1|x)p(y = ?1)
p(x|y = +1) =
p(x|y = ?1),
(1)
p(y = ?1|x)p(y = +1)
which can be further simplified when assuming equal priors p(y = +1) = p(y = ?1):
p(x|y = +1) =
p(y = +1|x)
p(x|y = ?1).
1 ? p(y = +1|x)
(2)
Reclassification Step:
training on the given training data + generated pseudo-negatives
Initial Classification
(given training data)
Convolutional Neural Networks
Classification
Final Classification
(given training data +
self-generated
pseudo-negatives)
Introspective Convolutional Networks:
?
?
Synthesis
Reclassification
Synthesis
Synthesis Step:
synthesize pseudo-negative samples
Figure 1: Schematic illustration of our reclassification-by-synthesis algorithm for ICN training. The top-left
figure shows the input training samples where the circles in red are positive samples and the crosses in blue are
the negatives. The bottom figures are the samples progressively self-generated by the classifier in the synthesis
steps and the top figures show the decision boundaries (in purple) progressively updated in the reclassification
steps. Pseudo-negatives (purple crosses) are gradually generated and help tighten the decision boundaries.
We make two interesting and important observations from Eqn. (2): 1) p(x|y = +1) is dependent
on the faithfulness of p(x|y = ?1), and 2) a classifier C to report p(y = +1|x) can be made
simultaneously generative and discriminative. However, there is a requirement: having an informative distribution for the negatives p(x|y = ?1) such that samples drawn x ? p(x|y = ?1)
3
have good coverage to the entire space of x ? Rm , especially for samples that are close to the
positives x ? p(x|y = +1), to allow the classifier to faithfully learn p(y = +1|x). There seems
to exist a dilemma. In supervised learning, we are only given a set of limited amount of training
data, and a classifier C is only focused on the decision boundary to separate the given samples and
the classification on the unseen data may not be accurate. This can be seen from the top left plot
in Figure 1. This motivates us to implement the synthesis part within learning ? make a learned
discriminative classifier generate samples that pass its own classification and see how different these
generated samples are to the given positive samples. This allows us to attain a single model that has
two aspects at the same time: a generative model for the positive samples and an improved classifier
for the classification.
Suppose we are given a training set S = {(xi , yi ), i = 1..n} and x ? Rm and y ? {?1, +1}. One
can directly train a discriminative classifier C, e.g. a convolutional neural networks [23] to learn
p(y = +1|x), which is always an approximation due to various reasons including insufficient training
samples, generalization error, and classifier limitations. Previous attempts to improve classification
by data augmentation were mostly done to add more positive samples [20, 11]; we instead argue
the importance of adding more negative samples to improve the classification performance. The
dilemma is that S = {(xi , yi ), i = 1..n} is limited to the given data. For clarity, we now use p? (x)
to represent p(x|y = ?1). Our goal is to augment the negative training set by generating confusing
pseudo-negatives to improve the classification (note that in the end pseudo-negative samples drawn
x ? p?
t (x) will become hard to distinguish from the given positive samples. Cross-validation
can be used to determine when using more pseudo-negatives is not reducing the validation error).
We call the samples drawn from x ? p?
t (x) pseudo-negatives (defined in [40]). We expand
t
0
S = {(xi , yi ), i = 1..n} by Set = S ? Spn
, where Spn
= ? and for t ? 1
t
Spn = {(xi , ?1), i = n + 1, ..., n + tl}.
t
Spn
includes all the pseudo-negative samples self-generated from our model up to time t. l indicates
the number of pseudo-negatives generated at each round. We define a reference distribution p?
r (x) =
U (x), where U (x) is a Gaussian distribution (e.g. N (0.0, 0.32 ) independently). We carry out
learning with t = 0...T to iteratively obtain qt (y = +1|x) and qt (y = ?1|x) by updating classifier
t
C t on Set = S ? Spn
. The initial classifier C 0 on Se0 = S reports discriminative probability
q0 (y = +1|x). The reason for using q is because it is an approximation to the true p due to limited
samples drawn in Rm . At each time t, we then compute
1 qt (y = +1|x) ?
p?
p (x),
(3)
t (x) =
Zt qt (y = ?1|x) r
R (y=+1|x) ?
where Zt = qqtt (y=?1|x)
pr (x)dx. Draw new samples xi ? p?
t (x) to expand the pseudo-negative
set:
t+1
t
Spn
= Spn
? {(xi , ?1), i = n + tl + 1, ..., n + (t + 1)l}.
(4)
We name the specific training algorithm for our introspective convolutional network (ICN) classifier
reclassification-by-synthesis, which is described in Algorithm 1. We adopt convolutional neural
networks (CNN) classifier to build an end-to-end learning framework with an efficient sampling
process (to be discussed in the next section).
3.2 Reclassification-by-synthesis
We present our reclassification-by-synthesis algorithm for ICN in this section. A schematic illustration
is shown in Figure 1. A single CNN classifier is being trained progressively which is simultaneously a
discriminator and a generator. With the pseudo-negatives being gradually generated, the classification
boundary gets tightened, and hence yields an improvement to the classifier?s performance. The
reclassification-by-synthesis method is described in Algorithm 1. The key to the algorithm includes
two steps: (1) reclassification-step, and (2) synthesis-step, which will be discussed in detail below.
3.2.1 Reclassification-step
t
The reclassification-step can be viewed as training a normal classifier on the training set Set = S ?Spn
0
t
where S = {(xi , yi ), i = 1..n} and Spn = ?. Spn = {(xi , ?1), i = n + 1, ..., n + tl} for t ? 1. We
use CNN as our base classifier. When training a classifier C t on Set , we denote the parameters to be
(0)
(1)
learned in C t by a high-dimensional vector Wt = (wt , wt ) which might consist of millions of
(1)
(0)
(0)
parameters. wt denotes the weights of the top layer combining the features ?(x; wt ) and wt
4
carries all the internal representations. Without loss of generality, we assume a sigmoid function for
the discriminative probability
(1)
qt (y|x; Wt ) = 1/(1 + exp{?ywt
(0)
? ?(x; wt )}),
(0)
(1)
(0)
where ?(x; wt ) defines the feature extraction function for x. Both wt and wt can be learned by
the standard stochastic gradient descent algorithm via backpropagation to minimize a cross-entropy
loss with an additional term on the pseudo-negatives:
L(Wt ) = ?
i=1..n
X
ln qt (yi |xi ; Wt ) ?
i=n+1..n+tl
X
ln qt (?1|xi ; Wt ).
(5)
t
(xi ,?1)?Spn
(xi ,yi )?S
Algorithm 1 Outline of the reclassification-by-synthesis algorithm for discriminative classifier
training.
Input: Given a set of training data S = {(xi , yi ), i = 1..n} with x ? Rm and y ? {?1, +1}.
0
Initialization: Obtain a reference distribution: p?
r (x) = U (x) and train an initial CNN binary classifier C
0
on S, q0 (y = +1|x). Spn
= ?. U (x) is a zero mean Gaussian distribution.
For t=0..T
1 qt (y=+1|x) ?
1. Update the model: p?
t (x) = Zt qt (y=?1|x) pr (x).
2. Synthesis-step: sample l pseudo-negative samples xi ? p?
t (x), i = n + tl + 1, ..., n + (t + 1)l from the
current model p?
t (x) using an SGD sampling procedure.
t+1
t
3. Augment the pseudo-negative set with Spn
= Spn
? {(xi , ?1), i = n + tl + 1, ..., n + (t + 1)l}.
t+1
t+1
on Set+1 = S ? Spn
, resulting in qt+1 (y = +1|x).
4. Reclassification-step: Update CNN classifier to C
5. t ? t + 1 and go back to step 1 until convergence (e.g. no improvement on the validation set).
End
3.2.2 Synthesis-step
In the reclassification step, we obtain qt (y|x; Wt ) which is then used to update p?
t (x) according to
Eqn. (3):
1 qt (y = +1|x; Wt ) ?
p (x).
(6)
p?
t (x) =
Zt qt (y = ?1|x; Wt ) r
?
In the synthesis-step, our goal is to draw fair samples from pt (x) (fair samples refer to typical
samples by a sampling process after convergence w.r.t the target distribution). In [40], various Markov
chain Monte Carlo techniques [28] including Gibbs sampling and Iterated Conditional Modes (ICM)
have been adopted, which are often slow. Motivated by the DeepDream code [32] and Neural
qt (y=+1|x;Wt )
Artistic Style work [9], we update a random sample x drawn from p?
r (x) by increasing qt (y=?1|x;Wt )
using backpropagation. Note that the partition function (normalization) Zt is a constant that is not
dependent on the sample x. Let
gt (x) =
qt (y = +1|x; Wt )
(1)
(0)
= exp{wt ? ?(x; wt )},
qt (y = ?1|x; Wt )
(7)
and take its ln, which is nicely turned into the logit of qt (y = +1|x; Wt )
(1)
ln gt (x) = wt
(0)
? ?(x; wt ).
(1)T
(8)
(0)
Starting from x drawn from p?
?(x; wt ) using stochastic gradient
r (x), we directly increase wt
ascent on x via backpropagation, which allows us to obtain fair samples subject to Eqn. (6). Gaussian
noise can be added to Eqn. (8) along the line of stochastic gradient Langevin dynamics [43] as
(1)
(0)
?x = ?(wt ? ?(x; wt )) + ?
2
where ? ? N (0, ) is a Gaussian distribution and is the step size that is annealed in the sampling
process.
Sampling strategies. When conducting experiments, we carry out several strategies using stochastic
gradient descent algorithm (SGD) and SGD Lagenvin including: i) early-stopping for the sampling
process after x becomes positive (aligned with contrastive divergence [4] where a short Markov chain
is simulated); ii) stopping at a large confidence for x being positive, and iii) sampling for a fixed,
large number of steps. Table 2 shows the results on these different options and no major differences
in the classification performance are observed.
5
Building connections between SGD and MCMC is an active area in machine learning [43, 5, 30]. In
[43], combining SGD and additional Gaussian noise under annealed stepsize results in a simulation
of Langevin dynamics MCMC. A recent work [30] further shows the similarity between constant
SGD and MCMC, along with analysis of SGD using momentum updates. Our progressively learned
discriminative classifier can be viewed as carving out the feature space on ?(x), which essentially
becomes an equivalent class for the positives; the volume of the equivalent class that satisfies
the condition is exponentially large, as analyzed in [46]. The probability landscape of positives
(equivalent class) makes our SGD sampling process not particularly biased towards a small limited
modes. Results in Figure 2 illustrates that large variation of the sampled/synthesized examples.
3.3 Analysis
t=? +
The convergence of p?
t (x) ? p (x) can be derived (see the supplementary material), inspired by
?
+
+
the proof from [40]: KL[p (x)||p?
t+1 (x)] ? KL[p (x)||pt (x)] where KL denotes the Kullback+
Leibler divergence and p(x|y = +1) ? p (x), under the assumption that classifier at t + 1 improves
over t.
Remark. Here we pay particular attention to the negative samples which live in a space that is
often much larger than the positive sample space. For the negative training samples, we have
yi = ?1 and xi ? Q? (x), where Q? (x) is a distribution on the given negative examples in
the original training set. Our reclassification-by-synthesis algorithm (Algorithm 1) essentially
PT ?1
constructs a mixture model p?(x) ? T1 t=0 p?
t (x) by sequentially generating pseudo-negative
samples to augment our training set. Our new distribution for augmented negative sample set thus
n
Tl
?
becomes Q?
?(x), where p?(x) encodes pseudo-negative samples that
new (x) ? n+T l Q (x) + n+T l p
are confusing and similar to (but are not) the positives. In the end, adding pseudo-negatives might
degrade the classification result since they become more and more similar to the positives. Crossvalidation can be used to decide when adding more pseudo-negatives is not helping the classification
task. How to better use the pseudo-negative samples that are increasingly faithful to the positives
is an interesting topic worth further exploring. Our overall algorithm thus is capable of enhancing
classification by self-generating confusing samples to improve CNN?s robustness.
3.4 Multi-class classification
One-vs-all. In the above section, we discussed the binary classification case. When dealing with
multi-class classification problems, such as MNIST and CIFAR-10, we will need to adapt our
proposed reclassification-by-synthesis scheme to the multi-class case. This can be done directly using
a one-vs-all strategy by training a binary classifier Ci using the i-th class as the positive class and
then combine the rest classes into the negative class, resulting in a total of K binary classifiers. The
training procedure then becomes identical to the binary classification case. If we have K classes,
then the algorithm will train K individual binary classifiers with
(0)1
< (wt
The prediction function is simply
(1)1
, wt
(0)K
), ..., (wt
(1)k
f (x) = arg max exp{wt
k
(1)K
, wt
)>.
(0)k
? ?(x; wt
)}.
The advantage of using the one-vs-all strategy is that the algorithm can be made nearly identical to
the binary case at the price of training K different neural networks.
Softmax function. It is also desirable to build a single CNN classifier to perform multi-class
classification directly. Here we propose a formulation to train an end-to-end multiclass classifier
directly. Since we are directly dealing with K classes, the pseudo-negative data set will be slightly
0
different and we introduce negatives for each individual class by Spn
= ? and:
t
Spn
= {(xi , ?k), k = 1, ..., K, i = n + (t ? 1) ? k ? l + 1, ..., n + t ? k ? l}
Suppose we are given a training set S = {(xi , yi ), i = 1..n} and x ? Rm and y ? {1, .., K}. We
want to train a single CNN classifier with
(0)
(1)1
Wt =< wt , wt
(1)K
, ..., wt
>
(0)
wt
(1)
where
denotes the internal feature and parameters for the single CNN, and wt k denotes the
top-layer weights for the k-th class. We therefore minimize an integrated objective function
L(Wt )=?(1??)
Pn
i=1
ln
(1)y
i ??(x ;w(0) )}
P
exp{wt
i
t
+? n+t?K?l
PK
i=n+1
(1)
(0)
exp{wt k ??(xi ;wt )}
k=1
6
(1)|y |
i ??(x ;w0 )})
i
t
ln(1+exp{wt
(9)
The first term in Eqn. (9) encourages a softmax loss on the original training set S. The second term
in Eqn. (9) encourages a good prediction on the individual pseudo-negative class generated for the
(1)|y |
k-th class (indexed by |yi | for wt i , e.g. for pseudo-negative samples belong to the k-th class,
|yi | = | ? k| = k). ? is a hyperparameter balancing the two terms. Note that we only need to build
(0)
a single CNN sharing wt for all the K classes. In particular, we are not introducing additional
model parameters here and we perform a direct K-class classification where the parameter setting is
identical to a standard CNN multi-class classification task; to compare, an additional ?not-real? class
is created in [36] and the classification task there [36] thus becomes a K + 1 class classification.
4
Experiments
Figure 2: Synthesized pseudo-negatives for the MNIST dataset by our ICN classifier. The top row shows some
training examples. As t increases, our classifier gradually synthesize pseudo-negative samples that become
increasingly faithful to the training samples.
We conduct experiments on three standard benchmark datasets, including MNIST, CIFAR-10 and
SVHN. We use MNIST as a running example to illustrate our proposed framework using a shallow
CNN; we then show competitive results using a state-of-the-art CNN classifier, ResNet [12] on
MNIST, CIFAR-10 and SVHN. In our experiments, for the reclassification step, we use the SGD
optimizer with mini-batch size of 64 (MNIST) or 128 (CIFAR-10 and SVHN) and momentum equal
to 0.9; for the synthesis step, we use the Adam optimizer [17] with momentum term ?1 equal to 0.5.
All results are obtained by averaging multiple rounds.
Training and test time. In general, the training time for ICN is around double that of the baseline
CNNs in our experiments: 1.8 times for MNIST dataset, 2.1 times for CIFAR-10 dataset and 1.7
times for SVHN dataset. The added overhead in training is mostly determined by the number of
generated pseudo-negative samples. For the test time, ICN introduces no additional overhead to the
baseline CNNs.
4.1 MNIST
We use the standard MNIST [24] dataset, which
consists of 55, 000 training, 5, 000 validation
and 10, 000 test samples. We adopt a simple
network, containing 4 convolutional layers, each
having a 5 ? 5 filter size with 64, 128, 256 and
512 channels, respectively. These convolutional
layers have stride 2, and no pooling layers are
used. LeakyReLU activations [29] are used after
each convolutional layer. The last convolutional
layer is flattened and fed into a sigmoid output
(in the one-vs-all case).
Table 1: Test errors on the MNIST dataset. We compare
our ICN method with the baseline CNN, Deep Belief
Network (DBN) [14], and CNN w/ Label Smoothing
(LS) [39]. Moreover, the two-step experiments combining CNN + GDL [40] and combining CNN + DCGAN
[35] are also reported, and see descriptions in text for
more details.
Method
One-vs-all (%) Softmax (%)
DBN
1.11
CNN (baseline)
0.87
0.77
CNN w/ LS
0.69
CNN + GDL
0.85
CNN + DCGAN
0.84
In the reclassification step, we run SGD (for 5
ICN-noise (ours)
0.89
0.77
t
epochs) on the current training data Se , includICN (ours)
0.78
0.72
ing previously generated pseudo-negatives. Our
initial learning rate is 0.025 and is decreased by
a factor of 10 at t = 25. In the synthesis step, we use the backpropagation sampling process as
discussed in Section 3.2.2. In Table 2, we compare different sampling strategies. Each time we
synthesize a fixed number (200 in our experiments) of pseudo-negative samples.
We show some synthesized pseudo-negatives from the MNIST dataset in Figure 2. The samples in
the top row are from the original training dataset. ICN gradually synthesizes pseudo-negatives, which
are increasingly faithful to the original data. Pseudo-negative samples will be continuously used
while improving the classification result.
7
Comparison of different sampling Table 2: Comparison of different sampling strategies in the
strategies. We perform SGD and SGD synthesis step in ICN.
Langevin (with injected Gaussians), and
Sampling Strategy
One-vs-all (%) Softmax (%)
try several options via backpropagation
SGD (option 1)
0.81
0.72
for the sampling strategies. Option 1:
SGD Langevin (option 1)
0.80
0.72
0.78
0.72
early-stopping once the generated sam- SGD (option 2)
SGD Langevin (option 2)
0.78
0.74
ples are classified as positive; option 2:
0.81
0.75
stopping at a high confidence for sam- SGD (option 3)
SGD
Langevin
(option
3)
0.80
0.73
ples being positive; option 3: stopping
after a large number of steps. Table 2
shows the results and we do not observe significant differences in these choices.
Ablation study. We experiment using random noise as synthesized pseudo-negatives in an ablation
study. From Table 1, we observe that our ICN outperforms the CNN baseline and the ICN-noise
method in both one-vs-all and softmax cases.
Figure 3: MNIST test error against the number of training examples (std dev. of the test error is also displayed).
The effect of ICN is more clear when having fewer training examples.
Effects on varying training sizes. To better understand the effectiveness of our ICN method, we
carry out an experiment by varying the number of training examples. We use training sets with
different sizes including 500, 2000, 10000, and 55000 examples. The results are reported in Figure 3.
ICN is shown to be particularly effective when the training set is relatively small, since ICN has the
capability to synthesize pseudo-negatives by itself to aid training.
Comparison with GDL and GAN. GDL [40] focuses on unsupervised learning; GAN [10] and
DCGAN [35] show results for unsupervised learning and semi-supervised classification. To apply
GDL and GAN to the supervised classification setting, we design an experiment to perform a two-step
implementation. For GDL, we ran the GDL code [40] and obtained the pseudo-negative samples
for each individual digit; the pseudo-negatives are then used as augmented negative samples to
train individual one-vs-all CNN classifiers (using an identical CNN architecture to ICN for a fair
comparison), which are combined to form a multi-class classifier in the end. To compare with
DCGAN [35], we follow the same procedure: each generator trained by DCGAN [35] using the
TensorFlow implementation [16] was used to generate positive samples, which are then augmented
to the negative set to train the individual one-vs-all CNN classifiers (also using an identical CNN
architecture to ICN), which are combined to create the overall multi-class classifier. CNN+GDL
achieves a test error of 0.85% and CNN+DCGAN achieves a test error of 0.84% on the MNIST
dataset, whereas ICN reports an error of 0.78% using the same CNN architecture. As the supervised
learning task was not directly specified in DCGAN [35], some care is needed to design the optimal
setting to utilize the generated samples from DCGAN in the two-step approach (we made attempts to
optimize the results). GDL [40] can be made into a discriminative classifier by utilizing the given
negative samples first but boosting [7] with manually designed features was adopted which may not
produce competitive results as CNN classifier does. Nevertheless, the advantage of ICN being an
integrated end-to-end supervised learning single-model framework can be observed.
To compare with generative model based deep learning approach, we report the classification result
of DBN [14] in Table 1. DBN achieves a test error of 1.11% using the softmax function. We also
compare with Label Smoothing (LS), which has been used in [39] as a regularization technique by
encouraging the model to be less confident. In LS, for a training example with ground-truth label,
the label distribution is replaced with a mixture of the original ground-truth distribution and a fixed
distribution. LS achieves a test error of 0.69% in the softmax case.
8
In addition, we also adopt ResNet-32 [13] (using the softmax function) as another baseline CNN
model, which achieves a test error of 0.50% on the MNIST dataset. Our ResNet-32 based ICN
achieves an improved result of 0.47%.
Robustness to external adversarial examples. To show the improved robustness of ICN in dealing
with confusing and challenging examples, we compare the baseline CNN with our ICN classifier on
adversarial examples generated using the ?fast gradient sign? method from [11]. This ?fast gradient
sign? method (with = 0.25) can cause a maxout network to misclassify 89.4% of adversarial
examples generated from the MNIST test set [11]. In our experiment, we set = 0.125. Starting
with 10, 000 MNIST test examples, we first determine those which are correctly classified by the
baseline CNN in order to generate adversarial examples from them. We find that 5, 111 generated
adversarial examples successfully fool the baseline CNN, however, only 3, 134 of these examples
can fool our ICN classifier, which is a 38.7% reduction in error against adversarial examples. Note
that the improvement is achieved without using any additional training data, nor knowing a prior
about how these adversarial examples are generated by the specific ?fast gradient sign method? [11].
On the contrary, of the 2, 679 adversarial examples generated from the ICN classifier side that fool
ICN using the same method, 2, 079 of them can still fool the baseline CNN classifier. This two-way
experiment shows the improved robustness of ICN over the baseline CNN.
4.2
CIFAR-10
Table 3: Test errors on the CIFAR-10 dataset. In both oneThe CIFAR-10 dataset [18] consists of vs-all and softmax cases, ICN shows improvement over the
60, 000 color images of size 32 ? 32. This baseline ResNet model. The result of convolutional DBN is
set of 60, 000 images is split into two sets, from [19].
50, 000 images for training and 10, 000 images for testing. We adopt ResNet [13] as
Method
One-vs-all (%) Softmax (%)
our baseline model [45]. For data augmen- w/o Data Augmentation
Convolutional DBN
21.1
tation, we follow the standard procedure
ResNet-32 (baseline)
13.44
12.38
in [26, 25, 13] by augmenting the dataset
12.65
ResNet-32 w/ LS
by zero-padding 4 pixels on each side; we
ResNet-32
+
DCGAN
12.99
also perform cropping and random flipping.
ICN-noise (ours)
13.28
11.94
The results are reported in Table 3. In
12.94
11.46
ICN (ours)
both one-vs-all and softmax cases, ICN
w/
Data
Augmentation
outperforms the baseline ResNet classifiers.
ResNet-32 (baseline)
6.70
7.06
Our proposed ICN method is orthogonal to
ResNet-32 w/ LS
6.89
many existing approaches which use vari6.75
ResNet-32 + DCGAN
ous improvements to the network structures
ICN-noise (ours)
6.58
6.90
in order to enhance the CNN performance.
ICN (ours)
6.52
6.70
We also compare ICN with Convolutional
DBN [19], ResNet-32 w/ Label Smoothing (LS) [39] and ResNet-32+DCGAN [35] methods as
described in the MNIST experiments. LS is shown to improve the baseline but is worse than our ICN
method in most cases except for the MNIST dataset.
4.3
SVHN
Table 4: Test errors on the SVHN dataset.
We use the standard SVHN [33] dataset. We combine the
training data with the extra data to form our training set
and use the test data as the test set. No data augmentation
has been applied. The result is reported in Table 4. ICN is
shown to achieve competitive results.
5
Method
ResNet-32 (baseline)
ResNet-32 w/ LS
ResNet-32 + DCGAN
ICN-noise (ours)
ICN (ours)
Softmax (%)
2.01
1.96
1.98
1.99
1.95
Conclusion
In this paper, we have proposed an introspective convolutional nets (ICN) algorithm that performs
internal introspection. We observe performance gains within supervised learning using state-of-the-art
CNN architectures on standard machine learning benchmarks.
Acknowledgement This work is supported by NSF IIS-1618477, NSF IIS-1717431, and a Northrop
Grumman Contextual Robotics grant. We thank Saining Xie, Weijian Xu, Fan Fan, Kwonjoon Lee,
Shuai Tang, and Sanjoy Dasgupta for helpful discussions.
9
References
[1] P. Baldi. Autoencoders, unsupervised learning, and deep architectures. In ICML Workshop on
Unsupervised and Transfer Learning, pages 37?49, 2012.
[2] L. Breiman. Random forests. Machine learning, 45(1):5?32, 2001.
[3] A. Brock, T. Lim, J. Ritchie, and N. Weston. Neural photo editing with introspective adversarial
networks. In ICLR, 2017.
[4] M. A. Carreira-Perpinan and G. Hinton. On contrastive divergence learning. In AISTATS,
volume 10, pages 33?40, 2005.
[5] T. Chen, E. B. Fox, and C. Guestrin. Stochastic gradient hamiltonian monte carlo. In ICML,
2014.
[6] S. Della Pietra, V. Della Pietra, and J. Lafferty. Inducing features of random fields. IEEE
transactions on pattern analysis and machine intelligence, 19(4):380?393, 1997.
[7] Y. Freund and R. E. Schapire. A Decision-theoretic Generalization of On-line Learning And
An Application to Boosting. Journal of computer and system sciences, 55(1):119?139, 1997.
[8] J. Friedman, T. Hastie, and R. Tibshirani. The elements of statistical learning, volume 1.
Springer series in statistics Springer, Berlin, 2001.
[9] L. A. Gatys, A. S. Ecker, and M. Bethge. A neural algorithm of artistic style. arXiv preprint
arXiv:1508.06576, 2015.
[10] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and
Y. Bengio. Generative adversarial nets. In NIPS, 2014.
[11] I. J. Goodfellow, J. Shlens, and C. Szegedy. Explaining and harnessing adversarial examples. In
ICLR, 2015.
[12] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. In CVPR,
2016.
[13] K. He, X. Zhang, S. Ren, and J. Sun. Identity mappings in deep residual networks. In European
Conference on Computer Vision, pages 630?645. Springer, 2016.
[14] G. E. Hinton, S. Osindero, and Y.-W. Teh. A fast learning algorithm for deep belief nets. Neural
computation, 18(7):1527?1554, 2006.
[15] T. Jebara. Machine learning: discriminative and generative, volume 755. Springer Science &
Business Media, 2012.
[16] T. Kim. DCGAN-tensorflow. https://github.com/carpedm20/DCGAN-tensorflow.
[17] D. Kingma and J. Ba. Adam: A method for stochastic optimization. In ICLR, 2015.
[18] A. Krizhevsky. Learning Multiple Layers of Features from Tiny Images. CS Dept., U Toronto,
Tech. Rep., 2009.
[19] A. Krizhevsky and G. Hinton. Convolutional deep belief networks on cifar-10. Unpublished
manuscript, 40, 2010.
[20] A. Krizhevsky, I. Sutskever, and G. E. Hinton. ImageNet Classification with Deep Convolutional
Neural Networks. In NIPS, 2012.
[21] J. Lazarow, L. Jin, and Z. Tu. Introspective neural networks for generative modeling. In ICCV,
2017.
[22] D. B. Leake. Introspective learning and reasoning. In Encyclopedia of the Sciences of Learning,
pages 1638?1640. Springer, 2012.
[23] Y. LeCun, B. Boser, J. S. Denker, D. Henderson, R. Howard, W. Hubbard, and L. Jackel.
Backpropagation applied to handwritten zip code recognition. In Neural Computation, 1989.
[24] Y. LeCun and C. Cortes. The MNIST database of handwritten digits, 1998.
[25] C.-Y. Lee, P. W. Gallagher, and Z. Tu. Generalizing pooling functions in convolutional neural
networks: Mixed, gated, and tree. In AISTATS, 2016.
[26] C.-Y. Lee, S. Xie, P. Gallagher, Z. Zhang, and Z. Tu. Deeply-supervised nets. In AISTATS,
2015.
[27] P. Liang and M. I. Jordan. An asymptotic analysis of generative, discriminative, and pseudolikelihood estimators. In ICML, 2008.
10
[28] J. S. Liu. Monte Carlo strategies in scientific computing. Springer Science & Business Media,
2008.
[29] A. L. Maas, A. Y. Hannun, and A. Y. Ng. Rectifier nonlinearities improve neural network
acoustic models. In ICML, 2013.
[30] S. Mandt, M. D. Hoffman, and D. M. Blei. Stochastic gradient descent as approximate bayesian
inference. arXiv preprint arXiv:1704.04289, 2017.
[31] C. Z. Mooney, R. D. Duval, and R. Duvall. Bootstrapping: A nonparametric approach to
statistical inference. Number 94-95. Sage, 1993.
[32] A. Mordvintsev, C. Olah, and M. Tyka. Deepdream - a code example for visualizing neural
networks. Google Research, 2015.
[33] Y. Netzer, T. Wang, A. Coates, A. Bissacco, B. Wu, and A. Y. Ng. Reading Digits in Natural
Images with Unsupervised Feature Learning. In NIPS Workshop on Deep Learning and
Unsupervised Feature Learning, 2011.
[34] J. R. Quinlan. Improved use of continuous attributes in c4. 5. Journal of artificial intelligence
research, 4:77?90, 1996.
[35] A. Radford, L. Metz, and S. Chintala. Unsupervised representation learning with deep convolutional generative adversarial networks. In ICLR, 2016.
[36] T. Salimans, I. Goodfellow, W. Zaremba, V. Cheung, A. Radford, and X. Chen. Improved
techniques for training gans. In NIPS, 2016.
[37] B. Settles. Active learning literature survey. University of Wisconsin, Madison, 52(55-66):11,
2010.
[38] A. Sinha, M. Sarkar, A. Mukherjee, and B. Krishnamurthy. Introspection: Accelerating neural
network training by learning weight evolution. arXiv preprint arXiv:1704.04959, 2017.
[39] C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, and Z. Wojna. Rethinking the inception architecture for computer vision. In CVPR, 2016.
[40] Z. Tu. Learning generative models via discriminative approaches. In CVPR, 2007.
[41] Z. Tu, K. L. Narr, P. Doll?r, I. Dinov, P. M. Thompson, and A. W. Toga. Brain anatomical
structure segmentation by hybrid discriminative/generative models. Medical Imaging, IEEE
Transactions on, 27(4):495?508, 2008.
[42] V. N. Vapnik. The nature of statistical learning theory. Springer-Verlag New York, Inc., 1995.
[43] M. Welling and Y. W. Teh. Bayesian learning via stochastic gradient langevin dynamics. In
ICML, 2011.
[44] M. Welling, R. S. Zemel, and G. E. Hinton. Self supervised boosting. In NIPS, 2002.
[45] Y. Wu. Tensorpack toolbox. https://github.com/ppwwyyxx/tensorpack/tree/
master/examples/ResNet.
[46] Y. N. Wu, S. C. Zhu, and X. Liu. Equivalence of julesz ensembles and frame models. International Journal of Computer Vision, 38(3), 2000.
[47] J. Xie, Y. Lu, S.-C. Zhu, and Y. N. Wu. Cooperative training of descriptor and generator
networks. arXiv preprint arXiv:1609.09408, 2016.
[48] J. Xie, Y. Lu, S.-C. Zhu, and Y. N. Wu. A theory of generative convnet. In ICML, 2016.
[49] J. Zhao, M. Mathieu, and Y. LeCun. Energy-based generative adversarial network. In ICLR,
2017.
[50] S. C. Zhu, Y. N. Wu, and D. Mumford. Minimax entropy principle and its application to texture
modeling. Neural Computation, 9(8):1627?1660, 1997.
[51] X. Zhu. Semi-supervised learning literature survey. Computer Science, University of WisconsinMadison, Technical Report 1530, 2005.
11
| 6684 |@word cnn:52 seems:1 logit:1 simulation:1 contrastive:2 sgd:20 carry:5 reduction:1 initial:4 liu:2 series:1 ours:8 past:2 existing:3 outperforms:2 current:2 icn:69 comparing:1 contextual:1 com:2 stemmed:2 activation:1 dx:1 partition:1 informative:1 grumman:1 designed:2 plot:1 progressively:4 update:5 v:21 generative:34 fewer:1 intelligence:2 hamiltonian:1 short:1 bissacco:1 blei:1 boosting:9 toronto:1 zhang:3 along:2 olah:1 direct:3 become:3 consists:3 combine:2 overhead:2 baldi:1 introduce:2 nor:1 gatys:1 multi:14 brain:1 inspired:3 encouraging:1 increasing:1 becomes:5 notation:1 underlying:1 moreover:1 medium:2 developed:1 bootstrapping:2 pseudo:36 absurd:2 zaremba:1 classifier:61 rm:5 internally:1 grant:1 medical:1 engages:1 positive:22 t1:1 mistake:1 tation:1 mandt:1 northrop:1 might:2 initialization:1 equivalence:1 challenging:1 limited:4 carving:1 faithful:3 lecun:3 testing:1 implement:1 backpropagation:8 digit:3 procedure:5 area:1 cascade:2 revealing:1 attain:1 confidence:2 get:1 close:1 live:1 optimize:2 equivalent:3 ecker:1 annealed:2 go:1 attention:1 starting:2 independently:1 l:10 focused:2 survey:2 thompson:1 pouget:1 rule:1 estimator:1 utilizing:1 borrow:1 shlens:2 variation:1 krishnamurthy:1 updated:1 diego:3 suppose:2 pt:3 target:1 goodfellow:3 synthesize:7 element:1 recognition:2 particularly:2 utilized:1 updating:2 std:1 mukherjee:1 database:1 cooperative:1 bottom:1 observed:2 preprint:4 wang:1 capture:1 sun:2 ran:1 deeply:1 warde:1 dynamic:3 trained:3 dilemma:2 various:2 train:11 distinct:1 fast:4 effective:1 monte:3 artificial:1 zemel:1 harnessing:1 supplementary:1 larger:1 cvpr:3 encoder:1 statistic:1 unseen:1 itself:4 final:1 sequence:2 advantage:2 net:5 propose:4 tu:6 turned:1 combining:5 aligned:1 ablation:2 achieve:2 representational:1 description:1 inducing:2 crossvalidation:1 sutskever:1 convergence:3 double:1 requirement:1 cropping:1 produce:1 generating:3 adam:2 resnet:19 help:2 illustrate:1 develop:1 augmenting:3 qt:18 coverage:1 c:1 attribute:1 cnns:3 stochastic:9 filter:1 human:3 settle:1 material:1 generalization:2 exploring:1 helping:1 around:1 ground:2 normal:1 exp:6 great:1 scope:1 mapping:1 major:1 achieves:6 optimizer:2 adopt:4 early:2 estimation:1 label:6 jackel:1 hubbard:1 faithfully:1 create:1 successfully:1 hoffman:1 always:1 gaussian:5 aim:2 pn:1 breiman:1 varying:2 derived:1 focus:6 improvement:9 indicates:1 tech:1 adversarial:19 baseline:18 kim:1 helpful:1 inference:2 dependent:2 stopping:5 entire:1 integrated:2 expand:2 pixel:1 overall:2 classification:49 arg:1 augment:3 development:1 art:4 softmax:13 special:1 uc:3 smoothing:3 field:1 equal:3 construct:1 having:4 beach:1 sampling:22 manually:3 extraction:1 nicely:1 identical:5 look:1 unsupervised:12 nearly:1 icml:6 introspection:4 report:5 mirza:1 employ:1 modern:1 simultaneously:3 divergence:3 individual:6 pietra:2 replaced:1 attempt:3 friedman:1 misclassify:1 interest:1 once:1 henderson:1 introduces:1 analyzed:1 mixture:2 farley:1 chain:2 accurate:1 capable:1 cifar10:1 netzer:1 orthogonal:1 fox:1 conduct:2 tree:3 indexed:1 ples:2 circle:1 sinha:1 classify:1 empowered:1 modeling:10 dev:1 retains:1 artistic:2 introducing:2 krizhevsky:3 osindero:1 reported:4 combined:2 confident:1 st:1 density:1 international:1 lee:3 enhance:2 bethge:1 synthesis:27 connecting:1 continuously:1 gans:2 augmentation:5 containing:1 worse:1 external:2 zhao:1 style:2 szegedy:2 nonlinearities:1 stride:1 ywt:1 summarized:1 includes:2 inc:1 explicitly:1 toga:1 later:1 try:2 red:1 competitive:4 bayes:3 maintains:2 capability:3 start:1 option:11 metz:1 minimize:2 purple:2 convolutional:26 descriptor:1 largely:1 conducting:1 ensemble:1 yield:1 landscape:1 weak:2 handwritten:2 bayesian:2 iterated:1 emphasizes:1 ren:2 carlo:3 lu:2 worth:2 mooney:1 classified:2 ztu:1 sharing:1 against:2 energy:1 gdl:19 proof:1 chintala:1 gain:2 sampled:1 dataset:16 color:1 lim:1 improves:1 segmentation:1 back:1 manuscript:1 originally:1 supervised:22 follow:2 xie:4 wisconsinmadison:1 improved:7 editing:1 formulation:6 done:2 generality:1 inception:1 shuai:1 until:1 autoencoders:1 eqn:6 google:1 defines:1 mode:2 quality:1 reveal:1 scientific:1 usa:1 building:3 name:1 concept:1 true:1 effect:2 evolution:1 hence:1 regularization:1 q0:2 iteratively:3 leibler:1 round:2 visualizing:1 self:10 encourages:2 outline:1 theoretic:1 leakyrelu:1 performs:1 svhn:8 reasoning:1 image:9 variational:1 common:1 endowing:1 sigmoid:2 exponentially:1 volume:4 million:1 discussed:4 belong:1 he:2 synthesized:4 refer:1 significant:1 gibbs:3 ritchie:1 automatic:1 dbn:7 carpedm20:1 similarity:1 gt:2 add:1 base:1 own:3 recent:7 verlag:1 binary:9 success:2 rep:1 accomplished:1 yi:11 ous:1 seen:1 guestrin:1 greater:1 additional:9 care:1 zip:1 determine:2 semi:5 ii:3 multiple:2 desirable:1 ing:1 technical:1 adapt:1 cross:4 long:2 cifar:9 devised:2 se0:1 schematic:2 prediction:2 basic:1 essentially:2 enhancing:1 vision:3 arxiv:8 represent:1 normalization:1 achieved:2 robotics:1 whereas:9 want:1 addition:1 addressed:1 decreased:1 biased:1 rest:1 extra:1 ascent:1 subject:1 pooling:2 contrary:1 lafferty:1 effectiveness:1 jordan:1 call:1 zhuowen:1 iii:1 split:1 bengio:1 architecture:7 hastie:1 knowing:1 multiclass:1 motivated:2 accelerating:1 padding:1 effort:1 york:1 cause:1 remark:1 deep:12 weijian:1 fake:1 se:1 clear:1 fool:4 julesz:1 amount:1 nonparametric:1 encyclopedia:1 category:1 generate:3 schapire:1 http:2 exist:1 nsf:2 coates:1 sign:3 correctly:1 tibshirani:1 blue:1 anatomical:1 hyperparameter:1 dasgupta:1 key:1 nevertheless:1 drawn:6 clarity:1 utilize:2 imaging:1 run:1 powerful:1 injected:1 ameliorate:1 master:1 family:2 decide:1 wu:6 draw:2 decision:5 confusing:4 layer:8 pay:1 distinguish:1 tackled:1 courville:1 fan:2 duval:1 encodes:1 aspect:4 relatively:1 according:1 combination:1 slightly:1 increasingly:3 sam:2 shallow:1 iccv:1 gradually:4 pr:2 pipeline:2 taken:1 ln:6 previously:1 hannun:1 discus:1 needed:1 fed:1 end:12 photo:1 adopted:3 generalizes:1 gaussians:1 doll:1 apply:1 observe:4 denker:1 salimans:1 generic:1 stepsize:1 batch:1 robustness:4 original:5 top:8 denotes:4 running:1 gan:14 maintaining:1 quinlan:1 madison:1 spn:17 build:5 especially:1 objective:3 already:1 added:2 flipping:1 mumford:1 strategy:12 primary:1 gradient:11 iclr:5 convnet:1 separate:3 thank:1 simulated:1 berlin:1 rethinking:1 w0:1 degrade:1 topic:1 argue:1 reason:2 ozair:1 assuming:1 code:4 relationship:1 illustration:2 insufficient:1 mini:1 liang:1 mostly:3 negative:53 sage:1 synthesizing:1 ba:1 wojna:1 design:2 implementation:2 motivates:1 zt:5 perform:8 teh:2 gated:1 observation:2 datasets:4 markov:2 howard:1 benchmark:5 jin:2 descent:4 displayed:1 introspective:13 immediate:1 langevin:7 hinton:5 frame:1 ucsd:3 perturbation:1 jebara:1 sarkar:1 introduced:1 namely:1 unpublished:1 specified:2 kl:3 connection:3 discriminator:8 faithfulness:1 imagenet:1 c4:1 acoustic:1 toolbox:1 learned:5 boser:1 tensorflow:3 kingma:1 nip:6 justin:1 able:2 alongside:1 below:3 pattern:1 saining:1 reading:1 including:7 max:1 belief:3 power:3 business:2 hybrid:2 natural:1 residual:2 tyka:1 zhu:5 minimax:4 scheme:2 improve:8 github:2 mathieu:1 created:3 auto:1 brock:1 text:1 prior:2 epoch:1 acknowledgement:1 literature:2 asymptotic:1 wisconsin:1 freund:1 loss:3 mixed:1 interesting:3 generation:1 limitation:1 generator:9 validation:4 vanhoucke:1 reclassification:21 consistent:1 principle:1 tightened:1 tiny:1 share:1 balancing:1 row:2 maas:1 supported:1 last:1 guide:1 allow:1 understand:1 side:2 pseudolikelihood:1 fall:1 explaining:1 benefit:1 boundary:4 computes:1 adopts:1 made:5 san:3 simplified:1 tighten:1 welling:2 transaction:2 approximate:1 emphasize:1 kullback:1 dealing:3 active:3 sequentially:2 ioffe:1 discriminative:27 xi:19 continuous:1 iterative:1 table:11 additionally:1 scratch:1 nature:1 learn:2 channel:1 ca:1 transfer:1 obtaining:1 improving:3 forest:2 synthesizes:1 european:1 aistats:3 significance:1 main:1 pk:1 noise:8 fair:4 icm:1 xu:2 augmented:3 tl:7 ng:2 slow:1 aid:1 momentum:3 perpinan:1 tang:1 specific:2 rectifier:1 abadie:1 cortes:1 consist:1 workshop:2 mnist:20 vapnik:1 adding:5 importance:2 ci:1 flattened:1 texture:1 gallagher:2 illustrates:1 chen:2 entropy:3 generalizing:1 simply:1 dcgan:14 springer:7 radford:2 truth:2 satisfies:1 relies:1 weston:1 conditional:1 goal:3 viewed:2 identity:1 mordvintsev:1 cheung:1 towards:1 maxout:1 price:1 hard:2 carreira:1 typical:1 determined:1 reducing:1 except:1 wt:49 averaging:1 total:1 sanjoy:1 pas:1 indicating:1 internal:4 support:1 meant:1 dept:1 mcmc:3 della:2 |
6,284 | 6,685 | Label Distribution Learning Forests
Wei Shen1,2 , Kai Zhao1 , Yilu Guo1 , Alan Yuille2
Key Laboratory of Specialty Fiber Optics and Optical Access Networks,
Shanghai Institute for Advanced Communication and Data Science,
School of Communication and Information Engineering, Shanghai University
2
Department of Computer Science, Johns Hopkins University
1
{shenwei1231,zhaok1206,gyl.luan0,alan.l.yuille}@gmail.com
Abstract
Label distribution learning (LDL) is a general learning framework, which assigns
to an instance a distribution over a set of labels rather than a single label or multiple
labels. Current LDL methods have either restricted assumptions on the expression
form of the label distribution or limitations in representation learning, e.g., to
learn deep features in an end-to-end manner. This paper presents label distribution
learning forests (LDLFs) - a novel label distribution learning algorithm based on
differentiable decision trees, which have several advantages: 1) Decision trees
have the potential to model any general form of label distributions by a mixture
of leaf node predictions. 2) The learning of differentiable decision trees can be
combined with representation learning. We define a distribution-based loss function
for a forest, enabling all the trees to be learned jointly, and show that an update
function for leaf node predictions, which guarantees a strict decrease of the loss
function, can be derived by variational bounding. The effectiveness of the proposed
LDLFs is verified on several LDL tasks and a computer vision application, showing
significant improvements to the state-of-the-art LDL methods.
1
Introduction
Label distribution learning (LDL) [6, 11] is a learning framework to deal with problems of label
ambiguity. Unlike single-label learning (SLL) and multi-label learning (MLL) [26], which assume an
instance is assigned to a single label or multiple labels, LDL aims at learning the relative importance
of each label involved in the description of an instance, i.e., a distribution over the set of labels. Such
a learning strategy is suitable for many real-world problems, which have label ambiguity. An example
is facial age estimation [8]. Even humans cannot predict the precise age from a single facial image.
They may say that the person is probably in one age group and less likely to be in another. Hence it is
more natural to assign a distribution of age labels to each facial image (Fig. 1(a)) instead of using a
single age label. Another example is movie rating prediction [7]. Many famous movie review web
sites, such as Netflix, IMDb and Douban, provide a crowd opinion for each movie specified by the
distribution of ratings collected from their users (Fig. 1(b)). If a system could precisely predict such a
rating distribution for every movie before it is released, movie producers can reduce their investment
risk and the audience can better choose which movies to watch.
Many LDL methods assume the label distribution can be represented by a maximum entropy model [2]
and learn it by optimizing an energy function based on the model [8, 11, 28, 6]. But, the exponential
part of this model restricts the generality of the distribution form, e.g., it has difficulty in representing
mixture distributions. Some other LDL methods extend the existing learning algorithms, e.g, by
boosting and support vector regression, to deal with label distributions [7, 27], which avoid making
this assumption, but have limitations in representation learning, e.g., they do not learn deep features
in an end-to-end manner.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
Figure 1: The real-world data which are suitable to be modeled by label distribution learning. (a)
Estimated facial ages (a unimodal distribution). (b) Rating distribution of crowd opinion on a movie
(a multimodal distribution).
In this paper, we present label distribution learning forests (LDLFs) - a novel label distribution
learning algorithm inspired by differentiable decision trees [20]. Extending differentiable decision
trees to deal with the LDL task has two advantages. One is that decision trees have the potential
to model any general form of label distributions by mixture of the leaf node predictions, which
avoid making strong assumption on the form of the label distributions. The second is that the split
node parameters in differentiable decision trees can be learned by back-propagation, which enables
a combination of tree learning and representation learning in an end-to-end manner. We define a
distribution-based loss function for a tree by the Kullback-Leibler divergence (K-L) between the
ground truth label distribution and the distribution predicted by the tree. By fixing split nodes, we
show that the optimization of leaf node predictions to minimize the loss function of the tree can
be addressed by variational bounding [19, 29], in which the original loss function to be minimized
gets iteratively replaced by a decreasing sequence of upper bounds. Following this optimization
strategy, we derive a discrete iterative function to update the leaf node predictions. To learn a forest,
we average the losses of all the individual trees to be the loss for the forest and allow the split nodes
from different trees to be connected to the same output unit of the feature learning function. In this
way, the split node parameters of all the individual trees can be learned jointly. Our LDLFs can be
used as a (shallow) stand-alone model, and can also be integrated with any deep networks, i.e., the
feature learning function can be a linear transformation and a deep network, respectively. Fig. 2
illustrates a sketch chart of our LDLFs, where a forest consists of two trees is shown.
We verify the effectiveness of our model on several LDL tasks, such as crowd opinion prediction on
movies and disease prediction based on human genes, as well as one computer vision application, i.e.,
facial age estimation, showing significant improvements to the state-of-the-art LDL methods. The
label distributions for these tasks include both unimodal distributions (e.g., the age distribution in
Fig. 1(a)) and mixture distributions (the rating distribution on a movie in Fig. 1(b)). The superiority
of our model on both of them verifies its ability to model any general form of label distributions
Figure 2: Illustration of a label distribution learning forest. The top circles denote the output units
of the function f parameterized by ?, which can be a feature vector or a fully-connected layer of
a deep network. The blue and green circles are split nodes and leaf nodes, respectively. Two index
function ?1 and ?2 are assigned to these two trees respectively. The black dash arrows indicate the
correspondence between the split nodes of these two trees and the output units of function f . Note
that, one output unit may correspond to the split nodes belonging to different trees. Each tree has
independent leaf node predictions q (denoted by histograms in leaf nodes). The output of the forest
is a mixture of the tree predictions. f (?; ?) and q are learned jointly in an end-to-end manner.
2
2
Related Work
Since our LDL algorithm is inspired by differentiable decision trees, it is necessary to first review
some typical techniques of decision trees. Then, we discuss current LDL methods.
Decision trees. Random forests or randomized decision trees [16, 1, 3, 4], are a popular ensemble
predictive model suitable for many machine learning tasks. In the past, learning of a decision tree was
based on heuristics such as a greedy algorithm where locally-optimal hard decisions are made at each
split node [1], and thus, cannot be integrated into in a deep learning framework, i.e., be combined
with representation learning in an end-to-end manner.
The newly proposed deep neural decision forests (dNDFs) [20] overcomes this problem by introducing
a soft differentiable decision function at the split nodes and a global loss function defined on a tree.
This ensures that the split node parameters can be learned by back-propagation and leaf node
predictions can be updated by a discrete iterative function.
Our method extends dNDFs to address LDL problems, but this extension is non-trivial, because
learning leaf node predictions is a constrained convex optimization problem. Although a step-size
free update function was given in dNDFs to update leaf node predictions, it was only proved to
converge for a classification loss. Consequently, it was unclear how to obtain such an update function
for other losses. We observed, however, that the update function in dNDFs can be derived from
variational bounding, which allows us to extend it to our LDL loss. In addition, the strategies used in
LDLFs and dNDFs to learning the ensemble of multiple trees (forests) are different: 1) we explicitly
define a loss function for forests, while only the loss function for a single tree was defined in dNDFs;
2) we allow the split nodes from different trees to be connected to the same output unit of the feature
learning function, while dNDFs did not; 3) all trees in LDLFs can be learned jointly, while trees in
dNDFs were learned alternatively. These changes in the ensemble learning are important, because as
shown in our experiments (Sec. 4.4), LDLFs can get better results by using more trees, but by using
the ensemble strategy proposed in dNDFs, the results of forests are even worse than those for a single
tree.
To sum up, w.r.t. dNDFs [20], the contributions of LDLFs are: first, we extend from classification [20]
to distribution learning by proposing a distribution-based loss for the forests and derive the gradient to
learn splits nodes w.r.t. this loss; second, we derived the update function for leaf nodes by variational
bounding (having observed that the update function in [20] was a special case of variational
bounding); last but not the least, we propose above three strategies to learning the ensemble of
multiple trees, which are different from [20], but we show are effective.
Label distribution learning. A number of specialized algorithms have been proposed to address the
LDL task, and have shown their effectiveness in many computer vision applications, such as facial
age estimation [8, 11, 28], expression recognition [30] and hand orientation estimation [10].
Geng et al. [8] defined the label distribution for an instance as a vector containing the probabilities
of the instance having each label. They also gave a strategy to assign a proper label distribution
to an instance with a single label, i.e., assigning a Gaussian or Triangle distribution whose peak
is the single label, and proposed an algorithm called IIS-LLD, which is an iterative optimization
process based on a two-layer energy based model. Yang et al. [28] then defined a three-layer energy
based model, called SCE-LDL, in which the ability to perform feature learning is improved by
adding the extra hidden layer and sparsity constraints are also incorporated to ameliorate the model.
Geng [6] developed an accelerated version of IIS-LLD, called BFGS-LDL, by using quasi-Newton
optimization. All the above LDL methods assume that the label distribution can be represented by a
maximum entropy model [2], but the exponential part of this model restricts the generality of the
distribution form.
Another way to address the LDL task, is to extend existing learning algorithms to deal with label
distributions. Geng and Hou [7] proposed LDSVR, a LDL method by extending support vector
regressor, which fit a sigmoid function to each component of the distribution simultaneously by a
support vector machine. Xing et al. [27] then extended boosting to address the LDL task by additive
weighted regressors. They showed that using the vector tree model as the weak regressor can lead to
better performance and named this method AOSO-LDLLogitBoost. As the learning of this tree model
is based on locally-optimal hard data partition functions at each split node, AOSO-LDLLogitBoost is
unable to be combined with representation learning. Extending current deep learning algorithms to
3
address the LDL task is an interesting topic. But, the existing such a method, called DLDL [5], still
focuses on maximum entropy model based LDL.
Our method, LDLFs, extends differentiable decision trees to address LDL tasks, in which the predicted
label distribution for a sample can be expressed by a linear combination of the label distributions
of the training data, and thus have no restrictions on the distributions (e.g., no requirement of the
maximum entropy model). In addition, thanks to the introduction of differentiable decision functions,
LDLFs can be combined with representation learning, e.g., to learn deep features in an end-to-end
manner.
3
Label Distribution Learning Forests
A forest is an ensemble of decision trees. We first introduce how to learn a single decision tree by
label distribution learning, then describe the learning of a forest.
3.1
Problem Formulation
Let X = Rm denote the input space and Y = {y1 , y2 , . . . , yC } denote the complete set of labels,
where C is the number of possible label values. We consider a label distribution learning (LDL)
problem, where for each input sample x ? X , there is a label distribution d = (dxy1 , dyx2 , . . . , dyxC )> ?
RC . Here dyxc expresses the probability of the sample x having the c-th label yc and thus has the
PC
constraints that dyxc ? [0, 1] and c=1 dyxc = 1. The goal of the LDL problem is to learn a mapping
function g : x ? d between an input sample x and its corresponding label distribution d.
Here, we want to learn the mapping function g(x) by a decision tree based model T . A decision
tree consists of a set of split nodes N and a set of leaf nodes L. Each split node n ? N defines
a split function sn (?; ?) : X ? [0, 1] parameterized by ? to determine whether a sample is sent
to the left or right subtree. Each leaf node ` ? L holds a distribution q` = (q`1 , q`2 , . . . , q`C )>
PC
over Y, i.e, q`c ? [0, 1] and c=1 q`c = 1. To build a differentiable decision tree, following [20],
we use a probabilistic split function sn (x; ?) = ?(f?(n) (x; ?)), where ?(?) is a sigmoid function,
?(?) is an index function to bring the ?(n)-th output of function f (x; ?) in correspondence with
split node n, and f : x ? RM is a real-valued feature learning function depending on the sample x
and the parameter ?, and can take any form. For a simple form, it can be a linear transformation
of x, where ? is the transformation matrix; For a complex form, it can be a deep network to
perform representation learning in an end-to-end manner, then ? is the network parameter. The
correspondence between the split nodes and the output units of function f , indicated by ?(?) that is
randomly generated before tree learning, i.e., which output units from ?f ? are used for constructing a
tree is determined randomly. An example to demonstrate ?(?) is shown in Fig. 2. Then, the probability
of the sample x falling into leaf node ` is given by
Y
l
r
p(`|x; ?) =
sn (x; ?)1(`?Ln ) (1 ? sn (x; ?))1(`?Ln ) ,
(1)
n?N
where 1(?) is an indicator function and Lln and Lrn denote the sets of leaf nodes held by the left and
right subtrees of node n, Tnl and Tnr , respectively. The output of the tree T w.r.t. x, i.e., the mapping
function g, is defined by
X
g(x; ?, T ) =
p(`|x; ?)q` .
(2)
`?L
3.2
Tree Optimization
Given a training set S = {(xi , di )}N
i=1 , our goal is to learn a decision tree T described in Sec. 3.1
which can output a distribution g(xi ; ?, T ) similar to di for each sample xi . To this end, a
straightforward way is to minimize the Kullback-Leibler (K-L) divergence between each g(xi ; ?, T )
and di , or equivalently to minimize the following cross-entropy loss:
R(q, ?; S) = ?
N C
N C
X
1 X X yc
1 X X yc
dxi log(gc (xi ; ?, T )) = ?
dxi log
p(`|xi ; ?)q`c , (3)
N i=1 c=1
N i=1 c=1
`?L
4
where q denote the distributions held by all the leaf nodes L and gc (xi ; ?, T ) is the c-th output unit
of g(xi ; ?, T ). Learning the tree T requires the estimation of two parameters: 1) the split node
parameter ? and 2) the distributions q held by the leaf nodes. The best parameters (?? , q? ) are
determined by
(?? , q? ) = arg min R(q, ?; S).
(4)
?,q
To solve Eqn. 4, we consider an alternating optimization strategy: First, we fix q and optimize
?; Then, we fix ? and optimize q. These two learning steps are alternatively performed, until
convergence or a maximum number of iterations is reached (defined in the experiments).
3.2.1
Learning Split Nodes
In this section, we describe how to learn the parameter ? for split nodes, when the distributions held
by the leaf nodes q are fixed. We compute the gradient of the loss R(q, ?; S) w.r.t. ? by the chain
rule:
N
?R(q, ?; S) X X ?R(q, ?; S) ?f?(n) (xi ; ?)
=
,
(5)
??
?f?(n) (xi ; ?)
??
i=1
n?N
where only the first term depends on the tree and the second term depends on the specific type of the
function f?(n) . The first term is given by
C
gc (xi ; ?, Tnl )
gc (xi ; ?, Tnr )
1 X yc
?R(q, ?; S)
=
dxi sn (xi ; ?)
? 1 ? sn (xi ; ?)
, (6)
?f?(n) (xi ; ?)
N c=1
gc (xi ; ?, T )
gc (xi ; ?, T )
P
P
where gc (xi ; ?, Tnl ) = `?Lln p(`|xi ; ?)q`c and g c (xi ; ?, Tnr ) = `?Lrn p(`|xi ; ?)q`c . Note that,
let Tn be the tree rooted at the node n, then we have gc (xi ; ?, Tn ) = gc (xi ; ?, Tnl ) + gc (xi ; ?, Tnr ).
This means the gradient computation in Eqn. 6 can be started at the leaf nodes and carried out in a
bottom up manner. Thus, the split node parameters can be learned by standard back-propagation.
3.2.2
Learning Leaf Nodes
Now, fixing the parameter ?, we show how to learn the distributions held by the leaf nodes q, which
is a constrained optimization problem:
min R(q, ?; S), s.t., ?`,
q
C
X
q`c = 1.
(7)
c=1
Here, we propose to address this constrained convex optimization problem by variational bounding [19, 29], which leads to a step-size free and fast-converged update rule for q. In variational
bounding, an original objective function to be minimized gets replaced by its bound in an iterative
manner. A upper bound for the loss function R(q, ?; S) can be obtained by Jensen?s inequality:
R(q, ?; S) = ?
N C
X
1 X X yc
dxi log
p(`|xi ; ?)q`c
N i=1 c=1
`?L
??
where ?` (q`c , xi ) =
1
N
N X
C
X
i=1 c=1
p(`|xi ;?)q`c
gc (xi ;?,T )
?(q, q
?) = ?
dyxci
X
?` (?
q`c , xi ) log
`?L
p(`|x ; ?)q
i
`c
,
?` (?
q`c , xi )
(8)
. We define
N C
p(`|x ; ?)q
1 X X yc X
i
`c
dxi
?` (?
q`c , xi ) log
.
N i=1 c=1
?` (?
q`c , xi )
(9)
`?L
? ) is an upper bound for R(q, ?; S), which has the property that for any q and q
Then ?(q, q
?,
?(q, q
?) ? R(q, ?; S), and ?(q, q) = R(q, ?; S). Assume that we are at a point q(t) corresponding
to the t-th iteration, then ?(q, q(t) ) is an upper bound for R(q, ?; S). In the next iteration, q(t+1)
is chosen such that ?(q(t+1) , q) ? R(q(t) , ?; S), which implies R(q(t+1) , ?; S) ? R(q(t) , ?; S).
5
Consequently, we can minimize ?(q, q
?) instead of R(q, ?; S) after ensuring that R(q(t) , ?; S) =
(t)
(t)
?(q , q
?), i.e., q
? = q . So we have
q(t+1) = arg min ?(q, q(t) ), s.t., ?`,
q
C
X
q`c = 1,
(10)
c=1
which leads to minimizing the Lagrangian defined by
?(q, q(t) ) = ?(q, q(t) ) +
X
?` (
`?L
where ?` is the Lagrange multiplier. By setting
?` =
(t+1)
Note that, q`c
(t+1)
? [0, 1] and
distributions held by the leaf nodes. The starting
(0)
distribution: q`c = C1 .
3.3
q`c ? 1),
(11)
c=1
??(q,q(t) )
?q`c
N C
1 X X yc
(t)
(t+1)
d ?` (q`c , xi ) and q`c
N i=1 c=1 xi
satisfies that q`c
C
X
= 0, we have
PN yc
(t)
dxi ?` (q`c , xi )
.
= PC i=1
PN yc
(t)
?
(q
,
x
)
d
x
`
i
i
c=1
i=1
`c
(12)
(t+1)
= 1. Eqn. 12 is the update scheme for
c=1 q`c
(0)
point q` can be simply initialized by the uniform
PC
Learning a Forest
A forest is an ensemble of decision trees F = {T1 , . . . , TK }. In the training stage, all trees in the
forest F use the same parameters ? for feature learning function f (?; ?) (but correspond to different
output units of f assigned by ?, see Fig. 2), but each tree has independent leaf node predictions
q. The loss function for a forest is given by averaging the loss functions for all individual trees:
PK
1
RF = K
k=1 RTk , where RTk is the loss function for tree Tk defined by Eqn. 3. To learn ? by
fixing the leaf node predictions q of all the trees in the forest F, based on the derivation in Sec. 3.2
and referring to Fig. 2, we have
N K
?f?k (n) (xi ; ?)
1 XX X
?RF
?RTk
=
,
??
K i=1
?f?k (n) (xi ; ?)
??
(13)
k=1 n?Nk
where Nk and ?k (?) are the split node set and the index function of Tk , respectively. Note that,
the index function ?k (?) for each tree is randomly assigned before tree learning, and thus split
nodes correspond to a subset of output units of f . This strategy is similar to the random subspace
method [17], which increases the randomness in training to reduce the risk of overfitting.
As for q, since each tree in the forest F has its own leaf node predictions q, we can update them
independently by Eqn. 12, given by ?. For implementational convenience, we do not conduct this
update scheme on the whole dataset S but on a set of mini-batches B. The training procedure of a
LDLF is shown in Algorithm. 1.
Algorithm 1 The training procedure of a LDLF.
Require: S: training set, nB : the number of mini-batches to update q
Initialize ? randomly and q uniformly, set B = {?}
while Not converge do
while |B| < nB do
Randomly select a mini-batch B from S
UpdateS
? by computing gradient (Eqn. 13) on B
B=B B
end while
Update q by iterating Eqn. 12 on B
B = {?}
end while
In the testing stage, the output of the forest F is given by averaging the predictions from all the
PK
1
individual trees: g(x; ?, F) = K
k=1 g(x; ?, Tk ).
6
4
Experimental Results
Our realization of LDLFs is based on ?Caffe? [18]. It is modular and implemented as a standard
neural network layer. We can either use it as a shallow stand-alone model (sLDLFs) or integrate it
with any deep networks (dLDLFs). We evaluate sLDLFs on different LDL tasks and compare it with
other stand-alone LDL methods. As dLDLFs can be learned from raw image data in an end-to-end
manner, we verify dLDLFs on a computer vision application, i.e., facial age estimation. The default
settings for the parameters of our forests are: tree number (5), tree depth (7), output unit number of
the feature learning function (64), iteration times to update leaf node predictions (20), the number of
mini-batches to update leaf node predictions (100), maximum iteration (25000).
4.1
Comparison of sLDLFs to Stand-alone LDL Methods
We compare our shallow model sLDLFs with other state-of-the-art stand-alone LDL methods.
For sLDLFs, the feature learning function f (x, ?) is a linear transformation of x, i.e., the i-th
output unit fi (x, ?i ) = ?i> x, where ?i is the i-th column of the transformation matrix ?. We
used 3 popular LDL datasets in [6], Movie, Human Gene and Natural Scene1 . The samples
in these 3 datasets are represented by numerical descriptors, and the ground truths for them are
the rating distributions of crowd opinion on movies, the diseases distributions related to human
genes and label distributions on scenes, such as plant, sky and cloud, respectively. The label
distributions of these 3 datasets are mixture distributions, such as the rating distribution shown in
Fig. 1(b). Following [7, 27], we use 6 measures to evaluate the performances of LDL methods,
which compute the average similarity/distance between the predicted rating distributions and the real
rating distributions, including 4 distance measures (K-L, Euclidean, S?rensen, Squared ?2 ) and two
similarity measures (Fidelity, Intersection).
We evaluate our shallow model sLDLFs on these 3 datasets and compare it with other state-of-the-art
stand-alone LDL methods. The results of sLDLFs and the competitors are summarized in Table 1.
For Movie we quote the results reported in [27], as the code of [27] is not publicly available. For the
results of the others two, we run code that the authors had made available. In all case, following [27, 6],
we split each dataset into 10 fixed folds and do standard ten-fold cross validation, which represents
the result by ?mean?standard deviation? and matters less how training and testing data get divided.
As can be seen from Table 1, sLDLFs perform best on all of the six measures.
Table 1: Comparison results on three LDL datasets [6]. ??? and ??? indicate the larger and the smaller
the better, respectively.
Dataset
Method
K-L ?
Euclidean ?
S?rensen ?
Squared ?2 ?
Fidelity ?
Intersection ?
Movie
sLDLF (ours)
AOSO-LDLogitBoost [27]
LDLogitBoost [27]
LDSVR [7]
BFGS-LDL [6]
IIS-LDL [11]
0.073?0.005
0.086?0.004
0.090?0.004
0.092?0.005
0.099?0.004
0.129?0.007
0.133?0.003
0.155?0.003
0.159?0.003
0.158?0.004
0.167?0.004
0.187?0.004
0.130?0.003
0.152?0.003
0.155?0.003
0.156?0.004
0.164?0.003
0.183?0.004
0.070?0.004
0.084?0.003
0.088?0.003
0.088?0.004
0.096?0.004
0.120?0.005
0.981?0.001
0.978?0.001
0.977?0.001
0.977?0.001
0.974?0.001
0.967?0.001
0.870?0.003
0.848?0.003
0.845?0.003
0.844?0.004
0.836?0.003
0.817?0.004
sLDLF (ours)
LDSVR [7]
BFGS-LDL [6]
IIS-LDL [11]
0.228?0.006
0.245?0.019
0.231?0.021
0.239?0.018
0.085?0.002
0.099?0.005
0.076?0.006
0.089?0.006
0.212?0.002
0.229?0.015
0.231?0.012
0.253?0.009
0.179?0.004
0.189?0.021
0.211?0.018
0.205?0.012
0.948?0.001
0.940?0.006
0.938?0.008
0.944?0.003
0.788?0.002
0.771?0.015
0.769?0.012
0.747?0.009
sLDLF (ours)
LDSVR [7]
BFGS-LDL [6]
IIS-LDL [11]
0.534?0.013
0.852?0.023
0.856?0.061
0.879?0.023
0.317?0.014
0.511?0.021
0.475?0.029
0.458?0.014
0.336?0.010
0.492?0.016
0.508?0.026
0.539?0.011
0.448?0.017
0.595?0.026
0.716?0.041
0.792?0.019
0.824?0.008
0.813?0.008
0.722?0.021
0.686?0.009
0.664?0.010
0.509?0.016
0.492?0.026
0.461?0.011
Human Gene
Natural Scene
4.2
Evaluation of dLDLFs on Facial Age Estimation
In some literature [8, 11, 28, 15, 5], age estimation is formulated as a LDL problem. We conduct
facial age estimation experiments on Morph [24], which contains more than 50,000 facial images
from about 13,000 people of different races. Each facial image is annotated with a chronological age.
To generate an age distribution for each face image, we follow the same strategy used in [8, 28, 5],
which uses a Gaussian distribution whose mean is the chronological age of the face image (Fig. 1(a)).
The predicted age for a face image is simply the age having the highest probability in the predicted
1
We download these datasets from http://cse.seu.edu.cn/people/xgeng/LDL/index.htm.
7
label distribution. The performance of age estimation is evaluated by the mean absolute error (MAE)
between predicted ages and chronological ages. As the current state-of-the-art result on Morph
is obtain by fine-tuning DLDL [5] on VGG-Face [23], we also build a dLDLF on VGG-Face, by
replacing the softmax layer in VGGNet by a LDLF. Following [5], we do standard 10 ten-fold cross
validation and the results are summarized in Table. 2, which shows dLDLF achieve the state-of-the-art
performance on Morph. Note that, the significant performance gain between deep LDL models
(DLDL and dLDLF) and non-deep LDL models (IIS-LDL, CPNN, BFGS-LDL) and the superiority
of dLDLF compared with DLDL verifies the effectiveness of end-to-end learning and our tree-based
model for LDL, respectively.
Table 2: MAE of age estimation comparison on Morph [24].
Method
IIS-LDL [11]
CPNN [11]
BFGS-LDL [6]
DLDL+VGG-Face [5]
dLDLF+VGG-Face (ours)
MAE
5.67?0.15
4.87?0.31
3.94?0.05
2.42?0.01
2.24?0.02
As the distribution of gender and ethnicity is very unbalanced in Morph, many age estimation methods [13, 14, 15] are evaluated on a subset of Morph, called Morph_Sub for short, which consists of
20,160 selected facial images to avoid the influence of unbalanced distribution. The best performance
reported on Morph_Sub is given by D2LDL [15], a data-dependent LDL method. As D2LDL used
the output of the ?fc7? layer in AlexNet [21] as the face image features, here we integrate a LDLF
with AlexNet. Following the experiment setting used in D2LDL, we evaluate our dLDLF and the
competitors, including both SLL and LDL based methods, under six different training set ratios (10%
to 60%). All of the competitors are trained on the same deep features used by D2LDL. As can be
seen from Table 3, our dLDLFs significantly outperform others for all training set ratios.
Note that, the generated age distri- Figure 3: MAE of age estimation comparison on
butions are unimodal distributions Morph_Sub.
and the label distributions used in
Training set ratio
Method
Sec. 4.1 are mixture distributions.
10%
20%
30%
40%
50%
60%
The proposed method LDLFs achieve
AAS [22]
4.9081
4.7616
4.6507
4.5553
4.4690
4.4061
the state-of-the-art results on both of
LARR [12]
4.7501
4.6112
4.5131
4.4273
4.3500
4.2949
IIS-ALDL [9]
4.1791
4.1683
4.1228
4.1107
4.1024
4.0902
them, which verifies that our model
D2LDL [15]
4.1080
3.9857
3.9204
3.8712
3.8560
3.8385
has the ability to model any general
dLDLF (ours)
3.8495
3.6220
3.3991
3.2401
3.1917
3.1224
form of label distributions.
4.3
Time Complexity
Let h and sB be the tree depth and the
batch size, respectively. Each tree has 2h?1 ? 1 split nodes and 2h?1 leaf nodes. Let D = 2h?1 ? 1.
For one tree and one sample, the complexity of a forward pass and a backward pass are O(D +
D + 1?C) = O(D?C) and O(D + 1?C + D?C) = O(D?C), respectively. So for K trees and
nB batches, the complexity of a forward and backward pass is O(D?C?K?nB ?sB ). The complexity of an iteration to update leaf nodes are O(nB ?sB ?K?C?D + 1) = O(D?C?K?nB ?sB ).
Thus, the complexity for the training procedure (one epoch, nB batches) and the testing procedure
(one sample) are O(D?C?K?nB ?sB ) and O(D?C?K), respectively. LDLFs are efficient: On
Morph_Sub (12636 training images, 8424 testing images), our model only takes 5250s for training
(25000 iterations) and 8s for testing all 8424 images.
4.4
Parameter Discussion
Now we discuss the influence of parameter settings on performance. We report the results of rating
prediction on Movie (measured by K-L) and age estimation on Morph_Sub with 60% training set
ratio (measured by MAE) for different parameter settings in this section.
Tree number. As a forest is an ensemble model, it is necessary to investigate how performances
change by varying the tree number used in a forest. Note that, as we discussed in Sec. 2, the
ensemble strategy to learn a forest proposed in dNDFs [20] is different from ours. Therefore, it is
necessary to see which ensemble strategy is better to learn a forest. Towards this end, we replace our
ensemble strategy in dLDLFs by the one used in dNDFs, and name this method dNDFs-LDL. The
corresponding shallow model is named by sNDFs-LDL. We fix other parameters, i.e., tree depth and
8
output unit number of the feature learning function, as the default setting. As shown in Fig. 4 (a), our
ensemble strategy can improve the performance by using more trees, while the one used in dNDFs
even leads to a worse performance than one for a single tree.
Observed from Fig. 4, the performance of LDLFs can be improved by using more trees, but the
improvement becomes increasingly smaller and smaller. Therefore, using much larger ensembles
does not yield a big improvement (On Movie, the number of trees K = 100: K-L = 0.070 vs K = 20:
K-L = 0.071). Note that, not all random forests based methods use a large number of trees, e.g.,
Shotton et al. [25] obtained very good pose estimation results from depth images by only 3 decision
trees.
Tree depth. Tree depth is another important parameter for decision trees. In LDLFs, there is an
implicit constraint between tree depth h and output unit number of the feature learning function ? :
? ? 2h?1 ? 1. To discuss the influence of tree depth to the performance of dLDLFs, we set ? = 2h?1
and fix tree number K = 1, and the performance change by varying tree depth is shown in Fig. 4 (b).
We see that the performance first improves then decreases with the increase of the tree depth. The
reason is as the tree depth increases, the dimension of learned features increases exponentially, which
greatly increases the training difficulty. So using much larger depths may lead to bad performance
(On Movie, tree depth h = 18: K-L = 0.1162 vs h = 9: K-L = 0.0831).
Figure 4: The performance change of age estimation on Morph_Sub and rating prediction on Movie
by varying (a) tree number and (b) tree depth. Our approach (dLDLFs/sLDLFs) can improve the
performance by using more trees, while using the ensemble strategy proposed in dNDFs (dNDFsLDL/sNDFs-LDL) even leads to a worse performance than one for a single tree.
5
Conclusion
We present label distribution learning forests, a novel label distribution learning algorithm inspired by
differentiable decision trees. We defined a distribution-based loss function for the forests and found
that the leaf node predictions can be optimized via variational bounding, which enables all the trees
and the feature they use to be learned jointly in an end-to-end manner. Experimental results showed
the superiority of our algorithm for several LDL tasks and a related computer vision application, and
verified our model has the ability to model any general form of label distributions.
Acknowledgement. This work was supported in part by the National Natural Science Foundation of
China No. 61672336, in part by ?Chen Guang? project supported by Shanghai Municipal Education
Commission and Shanghai Education Development Foundation No. 15CG43 and in part by ONR
N00014-15-1-2356.
References
[1] Y. Amit and D. Geman. Shape quantization and recognition with randomized trees. Neural Computation,
9(7):1545?1588, 1997.
[2] A. L. Berger, S. D. Pietra, and V. J. D. Pietra. A maximum entropy approach to natural language processing.
Computational Linguistics, 22(1):39?71, 1996.
[3] L. Breiman. Random forests. Machine Learning, 45(1):5?32, 2001.
[4] A. Criminisi and J. Shotton. Decision Forests for Computer Vision and Medical Image Analysis. Springer,
2013.
[5] B.-B. Gao, C. Xing, C.-W. Xie, J. Wu, and X. Geng. Deep label distribution learning with label ambiguity.
arXiv:1611.01731, 2017.
[6] X. Geng. Label distribution learning. IEEE Trans. Knowl. Data Eng., 28(7):1734?1748, 2016.
9
[7] X. Geng and P. Hou. Pre-release prediction of crowd opinion on movies by label distribution learning. In
Pro. IJCAI, pages 3511?3517, 2015.
[8] X. Geng, K. Smith-Miles, and Z. Zhou. Facial age estimation by learning from label distributions. In Proc.
AAAI, 2010.
[9] X. Geng, Q. Wang, and Y. Xia. Facial age estimation by adaptive label distribution learning. In Proc.
ICPR, pages 4465?4470, 2014.
[10] X. Geng and Y. Xia. Head pose estimation based on multivariate label distribution. In Proc. CVPR, pages
1837?1842, 2014.
[11] X. Geng, C. Yin, and Z. Zhou. Facial age estimation by learning from label distributions. IEEE Trans.
Pattern Anal. Mach. Intell., 35(10):2401?2412, 2013.
[12] G. Guo, Y. Fu, C. R. Dyer, and T. S. Huang. Image-based human age estimation by manifold learning and
locally adjusted robust regression. IEEE Trans. Image Processing, 17(7):1178?1188, 2008.
[13] G. Guo and G. Mu. Human age estimation: What is the influence across race and gender? In CVPR
Workshops, pages 71?78, 2010.
[14] G. Guo and C. Zhang. A study on cross-population age estimation. In Proc. CVPR, pages 4257?4263,
2014.
[15] Z. He, X. Li, Z. Zhang, F. Wu, X. Geng, Y. Zhang, M.-H. Yang, and Y. Zhuang. Data-dependent label
distribution learning for age estimation. IEEE Trans. on Image Processing, 2017.
[16] T. K. Ho. Random decision forests. In Proc. ICDAR, pages 278?282, 1995.
[17] T. K. Ho. The random subspace method for constructing decision forests. IEEE Trans. Pattern Anal. Mach.
Intell., 20(8):832?844, 1998.
[18] Y. Jia, E. Shelhamer, J. Donahue, S. Karayev, J. Long, R. Girshick, S. Guadarrama, and T. Darrell. Caffe:
Convolutional architecture for fast feature embedding. arXiv preprint arXiv:1408.5093, 2014.
[19] M. I. Jordan, Z. Ghahramani, T. S. Jaakkola, and L. K. Saul. An introduction to variational methods for
graphical models. Machine Learning, 37(2):183?233, 1999.
[20] P. Kontschieder, M. Fiterau, A. Criminisi, and S. R. Bul?. Deep neural decision forests. In Proc. ICCV,
pages 1467?1475, 2015.
[21] A. Krizhevsky, I. Sutskever, and G. E. Hinton. Imagenet classification with deep convolutional neural
networks. In Proc. NIPS, pages 1106?1114, 2012.
[22] A. Lanitis, C. Draganova, and C. Christodoulou. Comparing different classifiers for automatic age
estimation. IEEE Trans. on Cybernetics,, 34(1):621?628, 2004.
[23] O. M. Parkhi, A. Vedaldi, and A. Zisserman. Deep face recognition. In Proc. BMVC, pages 41.1?41.12,
2015.
[24] K. Ricanek and T. Tesafaye. MORPH: A longitudinal image database of normal adult age-progression. In
Proc. FG, pages 341?345, 2006.
[25] J. Shotton, A. W. Fitzgibbon, M. Cook, T. Sharp, M. Finocchio, R. Moore, A. Kipman, and A. Blake.
Real-time human pose recognition in parts from single depth images. In Proc. CVPR, pages 1297?1304,
2011.
[26] G. Tsoumakas and I. Katakis. Multi-label classification: An overview. International Journal of Data
Warehousing and Mining, 3(3):1?13, 2007.
[27] C. Xing, X. Geng, and H. Xue. Logistic boosting regression for label distribution learning. In Proc. CVPR,
pages 4489?4497, 2016.
[28] X. Yang, X. Geng, and D. Zhou. Sparsity conditional energy label distribution learning for age estimation.
In Proc. IJCAI, pages 2259?2265, 2016.
[29] A. L. Yuille and A. Rangarajan. The concave-convex procedure. Neural Computation, 15(4):915?936,
2003.
[30] Y. Zhou, H. Xue, and X. Geng. Emotion distribution recognition from facial expressions. In Proc. MM,
pages 1247?1250, 2015.
10
| 6685 |@word version:1 eng:1 contains:1 ours:6 longitudinal:1 past:1 existing:3 current:4 com:1 guadarrama:1 comparing:1 gmail:1 assigning:1 john:1 hou:2 numerical:1 additive:1 partition:1 shape:1 enables:2 update:18 v:2 alone:6 greedy:1 leaf:31 selected:1 cook:1 smith:1 short:1 boosting:3 node:57 cse:1 zhang:3 rc:1 consists:3 introduce:1 manner:11 multi:2 inspired:3 decreasing:1 sce:1 becomes:1 distri:1 xx:1 project:1 katakis:1 alexnet:2 what:1 developed:1 proposing:1 transformation:5 guarantee:1 sky:1 every:1 concave:1 chronological:3 rm:2 classifier:1 unit:14 medical:1 superiority:3 before:3 t1:1 engineering:1 tnr:4 mach:2 black:1 china:1 specialty:1 testing:5 investment:1 fitzgibbon:1 procedure:5 significantly:1 vedaldi:1 pre:1 get:4 cannot:2 convenience:1 nb:8 risk:2 influence:4 restriction:1 optimize:2 lagrangian:1 straightforward:1 starting:1 independently:1 convex:3 assigns:1 rule:2 population:1 embedding:1 updated:1 user:1 us:1 recognition:5 geman:1 database:1 observed:3 bottom:1 cloud:1 preprint:1 wang:1 ensures:1 connected:3 decrease:2 highest:1 disease:2 mu:1 complexity:5 trained:1 predictive:1 yuille:2 imdb:1 triangle:1 multimodal:1 htm:1 represented:3 fiber:1 derivation:1 fast:2 effective:1 describe:2 crowd:5 caffe:2 whose:2 heuristic:1 kai:1 valued:1 solve:1 say:1 modular:1 larger:3 cvpr:5 ability:4 jointly:5 advantage:2 differentiable:11 sequence:1 karayev:1 propose:2 sll:2 realization:1 achieve:2 description:1 guo1:1 sutskever:1 convergence:1 ijcai:2 requirement:1 extending:3 darrell:1 rangarajan:1 tk:4 derive:2 depending:1 pose:3 fixing:3 measured:2 school:1 strong:1 implemented:1 predicted:6 indicate:2 implies:1 annotated:1 criminisi:2 human:8 opinion:5 tsoumakas:1 education:2 require:1 assign:2 fix:4 adjusted:1 extension:1 hold:1 mm:1 ground:2 normal:1 blake:1 mapping:3 predict:2 released:1 estimation:26 proc:13 label:69 quote:1 knowl:1 weighted:1 gaussian:2 aim:1 rather:1 avoid:3 pn:2 zhou:4 breiman:1 varying:3 jaakkola:1 derived:3 focus:1 release:1 improvement:4 tnl:4 greatly:1 dependent:2 sb:5 integrated:2 hidden:1 quasi:1 arg:2 classification:4 orientation:1 fidelity:2 denoted:1 development:1 art:7 constrained:3 special:1 initialize:1 softmax:1 emotion:1 having:4 beach:1 represents:1 geng:14 minimized:2 others:2 report:1 producer:1 randomly:5 simultaneously:1 divergence:2 seu:1 individual:4 national:1 mll:1 pietra:2 replaced:2 intell:2 investigate:1 mining:1 evaluation:1 mixture:7 pc:4 lrn:2 held:6 chain:1 subtrees:1 fu:1 necessary:3 facial:16 shen1:1 tree:94 conduct:2 euclidean:2 initialized:1 circle:2 girshick:1 instance:6 column:1 soft:1 implementational:1 introducing:1 deviation:1 subset:2 uniform:1 krizhevsky:1 reported:2 commission:1 morph:7 lanitis:1 xue:2 combined:4 referring:1 person:1 st:1 peak:1 randomized:2 thanks:1 international:1 probabilistic:1 regressor:2 hopkins:1 squared:2 ambiguity:3 aaai:1 containing:1 choose:1 huang:1 ldl:56 worse:3 li:1 potential:2 bfgs:6 sec:5 summarized:2 matter:1 explicitly:1 race:2 depends:2 performed:1 reached:1 netflix:1 xing:3 jia:1 contribution:1 minimize:4 chart:1 publicly:1 convolutional:2 descriptor:1 ensemble:14 correspond:3 yield:1 weak:1 famous:1 raw:1 cybernetics:1 randomness:1 converged:1 competitor:3 energy:4 involved:1 di:3 dxi:6 gain:1 newly:1 proved:1 dataset:3 popular:2 improves:1 back:3 xie:1 follow:1 zisserman:1 wei:1 improved:2 bmvc:1 formulation:1 evaluated:2 generality:2 stage:2 implicit:1 until:1 sketch:1 hand:1 eqn:7 web:1 replacing:1 propagation:3 defines:1 logistic:1 indicated:1 usa:1 name:1 verify:2 y2:1 multiplier:1 hence:1 assigned:4 alternating:1 laboratory:1 leibler:2 iteratively:1 mile:1 moore:1 deal:4 rooted:1 butions:1 complete:1 demonstrate:1 tn:2 bring:1 pro:1 image:20 variational:9 novel:3 fi:1 sigmoid:2 specialized:1 overview:1 shanghai:4 exponentially:1 extend:4 discussed:1 he:1 mae:5 significant:3 tuning:1 automatic:1 language:1 had:1 access:1 similarity:2 fc7:1 multivariate:1 own:1 showed:2 optimizing:1 n00014:1 inequality:1 onr:1 seen:2 converge:2 determine:1 ii:8 multiple:4 unimodal:3 alan:2 aoso:3 cross:4 long:2 divided:1 ensuring:1 prediction:23 regression:3 vision:6 arxiv:3 histogram:1 iteration:7 audience:1 c1:1 addition:2 want:1 fine:1 addressed:1 extra:1 unlike:1 strict:1 probably:1 sent:1 effectiveness:4 jordan:1 yang:3 split:27 shotton:3 ethnicity:1 fit:1 gave:1 architecture:1 reduce:2 cn:1 vgg:4 whether:1 expression:3 six:2 finocchio:1 deep:18 iterating:1 locally:3 ten:2 generate:1 http:1 outperform:1 rensen:2 restricts:2 estimated:1 rtk:3 blue:1 discrete:2 express:1 group:1 key:1 falling:1 verified:2 backward:2 sum:1 run:1 parameterized:2 ameliorate:1 named:2 extends:2 wu:2 decision:31 bound:5 layer:7 dash:1 correspondence:3 ldlf:4 fold:3 optic:1 precisely:1 constraint:3 scene:2 min:3 optical:1 department:1 icpr:1 combination:2 belonging:1 smaller:3 across:1 lld:2 increasingly:1 shallow:5 making:2 restricted:1 iccv:1 ln:2 discus:3 icdar:1 dyer:1 end:24 available:2 progression:1 douban:1 batch:7 ho:2 original:2 top:1 include:1 linguistics:1 graphical:1 newton:1 ghahramani:1 build:2 amit:1 objective:1 strategy:14 unclear:1 gradient:4 subspace:2 distance:2 unable:1 topic:1 manifold:1 collected:1 trivial:1 reason:1 code:2 modeled:1 index:5 illustration:1 mini:4 minimizing:1 ratio:4 berger:1 equivalently:1 christodoulou:1 warehousing:1 zhao1:1 anal:2 proper:1 perform:3 upper:4 datasets:6 enabling:1 extended:1 communication:2 precise:1 incorporated:1 y1:1 gc:11 head:1 hinton:1 sharp:1 download:1 rating:11 specified:1 kipman:1 optimized:1 imagenet:1 learned:11 nip:2 trans:6 address:7 adult:1 pattern:2 yc:10 sparsity:2 rf:2 green:1 including:2 suitable:3 natural:5 difficulty:2 indicator:1 advanced:1 representing:1 scheme:2 improve:2 movie:18 zhuang:1 vggnet:1 started:1 carried:1 sn:6 review:2 literature:1 epoch:1 acknowledgement:1 relative:1 loss:22 fully:1 plant:1 interesting:1 limitation:2 age:37 validation:2 foundation:2 integrate:2 shelhamer:1 supported:2 last:1 free:2 lln:2 allow:2 institute:1 saul:1 face:9 absolute:1 fg:1 xia:2 default:2 depth:15 world:2 stand:6 dimension:1 author:1 made:2 forward:2 regressors:1 adaptive:1 kullback:2 gene:4 overcomes:1 global:1 overfitting:1 xi:37 alternatively:2 iterative:4 table:6 learn:15 robust:1 ca:1 forest:38 complex:1 constructing:2 did:1 pk:2 arrow:1 bounding:8 whole:1 big:1 guang:1 verifies:3 fig:13 site:1 exponential:2 donahue:1 bad:1 specific:1 showing:2 jensen:1 workshop:1 quantization:1 adding:1 importance:1 subtree:1 illustrates:1 nk:2 chen:1 entropy:6 intersection:2 yin:1 simply:2 likely:1 gao:1 lagrange:1 expressed:1 watch:1 springer:1 gender:2 aa:1 truth:2 satisfies:1 conditional:1 goal:2 formulated:1 consequently:2 bul:1 towards:1 replace:1 hard:2 change:4 parkhi:1 typical:1 determined:2 uniformly:1 kontschieder:1 averaging:2 called:5 pas:3 experimental:2 select:1 support:3 people:2 guo:3 unbalanced:2 accelerated:1 evaluate:4 |
6,285 | 6,686 | Unsupervised learning of object frames by dense
equivariant image labelling
James Thewlis1
Hakan Bilen2
1
Andrea Vedaldi1
2
Visual Geometry Group
University of Oxford
{jdt,vedaldi}@robots.ox.ac.uk
School of Informatics
University of Edinburgh
[email protected]
Abstract
One of the key challenges of visual perception is to extract abstract models of 3D
objects and object categories from visual measurements, which are affected by
complex nuisance factors such as viewpoint, occlusion, motion, and deformations.
Starting from the recent idea of viewpoint factorization, we propose a new approach
that, given a large number of images of an object and no other supervision, can
extract a dense object-centric coordinate frame. This coordinate frame is invariant
to deformations of the images and comes with a dense equivariant labelling neural
network that can map image pixels to their corresponding object coordinates.
We demonstrate the applicability of this method to simple articulated objects
and deformable objects such as human faces, learning embeddings from random
synthetic transformations or optical flow correspondences, all without any manual
supervision.
1
Introduction
Humans can easily construct mental models of complex 3D objects and object categories from
visual observations. This is remarkable because the dependency between an object?s appearance
and its structure is tangled in a complex manner with extrinsic nuisance factors such as viewpoint,
illumination, and articulation. Therefore, learning the intrinsic structure of an object from images
requires removing these unwanted factors of variation from the data.
The recent work of [37] has proposed an unsupervised approach to do so, based on on the concept
of viewpoint factorization. The idea is to learn a deep Convolutional Neural Network (CNN) that
can, given an image of the object, detect a discrete set of object landmarks. Differently from
traditional approaches to landmark detection, however, landmarks are neither defined nor supervised
manually. Instead, the detectors are learned using only the requirement that the detected points must
be equivariant (consistent) with deformations of the input images. The authors of [37] show that this
constraint is sufficient to learn landmarks that are ?intrinsic? to the objects and hence capture their
structure; remarkably, due to the generalization ability of CNNs, the landmark points are detected
consistently not only across deformations of a given object instance, which are observed during
training, but also across different instances. This behaviour emerges automatically from training on
thousands of single-instance correspondences.
In this paper, we take this idea further, moving beyond a sparse set of landmarks to a dense model
of the object structure (section 3). Our method relates each point on an object to a point in a low
dimensional vector space in a way that is consistent across variation in motion and in instance identity.
This gives rise to an object-centric coordinate system, which allows points on the surface of an object
to be indexed semantically (figure 1). As an illustrative example, take the object category of a face
and the vector space R3 . Our goal is to semantically map out the object such that any point on a
face, such as the left eye, lives at a canonical position in this ?label space?. We train a CNN to learn
the function that projects any face image into this space, essentially ?coloring? each pixel with its
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
Figure 1: Dense equivariant image labelling. Left: Given an image x of an object or object category
and no other supervision, our goal is to find a common latent space Z, homeomorphic to a sphere,
which attaches a semantically-consistent coordinate frame to the object points. This is done by
learning a dense labelling function that maps image pixels to their corresponding coordinate in the
Z space. This mapping function is equivariant (compatible) with image warps or object instance
variations. Right: An equivariant dense mapping learned in an unsupervised manner from a large
dataset of faces. (Results of S IMPLE network, Ldist , ? = 0.5)
corresponding label. As a result of our learning formulation, the label space has the property of being
locally smooth: points nearby in the image are nearby in the label space. In an ideal case, we could
imagine the surface of an object to be mapped to a sphere.
In order to achieve these results, we contribute several technical innovations (section 3.2). First,
we show that, in order to learn a non-trivial object coordinate frame, the concept of equivariance
must be complemented with the one of distinctiveness of the embedding. Then, we propose a CNN
implementation of this concept that can explicitly express uncertainty in the labelling of the object
points. The formulation is used in combination with a probabilistic loss, which is augmented with a
robust geometric distance to encourage better alignment of the object features.
We show that this framework can be used to learn meaningful object coordinate frames in a purely
unsupervised manner, by analyzing thousands of deformations of visual objects. While [37] proposed
to use Thin Plate Spline image warps for training, here we also consider simple synthetic articulated
objects having frames related by known optical flow (section 4).
We conclude the paper with a summary of our finding (section 5).
2
Related Work
Learning the structure of visual objects. Modeling the structure of visual objects is a widelystudied (e.g. [6, 7, 11, 39, 12]) computer vision problem with important applications such as facial
landmark detection and human body pose estimation. Much of this work is supervised and aimed
at learning detectors of objects or their parts, often using deep learning. A few approaches such as
spatial transformer networks [20] can learn geometric transformations without explicit geometric
supervision, but do not build explicit geometric models of visual objects.
More related to our work, WarpNet [21] and geometric matching networks [34] learn a neural network
that predicts Thin Plate Spline [3] transformations between pairs of images of an object, including
synthetic warps. Deep Deformation Network [42] improves WarpNet by using a Point Transformer
Network to refine the computed landmarks, but it requires manual supervision. None of these works
look at the problem of learning an invariant geometric embedding for the object.
Our work builds on the idea of viewpoint factorization (section 3.1), recently introduced in [37, 31].
However, we extend [37] in several significant ways. First, we construct a dense rather than discrete
embedding, where all pixels of an object are mapped to an invariant object-centric coordinate instead
of just a small set of selected landmarks. Second, we show that the equivariance constraint proposed
in [37] is not quite enough to learn such an embedding; it must be complemented with the concept of
a distinctive embedding (section 3.1). Third, we introduce a new neural network architecture and
corresponding training objective that allow such an embedding to be learned in practice (section 3.2).
Optical/semantic flow. A common technique to find correspondences between temporally related
video frames is optical flow [18]. The state-of-the-art methods [14, 38, 19] typically employ convolu2
tional neural networks to learn pairwise dense correspondences between the same object instances
at subsequent frames. The SIFT Flow method [25] extends the between-instance correspondences
to cross-instance mappings by matching SIFT features [27] between semantically similar object
instances. Learned-Miller [24] extends the pairwise correspondences to multiple images by posing
a problem of alignment among the images of a set. Collection Flow [22] and Mobahi et al. [29]
project objects onto a low-rank space that allow for joint alignment. FlowWeb [50], and Zhou et
al. [49] construct fully connected graphs to maximise cycle consistency between each image pair and
synthethic data as an intermediary by training a CNN. In our experiments (section 4) flow is known
from synthetic warps or motion, but our work could build on any unsupervised optical flow method.
Unsupervised learning. Classical unsupervised learning methods such as autoencoders [4, 2, 17]
and denoising autoencoders aim to learn useful feature representations from an input by simply
reconstructing it after a bottleneck. Generative adversarial networks [16] target producing samples of
realistic images by training generative models. These models when trained joint with image encoders
are also shown to learn good feature representations [9, 10]. More recently several studies have
emerged that train neural networks by learning auxiliary or pseudo tasks. These methods exploit
typically some existing information in input as ?self-supervision? without any manual labeling by
removing or perturbing some information from an input and requiring a network to reconstruct it.
For instance, Doersch et al. [8], and Noroozi and Favaro [30] train a network to predict the relative
locations of shuffled image patches. Other self-supervised tasks include colorizing images [44],
inpainting [33], ranking frames of a video in temporally correct order [28, 13]. More related to our
approach, Agrawal et al. [1] use egomotion as supervisory signal to learn feature representations in a
Siamese network by predicting camera transformations from image pairs, [32] learn to group pixels
that move together in a video. [48, 15] use a warping-based loss to learn depth from video.
3
Method
This section discusses our method in detail, first introducing the general idea of dense equivariant
labelling (section 3.1), and then presenting a concrete implementation of the latter using a novel deep
CNN architecture (section 3.2).
3.1
Dense equivariant labelling
Consider a 3D object S ? R3 or a class of such objects S that are topologically isomorphic to
a sphere Z ? R3 (i.e. the objects are simple closed surfaces without holes). We can construct a
homeomorphism p = ?S (q) mapping points of the sphere q ? Z to points p ? S of the objects.
Furthermore, if the objects belong to the same semantic category (e.g. faces), we can assume that
these isomorphisms are semantically consistent, in the sense that ?S 0 ? ?S?1 : S ? S 0 maps points of
object S to semantically-analogous points in object S 0 (e.g. for human faces the right eye in one face
should be mapped to the right eye in another [37]).
While this construction is abstract, it shows that we can endow the object (or object category) with
a spherical reference system Z. The authors of [37] build on this construction to define a discrete
system of object landmarks by considering a finite number of points zk ? Z. Here, we take the
geometric embedding idea more literally and propose to explicitly learn a dense mapping from images
of the object to the object-centric coordinate space Z. Formally, we wish to learn a labelling function
? : (x, u) 7? z that takes a RGB image x : ? ? R3 , ? ? R3 and a pixel u ? ? to the object point
z ? Z which is imaged at u (figure 1).
Similarly to [37], this mapping must be compatible or equivariant with image deformations. Namely,
let g : ? ? ? be a deformation of the image domain, either synthetic or due to a viewpoint change
or other motion. Furthermore, let gx = x ? g ?1 be the action of g on the image (obtained by inverse
warp). Barring occlusions and boundary conditions, pixel u in image x must receive the same label
as pixel gu in image gx, which results in the invariance constraint:
?x, u :
?(x, u) = ?(gx, gu).
(1)
Equivalently, we can view the network as a functional x 7? ?(x, ?) that maps the image to a
corresponding label map. Since the label map is an image too, g acts on it by inverse warp.1 Using
1
In the sense that g?(x, ?) = ?(x, ?) ? g ?1 .
3
this, the constraint (1) can be rewritten as the equivariance relation g?(x, ?) = ?(gx, ?). This can be
visualized by noting that the label image deforms in the same way as the input image, as show for
example in figure 3.
For learning, constraint (1) can be incorporated in a loss function as follows:
Z
1
2
L(?|?) =
k?(x, u) ? ?(gx, gu)k du.
|?| ?
However, minimizing this loss has the significant drawback that a global optimum is obtained by
simply setting ?(x, u) = const. The reason for this issue is that (1) is not quite enough to learn a
useful object representation. In order to do so, we must require the labels not only to be equivariant,
but also distinctive, in the sense that
?(x, u) = ?(gx, v)
?
v = gu.
We can encode this requirement as a loss in different ways. For example, by using the fact that points
?(x, u) are on the unit sphere, we can use the loss:
Z
1
2
0
kgu ? argmaxv h?(x, u), ?(gx, v)ik du.
(2)
L (?|x, g) =
|?| ?
By doing so, the labels ?(x, u) must be able to discriminate between different object points, so that a
constant labelling would receive a high penalty.
Relationship with learning invariant visual descriptors. As an alternative to loss (2), we could
have used a pairwise loss2 to encourage the similarity h?(x, u), ?(x0 , gu)i of the labels assigned
to corresponding pixels u and gu to be larger than the similarity h?(x, u), ?(x0 , v)i of the labels
assigned to pixels u and v that do not correspond. Formally, this would result in a pairwise loss
similar to the ones often used to learn invariant visual descriptors for image matching. The reason
why our method learns an object representation instead of a generic visual descriptor is that the
dimensionality of the label space Z is just enough to represent a point on a surface. If we replace
Z with a larger space such as Rd , d 2, we can expect ?(x, u) to learn to extract generic visual
descriptors like SIFT instead. This establishes an interesting relationship between visual descriptors
and object-specific coordinate vectors and suggests that it is possible to transition between the two by
controlling their dimensionality.
3.2
Concrete learning formulation
In this section we introduce a concrete implementation of our method (figure 2). For the mapping
?, we use a CNN that receives as input an image tensor x ? RH?W ?C and produces as output a
label tensor z ? RH?W ?L . We use the notation ?u (x) to indicate the L-dimensional label vector
extracted at pixel u from the label image computed by the network.
The dimension of the label vectors is set to L = 3 (instead of L = 2) in order to allow the network
to express uncertainty about the label assigned to a pixel. The network can do so by modulating
the norm of ?u (x). In fact, correspondences are expressed probabilistically by computing the inner
product of label vectors followed by the softmax operator. Formally, the probability that pixel v in
image x0 corresponds to pixel u in image x is expressed as:
0
eh?u (x),?v (x )i
p(v|u; x, x0 , ?) = P h? (x),? (x0 )i .
u
z
ze
(3)
In this manner, a shorter vector ?u results in a more diffuse probability distribution.
Next, we wish to define a loss function for learning ? from data. To this end, we consider a triplet
? = (x, x0 , g), where x0 = gx is an image that corresponds to x up to transformation g (the nature
2
Formally, this is achieved by the loss
Z
n
o
1
L00 (?|x, g) =
max 0, max ?(u, v) + h?(x, u), ?(gx, v)i ? h?(x, u), ?(gx, gu)i du,
v
|?| ?
where ?(u, v) ? 0 is an error-dependent margin.
4
Optical
flow
Figure 2: Unsupervised dense correspondence network. From left to right: The network ? extracts
label maps ?u (x) and ?v (x0 ) from the image pair x and x0 . An optical flow module (or ground truth
for synthetic transformation) computes the warp (correspondence field) g such that x0 = gx. Then
the label of each point u in the first image is correlated to each point v in the second, obtaining a
number of score maps. The loss evaluates how well the score maps predict the warp g.
of the data is discussed below). We then assess the performance of the network ? on the triplet ?
using two losses. The first loss is the negative log-likelihood of the ground-truth correspondences:
1 X
Llog (?|x, x0 , g) = ?
log p(gu|u; x, x0 , ?).
(4)
HW u
This loss has the advantage that it explicitly learns (3) as the probability of a match. However, it is
not sensitive to the size of a correspondence error v ? gu. In order to address this issue, we also
consider the loss
1 XX
Ldist (?|x, x0 , g) =
kv ? guk?2 p(v|u; x, x0 , ?).
(5)
HW u v
Here ? > 0 is an exponent used to control the robustness of the distance measure, which we set to
? = 0.5, 1.
Nework details. We test two architecture. The first one, denoted S IMPLE, is the same as [47, 37]
and is a chain (5, 20)+ , (2, mp), ?2 , (5, 48)+ , (3, 64)+ , (3, 80)+ , (3, 256)+ , (1, 3) where (h, c) is
a bank of c filters of size h ? h, + denotes ReLU, (h, mp) is h ? h max-pooling, ?s is
s? downsampling. Better performance can be obtained by increasing the support of the filters in the network; for this, we consider a second network D ILATIONS (5, 20)+ , (2, mp), ?2
, (5, 48)+ , (5, 64, 2)+ , (3, 80, 4)+ , (3, 256, 2)+ , (1, 3) where (h, c, d) is a filter with ?d dilation [41].
3.3
Learning from synthetic and true deformations
Losses (4) and (5) learn from triplets ? = (x, x0 , g). Here x0 can be either generated synthetically by
applying a random transformation g to a natural image x [37, 21], or it can be obtained by observing
image pairs (x, x0 ) containing true object deformations arising from a viewpoint change or an object
motion or deformation.
The use of synthetic transformations enables training even on static images and was considered
in [37], who showed it to be sufficient to learn meaningful landmarks for a number of real-world
object such as human and cat faces. Here, in addition to using synthetic deformations, we also
consider using animated image pairs x and x0 . In principle, the learning formulation can be modified
so that knowledge of g is not required; instead, images and their warps can be compared and aligned
directly based on the brightness constancy principle. In our toy video examples we obtain g from the
rendering engine, but it can in theory be obtained using an off-the-shelf optical flow algorithm which
would produce a noisy version of g.
4
Experiments
This section assesses our unsupervised method for dense object labelling on two representative tasks:
two toy problems (sections 4.1 and 4.2) and human and cat faces (section 4.3).
5
Figure 3: Roboarm equivariant labelling. Top: Original video frames of a simple articulated object.
Middle and bottom: learned labels, which change equivariantly with the arm, learned using Llog and
Ldist , respectively. Different colors denote different points of the spherical object frame.
4.1
Roboarm example
In order to illustrate our method we consider a toy problem consisting of a simple articulated object,
namely an animated robotic arm (figure 3) created using a 2D physics engine [36]. We do so for two
reasons: to show that the approach is capable of labelling correctly deformable/articulated objects and
to show that the spherical model Z is applicable also to thin objects, that have mainly a 1D structure.
Dataset details. The arm is anchored to the bottom left corner and is made up of colored capsules
connected with joints having reasonable angle limits to prevent unrealistic contortion and selfocclusion. Motion is achieved by varying the gravity vector, sampling each element from a Gaussian
with standard deviation 15 m s?2 every 100 iterations. Frames x of size 90 ? 90 pixels and the
corresponding flow fields g : x 7? x0 are saved every 20 iterations. We also save the positions of the
capsule centers. The final dataset has 23999 frames.
Learning. Using the correspondences ? = (x, x0 , g) provided by the flow fields, we use our method
to learn an object centric coordinate frame Z and its corresponding labelling function ?u (x). We
test learning ? using the probabilistic loss (4) and distance-based loss (5). In the loss we ignore
areas with zero flow, which automatically removes the background. We use the S IMPLE network
architecture (section 3.2).
Results. Figure 3 provides some qualitative results, showing by means of colormaps the labels ?u (x)
associated to different pixels of each input image. It is easy to see that the method attaches consistent
labels to the different arm elements. The distance-based loss produces a more uniform embedding, as
may be expected. The embeddings are further visualized in Figure 4 by projecting a number of video
frames back to the learned coordinate spaces Z. It can be noted that the space is invariant, in the sense
that the resulting figure is approximately the same despite the fact that the object deforms significantly
in image space. This is true for both embeddings, but the distance-based ones are geometrically more
consistent.
2
2
2
1
1
1
0
0
0
-1
-1
-1
-2
-2
-4
-2
0
2
4
-2
-4
-2
0
2
4
2
2
2
1
1
1
0
0
0
-1
-1
-1
-2
-2
-4
-2
0
2
4
-4
-2
0
2
4
-4
-2
0
2
4
-2
-4
-2
0
2
4
Figure 4: Invariance of the object-centric coordinate space for Roboarm. The plot projects
frames 3,6,9 of figure 3 on the object-centric coordinate space Z, using the embedding functions
learned by means of the probabilistic (top) and distance (bottom) based losses. The sphere is then
unfolded, plotting latitude and longitude (in radians) along the vertical and horizontal axes.
6
Log
Dist 0.5
Dist 1
Ground truth
Figure 5: Left: Embedding spaces of different dimension. Spherical embedding (from the 3D
embedding function ?u (x) ? R3 ) learned using the distance loss compared to a circular embedding
with one dimension less. Right: Capsule center prediction for different losses.
Predicting capsule centers. We evaluate quantitatively the ability of our object frames to localise the
capsule centers. If our assumption is correct and a coordinate system intrinsic to the object has been
learned, then we should expect there to be a specific 3-vector in Z corresponding to each center, and
our job is to find these vectors. Various strategies could be used, such as averaging the object-centric
coordinates given to the centers over the training set, but we choose to incorporate the problem into
the learning framework. This is done using the negative log-likelihood in much the same way as
(4), limiting our vectors u to the centers. This is done as an auxiliary layer with no backpropagation
to the rest of the network, so that the embedding remains unsupervised. The error reported is the
Euclidean distance as a percentage of the image width.
Results are given for the different loss functions used for unsupervised training in Table 1 and
visualized in Figure 5 right, showing that the object centers can be located to a high degree of
accuracy. The negative log likelihood performs best while the two losses incorporating distance
perform similarly.
We also perform experiments varying the dimensionality L of the label space Z (Table 2). Perhaps
most interestingly, given the almost one-dimensional nature of the arm, is the case of L = 2, which
would correspond to an approximately circular space (since the length of vectors is used to code
for uncertainty). As seen in the right of Figure 5 left, the segments are represented almost perfectly
on the boundary of a circle, with the exception of the bifurcation which it is unable to accurately
represent. This is manifested by the light blue segment trying, and failing, to be in two places at once.
Unsupervised Loss
Error
Llog
0.97 %
Ldist , ? = 1
1.13 %
Ldist , ? = 0.5
1.14 %
Table 1: Predicting capsule centers.
Error as percent of image width.
4.2
Textured sphere example
Descriptor Dimension
2
3
5
20
Error
1.29 %
1.14 %
1.16 %
1.28 %
Table 2: Descriptor dimension
(Ldist , ? = 0.5). L>3 shows no improvement,
suggesting L=3 is the natural manifold of the arm.
The experiment of Figure 6 tests the ability of the method to understand a complete rotation of a 3D
object, a simple textured sphere. Despite the fact that the method is trained on pairs of adjacent video
frames (and corresponding optical flow), it still learns a globally-consistent embedding. However,
this required switching from from the S IMPLE to the D ILATIONS architecture (section 3.2).
4.3
Faces
After testing our method on a toy problem, we move to a much harder task and apply our method to
generate an object-centric reference frame Z for the category of human faces. In order to generate
an image pair and corresponding flow field for training we warp each face synthetically using Thin
Plate Spline warps in a manner similar to [37]. We train our models on the extensive CelebA [26]
dataset of over 200k faces as in [37], excluding MAFL [47] test overlap from the given training split.
It has annotations of the eyes, nose and mouth corners. Note that we do not use these to train our
model. We also use AFLW [23], testing on 2995 faces [47, 40, 46] with 5 landmarks. Like [37] we
use 10,122 faces for training. We additionally evaluate qualitatively on a dataset of cat faces [45],
using 8609 images for training.
Qualitative assessment. We find that for network S IMPLE the negative log-likelihood loss, while
performing best for the simple example of the arm, performs poorly on faces. Specifically, this model
7
Figure 6: Sphere equivariant labelling. Top: video frames of a rotating textured sphere. Middle:
learned dense labels, which change equivariantly with the sphere. Bottom: re-projection of the video
frames on the object frame (also spherical). Except for occlusions, the reprojections are approximately
invariant, correctly mapping the blue and orange sides to different regions of the label space
fails to disambiguate the left and right eye, as shown in Figure 9 (right). The distance-based loss (5)
produces a more coherent embedding, as seen in Figure 9 (left). Using D ILATIONS this problem
disappears, giving qualitatively smooth and unambiguous labels for both the distance loss (Figure 7)
and the log-likelihood loss (Figure 8). For cats our method is able to learn a consistent object frame
despite large variations in appearance (Figure 8).
Figure 7: Faces. D ILATIONS network with Ldist , ? = 0.5. Top: Input images, Middle: Predicted
dense labels mapped to colours, Bottom: Image pixels mapped to label sphere and flattened.
Figure 8: Cats. D ILATIONS network with Llog . Top: Input images, Middle: Labels mapped to
colours, Bottom: Images mapped to the spherical object frames.
8
Figure 9: Annotated landmark prediction from the shown unsupervised label maps (S IMPLE
network). Left: Trained with Ldist , ? = 0.5, Right: Failure to disambiguate eyes with Llog .
(Prediction: green, Ground truth: Blue)
Regressing semantic landmarks. We would like to quantify the accuracy of our model in terms of
ability to consistently locate manually annotated points, specifically the eyes, nose, and mouth corners
given in the CelebA dataset. We use the standard test split for evaluation of the MAFL dataset [47],
containing 1000 images. We also use the MAFL training subset of 19k images for learning to predict
the ground truth landmarks, which gives a quantitative measure of the consistency of our object frame
for detecting facial features. These are reported as Euclidean error normalized as a percentage of
inter-ocular distance.
In order to map the object frame to the semantic landmarks, as in the case of the robot arm centers, we
learn the vectors zk ? Z corresponding to the position of each point in our canonical reference space
and then, for any given image, find the nearest z and its corresponding pixel location u. We report
the localization performance of this model in Table 3 (?Error Nearest?). We empirically validate
that with the S IMPLE network the negative log-likelihood is not ideal for this task (Figure 9) and
we obtain higher performance for the robust distance with power 0.5. However, after switching to
D ILATIONS to increase the receptive field both methods perform comparably.
The method of [37] learns to regress P ground truth coordinates based on M > P unsupervised
landmarks. By regressing from multiple points it is not limited to integer pixel coordinates. While we
are not predicting landmarks as network output, we can emulate this method by allowing multiple
points in our object coordinate space to be predictive for a single ground truth landmark. We learn
one regressor per ground truth point, each formulated as a linear regressor R2M ? R2 on top of
coordinates from M = 50 learned intermediate points. This allows the regression to say which points
in Z are most useful for predicting each ground truth point.
We also report results after unsupervised finetuning of a CelebA network to the more challenging
AFLW followed by regressor training on AFLW. As shown in Tables 3 and 4, we outperform other
unsupervised methods on both datasets, and are comparable to fully supervised methods.
Network
Unsup. Loss
S IMPLE
S IMPLE
S IMPLE
Llog
Ldist , ? = 1
Ldist , ? = 0.5
Llog
Ldist , ? = 0.5
Error
Nearest
Error
Regress
?
7.94 %
7.18 %
D ILATIONS
5.83 %
D ILATIONS
5.87 %
[37]
6.67 %
Table 3: Nearest neighbour and regression
landmark prediction on MAFL
5
75.02 %
14.57 %
13.29 %
11.05 %
10.53 %
Method
Error
RCPR [5]
11.6 %
Cascaded CNN [35]
8.97 %
CFAN [43]
10.94 %
TCDCN [47]
7.65 %
RAR [40]
7.23 %
Unsup. Landmarks [37]
10.53 %
D ILATIONS Ldist , ? = 0.5
8.80 %
Table 4: Comparison with supervised and unsupervised methods on AFLW
Conclusions
Building on the idea of viewpoint factorization, we have introduce a new method that can endow
an object or object category with an invariant dense geometric embedding automatically, by simply
observing a large dataset of unlabelled images. Our learning framework combines in a novel way
the concept of equivariance with the one of distinctiveness. We have also proposed a concrete
implementation using novel losses to learn a deep dense image labeller. We have shown empirically
that the method can learn a consistent geometric embedding for a simple articulated synthetic robotic
arm as well as for a 3D sphere model and real faces. The resulting embeddings are invariant to
deformations and, importantly, to intra-category variations.
Acknowledgments: This work acknowledges the support of the AIMS CDT (EPSRC EP/L015897/1) and ERC
677195-IDIU. Clipart: FreePik.
9
References
[1] Pulkit Agrawal, Joao Carreira, and Jitendra Malik. Learning to see by moving. In Proc. ICCV,
2015.
[2] Yoshua Bengio. Learning deep architectures for AI. Foundations and trends in Machine
Learning, 2009.
[3] Fred L. Bookstein. Principal Warps: Thin-Plate Splines and the Decomposition of Deformations.
PAMI, 1989.
[4] H Bourlard and Y Kamp. Auto-Association by Multilayer Perceptrons and Singular Value
Decomposition. Biological Cybernetics, 1988.
[5] Xavier P. Burgos-Artizzu, Pietro Perona, and Piotr Doll?r. Robust face landmark estimation
under occlusion-supp. mat. In Proc. ICCV, 2013.
[6] T F Cootes, C J Taylor, D H Cooper, and J Graham. Active shape models: their training and
application. CVIU, 1995.
[7] Navneet Dalal and Bill Triggs. Histograms of Oriented Gradients for Human Detection. In
Proc. CVPR, 2005.
[8] Carl Doersch, Abhinav Gupta, and Alexei A Efros. Unsupervised Visual Representation
Learning by Context Prediction. In Proc. ICCV, 2015.
[9] Jeff Donahue, Philipp Kr?henb?hl, and Trevor Darrell. Adversarial feature learning. Proc.
ICLR, 2017.
[10] Vincent Dumoulin, Ishmael Belghazi, Ben Poole, Alex Lamb, Martin Arjovsky, Olivier Mastropietro, and Aaron Courville. Adversarially learned inference. Proc. ICLR, 2017.
[11] Pedro F. Felzenszwalb, Ross B. Girshick, David McAllester, and Deva Ramanan. Object
Detection with Discriminatively Trained Part Based Models. PAMI, 2010.
[12] Rob Fergus, Pietro Perona, and Andrew Zisserman. Object class recognition by unsupervised
scale-invariant learning. In Proc. CVPR, 2003.
[13] Basura Fernando, Hakan Bilen, Efstratios Gavves, and Stephen Gould. Self-supervised video
representation learning with odd-one-out networks. In Proc. CVPR, 2017.
[14] Philipp Fischer, Alexey Dosovitskiy, Eddy Ilg, Philip H?usser, Caner Haz?rba?s, Vladimir Golkov,
Patrick van der Smagt, Daniel Cremers, and Thomas Brox. FlowNet: Learning Optical Flow
with Convolutional Networks. In Proc. ICCV, 2015.
[15] Ravi Garg, Gustavo Carneiro, and Ian Reid. Unsupervised cnn for single view depth estimation:
Geometry to the rescue. In Proc. ECCV, pages 740?756, 2016.
[16] Ian Goodfellow, Yoshua Bengio, and Aaron Courville. Deep Learning. MIT Press, 2016.
http://www.deeplearningbook.org.
[17] G E Hinton and R R Salakhutdinov. Reducing the Dimensionality of Data with Neural Networks.
Science, 2006.
[18] Berthold K.P. Horn and Brian G. Schunck. Determining optical flow. Artificial Intelligence,
1981.
[19] Eddy Ilg, Nikolaus Mayer, Tonmoy Saikia, Margret Keuper, Alexey Dosovitskiy, and Thomas
Brox. FlowNet 2.0: Evolution of Optical Flow Estimation with Deep Networks. arXiv preprint
arXiv:1612.01925, 2016.
[20] Max Jaderberg, Karen Simonyan, Andrew Zisserman, and Koray Kavukcuoglu. Spatial Transformer Networks. In Proc. NIPS, 2015.
[21] A. Kanazawa, D. W. Jacobs, and M. Chandraker. WarpNet: Weakly supervised matching for
single-view reconstruction. In Proc. CVPR, 2016.
[22] Ira Kemelmacher-Shlizerman and Steven M. Seitz. Collection flow. In Proc. CVPR, 2012.
[23] Martin Koestinger, Paul Wohlhart, Peter M. Roth, and Horst Bischof. Annotated facial landmarks in the wild: A large-scale, real-world database for facial landmark localization. In First
10
IEEE International Workshop on Benchmarking Facial Image Analysis Technologies, 2011.
[24] Erik G Learned-Miller. Data driven image models through continuous joint alignment. IEEE
Transactions on Pattern Analysis and Machine Intelligence, 2006.
[25] Ce Liu, Jenny Yuen, and Antonio Torralba. SIFT Flow: Dense correspondence across scenes
and its applications. PAMI, 2011.
[26] Ziwei Liu, Ping Luo, Xiaogang Wang, and Xiaoou Tang. Deep learning face attributes in the
wild. In Proc. ICCV, 2015.
[27] David G Lowe. Distinctive image features from scale-invariant keypoints. International journal
of computer vision, 60(2):91?110, 2004.
[28] Ishan Misra, C Lawrence Zitnick, and Martial Hebert. Shuffle and learn: unsupervised learning
using temporal order verification. In Proc. ECCV, 2016.
[29] Hossein Mobahi, Ce Liu, and William T. Freeman. A Compositional Model for LowDimensional Image Set Representation. Proc. CVPR, 2014.
[30] Mehdi Noroozi and Paolo Favaro. Unsupervised learning of visual representations by solving
jigsaw puzzles. In Proc. ECCV, 2016.
[31] D. Novotny, D. Larlus, and A. Vedaldi. Learning 3d object categories by looking around them.
In Proc. ICCV, 2017.
[32] Deepak Pathak, Ross Girshick, Piotr Doll?r, Trevor Darrell, and Bharath Hariharan. Learning
features by watching objects move. In Proc. CVPR, 2017.
[33] Deepak Pathak, Philipp Krahenbuhl, Jeff Donahue, Trevor Darrell, and Alexei A Efros. Context
Encoders: Feature Learning by Inpainting. In Proc. CVPR, 2016.
[34] I. Rocco, R. Arandjelovi?c, and J. Sivic. Convolutional neural network architecture for geometric
matching. In Proc. CVPR, 2017.
[35] Yi Sun, Xiaogang Wang, and Xiaoou Tang. Deep convolutional network cascade for facial point
detection. In Proc. CVPR, 2013.
[36] Yuval Tassa. CapSim - the MATLAB physics engine. https://mathworks.com/
matlabcentral/fileexchange/29249-capsim-the-matlab-physics-engine.
[37] James Thewlis, Hakan Bilen, and Andrea Vedaldi. Unsupervised learning of object landmarks
by factorized spatial embeddings. In Proc. ICCV, 2017.
[38] James Thewlis, Shuai Zheng, Philip H. S. Torr, and Andrea Vedaldi. Fully-Trainable Deep
Matching. In Proc. BMVC, 2016.
[39] Markus Weber, Max Welling, and Pietro Perona. Towards automatic discovery of object
categories. In Proc. CVPR, 2000.
[40] Shengtao Xiao, Jiashi Feng, Junliang Xing, Hanjiang Lai, Shuicheng Yan, and Ashraf Kassim.
Robust Facial Landmark Detection via Recurrent Attentive-Refinement Networks. In Proc.
ECCV, 2016.
[41] Fisher Yu and Vladlen Koltun. Multi-scale context aggregation by dilated convolutions. In Proc.
ICLR, 2016.
[42] Xiang Yu, Feng Zhou, and Manmohan Chandraker. Deep Deformation Network for Object
Landmark Localization. In Proc. ECCV, Cham, 2016.
[43] Jie Zhang, Shiguang Shan, Meina Kan, and Xilin Chen. Coarse-to-fine auto-encoder networks
(CFAN) for real-time face alignment. In Proc. ECCV, 2014.
[44] Richard Zhang, Phillip Isola, and Alexei A Efros. Colorful Image Colorization. In Proc. ECCV,
2016.
[45] Weiwei Zhang, Jian Sun, and Xiaoou Tang. Cat head detection - How to effectively exploit
shape and texture features. In Proc. ECCV, 2008.
[46] Zhanpeng Zhang, Ping Luo, Chen Change Loy, and Xiaoou Tang. Facial landmark detection by
deep multi-task learning. In Proc. ECCV, 2014.
11
[47] Zhanpeng Zhang, Ping Luo, Chen Change Loy, and Xiaoou Tang. Learning Deep Representation
for Face Alignment with Auxiliary Attributes. PAMI, 2016.
[48] Tinghui Zhou, Matthew Brown, Noah Snavely, and David G Lowe. Unsupervised learning of
depth and ego-motion from video. In Proc. CVPR, 2017.
[49] Tinghui Zhou, Philipp Kr?henb?hl, Mathieu Aubry, Qixing Huang, and Alexei A. Efros.
Learning Dense Correspondences via 3D-guided Cycle Consistency. In Proc. CVPR, 2016.
[50] Tinghui Zhou, Yong Jae Lee, Stella X. Yu, and Alexei A. Efros. FlowWeb: Joint image set
alignment by weaving consistent, pixel-wise correspondences. In Proc. CVPR, 2015.
12
| 6686 |@word cnn:8 middle:4 version:1 dalal:1 norm:1 triggs:1 seitz:1 shuicheng:1 rgb:1 decomposition:2 jacob:1 brightness:1 inpainting:2 harder:1 liu:3 score:2 daniel:1 interestingly:1 animated:2 existing:1 com:1 luo:3 must:7 gavves:1 subsequent:1 realistic:1 shape:2 enables:1 remove:1 plot:1 localise:1 generative:2 selected:1 intelligence:2 colored:1 mental:1 provides:1 detecting:1 contribute:1 location:2 gx:11 philipp:4 org:1 coarse:1 zhang:5 favaro:2 along:1 ik:1 koltun:1 qualitative:2 combine:1 wild:2 introduce:3 manner:5 x0:20 pairwise:4 inter:1 expected:1 andrea:3 equivariant:12 nor:1 dist:2 multi:2 salakhutdinov:1 globally:1 spherical:6 freeman:1 automatically:3 unfolded:1 considering:1 increasing:1 project:3 xx:1 notation:1 joao:1 provided:1 factorized:1 nework:1 keuper:1 finding:1 transformation:8 pseudo:1 quantitative:1 every:2 temporal:1 act:1 unwanted:1 gravity:1 uk:2 control:1 unit:1 ramanan:1 haz:1 colorful:1 producing:1 reid:1 maximise:1 limit:1 switching:2 despite:3 oxford:1 analyzing:1 approximately:3 pami:4 alexey:2 garg:1 suggests:1 challenging:1 factorization:4 limited:1 acknowledgment:1 camera:1 horn:1 testing:2 practice:1 backpropagation:1 area:1 deforms:2 yan:1 significantly:1 vedaldi:4 matching:6 projection:1 cascade:1 onto:1 operator:1 context:3 transformer:3 applying:1 tangled:1 www:1 bill:1 map:12 center:10 roth:1 starting:1 importantly:1 embedding:18 coordinate:21 variation:5 analogous:1 limiting:1 imagine:1 target:1 construction:2 controlling:1 olivier:1 carl:1 goodfellow:1 element:2 trend:1 ze:1 recognition:1 located:1 ego:1 predicts:1 database:1 observed:1 constancy:1 module:1 bottom:6 epsrc:1 ep:1 capture:1 preprint:1 thousand:2 wang:2 region:1 connected:2 cycle:2 sun:2 shuffle:1 imple:10 trained:4 weakly:1 deva:1 segment:2 solving:1 predictive:1 purely:1 distinctive:3 localization:3 unsup:2 textured:3 gu:9 easily:1 joint:5 finetuning:1 differently:1 xiaoou:5 cat:6 various:1 represented:1 emulate:1 carneiro:1 articulated:6 train:5 detected:2 artificial:1 labeling:1 basura:1 quite:2 emerged:1 larger:2 cvpr:14 say:1 reconstruct:1 encoder:1 ability:4 simonyan:1 fischer:1 noisy:1 final:1 agrawal:2 advantage:1 propose:3 reconstruction:1 lowdimensional:1 product:1 aligned:1 poorly:1 achieve:1 deformable:2 kv:1 validate:1 requirement:2 optimum:1 darrell:3 produce:4 ben:1 object:95 illustrate:1 andrew:2 ac:2 pose:1 recurrent:1 nearest:4 odd:1 school:1 job:1 longitude:1 auxiliary:3 homeomorphism:1 predicted:1 come:1 indicate:1 hakan:3 quantify:1 guided:1 drawback:1 correct:2 attribute:2 cnns:1 filter:3 saved:1 annotated:3 human:8 mcallester:1 caner:1 rar:1 require:1 behaviour:1 generalization:1 yuen:1 biological:1 brian:1 around:1 considered:1 ground:9 lawrence:1 mapping:8 predict:3 puzzle:1 matthew:1 efros:5 torralba:1 failing:1 estimation:4 proc:35 intermediary:1 applicable:1 label:32 ross:2 sensitive:1 modulating:1 ilg:2 establishes:1 cdt:1 mit:1 gaussian:1 rba:1 aim:2 modified:1 rather:1 zhou:5 shelf:1 varying:2 probabilistically:1 endow:2 encode:1 ax:1 ira:1 improvement:1 consistently:2 rank:1 likelihood:6 mainly:1 adversarial:2 detect:1 sense:4 tional:1 inference:1 flowweb:2 xilin:1 dependent:1 typically:2 perona:3 relation:1 smagt:1 pixel:20 issue:2 among:1 hossein:1 denoted:1 exponent:1 spatial:3 art:1 softmax:1 brox:2 bifurcation:1 orange:1 construct:4 field:5 having:2 beach:1 barring:1 manually:2 sampling:1 once:1 piotr:2 look:1 unsupervised:24 adversarially:1 thin:5 koray:1 celeba:3 yu:3 report:2 spline:4 quantitatively:1 jdt:1 few:1 employ:1 yoshua:2 oriented:1 neighbour:1 dosovitskiy:2 richard:1 geometry:2 occlusion:4 consisting:1 william:1 detection:8 circular:2 intra:1 alexei:5 regressing:2 guk:1 alignment:7 evaluation:1 bilen:2 zheng:1 light:1 chain:1 encourage:2 capable:1 shorter:1 facial:8 literally:1 indexed:1 pulkit:1 euclidean:2 taylor:1 rotating:1 circle:1 re:1 deformation:15 girshick:2 instance:10 kemelmacher:1 modeling:1 ishmael:1 applicability:1 introducing:1 deviation:1 subset:1 uniform:1 jiashi:1 too:1 reported:2 arandjelovi:1 dependency:1 encoders:2 synthetic:10 st:1 international:2 probabilistic:3 off:1 informatics:1 physic:3 regressor:3 lee:1 together:1 concrete:4 flownet:2 containing:2 choose:1 huang:1 watching:1 corner:3 toy:4 supp:1 suggesting:1 dilated:1 jitendra:1 cremers:1 explicitly:3 ranking:1 mp:3 view:3 lowe:2 closed:1 dumoulin:1 doing:1 observing:2 jigsaw:1 xing:1 aggregation:1 annotation:1 ass:2 hariharan:1 accuracy:2 convolutional:4 descriptor:7 who:1 kassim:1 miller:2 correspond:2 kamp:1 vincent:1 kavukcuoglu:1 accurately:1 comparably:1 steven:1 none:1 cybernetics:1 bharath:1 detector:2 ping:3 manual:3 ed:1 trevor:3 evaluates:1 failure:1 attentive:1 ocular:1 james:3 regress:2 associated:1 static:1 radian:1 dataset:8 aubry:1 knowledge:1 emerges:1 improves:1 dimensionality:4 color:1 eddy:2 usser:1 back:1 coloring:1 centric:9 higher:1 supervised:7 zisserman:2 bmvc:1 formulation:4 done:3 ox:1 furthermore:2 just:2 shuai:1 autoencoders:2 receives:1 horizontal:1 mehdi:1 assessment:1 perhaps:1 supervisory:1 usa:1 building:1 phillip:1 concept:5 requiring:1 true:3 normalized:1 evolution:1 hence:1 assigned:3 shuffled:1 xavier:1 imaged:1 semantic:4 adjacent:1 during:1 self:3 nuisance:2 width:2 unambiguous:1 illustrative:1 noted:1 trying:1 plate:4 presenting:1 complete:1 demonstrate:1 performs:2 motion:7 percent:1 image:68 weber:1 wise:1 novel:3 recently:2 common:2 rotation:1 deeplearningbook:1 ishan:1 functional:1 empirically:2 perturbing:1 tassa:1 extend:1 belong:1 discussed:1 association:1 shlizerman:1 measurement:1 significant:2 ai:1 doersch:2 rd:1 consistency:3 automatic:1 similarly:2 erc:1 moving:2 robot:2 supervision:6 surface:4 similarity:2 patrick:1 recent:2 showed:1 driven:1 misra:1 manifested:1 life:1 der:1 yi:1 cham:1 argmaxv:1 seen:2 arjovsky:1 tinghui:3 isola:1 fernando:1 signal:1 jenny:1 relates:1 multiple:3 siamese:1 stephen:1 keypoints:1 smooth:2 technical:1 match:1 unlabelled:1 cross:1 long:1 sphere:13 lai:1 prediction:5 regression:2 multilayer:1 vision:2 essentially:1 arxiv:2 iteration:2 represent:2 histogram:1 achieved:2 receive:2 addition:1 remarkably:1 background:1 fine:1 singular:1 jian:1 rest:1 pooling:1 flow:21 integer:1 golkov:1 noting:1 ideal:2 mastropietro:1 synthetically:2 embeddings:5 enough:3 rendering:1 easy:1 split:2 relu:1 intermediate:1 bengio:2 architecture:7 perfectly:1 inner:1 idea:7 bottleneck:1 isomorphism:1 colour:2 labeller:1 penalty:1 peter:1 henb:2 karen:1 wohlhart:1 action:1 compositional:1 deep:14 antonio:1 useful:3 matlab:2 jie:1 aimed:1 locally:1 visualized:3 category:11 generate:2 http:2 outperform:1 percentage:2 canonical:2 rescue:1 extrinsic:1 arising:1 correctly:2 per:1 blue:3 discrete:3 mat:1 affected:1 express:2 group:2 key:1 paolo:1 prevent:1 neither:1 ce:2 ravi:1 graph:1 geometrically:1 pietro:3 inverse:2 angle:1 uncertainty:3 cootes:1 topologically:1 extends:2 almost:2 reasonable:1 place:1 lamb:1 patch:1 weiwei:1 krahenbuhl:1 comparable:1 graham:1 layer:1 shan:1 followed:2 courville:2 correspondence:15 refine:1 xiaogang:2 noah:1 constraint:5 alex:1 scene:1 diffuse:1 yong:1 markus:1 nearby:2 performing:1 optical:12 saikia:1 martin:2 gould:1 combination:1 vladlen:1 across:4 reconstructing:1 rob:1 larlus:1 hl:2 projecting:1 invariant:11 iccv:7 bookstein:1 remains:1 discus:1 r3:6 mathworks:1 kgu:1 nose:2 weaving:1 end:1 rewritten:1 doll:2 apply:1 generic:2 nikolaus:1 r2m:1 save:1 alternative:1 robustness:1 original:1 thomas:2 denotes:1 top:6 include:1 const:1 homeomorphic:1 exploit:2 giving:1 build:4 classical:1 feng:2 warping:1 objective:1 move:3 tensor:2 malik:1 manmohan:1 strategy:1 receptive:1 rocco:1 snavely:1 traditional:1 gradient:1 iclr:3 distance:13 unable:1 mapped:7 landmark:28 warpnet:3 philip:2 manifold:1 trivial:1 reason:3 erik:1 length:1 code:1 relationship:2 colorization:1 minimizing:1 downsampling:1 innovation:1 equivalently:1 vladimir:1 loy:2 negative:5 rise:1 ziwei:1 implementation:4 shiguang:1 perform:3 allowing:1 vertical:1 observation:1 convolution:1 datasets:1 finite:1 hinton:1 incorporated:1 excluding:1 looking:1 frame:27 locate:1 head:1 hbilen:1 introduced:1 david:3 pair:8 namely:2 required:2 extensive:1 mayer:1 bischof:1 sivic:1 engine:4 coherent:1 learned:14 nip:2 address:1 beyond:1 able:2 poole:1 below:1 perception:1 pattern:1 articulation:1 latitude:1 challenge:1 including:1 max:5 video:12 green:1 mouth:2 unrealistic:1 overlap:1 pathak:2 natural:2 eh:1 power:1 predicting:5 cascaded:1 bourlard:1 arm:9 technology:1 eye:7 temporally:2 abhinav:1 disappears:1 martial:1 created:1 brown:1 acknowledges:1 mathieu:1 stella:1 extract:4 auto:2 geometric:10 discovery:1 determining:1 relative:1 xiang:1 loss:32 fully:3 expect:2 discriminatively:1 interesting:1 attache:2 remarkable:1 foundation:1 degree:1 sufficient:2 consistent:10 verification:1 xiao:1 principle:2 viewpoint:8 bank:1 egomotion:1 plotting:1 navneet:1 eccv:9 compatible:2 summary:1 hebert:1 side:1 allow:3 understand:1 warp:12 distinctiveness:2 face:24 felzenszwalb:1 deepak:2 sparse:1 edinburgh:1 van:1 boundary:2 depth:3 dimension:5 transition:1 world:2 fred:1 computes:1 berthold:1 author:2 collection:2 made:1 qualitatively:2 horst:1 refinement:1 welling:1 transaction:1 ignore:1 l00:1 jaderberg:1 belghazi:1 global:1 robotic:2 active:1 chandraker:2 conclude:1 fergus:1 continuous:1 latent:1 triplet:3 anchored:1 why:1 dilation:1 additionally:1 disambiguate:2 table:8 capsule:6 learn:28 zk:2 robust:4 ca:1 synthethic:1 nature:2 obtaining:1 du:3 posing:1 complex:3 domain:1 equivariance:4 zitnick:1 dense:20 rh:2 paul:1 jae:1 body:1 augmented:1 representative:1 benchmarking:1 cooper:1 fails:1 position:3 explicit:2 wish:2 third:1 learns:4 hw:2 donahue:2 ian:2 removing:2 loss2:1 tang:5 specific:2 sift:4 mobahi:2 showing:2 r2:1 gupta:1 intrinsic:3 incorporating:1 kanazawa:1 gustavo:1 workshop:1 effectively:1 flattened:1 kr:2 texture:1 labelling:14 illumination:1 hole:1 margin:1 cviu:1 chen:3 simply:3 appearance:2 visual:15 schunck:1 expressed:2 pedro:1 corresponds:2 truth:9 kan:1 complemented:2 extracted:1 identity:1 goal:2 formulated:1 towards:1 jeff:2 replace:1 fisher:1 change:6 carreira:1 specifically:2 except:1 reducing:1 semantically:6 averaging:1 llog:7 denoising:1 principal:1 yuval:1 colormaps:1 torr:1 isomorphic:1 invariance:2 discriminate:1 meaningful:2 perceptrons:1 exception:1 formally:4 aaron:2 qixing:1 support:2 latter:1 ashraf:1 incorporate:1 evaluate:2 trainable:1 correlated:1 |
6,286 | 6,687 | Compression-aware Training of Deep Networks
Mathieu Salzmann
EPFL - CVLab
Lausanne, Switzerland
[email protected]
Jose M. Alvarez
Toyota Research Institute
Los Altos, CA 94022
[email protected]
Abstract
In recent years, great progress has been made in a variety of application domains
thanks to the development of increasingly deeper neural networks. Unfortunately,
the huge number of units of these networks makes them expensive both computationally and memory-wise. To overcome this, exploiting the fact that deep networks
are over-parametrized, several compression strategies have been proposed. These
methods, however, typically start from a network that has been trained in a standard manner, without considering such a future compression. In this paper, we
propose to explicitly account for compression in the training process. To this end,
we introduce a regularizer that encourages the parameter matrix of each layer to
have low rank during training. We show that accounting for compression during
training allows us to learn much more compact, yet at least as effective, models
than state-of-the-art compression techniques.
1
Introduction
With the increasing availability of large-scale datasets, recent years have witnessed a resurgence of
interest for Deep Learning techniques. Impressive progress has been made in a variety of application
domains, such as speech, natural language and image processing, thanks to the development of new
learning strategies [15, 53, 30, 45, 26, 3] and of new architectures [31, 44, 46, 23]. In particular, these
architectures tend to become ever deeper, with hundreds of layers, each of which containing hundreds
or even thousands of units.
While it has been shown that training such very deep architectures was typically easier than smaller
ones [24], it is also well-known that they are highly over-parameterized. In essence, this means that
equally good results could in principle be obtained with more compact networks. Automatically
deriving such equivalent, compact models would be highly beneficial in runtime- and memorysensitive applications, e.g., to deploy deep networks on embedded systems with limited hardware
resources. As a consequence, many methods have been proposed to compress existing architectures.
An early trend for such compression consisted of removing individual parameters [33, 22] or entire
units [36, 29, 38] according to their influence on the output. Unfortunately, such an analysis of
individual parameters or units quickly becomes intractable in the presence of very deep networks.
Therefore, currently, one of the most popular compression approaches amounts to extracting low-rank
approximations either of individual units [28] or of the parameter matrix/tensor of each layer [14].
This latter idea is particularly attractive, since, as opposed to the former one, it reduces the number of
units in each layer. In essence, the above-mentioned techniques aim to compress a network that has
been pre-trained. There is, however, no guarantee that the parameter matrices of such pre-trained
networks truly have low-rank. Therefore, these methods typically truncate some of the relevant
information, thus resulting in a loss of prediction accuracy, and, more importantly, do not necessarily
achieve the best possible compression rates.
In this paper, we propose to explicitly account for compression while training the initial deep network.
Specifically, we introduce a regularizer that encourages the parameter matrix of each layer to have
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
low rank in the training loss, and rely on a stochastic proximal gradient descent strategy to optimize
the network parameters. In essence, and by contrast with methods that aim to learn uncorrelated units
to prevent overfitting [5, 54, 40], we seek to learn correlated ones, which can then easily be pruned
in a second phase. Our compression-aware training scheme therefore yields networks that are well
adapted to the following post-processing stage. As a consequence, we achieve higher compression
rates than the above-mentioned techniques at virtually no loss in prediction accuracy.
Our approach constitutes one of the very few attempts at explicitly training a compact network
from scratch. In this context, the work of [4] has proposed to learn correlated units by making
use of additional noise outputs. This strategy, however, is only guaranteed to have the desired
effect for simple networks and has only been demonstrated on relatively shallow architectures. In
the contemporary work [51], units are coordinated via a regularizer acting on all pairs of filters
within a layer. While effective, exploiting all pairs can quickly become cumbersome in the presence
of large numbers of units. Recently, group sparsity has also been employed to obtain compact
networks [2, 50]. Such a regularizer, however, acts on individual units, without explicitly aiming to
model their redundancies. Here, we show that accounting for interactions between the units within a
layer allows us to obtain more compact networks. Furthermore, using such a group sparsity prior in
conjunction with our compression-aware strategy lets us achieve even higher compression rates.
We demonstrate the benefits of our approach on several deep architectures, including the 8-layers
DecomposeMe network of [1] and the 50-layers ResNet of [23]. Our experiments on ImageNet and
ICDAR show that we can achieve compression rates of more than 90%, thus hugely reducing the
number of required operations at inference time.
2
Related Work
It is well-known that deep neural networks are over-parametrized [13]. While, given sufficient
training data, this seems to facilitate the training procedure, it also has two potential drawbacks. First,
over-parametrized networks can easily suffer from overfitting. Second, even when they can be trained
successfully, the resulting networks are expensive both computationally and memory-wise, thus
making their deployment on platforms with limited hardware resources, such as embedded systems,
challenging. Over the years, much effort has been made to overcome these two drawbacks.
In particular, much progress has been made to reduce overfitting, for example by devising new
optimization strategies, such as DropOut [45] or MaxOut [16]. In this context, other works have
advocated the use of different normalization strategies, such as Batch Normalization [26], Weight
Normalization [42] and Layer Normalization [3]. Recently, there has also been a surge of methods
aiming to regularize the network parameters by making the different units in each layer less correlated.
This has been achieved by designing new activation functions [5], by explicitly considering the
pairwise correlations of the units [54, 37, 40] or of the activations [9, 52], or by constraining the
weight matrices of each layer to be orthonormal [21].
In this paper, we are more directly interested in addressing the second drawback, that is, the large
memory and runtime required by very deep networks. To tackle this, most existing research has
focused on pruning pre-trained networks. In this context, early works have proposed to analyze the
saliency of individual parameters [33, 22] or units [36, 29, 38, 34], so as to measure their impact on
the output. Such a local analysis, however, quickly becomes impractically expensive when dealing
with networks with millions of parameters.
As a consequence, recent works have proposed to focus on more global methods, which analyze
larger groups of parameters simultaneously. In this context, the most popular trend consists of
extracting low-rank approximations of the network parameters. In particular, it has been shown that
individual units can be replaced by rank 1 approximations, either via a post-processing step [28, 46]
or directly during training [1, 25]. Furthermore, low-rank approximations of the complete parameter
matrix/tensor of each layer were computed in [14], which has the benefit of reducing the number of
units in each layer. The resulting low-rank representation can then be fine-tuned [32], or potentially
even learned from scratch [47], given the rank of each layer in the network. With the exception of
this last work, which assumes that the ranks are known, these methods, however, aim to approximate
a given pre-trained model. In practice, however, the parameter matrices of this model might not have
low rank. Therefore, the resulting approximations yield some loss of accuracy and, more importantly,
2
will typically not correspond to the most compact networks. Here, we propose to explicitly learn a
low-rank network from scratch, but without having to manually define the rank of each layer a priori.
To this end, and in contrast with the above-mentioned methods that aim to minimize correlations,
we rather seek to maximize correlations between the different units within each layer, such that
many of these units can be removed in a post-processing stage. In [4], additional noise outputs were
introduced in a network to similarly learn correlated filters. This strategy, however, is only justified for
simple networks and was only demonstrated on relatively shallow architectures. The contemporary
work [51] introduced a penalty during training to learn correlated units. This, however, was achieved
by explicitly computing all pairwise correlations, which quickly becomes cumbersome in very deep
networks with wide layers. By contrast, our approach makes use of a low-rank regularizer that can
effectively be optimized by proximal stochastic gradient descent.
Our approach belongs to the relatively small group of methods that explicitly aim to learn a compact
network during training, i.e., not as a post-processing step. Other methods have proposed to make
use of sparsity-inducing techniques to cancel out individual parameters [49, 10, 20, 19, 35] or
units [2, 50, 55]. These methods, however, act, at best, on individual units, without considering the
relationships between multiple units in the same layer. Variational inference [17] has also been used
to explicitly compress the network. However, the priors and posteriors used in these approaches
will typically zero out individual weights. Our experiments demonstrate that accounting for the
interactions between multiple units allows us to obtain more compact networks.
Another line of research aims to quantize the weights of deep networks [48, 12, 18]. Note that,
in a sense, this research direction is orthogonal to ours, since one could still further quantize our
compact networks. Furthermore, with the recent progress in efficient hardware handling floating-point
operations, we believe that there is also high value in designing non-quantized compact networks.
3
Compression-aware Training of Deep Networks
In this section, we introduce our approach to explicitly encouraging compactness while training a
deep neural network. To this end, we propose to make use of a low-rank regularizer on the parameter
matrix in each layer, which inherently aims to maximize the compression rate when computing a
low-rank approximation in a post-processing stage. In the following, we focus on convolutional neural
networks, because the popular visual recognition models tend to rely less and less on fully-connected
layers, and, more importantly, the inference time of such models is dominated by the convolutions in
the first few layers. Note, however, that our approach still applies to fully-connected layers.
To introduce our approach, let us first consider the l-th layer of a convolutional network, and denote
H
W
its parameters by ?l ? RKl ?Cl ?dl ?dl , where Cl and Kl are the number of input and output channels,
H
W
respectively, and dl and dl are the height and width of each convolutional kernel. Alternatively,
these parameters can be represented by a matrix ??l ? RKl ?Sl with Sl = Cl dlH dlW . Following [14], a
network can be compacted via a post-processing step performing a singular value decomposition of
??l and truncating the 0, or small, singular values. In essence, after this step, the parameter matrix
can be approximated as ??l ? Ul MlT , where Ul is a Kl ? rl matrix representing the basis kernels, with
rl ? min(Kl , Sl ), and Ml is an Sl ? rl matrix that mixes the activations of these basis kernels.
By making use of a post-processing step on a network trained in the usual way, however, there is
no guarantee that, during training, many singular values have become near-zero. Here, we aim to
explicitly account for this post-processing step during training, by seeking to obtain a parameter
matrix such that rl << min(Kl , Sl ). To this end, given N training input-output pairs (xi , yi ), we
formulate learning as the regularized minimization problem
min
?
1 N
? `(yi , f (xi , ?)) + r(?) ,
N i=1
(1)
where ? encompasses all network parameters, `(?, ?) is a supervised loss, such as the cross-entropy,
and r(?) is a regularizer encouraging the parameter matrix in each layer to have low rank.
Since explicitly minimizing the rank of a matrix is NP-hard, following the matrix completion
literature [7, 6], we make use of a convex relaxation in the form of the nuclear norm. This lets us
3
write our regularizer as
L
r(?) = ? ? k??l k? ,
(2)
l=1
where ? is a hyper-parameter setting the influence of the regularizer, and the nuclear norm is defined
rank(??l ) j
as k??l k? = ?
? , with ? j the singular values of ??l .
j=1
l
l
In practice, to minimize (1), we make use of proximal stochastic gradient descent. Specifically,
this amounts to minimizing the supervised loss only for one epoch, with learning rate ?, and then
applying the proximity operator of our regularizer. In our case, this can be achieved independently
for each layer. For layer l, this proximity operator corresponds to solving
?l? = argmin
??l
1 ?
k?l ? ??l k2F + ?k??l k? ,
2?
(3)
where ??l is the current estimate of the parameter matrix for layer l. As shown in [6], the solution to
this problem can be obtained by soft-thresholding the singular values of ??l , which can be written as
?l? = Ul ?l (??)VlT ,
rank(??l )
where ?l (??) = diag([(?l1 ? ??)+ , . . . , (?l
? ??)+ ]),
(4)
Ul and Vl are the left - and right-singular vectors of ??l , and (?)+ corresponds to taking the maximum
between the argument and 0.
3.1
Low-rank and Group-sparse Layers
While, as shown in our experiments, the low-rank solution discussed above significantly reduces
the number of parameters in the network, it does not affect the original number of input and output
channels Cl and Kl . By contrast, the group-sparsity based methods [2, 50] discussed in Section 2
cancel out entire units, thus reducing these numbers, but do not consider the interactions between
multiple units in the same layer, and would therefore typically not benefit from a post-processing
step such as the one of [14]. Here, we propose to make the best of both worlds to obtain low-rank
parameter matrices, some of whose units have explicitly been removed.
To this end, we combine the sparse group Lasso regularizer used in [2] with the low-rank one
described above. This lets us re-define the regularizer in (1) as
!
L
L
p Kl
n
r(?) = ? (1 ? ?)?l Pl ? k?l k2 + ??l k?l k1 + ? ? k??l k? ,
(5)
l=1
n=1
l=1
?ln
where Kl is the number of units in layer l,
denotes the vector of parameters for unit n in layer
l, Pl is the size of this vector (the same for all units in a layer), ? ? [0, 1] balances the influence
of sparsity terms on groups vs. individual parameters, and ?l is a layer-wise hyper-parameter. In
practice, following [2], we use only two different values of ?l ; one for the first few layers and one for
the remaining ones.
To learn our model with this new regularizer consisting of two main terms, we make use of the
incremental proximal descent approach proposed in [39], which has the benefit of having a lower
memory footprint than parallel proximal methods. The proximity operator for the sparse group Lasso
regularizer also has a closed form solution derived in [43] and provided in [2].
3.2
Benefits at Inference
Once our model is trained, we can obtain a compact network for faster and more memory-efficient
inference by making use of a post-processing step. In particular, to account for the low rank of the
parameter matrix of each layer, we make use of the SVD-based approach of [14]. Specifically, for
each layer l, we compute the SVD of the parameter matrix as ??l = U? l ?? l V?l and only keep the rl singular
values that are either non-zero, thus incurring no loss, or larger than a pre-defined threshold, at some
H W
potential loss. The parameter matrix can then be represented as ??l = Ul Ml , with Ul ? RCl dl dl ?rl and
Ml = ?l Vl ? Rrl ?Kl ) . In essence, every layer is decomposed into two layers. This incurs significant
memory and computational savings if rl (Cl dlH dlW + Kl ) << (Cl dlH dlW Kl ).
4
Furthermore, additional savings can be achieved when using the sparse group Lasso regularizer
discussed in Section 3.1. Indeed, in this case, the zeroed-out units can explicitly be removed, thus
yielding only K? l filters, with K? l < Kl . Note that, except for the first layer, units have also been
removed from the previous layer, thus reducing Cl to a lower C?l . Furthermore, thanks to our low-rank
regularizer, the remaining, non-zero, units will form a parameter matrix that still has low rank, and
can thus also be decomposed. This results in a total of rl (C?l dlH dlW + K? l ) parameters.
In our experiments, we select the rank rl based on the percentage el of the energy (i.e., the sum of
singular values) that we seek to capture by our low-rank approximation. This percentage plays an
important role in the trade-off between runtime/memory savings and drop of prediction accuracy. In
our experiments, we use the same percentage for all layers.
4
Experimental Settings
Datasets: For our experiments, we used two image classification datasets: ImageNet [41] and
ICDAR, the character recognition dataset introduced in [27]. ImageNet is a large-scale dataset
comprising over 15 million labeled images split into 22, 000 categories. We used the ILSVRC2012 [41] subset consisting of 1000 categories, with 1.2 million training images and 50, 000 validation
images. The ICDAR dataset consists of 185,639 training samples combining real and synthetic
characters and 5,198 test samples coming from the ICDAR2003 training set after removing all
non-alphanumeric characters. The images in ICDAR are split into 36 categories. The use of ICDAR
here was motivated by the fact that it is fairly large-scale, but, in contrast with ImageNet, existing
architectures haven?t been heavily tuned to this data. As such, one can expect our approach consisting
of training a compact network from scratch to be even more effective on this dataset.
Network Architectures: In our experiments, we make use of architectures where each kernel
in the convolutional layers has been decomposed into two 1D kernels [1], thus inherently having
rank-1 kernels. Note that this is orthogonal to the purpose of our low-rank regularizer, since, here,
we essentially aim at reducing the number of kernels, not the rank of individual kernels. The
decomposed layers yield even more compact architectures that require a lower computational cost for
training and testing while maintaining or even improving classification accuracy. In the following, a
convolutional layer refers to a layer with 1D kernels, while a decomposed layer refers to a block of
two convolutional layers using 1D vertical and horizontal kernels, respectively, with a non-linearity
and batch normalization after each convolution.
Let us consider a decomposed layer consisting of C and K input and output channels, respectively.
Let v? and h? T be vectors of length d v and d h , respectively, representing the kernel size of each 1D
feature map. In this paper, we set d h = d v ? d. Furthermore, let ?(?) be a non-linearity, and xc denote
the c-th input channel of the layer. In this setting, the activation of the i-th output channel fi can be
written as
L
C
fi = ?(bhi + ? h? Til ? [?(bvl + ? v?lc ? xc )]),
(6)
c=1
l=1
where L is the number of vertical filters, corresponding to the number of input channels for the
horizontal filters, and bvl and bhl are biases.
We report results with two different models using such decomposed layers: DecomposeMe [1] and
ResNets [23]. In all cases, we make use of batch-normalization after each convolutional layer 1 . We
rely on rectified linear units (ReLU) [31] as non-linearities, although some initial experiments suggest
that slightly better performance can be obtained with exponential linear units [8]. For DecomposeMe,
we used two different Dec8 architectures, whose specific number of units are provided in Table 1. For
residual networks, we used a decomposed ResNet-50, and empirically verified that the use of 1D
kernels instead of the standard ones had no significant impact on classification accuracy.
Implementation details: For the comparison to be fair, all models, including the baselines, were
trained from scratch on the same computer using the same random seed and the same framework.
More specifically, we used the torch-7 multi-gpu framework [11].
1
We empirically found the use of batch normalization after each convolutional layer to have more impact with
our low-rank regularizer than with group sparsity or with no regularizer, in which cases the computational cost
can be reduced by using a single batch normalization after each decomposed layer.
5
Dec256
8
Dec512
8
Dec512
3
1v
32/11
32/11
48/9
1h
64/11
64/11
96/9
2v
128/5
128/5
160/9
2h
192/5
192/5
256/9
3v
256/3
256/3
512/8
3h
384/3
384/3
512/8
4v
256/3
256/3
?
4h
256/3
256/3
?
5v
256/3
512/3
?
5h
256/3
512/3
?
6v
256/3
512/3
?
6h
256/3
512/3
?
7v
256/3
512/3
?
7h
256/3
512/3
?
8v
256/3
512/3
?
8h
256/3
512/3
?
Table 1: Different DecomposeMe architectures used on ImageNet and ICDAR. Each entry represents the number of filters and their dimension.
Layer / Conf?
1v to 2h
3v to 8h
0
?
?
1
0.0127
0.0127
2
0.051
0.051
3
0.204
0.204
4
0.255
0.255
5
0.357
0.357
6
0.051
0.357
7
0.051
0.408
8
0.051
0.510
9
0.051
0.255
10
0.051
0.765
11
0.153
0.51
Table 2: Sparse group Lasso hyper-parameter configurations. The first row provides ? for the
first four convolutional layers, while the second one shows ? for the remaining layers. The first five
configurations correspond to using the same regularization penalty for all the layers, while the later
ones define weaker penalties on the first two layers, as suggested in [2].
For ImageNet, training was done on a DGX-1 node using two-P100 GPUs in parallel. We used
stochastic gradient descent with a momentum of 0.9 and a batch size of 180 images. The models were
trained using an initial learning rate of 0.1 multiplied by 0.1 every 20 iterations for the small models
512
(Dec256
8 in Table 1) and every 30 iterations for the larger models (Dec8 in Table 1). For ICDAR, we
trained each network on a single TitanX-Pascal GPU for a total of 55 epochs with a batch size of
256 and 1,000 iterations per epoch. We follow the same experimental setting as in [2]: The initial
learning rate was set to an initial value of 0.1 and multiplied by 0.1. We used a momentum of 0.9.
For DecomposeMe networks, we only performed basic data augmentation consisting of using random
crops and random horizontal flips with probability 0.5. At test time, we used a single central crop.
For ResNets, we used the standard data augmentation advocated for in [23]. In practice, in all models,
we also included weight decay with a penalty strength of 1e?4 in our loss function. We observed
empirically that adding this weight decay prevents the weights to overly grow between every two
computations of the proximity operator.
In terms of hyper-parameters, for our low-rank regularizer, we considered four values: ? ? {0, 1, 5, 10}.
For the sparse group Lasso term, we initially set the same ? to every layer to analyze the effect of
combining both types of regularization. Then, in a second experiment, we followed the experimental
set-up proposed in [2], where the first two decomposed layers have a lower penalty. In addition, we
set ? = 0.2 to favor promoting sparsity at group level rather than at parameter level. The sparse group
Lasso hyper-parameter values are summarized in Table 2.
Computational cost: While a convenient measure of computational cost is the forward time,
this measure is highly hardware-dependent. Nowadays, hardware is heavily optimized for current
architectures and does not necessarily reflect the concept of any-time-computation. Therefore, we
focus on analyzing the number of multiply-accumulate operations (MAC). Let a convolution be
defined as fi = ?(bi + ?Cj=1 Wi j ? x j ), where each Wi j is a 2D kernel of dimensions d H ? dW and
i ? [1, . . . K]. Considering a naive convolution algorithm, the number of MACs for a convolutional
layer is equal to PCKd h dW where P is the number of pixels in the output feature map. Therefore,
it is important to reduce CK whenever P is large. That is, reducing the number of units in the first
convolutional layers has more impact than in the later ones.
5
Experimental Results
Parameter sensitivity and comparison to other methods on ImagNet: We first analyze the
effect of our low-rank regularizer on its own and jointly with the sparse group Lasso one on MACs
and accuracy. To this end, we make use of the Dec256
8 model on ImageNet, and measure the impact
of varying both ? and ? in Eq. 5. Note that using ? = ? = 0 corresponds to the standard model,
and ? = 0 and ? 6= 0 to the method of [2]. Below, we report results obtained without and with the
post-processing step described in Section 3.2. Note that applying such a post-processing on the
standard model corresponds to the compression technique of [14]. Fig. 1 summarizes the results of
this analysis.
6
(a)
(b)
(c)
Figure 1: Parameter sensitivity for Dec256
on ImageNet. (a) Accuracy as a function of the
8
regularization strength. (b) MACs directly after training. (c) MACS after the post-processing step
of Section 3.2 for el = {100%, 80%}. In all the figures, isolated points represent the models trained
without sparse group Lasso regularizer. The red point corresponds to the baseline, where no low-rank
or sparsity regularization was applied. The specific sparse group Lasso hyper-parameters for each
configuration Conf? are given in Table 2.
Figure 2: Effect of the low-rank regularizer on its own on Dec256
8 on ImageNet. (Left) Number
of units per layer. (Right) Effective rank per layer for (top) el =100% and (bottom) el =80%. Note that,
on its own, our low-rank regularizer already helps cancel out entire units, thus inherently performing
model selection.
In Fig. 1(a), we can observe that accuracy remains stable for a wide range of values of ? and ? . In
fact, there are even small improvements in accuracy when a moderate regularization is applied.
Figs. 1(b,c) depict the MACs without and with applying the post-processing step discussed in
Section 3.2. As expected, the MACs decrease as the weights of the regularizers increase. Importantly,
however, Figs. 1(a,b) show that several models can achieve a high compression rate at virtually no
loss in accuracy. In Fig. 1(c), we provide the curves after post-processing with two different energy
percentages el = {100%, 80%}. Keeping all the energy tends to incur an increase in MAC, since
the inequality defined in Section 3.2 is then not satisfied anymore. Recall, however, that, without
post-processing, the resulting models are still more compact than and as accurate as the baseline one.
With el = 80%, while a small drop in accuracy typically occurs, the gain in MAC is significantly
larger. Altogether, these experiments show that, by providing more compact models, our regularizer
lets us consistently reduce the computational cost over the baseline.
Interestingly, by looking at the case where Conf? = 0 in Fig. 1(b), we can see that we already
significantly reduce the number of operations when using our low-rank regularizer only, even without
post-processing. This is due to the fact that, even in this case, a significant number of units are
automatically zeroed-out. Empirically, we observed that, for moderate values of ?, the number of
zeroed-out singular values corresponds to complete units going to zero. This can be observed in
Fig. 2(left), were we show the number of non-zero units for each layer. In Fig. 2(right), we further
show the effective rank of each layer before and after post-processing.
7
hyper-params
# Params
top-1
Baseline
?
3.7M
88.6%
[14]
el = 90%
3.6M
88.5%
[2]
?
525K
89.6%
Ours
? = 15, el = 90%
728K
88.8%
Ours+[2]
? = 15, el = 90%
318K
89.7%
Ours+[2]
? = 15, el = 100%
454K
90.5%
Table 3: Comparison to other methods on ICDAR.
Imagenet
Dec512
8 -el =80%
Dec512
8 -el =100%
Top-1
66.8
67.6
Low-Rank appx.
Params MAC
-53.5
-46.2
-21.1
-4.8
no SVD
Params MAC
-39.5
-25.3
-39.5
-25.3
ICDAR
Dec512
3 -el =80%
Dec512
3 -el =100%
Top-1
89.6
90.8
Low-Rank appx.
Params MAC
-91.9
-92.9
-85.3
-86.8
no SVD
Params MAC
-89.2
-81.6
-89.2
-81.6
512
Table 4: Accuracy and compression rates for Dec512
8 models on ImageNet (left) and Dec3 on
ICDAR (right). The number of parameters and MACs are given in % relative to the baseline model
(i.e., without any regularizer). A negative value indicates reduction with respect to the baseline. The
accuracy of the baseline is 67.0 for ImageNet and 89.3 for ICDAR.
Dec256
8 -? = 1
Dec256
8 -? = 5
Dec256
8 -? = 10
el = 80%
97.33
88.33
85.78
el = 100%
125.44
119.27
110.35
no SVD
94.60
90.55
91.36
baseline (? = 0)
94.70
94.70
94.70
Table 5: Forward time in milliseconds (ms) using a Titan X (Pascal). We report the average over
50 forward passes using a batch size of 256. A large batch size minimizes the effect of memory
overheads due to non-hardware optimizations.
Comparison to other approaches on ICDAR: We now compare our results with existing approaches on the ICDAR dataset. As a baseline, we consider the Dec512
3 trained using SGD and L2
regularization for 75 epochs. For comparison, we consider the post-processing approach in [14]
with el = 90%, the group-sparsity regularization approach proposed in [2] and three different instances of our model. First, using ? = 15, no group-sparsity and el = 90%. Then, two instances
combining our low-rank regularizer with group-sparsity (Section 3.1) with el = 90% and el = 100%.
In this case, the models are trained for 55 epochs and then reloaded and fine tuned for 20 more
epochs. Table 3 summarizes these results. The comparison with [14] clearly evidences the benefits
of our compression-aware training strategy. Furthermore, these results show the benefits of further
combining our low-rank regularizer with the groups-sparsity one of [2].
In addition, we also compare our approach with L1 and L2 regularizers on the same dataset and with
the same experimental setup. Pruning the weights of the baseline models with a threshold of 1e ? 4
resulted in 1.5M zeroed-out parameters for the L2 regularizer and 2.8M zeroed-out parameters for
the L1 regularizer. However, these zeroed out weights are sparsely located within units (neurons).
Applying our post-processing step (low-rank approximation with el = 100%) to these results yielded
models with 3.6M and 3.2M parameters for L2 and L1 regularizers, respectively. The top-1 accuracy
for these two models after post-processing was 87% and 89%, respectively. Using a stronger L1
regularizer resulted in lower top-1 accuracy. By comparison, our approach yields a model with 3.4M
zeroed-out parameters after post-processing and a top-1 accuracy of 90%. Empirically, we found the
benefits of our approach to hold for varying regularizer weights.
Results with larger models: In Table 4, we provide the accuracies and MACs for our approach and
the baseline on ImageNet and ICDAR for Dec512
8 models. Note that using our low-rank regularizer
yields more compact networks than the baselines for similar or higher accuracies. In particular, for
ImageNet, we achieve reductions in parameter number of more than 20% and more than 50% for
el = 100% and el = 80%, respectively. For ICDAR, these reductions are around 90% in both cases.
We now focus on our results with a ResNet-50 model on ImageNet. For post-processing we used
el = 90% for all these experiments which resulted in virtually no loss of accuracy. The baseline
corresponds to a top-1 accuracy of 74.7% and 18M parameters. Applying the post-processing step on
this baseline resulted in a compression rate of 4%. By contrast, our approach with low-rank yields
a top-1 accuracy of 75.0% for a compression rate of 20.6%, and with group sparsity and low-rank
8
Baseline
r5
r15
r25
r35
r45
r55
r65
Epoch
reload
?
5
15
25
35
45
55
65
Num. parameters
Total no SVD
3.7M
?
3.2M
3.71M
210K
2.08M
218K
1.60M
222K
1.52M
324K
1.24M
388K
1.24M
414K
1.23M
accuracy
top-1
88.4%
89.8%
90.0%
90.0%
89.0%
90.1%
89.2%
87.7%
Total
train-time
1.69h
1.81h
0.77h
0.88h
0.99h
1.12h
1.26h
1.36h
Figure 3: Forward-Backward training time in milliseconds when varying the reload epoch for
Dec512
3 on ICDAR. (Left) Forward-backward time per batch in milliseconds (with a batch size of
32). (Right) Summary of the results of each experiment. Note that we could reduce the training time
from 1.69 hours (baseline) to 0.77 hours by reloading the model at the 15th epoch. This corresponds
to a relative training-time speed up of 54.5% and yields a 2% improvement in top-1 accuracy.
jointly, a top-1 accuracy of 75.2% for a compression rate of 27%. By comparison, applying [2] to the
same model yields an accuracy of 74.5% for a compression rate of 17%.
Inference time: While MACs represent the number of operations, we are also interested in the
inference time of the resulting models. Table 5 summarizes several representative inference times
for different instances of our experiments. Interestingly, there is a significant reduction in inference
time when we only remove the zeroed-out neurons from the model. This is a direct consequence
of the pruning effect, especially in the first layers. However, there is no significant reduction in
inference time when post-processing our model via a low-rank decomposition. The main reason for
this is that modern hardware is designed to compute convolutions with much fewer operations than
a naive algorithm. Furthermore, the actual computational cost depends not only on the number of
floating point operations but also on the memory bandwidth. In modern architectures, decomposing a
convolutional layer into a convolution and a matrix multiplication involves (with current hardware)
additional intermediate computations, as one cannot reuse convolutional kernels. Nevertheless, we
believe that our approach remains beneficial for embedded systems using customized hardware, such
as FPGAs.
Additional benefits at training time: So far, our experiments have demonstrated the effectiveness
of our approach at test time. Empirically, we found that our approach is also beneficial for training,
by pruning the network after only a few epochs (e.g., 15) and reloading and training the pruned
network, which becomes much more efficient. Specifically, Table 3 summarizes the effect of varying
the reload epoch for a model relying on both low-rank and group-sparsity. We were able to reduce the
training time (with a batch size of 32 and training for 100 epochs) from 1.69 to 0.77 hours (relative
speedup of 54.5%). The accuracy also improved by 2% and the number of parameters reduced from
3.7M (baseline) to 210K (relative 94.3% reduction). We found this behavior to be stable across a
wide range of regularization parameters. If we seek to maintain accuracy compared to the baseline,
we found that we could achieve a compression rate of 95.5% (up to 96% for an accuracy drop of
0.5%), which corresponds to a training time reduced by up to 60%.
6
Conclusion
In this paper, we have proposed to explicitly account for a post-processing compression stage when
training deep networks. To this end, we have introduced a regularizer in the training loss to encourage
the parameter matrix of each layer to have low rank. We have further studied the case where this
regularizer is combined with a sparsity-inducing one to achieve even higher compression. Our
experiments have demonstrated that our approach can achieve higher compression rates than state-ofthe-art methods, thus evidencing the benefits of taking compression into account during training. The
SVD-based technique that motivated our approach is only one specific choice of compression strategy.
In the future, we will therefore study how regularizers corresponding to other such compression
mechanisms can be incorporated in our framework.
9
References
[1] J. M. Alvarez and L. Petersson. Decomposeme: Simplifying convnets for end-to-end learning.
CoRR, abs/1606.05426, 2016.
[2] J. M. Alvarez and M. Salzmann. Learning the number of neurons in neural networks. In NIPS,
2016.
[3] L. J. Ba, R. Kiros, and G. E. Hinton. Layer normalization. CoRR, abs/1607.06450, 2016.
[4] M. Babaeizadeh, P. Smaragdis, and R. H. Campbell. Noiseout: A simple way to prune neural
networks. In emdnn Nips workshops, 2016.
[5] Y. Bengio and J. S. Bergstra. Slow, decorrelated features for pretraining complex cell-like
networks. In NIPS, pages 99?107. 2009.
[6] J.-F. Cai, E. J. Cand?s, and Z. Shen. A singular value thresholding algorithm for matrix
completion. SIAM J. on Optimization, 20(4):1956?1982, Mar. 2010.
[7] E. J. Cand?s and B. Recht. Exact matrix completion via convex optimization. CoRR,
abs/0805.4471, 2008.
[8] D. Clevert, T. Unterthiner, and S. Hochreiter. Fast and accurate deep network learning by
exponential linear units (elus). CoRR, abs/1511.07289, 2015.
[9] M. Cogswell, F. Ahmed, R. Girshick, L. Zitnick, and D. Batra. Reducing overfitting in deep
networks by decorrelating representations. In ICLR, 2016.
[10] M. D. Collins and P. Kohli. Memory Bounded Deep Convolutional Networks. In CoRR, volume
abs/1412.1442, 2014.
[11] R. Collobert, K. Kavukcuoglu, and C. Farabet. Torch7: A matlab-like environment for machine
learning. In BigLearn, NIPS Workshop, 2011.
[12] M. Courbariaux and Y. Bengio. Binarynet: Training deep neural networks with weights and
activations constrained to +1 or -1. CoRR, abs/1602.02830, 2016.
[13] M. Denil, B. Shakibi, L. Dinh, M. Ranzato, and N. de Freitas. Predicting parameters in deep
learning. CoRR, abs/1306.0543, 2013.
[14] E. L. Denton, W. Zaremba, J. Bruna, Y. LeCun, and R. Fergus. Exploiting linear structure
within convolutional networks for efficient evaluation. In NIPS. 2014.
[15] J. Duchi, E. Hazan, and Y. Singer. Adaptive subgradient methods for online learning and
stochastic optimization. Technical Report UCB/EECS-2010-24, EECS Department, University
of California, Berkeley, Mar 2010.
[16] I. J. Goodfellow, D. Warde-farley, M. Mirza, A. Courville, and Y. Bengio. Maxout networks. In
ICML, 2013.
[17] A. Graves. Practical variational inference for neural networks. In NIPS, 2011.
[18] S. Gupta, A. Agrawal, K. Gopalakrishnan, and P. Narayanan. Deep learning with limited
numerical precision. CoRR, abs/1502.02551, 2015.
[19] S. Han, H. Mao, and W. J. Dally. Deep compression: Compressing deep neural networks with
pruning, trained quantization and huffman coding. ICLR, 2016.
[20] S. Han, J. Pool, J. Tran, and W. Dally. Learning both weights and connections for efficient
neural network. In NIPS, 2015.
[21] M. Harandi and B. Fernando. Generalized backpropagation, ?tude de cas: Orthogonality. CoRR,
abs/1611.05927, 2016.
[22] B. Hassibi, D. G. Stork, and G. J. Wolff. Optimal brain surgeon and general network pruning.
In ICNN, 1993.
10
[23] K. He, X. Zhang, S. Ren, and J. Sun. Deep Residual Learning for Image Recognition. In CoRR,
volume abs/1512.03385, 2015.
[24] V. O. Hinton, G. E. and J. Dean. Distilling the knowledge in a neural network. In arXiv, 2014.
[25] Y. Ioannou, D. P. Robertson, J. Shotton, R. Cipolla, and A. Criminisi. Training cnns with
low-rank filters for efficient image classification. CoRR, abs/1511.06744, 2015.
[26] S. Ioffe and C. Szegedy. Batch normalization: Accelerating deep network training by reducing
internal covariate shift. CoRR, 2015.
[27] M. Jaderberg, A. Vedaldi, and Zisserman. Deep features for text spotting. In ECCV, 2014.
[28] M. Jaderberg, A. Vedaldi, and A. Zisserman. Speeding up convolutional neural networks with
low rank expansions. In British Machine Vision Conference, 2014.
[29] C. Ji, R. R. Snapp, and D. Psaltis. Generalizing smoothness constraints from discrete samples.
Neural Computation, 2(2):188?197, June 1990.
[30] D. P. Kingma and J. Ba. Adam: A method for stochastic optimization. CoRR, abs/1412.6980,
2014.
[31] A. Krizhevsky, I. Sutskever, and G. E. Hinton. Imagenet classification with deep convolutional
neural networks. In NIPS, 2012.
[32] V. Lebedev, Y. Ganin, M. Rakhuba, I. V. Oseledets, and V. S. Lempitsky. Speeding-up convolutional neural networks using fine-tuned cp-decomposition. CoRR, abs/1412.6553, 2014.
[33] Y. LeCun, J. S. Denker, S. Solla, R. E. Howard, and L. D. Jackel. Optimal brain damage. In
NIPS, 1990.
[34] B. Liu, M. Wang, H. Foroosh, M. Tappen, and M. Penksy. Sparse convolutional neural networks.
In CVPR, 2015.
[35] P. Molchanov, S. Tyree, T. Karras, T. Aila, and J. Kautz. Pruning convolutional neural networks
for resource efficient transfer learning. CoRR, abs/1611.06440, 2016.
[36] M. Mozer and P. Smolensky. Skeletonization: A technique for trimming the fat from a network
via relevance assessment. In NIPS, 1988.
[37] H. Pan and H. Jiang. Learning convolutional neural networks using hybrid orthogonal projection
and estimation. CoRR, abs/1606.05929, 2016.
[38] R. Reed. Pruning algorithms-a survey. IEEE Transactions on Neural Networks, 4(5):740?747,
Sep 1993.
[39] E. Richard, P. andre Savalle, and N. Vayatis. Estimation of simultaneously sparse and low rank
matrices. In ICML, 2012.
[40] P. Rodr?guez, J. Gonzalez, G. Cucurull, and J. M. G. andXavier Roca. Regularizing cnns with
locally constrained decorrelations. In ICLR, 2017.
[41] O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy,
A. Khosla, M. S. Bernstein, A. C. Berg, and F.-F. Li. Imagenet large scale visual recognition
challenge. CoRR, abs/1409.0575, 2014.
[42] T. Salimans and D. P. Kingma. Weight normalization: A simple reparameterization to accelerate
training of deep neural networks. CoRR, abs/1602.07868, 2016.
[43] N. Simon, J. Friedman, T. Hastie, and R. Tibshirani. A sparse-group lasso. Journal of
Computational and Graphical Statistics, 2013.
[44] K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image
recognition. CoRR, abs/1409.1556, 2014.
11
[45] N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov. Dropout: A
simple way to prevent neural networks from overfitting. Journal of Machine Learning Research,
15:1929?1958, 2014.
[46] C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and
A. Rabinovich. Going deeper with convolutions. In CVPR, 2015.
[47] C. Tai, T. Xiao, X. Wang, and W. E. Convolutional neural networks with low-rank regularization.
CoRR, abs/1511.06067, 2015.
[48] K. Ullrich, E. Meeds, and M. Welling. Soft weight-sharing for neural network compression.
CoRR, abs/1702.04008, 2016.
[49] A. S. Weigend, D. E. Rumelhart, and B. A. Huberman. Generalization by weight-elimination
with application to forecasting. In NIPS, 1991.
[50] W. Wen, C. Wu, Y. Wang, Y. Chen, and H. Li. Learning structured sparsity in deep neural
networks. In NIPS, 2016.
[51] W. Wen, C. Xu, C. Wu, Y. Wang, Y. Chen, and H. Li. Coordinating filters for faster deep neural
networks. CoRR, abs/1703.09746, 2017.
[52] W. Xiong, B. Du, L. Zhang, R. Hu, and D. Tao. Regularizing deep convolutional neural networks
with a structured decorrelation constraint. In IEEE Int. Conf. on Data Mining (ICDM), 2016.
[53] M. D. Zeiler. ADADELTA: an adaptive learning rate method. CoRR, abs/1212.5701, 2012.
[54] S. Zhang and H. Jiang. Hybrid orthogonal projection and estimation (HOPE): A new framework
to probe and learn neural networks. CoRR, abs/1502.00702, 2015.
[55] H. Zhou, J. M. Alvarez, and F. Porikli. Less is more: Towards compact cnns. In ECCV, 2006.
12
| 6687 |@word kohli:1 compression:34 stronger:1 seems:1 norm:2 hu:1 seek:4 accounting:3 decomposition:3 simplifying:1 sgd:1 incurs:1 reduction:6 initial:5 configuration:3 liu:2 salzmann:3 tuned:4 ours:4 interestingly:2 existing:4 freitas:1 current:3 activation:5 yet:1 guez:1 written:2 gpu:2 numerical:1 alphanumeric:1 remove:1 drop:3 designed:1 depict:1 v:1 fewer:1 devising:1 num:1 provides:1 quantized:1 node:1 zhang:3 five:1 height:1 direct:1 become:3 consists:2 combine:1 overhead:1 introduce:4 manner:1 pairwise:2 expected:1 indeed:1 behavior:1 cand:2 surge:1 kiros:1 multi:1 brain:2 salakhutdinov:1 titanx:1 decomposed:10 relying:1 automatically:2 encouraging:2 actual:1 considering:4 increasing:1 becomes:4 provided:2 linearity:3 bounded:1 alto:1 argmin:1 minimizes:1 savalle:1 porikli:1 guarantee:2 berkeley:1 every:5 act:2 tackle:1 runtime:3 zaremba:1 fat:1 k2:1 unit:44 before:1 local:1 tends:1 consequence:4 aiming:2 analyzing:1 jiang:2 might:1 studied:1 lausanne:1 challenging:1 deployment:1 limited:3 bi:1 range:2 practical:1 lecun:2 testing:1 practice:4 block:1 backpropagation:1 footprint:1 procedure:1 reloading:2 significantly:3 vedaldi:2 convenient:1 projection:2 pre:5 refers:2 r65:1 suggest:1 cannot:1 selection:1 operator:4 context:4 influence:3 applying:6 optimize:1 equivalent:1 map:2 demonstrated:4 dean:1 hugely:1 truncating:1 convex:2 focused:1 formulate:1 independently:1 shen:1 survey:1 importantly:4 deriving:1 regularize:1 orthonormal:1 nuclear:2 dw:2 reparameterization:1 oseledets:1 deploy:1 play:1 heavily:2 exact:1 designing:2 goodfellow:1 trend:2 robertson:1 expensive:3 particularly:1 recognition:5 approximated:1 located:1 tappen:1 rumelhart:1 sparsely:1 adadelta:1 labeled:1 observed:3 role:1 bottom:1 wang:4 capture:1 thousand:1 compressing:1 connected:2 ilsvrc2012:1 sun:1 ranzato:1 solla:1 trade:1 contemporary:2 removed:4 decrease:1 mentioned:3 mozer:1 environment:1 warde:1 trained:15 solving:1 incur:1 surgeon:1 meed:1 basis:2 easily:2 sep:1 accelerate:1 represented:2 regularizer:36 train:1 evidencing:1 fast:1 effective:5 vlt:1 hyper:7 dec8:2 whose:2 larger:5 cvpr:2 favor:1 statistic:1 simonyan:1 jointly:2 online:1 agrawal:1 cai:1 propose:5 tran:1 interaction:3 coming:1 clevert:1 relevant:1 combining:4 achieve:9 inducing:2 los:1 exploiting:3 sutskever:2 foroosh:1 incremental:1 adam:1 resnet:3 help:1 completion:3 ganin:1 advocated:2 progress:4 eq:1 involves:1 elus:1 distilling:1 switzerland:1 direction:1 drawback:3 filter:8 stochastic:6 criminisi:1 cnns:3 elimination:1 require:1 generalization:1 icnn:1 pl:2 hold:1 proximity:4 around:1 considered:1 great:1 seed:1 early:2 purpose:1 estimation:3 psaltis:1 currently:1 jackel:1 successfully:1 minimization:1 hope:1 clearly:1 biglearn:1 aim:9 rrl:1 rather:2 ck:1 denil:1 zhou:1 varying:4 conjunction:1 derived:1 focus:4 june:1 improvement:2 consistently:1 rank:56 indicates:1 contrast:6 baseline:19 sense:1 inference:11 dependent:1 el:24 epfl:2 vl:2 entire:3 typically:7 torch:1 initially:1 compactness:1 going:2 interested:2 comprising:1 tao:1 pixel:1 rodr:1 classification:5 ullrich:1 pascal:2 priori:1 development:2 art:2 platform:1 fairly:1 constrained:2 equal:1 aware:5 once:1 having:3 beach:1 saving:3 manually:1 r5:1 represents:1 denton:1 constitutes:1 cancel:3 k2f:1 icml:2 future:2 np:1 report:4 mirza:1 haven:1 few:4 wen:2 modern:2 richard:1 simultaneously:2 resulted:4 individual:11 floating:2 replaced:1 phase:1 consisting:5 decomposeme:6 maintain:1 attempt:1 ab:23 friedman:1 huge:1 interest:1 trimming:1 highly:3 mining:1 multiply:1 evaluation:1 truly:1 yielding:1 farley:1 regularizers:4 accurate:2 nowadays:1 encourage:1 orthogonal:4 unterthiner:1 desired:1 re:1 isolated:1 girshick:1 witnessed:1 instance:3 soft:2 rabinovich:1 cost:6 mac:16 addressing:1 subset:1 entry:1 hundred:2 r25:1 krizhevsky:2 eec:2 proximal:5 synthetic:1 params:6 combined:1 thanks:3 st:1 recht:1 sensitivity:2 siam:1 off:1 pool:1 quickly:4 lebedev:1 augmentation:2 central:1 reflect:1 satisfied:1 containing:1 opposed:1 huang:1 conf:4 til:1 li:3 szegedy:2 account:6 potential:2 de:2 bergstra:1 rcl:1 summarized:1 availability:1 coding:1 int:1 titan:1 coordinated:1 explicitly:15 depends:1 collobert:1 later:2 performed:1 closed:1 dally:2 analyze:4 hazan:1 red:1 start:1 compacted:1 parallel:2 kautz:1 simon:1 jia:1 minimize:2 shakibi:1 accuracy:29 convolutional:24 yield:8 saliency:1 correspond:2 ofthe:1 kavukcuoglu:1 ren:1 rectified:1 russakovsky:1 cumbersome:2 whenever:1 decorrelated:1 farabet:1 andre:1 sharing:1 energy:3 r45:1 gain:1 dataset:6 popular:3 dlw:4 recall:1 knowledge:1 cj:1 campbell:1 higher:5 supervised:2 follow:1 molchanov:1 zisserman:3 alvarez:5 improved:1 decorrelating:1 done:1 mar:2 furthermore:8 stage:4 correlation:4 convnets:1 horizontal:3 su:1 assessment:1 believe:2 usa:1 effect:7 facilitate:1 consisted:1 concept:1 former:1 regularization:9 mlt:1 attractive:1 during:8 width:1 encourages:2 essence:5 m:1 generalized:1 complete:2 demonstrate:2 duchi:1 l1:5 cp:1 dgx:1 image:10 wise:3 variational:2 recently:2 fi:3 rl:9 empirically:6 ji:1 stork:1 volume:2 million:3 discussed:4 he:1 accumulate:1 significant:5 dinh:1 anguelov:1 smoothness:1 similarly:1 language:1 had:1 bruna:1 stable:2 han:2 impressive:1 posterior:1 own:3 recent:4 belongs:1 moderate:2 inequality:1 yi:2 additional:5 employed:1 prune:1 deng:1 maximize:2 fernando:1 multiple:3 mix:1 reduces:2 technical:1 faster:2 ahmed:1 cross:1 long:1 icdm:1 post:26 equally:1 impact:5 prediction:3 basic:1 crop:2 essentially:1 vision:1 arxiv:1 resnets:2 normalization:11 kernel:14 iteration:3 represent:2 achieved:4 cell:1 hochreiter:1 justified:1 addition:2 huffman:1 fine:3 vayatis:1 krause:1 singular:10 grow:1 rakhuba:1 tri:1 pass:1 tend:2 virtually:3 effectiveness:1 extracting:2 near:1 presence:2 constraining:1 split:2 intermediate:1 bengio:3 shotton:1 variety:2 affect:1 relu:1 cogswell:1 bernstein:1 architecture:15 lasso:10 appx:2 bandwidth:1 reduce:6 idea:1 hastie:1 shift:1 motivated:2 reuse:1 ul:6 effort:1 torch7:1 penalty:5 accelerating:1 suffer:1 forecasting:1 speech:1 pretraining:1 matlab:1 deep:32 tude:1 karpathy:1 amount:2 locally:1 hardware:9 narayanan:1 category:3 reduced:3 sl:5 percentage:4 millisecond:3 coordinating:1 overly:1 per:4 tibshirani:1 write:1 discrete:1 group:25 redundancy:1 four:2 threshold:2 nevertheless:1 prevent:2 verified:1 backward:2 relaxation:1 subgradient:1 year:3 sum:1 weigend:1 jose:2 parameterized:1 wu:2 gonzalez:1 summarizes:4 dropout:2 layer:71 guaranteed:1 followed:1 courville:1 smaragdis:1 yielded:1 adapted:1 rkl:2 strength:2 orthogonality:1 constraint:2 r15:1 dominated:1 speed:1 argument:1 min:3 pruned:2 performing:2 dec3:1 relatively:3 gpus:1 speedup:1 department:1 structured:2 according:1 imagnet:1 truncate:1 smaller:1 beneficial:3 increasingly:1 character:3 slightly:1 wi:2 across:1 shallow:2 aila:1 making:5 pan:1 computationally:2 resource:3 ln:1 remains:2 tai:1 icdar:16 mechanism:1 singer:1 flip:1 end:9 operation:7 incurring:1 decomposing:1 multiplied:2 promoting:1 observe:1 denker:1 salimans:1 probe:1 icdar2003:1 anymore:1 skeletonization:1 batch:13 xiong:1 altogether:1 fpgas:1 original:1 compress:3 assumes:1 denotes:1 remaining:3 top:12 zeiler:1 graphical:1 maintaining:1 ioannou:1 xc:2 k1:1 especially:1 tensor:2 seeking:1 already:2 occurs:1 strategy:10 damage:1 usual:1 gradient:4 iclr:3 parametrized:3 reason:1 gopalakrishnan:1 length:1 relationship:1 reed:2 providing:1 minimizing:2 balance:1 sermanet:1 bhi:1 unfortunately:2 setup:1 potentially:1 negative:1 resurgence:1 ba:2 implementation:1 satheesh:1 vertical:2 convolution:7 neuron:3 datasets:3 howard:1 descent:5 hinton:4 ever:1 looking:1 incorporated:1 introduced:4 pair:3 required:2 kl:11 optimized:2 imagenet:17 connection:1 california:1 learned:1 binarynet:1 hour:3 kingma:2 nip:13 able:1 suggested:1 spotting:1 below:1 smolensky:1 sparsity:16 challenge:1 encompasses:1 including:2 memory:10 decorrelation:1 natural:1 rely:3 regularized:1 predicting:1 hybrid:2 residual:2 customized:1 representing:2 scheme:1 mathieu:2 naive:2 speeding:2 text:1 prior:2 literature:1 epoch:12 l2:4 multiplication:1 relative:4 graf:1 embedded:3 loss:12 fully:2 expect:1 validation:1 vanhoucke:1 sufficient:1 xiao:1 principle:1 thresholding:2 zeroed:8 courbariaux:1 uncorrelated:1 tyree:1 row:1 eccv:2 summary:1 petersson:1 last:1 keeping:1 bias:1 weaker:1 deeper:3 institute:1 wide:3 taking:2 sparse:13 benefit:10 overcome:2 dimension:2 curve:1 world:1 forward:5 made:4 adaptive:2 far:1 erhan:1 welling:1 transaction:1 pruning:8 compact:18 approximate:1 jaderberg:2 keep:1 dealing:1 ml:3 global:2 overfitting:5 ioffe:1 xi:2 fergus:1 alternatively:1 khosla:1 table:14 learn:10 channel:6 transfer:1 ca:3 inherently:3 p100:1 correlated:5 improving:1 du:1 quantize:2 expansion:1 necessarily:2 cl:7 complex:1 domain:2 diag:1 zitnick:1 main:2 noise:2 snapp:1 fair:1 cvlab:1 xu:1 fig:8 representative:1 slow:1 lc:1 precision:1 hassibi:1 momentum:2 mao:1 exponential:2 toyota:1 reloaded:1 removing:2 british:1 specific:3 harandi:1 covariate:1 decay:2 reload:3 gupta:1 evidence:1 dl:6 intractable:1 workshop:2 quantization:1 adding:1 effectively:1 corr:24 roca:1 chen:2 easier:1 entropy:1 generalizing:1 visual:2 prevents:1 applies:1 cipolla:1 ch:1 corresponds:9 srivastava:1 ma:1 lempitsky:1 towards:1 maxout:2 hard:1 included:1 specifically:5 except:1 reducing:8 huberman:1 acting:1 impractically:1 wolff:1 total:4 batra:1 svd:7 experimental:5 ucb:1 exception:1 select:1 berg:1 internal:1 latter:1 collins:1 relevance:1 regularizing:2 scratch:5 handling:1 |
6,287 | 6,688 | Multiscale Semi-Markov Dynamics for
Intracortical Brain-Computer Interfaces
Daniel J. Milstein ?
[email protected]
John D. Simeral ? ?
[email protected]
Jason L. Pacheco ?
Leigh R. Hochberg ? ? ?
[email protected]
[email protected]
Beata Jarosiewicz k ? ??
[email protected]
Erik B. Sudderth ?? ?
[email protected]
Abstract
Intracortical brain-computer interfaces (iBCIs) have allowed people with tetraplegia
to control a computer cursor by imagining the movement of their paralyzed arm
or hand. State-of-the-art decoders deployed in human iBCIs are derived from a
Kalman filter that assumes Markov dynamics on the angle of intended movement,
and a unimodal dependence on intended angle for each channel of neural activity.
Due to errors made in the decoding of noisy neural data, as a user attempts to
move the cursor to a goal, the angle between cursor and goal positions may change
rapidly. We propose a dynamic Bayesian network that includes the on-screen goal
position as part of its latent state, and thus allows the person?s intended angle of
movement to be aggregated over a much longer history of neural activity. This
multiscale model explicitly captures the relationship between instantaneous angles
of motion and long-term goals, and incorporates semi-Markov dynamics for motion
trajectories. We also introduce a multimodal likelihood model for recordings
of neural populations which can be rapidly calibrated for clinical applications.
In offline experiments with recorded neural data, we demonstrate significantly
improved prediction of motion directions compared to the Kalman filter. We derive
an efficient online inference algorithm, enabling a clinical trial participant with
tetraplegia to control a computer cursor with neural activity in real time. The
observed kinematics of cursor movement are objectively straighter and smoother
than prior iBCI decoding models without loss of responsiveness.
1
Introduction
Paralysis of all four limbs from injury or disease, or tetraplegia, can severely limit function, independence, and even sometimes communication. Despite its inability to effect movement in muscles,
neural activity in motor cortex still modulates according to people?s intentions to move their paralyzed
arm or hand, even years after injury [Hochberg et al., 2006, Simeral et al., 2011, Hochberg et al.,
?
Department of Computer Science, Brown University, Providence, RI, USA.
Computer Science and Artificial Intelligence Laboratory, MIT, Cambridge, MA, USA.
?
School of Engineering, Brown University, Providence, RI, USA; and Department of Neurology, Massachusetts General Hospital, Boston, MA, USA.
?
Rehabilitation R&D Service, Department of Veterans Affairs Medical Center, Providence, RI, USA; and
Brown Institute for Brain Science, Brown University, Providence, RI, USA.
?
Department of Neurology, Harvard Medical School, Boston, MA, USA.
k
Department of Neuroscience, Brown University, Providence, RI, USA.
??
Present affiliation: Dept. of Neurosurgery, Stanford University, Stanford, CA, USA.
??
Department of Computer Science, University of California, Irvine, CA, USA.
?
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
Figure 1: A microelectrode array (left) is implanted in the motor cortex (center) to record electrical activity.
Via this activity, a clinical trial participant (right, lying on his side in bed) then controls a computer cursor with
an iBCI. A cable connected to the electrode array via a transcutaneous connector (gray box) sends neural signals
to the computer for decoding. Center drawing from Donoghue et al. [2011] and used with permission of the
author. The right image is a screenshot of a video included in the supplemental material that demonstrates real
time decoding via our MSSM model.
2012, Collinger et al., 2013]. Intracortical brain-computer interfaces (iBCIs) utilize neural signals
recorded from implanted electrode arrays to extract information about movement intentions. They
have enabled individuals with tetraplegia to control a computer cursor to engage in tasks such as
on-screen typing [Bacher et al., 2015, Jarosiewicz et al., 2015, Pandarinath et al., 2017], and to regain
volitional control of their own limbs [Ajiboye et al., 2017].
Current iBCIs are based on a Kalman filter that assumes the vector of desired cursor movement
evolves according to Gaussian random walk dynamics, and that neural activity is a Gaussian-corrupted
linear function of this state [Kim et al., 2008]. In Sec. 2, we review how the Kalman filter is applied
to neural decoding, and studies of the motor cortex by Georgopoulos et al. [1982] that justify its use.
In Sec. 3, we improve upon the Kalman filter?s linear observation model by introducing a flexible,
multimodal likelihood inspired by more recent research [Amirikian and Georgopulos, 2000]. Sec. 4
then proposes a graphical model (a dynamic Bayesian network [Murphy, 2002]) for the relationship
between the angle of intended movement and the intended on-screen goal position. We derive an
efficient inference algorithm via an online variant of the junction tree algorithm [Boyen and Koller,
1998]. In Sec. 5, we use recorded neural data to validate the components of our multiscale semiMarkov (MSSM) model, and demonstrate significantly improved prediction of motion directions in
offline analysis. Via a real time implementation of the inference algorithm on a constrained embedded
system, we then evaluate online decoding performance as a participant in the BrainGate21 iBCI pilot
clinical trial uses the MSSM model to control a computer cursor with his neural activity.
2
Neural decoding via a Kalman filter
The Kalman filter is the current state-of-the-art for iBCI decoding. There are several configurations
of the Kalman filter used to enable cursor control in contemporary iBCI systems [Pandarinath et al.,
2017, Jarosiewicz et al., 2015, Gilja et al., 2015] and there is no broad consensus in the iBCI field on
which is most suited for clinical use. In this paper, we focus on the variant described by Jarosiewicz
et al. [2015].
Participants in the BrainGate2 clinical trial receive one or two microelectrode array implants in the
motor cortex (see Fig. 1). The electrical signals recorded by this electrode array are then transformed
(via signal processing methods designed to reduce noise) into a D-dimensional neural activity vector
zt ? RD , sampled at 50 Hz. From the sequence of neural activity, the Kalman filter estimates the
latent state xt ? R2 , a vector pointing in the intended direction of cursor motion. The Kalman filter
assumes a jointly Gaussian model for cursor dynamics and neural activity,
xt | xt?1 ? N (Axt?1 , W ),
zt | xt ? N (b + Hxt , Q),
2?2
(1)
2?2
with cursor dynamics A ? R , process noise covariance W ? R , and (typically non-diagonal)
observation covariance Q ? RD?D . At each time step, the on-screen cursor?s position is moved by
the estimated latent state vector (decoder output) scaled by a constant, the speed gain.
The function relating neural activity to some measurable quantity of interest is called a tuning curve.
A common model of neural activity in the motor cortex assumes that each neuron?s activity is highest
1
Caution: Investigational Device. Limited by Federal Law to Investigational Use.
2
for some preferred direction of motion, and lowest in the opposite direction, with intermediate activity
often resembling a cosine function. This cosine tuning model is based on pioneering studies of
the motor cortex of non-human primates [Georgopoulos et al., 1982], and is commonly used (or
implicitly assumed) in iBCI systems because of its mathematical simplicity and tractability.
Expressing the inner product between vectors via the cosine of the angle between them, the expected
neural activity of the j th component of Eq. (1) can be written as
hj2
T
E[ztj | xt ] = bj + hj xt = bj + ||xt || ? ||hj || ? cos ?t ? atan
,
(2)
hj1
where ?t is the intended angle of movement at timestep t, bj is the baseline activity rate for channel j,
and hj is the j th row of the observation matrix H = (hT1 , . . . , hTD )T . If xt is further assumed
to be a unit vector (a constraint not enforced by the Kalman filter), Eq. (2) simplifies to hTj xt =
mj cos(?t ? pj ), where mj is the modulation of the tuning curve and pj specifies the angular location
of the peak of the cosine tuning curve (the preferred direction). Thus, cosine tuning models are linear.
To collect labeled training data for decoder calibration, the participant is asked to attempt to move
a cursor to prompted target locations. We emphasize that although the clinical random target
task displays only one target at a time, this target position is unknown to the decoder. Labels are
constructed for the neural activity patterns by assuming that at each 20ms time step, the participant
intends to move the cursor straight to the target [Jarosiewicz et al., 2015, Gilja et al., 2015]. These
labeled data are used to fit the observation matrix H and neuron baseline rates (biases) b via ridge
regression. The observation noise covariance Q is estimated as the empirical covariance of the
residuals. The state dynamics matrix A and process covariance matrix W may be tuned to adjust the
responsiveness of the iBCI system.
3
Flexible tuning likelihoods
The cosine tuning model reviewed in the previous section has several shortcomings. First, motor
cortical neurons that have unimodal tuning curves often have narrower peaks that are better described
by von Mises distributions [Amirikian and Georgopulos, 2000]. Second, tuning can be multimodal.
Third, neural features used for iBCI decoding may capture the pooled activity of several neurons,
not just one [Fraser et al., 2009]. While bimodal von Mises models were introduced by Amirikian
and Georgopulos [2000], up to now iBCI decoders based on von Mises tuning curves have only
employed unimodal mean functions proportional to a single von Mises density [Koyama et al., 2010].
In contrast, we introduce a multimodal likelihood proportional to an arbitrary number of regularly
spaced von Mises densities and incorporate this likelihood into an iBCI decoder. Moreover, we can
efficiently fit parameters of this new likelihood via ridge regression. Computational efficiency is
crucial to allow rapid calibration in clinical applications.
Let ?t ? [0, 2?) denote the intended angle of cursor movement at time t. The flexible tuning likelihood
captures more complex neural activity distributions via a regression model with nonlinear features:
zt | ?t ? N b + wT ?(?t ), Q ,
?k (?t ) = exp [ cos (?t ? ?k )] .
(3)
The features are a set of K von Mises basis functions ?(?) = (?1 (?), . . . , ?K (?))T . Basis functions
?k (x) are centered on a regular grid of angles ?k , and have tunable concentration .
Using human neural data recorded during cued target tasks, we compare regression fits for the flexible
tuning model to the standard cosine tuning model (Fig. 2). In addition to providing better fits for
channels with complex or multimodal activity, the flexible tuning model also provides good fits to
apparently cosine-tuned signals. This leads to higher predictive likelihoods for held-out data, and as
we demonstrate in Sec. 5, more accurate neural decoding algorithms.
4
Multiscale Semi-Markov Dynamical Models
The key observation underlying our multiscale dynamical model is that the sampling rate used for
neural decoding (typically around 50 Hz) is much faster than the rate that the goal position changes
(under normal conditions, every few seconds). In addition, frequent but small adjustments of cursor
aim angle are required to maintain a steady heading. State-of-the-art Kalman filter approaches to iBCIs
3
8
5.8
5.75
Spike power measurement
13
12.5
Spike power measurement
Spike power measurement
7.5
12
11.5
7
6.5
11
10.5
6
5.5
-180 -135
Data
Cosine fit
Flexible fit
-90
-45
0
45
Angle (degrees)
90
135
180
5.6
5.55
10
Data
Cosine fit
Flexible fit
9.5
9
-180 -135
5.7
5.65
-90
-45
0
45
Angle (degrees)
90
135
5.5
Data
Cosine fit
Flexible fit
5.45
180
5.4
-180 -135
-90
-45
0
45
Angle (degrees)
90
135
180
Figure 2: Flexible tuning curves. Each panel shows the empirical mean and standard deviation (red) of
example neural signals recorded from a single intracortical electrode while a participant is moving within
45 degrees of a given direction in a cued target task. These signals can violate the assumptions of a cosine
tuning model (black), as evident in the left two examples. The flexible regression likelihood (cyan) captures
neural activity with varying concentration (left) and multiple tuning directions (center), as well as cosine-tuned
signals (right). Because neural activity from individual electrodes is very noisy (the standard deviation within
each angular bin exceeds the change in mean activity across angles), information from multiple electrodes is
aggregated over time for effective decoding.
are incapable of capturing these multiscale dynamics since they assume first-order Markov dependence
across time and do not explicitly represent goal position. To cope with this, hyperparameters of the
linear Gaussian dynamics must be tuned to simultaneously remain sensitive to frequent directional
adjustments, but not so sensitive that cursor dynamics are dominated by transient neural activity.
Our proposed MSSM decoder, by contrast, explicitly represents goal position in addition to cursor
aim angle. Through the use of semi-Markov dynamics, the MSSM enables goal position to evolve
at a different rate than cursor angle while allowing for a high rate of neural data acquisition. In this
way, the MSSM can integrate across different timescales to more robustly infer the (unknown) goal
position and the (also unknown) cursor aim. We introduce the model in Sec. 4.1 and 4.2. We derive
an efficient decoding algorithm, based on an online variant of the junction tree algorithm, in Sec. 4.3.
4.1
Modeling Goals and Motion via a Dynamic Bayesian Network
The MSSM directed graphical model (Fig. 3) uses a structured latent state representation, sometimes
referred to as a dynamic Bayesian network [Murphy, 2002]. This factorization allows us to discretize
latent state variables, and thereby support non-Gaussian dynamics and data likelihoods. At each time
t we represent discrete cursor aim ?t as 72 values in [0, 2?) and goal position gt as a regular grid
of 40 ? 40 = 1600 locations (see Fig. 4). Each cell of the grid is small compared to elements of a
graphical interface. Cursor aim dynamics are conditioned on goal position and evolve according to a
smoothed von Mises distribution:
vMS(?t | gt , pt ) , ?/2? + (1 ? ?)vonMises(?t | a(gt , pt ), ?
? ).
(4)
?1
Here, a(g, p) = tan ((gy ? py )/(gx ? px )) is the angle from the cursor p = (px , py ) to the goal
g = (gx , gy ), and the concentration parameter ?
? encodes the expected accuracy of user aim. Neural
activity from some participants has short bursts of noise during which the learned angle likelihood is
inaccurate; the outlier weight 0 < ? < 1 adds robustness to these noise bursts.
4.2
Multiscale Semi-Markov Dynamics
The first-order Markov assumption made by existing iBCI decoders (see Eq. (1)) imposes a geometric
decay in state correlation over time. For example, consider a scalar Gaussian state-space model:
xt = ?xt?1 + v, v ? N (0, ? 2 ). For time lag k > 0, the covariance between two states cov(xt , xt+k )
decays as ? ?k . This weak temporal dependence is highly problematic in the iBCI setting due to the
mismatch between downsampled sensor acquisition rates used for decoding (typically around 50Hz,
or 20ms per timestep) and the time scale at which the desired goal position changes (seconds).
We relax the first-order Markov assumption via a semi-Markov model of state dynamics [Yu, 2010].
Semi-Markov models, introduced by Levy [1954] and Smith [1955], divide the state evolution into
contiguous segments. A segment is a contiguous series of timesteps during which a latent variable is
unchanged. The conditional distribution over the state at time xt depends not only on the previous
state xt?1 , but also on a duration dt which encodes how long the state is to remain unchanged:
4
Goal
reconsideration
counter
Goal position
Aim adjustment
counter
Angle of aim
Observation
Cursor position
Multiscale Semi-Markov Model
Junction Tree for Online Decoding
Figure 3: Multiscale semi-Markov dynamical model. Left: The multiscale directed graphical model of how
goal positions gt , angles of aim ?t , and observed cursor positions pt evolve over three time steps. Dashed nodes
are counter variables enabling semi-Markov dynamics. Right: Illustration of the junction tree used to compute
marginals for online decoding, as in Boyen and Koller [1998]. Dashed edges indicate cliques whose potentials
depend on the marginal approximations at time t ? 1. The inference uses an auxiliary variable rt , a(gt , pt ),
the angle from the cursor to the current goal, to reduce computation and allow inference to operate in real time.
p(xt | xt?1 , dt ). Duration is modeled via a latent counter variable, which is drawn at the start of
each segment and decremented deterministically until it reaches zero, at which point it is resampled.
In this way the semi-Markov model is capable of integrating information over longer time horizons,
and thus less susceptible to intermittent bursts of sensor noise.
We define separate semi-Markov dynamical models for the goal position and the angle of intended
movement. As detailed in the supplement, in experiments our duration distributions were uniform,
with parameters informed by knowledge about typical trajectory durations and reaction times.
Goal Dynamics A counter ct encodes the temporal evolution of the semi-Markov dynamics on
goal positions: ct is drawn from a discrete distribution p(c) at the start of each trajectory, and then
decremented deterministically until it reaches zero. (During decoding we do not know the value of
the counter, and maintain a posterior probability distribution over its value.) The goal position gt
remains unchanged until the goal counter reaches zero, at which point with probability ? we resample
a new goal, and we keep the same goal with the remaining probability 1 ? ?:
(
1,
ct = ct?1 ? 1, ct?1 > 0, Decrement
p(ct ), ct?1 = 0,
Sample new counter
p(ct | ct?1 ) =
(5)
0,
Otherwise
?
1,
ct?1 > 0, gt = gt?1 , Goal position unchanged
?
? 1
? G + (1 ? ?), ct?1 = 0, gt = gt?1 , Sample same goal position
p(gt | ct?1 , gt?1 ) =
(6)
1
ct?1 = 0, gt 6= gt?1 , Sample new goal position
?
? ?G,
0,
Otherwise
Cursor Angle Dynamics We define similar semi-Markov dynamics for the cursor angle via an aim
counter bt . Once the counter reaches zero, we sample a new aim counter value from the discrete
distribution p(b), and a new cursor aim angle from the smoothed von Mises distribution of Eq. (4):
(
1,
bt = bt?1 ? 1, bt?1 > 0, Decrement
p(bt ), bt?1 = 0,
Sample new counter
p(bt | bt?1 ) =
(7)
0,
Otherwise
?t?1
bt?1 > 0, Keep cursor aim
p(?t | bt?1 , ?t?1 , pt , gt ) =
(8)
vMS(?t | gt , pt ) bt?1 = 0, Sample new cursor aim
4.3
Decoding via Approximate Online Inference
Efficient decoding is possible via an approximate variant of the junction tree algorithm [Boyen and
Koller, 1998]. We approximate the full posterior at time t via a partially factorized posterior:
p(gt , ct , ?t , bt | z1...t ) ? p(gt , ct | z1...t )p(?t , bt | z1...t ).
(9)
5
Goal Positions
0s
0.5s
1s
1.5s
2s
Figure 4: Decoding goal positions. The MSSM represents goal position via a regular grid of 40 ? 40 locations
(upper left). For one real sequence of recorded neural data, the above panels illustrate the motion of the cursor
(white dot) to the user?s target (red circle). Panels show the marginal posterior distribution over goal positions at
0.5s intervals (25 discrete time steps of graphical model inference). Yellow goal states have highest probability,
dark blue goal states have near-zero probability. Note the temporal aggregation of directional cues.
Here p(gt , ct | z1...t ) is the marginal on the goal position and goal counter, and p(?t , bt | z1...t ) is the
marginal on the angle of aim and the aim counter. Note that in this setting goal position gt and cursor
aim ?t , as well as their respective counters ct and bt , are unknown and must be inferred from neural
data. At each inference step we use the junction tree algorithm to compute state marginals at time t,
conditioned on the factorized posterior approximation from time step t ? 1 (see Fig. 3). Boyen and
Koller [1998] show that this technique has bounded approximation error over time, and Murphy and
Weiss [2001] show this as a special case of loopy belief propagation.
Detailed inference equations are derived in the supplemental material. Given G goal positions
and A discrete angle states, each temporal update for our online decoder requires O(GA + A2 )
operations. In contrast, the exact junction tree algorithm would require O(G2 A2 ) operations; for
practical numbers of goals G, realtime implementation of this exact decoder is infeasible.
Figure 4 shows several snapshots of the marginal posterior
over
h
i goal position. At each time the
?pt
MSSM decoder moves the cursor along the vector E kggtt ?p
, computed by taking an average of
tk
the directions needed to get to each possible goal, weighted by the inferred probability that each goal
is the participant?s true target. This vector is smaller in magnitude when the decoder is less certain
about the direction in which the intended goal lies, which has the practical benefit of allowing the
participant to slow down near the goal.
5
Experiments
We evaluate all decoders under a variety of conditions and a range of configurations for each decoder.
Controlled offline evaluations allow us to assess the impact of each proposed innovation. To analyze
the effects of our proposed likelihood and multiscale dynamics in isolation, we construct a baseline
hidden Markov model (HMM) decoder using the same discrete representation of angles as the
MSSM, and either cosine-tuned or flexible likelihoods. Our findings show that the offline decoding
performance of the MSSM is superior in all respects to baseline models.
We also evaluate the MSSM decoder in two online clinical research sessions, and compare headto-head performance with the Kalman filter. Previous studies have tested the Kalman filter under a
variety of responsive parameter configurations and found a tradeoff between slow, smooth control
versus fast, meandering control [Willett et al., 2016, 2017]. Through comparisons to the Kalman, we
demonstrate that the MSSM decoder maintains smoother and more accurate control at comparable
speeds. These realtime results are preliminary since we have yet to evaluate the MSSM decoder on
other clinical metrics such as communication rate.
6
0.25
Mean squared error
0.2
0.15
0.1
0.05
0.05
(ra
K w)
os alm
a
in
Fl e H n
ex
ib MM
le
H
M
Ka
M
lm
an
BC
Ka (ra
w
l
C
os man )
in
B
Fl e M C
ex
ib SS
le
M
M
SS
M
an
C
Ka
lm
an
lm
Ka
0.1
0
(ra
Ka w)
l
os
m
a
in
Fl e H n
ex
ib MM
le
H
M
Ka
M
lm
an
BC
Ka (ra
w
l
C
os man )
in
B
Fl e M C
ex
ib SS
le
M
M
SS
M
0
0.2
0.15
C
Mean squared error
0.25
Figure 5: Offline decoding. Mean squared error of angular prediction for a variety of decoders, where each
decoder processes the same sets of recorded data. We analyze 24 minutes (eight 3-minute blocks) of neural data
recorded from participant T9 on trial days 546 and 552. We use one block for testing and the remainder for
training, and average errors across the choice of test block. On the left, we report errors over all time points. On
the right, we report errors on time points during which the cursor was outside a fixed distance from the target.
For both analyses, we exclude the initial 1s after target acquisition, during which the ground truth is unreliable.
To isolate preprocessing effects, the plots separately report the Kalman without preprocessing (?raw?). Dynamics
effects are isolated by separately evaluating HMM dynamics (?HMM?), and likelihood effects are isolated by
separately evaluating flexible likelihood and cosine tuning in each configuration. ?KalmanBC? denotes the
Kalman filter with an additional kinematic bias-correction heuristic [Jarosiewicz et al., 2015].
5.1
Offline evaluation
We perform offline analysis using previously recorded data from two historical sessions of iBCI use
with a single participant (T9). During each session the participant is asked to perform a cued target
task in which a target appears at a random location on the screen and the participant attempts to move
the cursor to the target. Once the target is acquired or after a timeout (10 seconds), a new target is
presented at a different location. Each session is composed of several 3 minute segments or blocks.
To evaluate the effect of each innovation we compare to an HMM decoder. This HMM baseline
isolates the effect of our flexible likelihood since, like the Kalman filter, it does not model goal
positions and assumes first-order Markov dynamics. Let ?t be the latent angle state at time t and
T
x(?) = (cos(?), sin(?)) the corresponding unit vector. We implement a pair of HMM decoders for
cosine tuning and our proposed flexible tuning curves,
zt | ?t ? N b + wT ?(?t ), Q
|
{z
}
zt | ?t ? N (b + Hx(?t ), Q),
{z
}
|
Cosine HMM
Flexible HMM
Here, ?(?) are the basis vectors defined in Eq. (3). The state ?t is discrete, taking one of 72 angular
values equally spaced in [0, 2?), the same discretization used by the MSSM. Continuous densities
are appropriately normalized. Unlike the linear Gaussian state-space model, the HMMs constrain
latent states to be valid angles (equivalently, unit vectors) rather than arbitrary vectors in R2 .
We analyze decoder accuracy within each session using a leave-one-out approach. Specifically, we
test the decoder on each held-out block using the remaining blocks in the same session for training.
We report MSE of the predicted cursor direction, using the unit vector from the cursor to the target
as ground truth, and normalizing decoder output vectors. We used the same recorded data for each
decoder. See the supplement for further details.
Figure 5 summarizes the findings of the offline comparisons for a variety of decoder configurations.
First, we evaluate the effect of preprocessing the data by taking the square root, applying a low-pass
IIR filter, and clipping the data outside a 5? threshold, where ? is the empirical standard deviation of
training data. This preprocessing significantly improves accuracy for all decoders. The MSSM model
compares favorably to all configurations of the Kalman decoders. The majority of benefit comes from
the semi-Markov dynamical model, but additional gains are observed when including the flexible
tuning likelihood. Finally, it has been observed that the Kalman decoder is sensitive to outliers for
which Jarosiewicz et al. [2015] propose a correction to avoid biased estimates. We test the Kalman
filter with and without this correction.
7
Session 1
3
Session 2
3
Time
Time
1.5
1
0
0
SS
M
M
lm
Ka
SS
M
M
0.5
M
SS
M
Ka
lm
an
M
SS
M
Ka
lm
an
0.5
M
SS
M
Ka
lm
an
M
SS
M
Ka
lm
an
1
an
1.5
2
lm
2
Ka
Squared error
2.5
an
Squared error
2.5
Figure 6: Realtime decoding. A realtime comparison of the Kalman filter and MSSM with flexible likelihoods
from two sessions with clinical trial participant T10. Left: Box plots of squared error between unit vectors from
cursor to target and normalized (unit vector) decoder output for each four-minute comparison block in a session.
MSSM errors are consistently smaller. Right: Two metrics that describe the smoothness of cursor trajectories,
introduced in MacKenzie et al. [2001] and commonly used to quantify iBCI performance [Kim et al., 2008,
Simeral et al., 2011]. The task axis for a trajectory is the straight line from the cursor?s starting position at the
beginning of a trajectory to a goal. Orthogonal directional changes measure the number of direction changes
towards or away from the goal, and movement direction changes measure the number of direction changes
towards or away from the task axis. The MSSM shows significantly fewer direction changes according to both
metrics.
5.2
Realtime evaluation
Next, we examined whether the MSSM method was effective for realtime iBCI control by a clinical
trial participant. On two different days, a clinical trial participant (T10) completed six four-minute
comparison blocks. In these blocks, we alternated using an MSSM decoder with flexible likelihoods
and novel preprocessing, or a standard Kalman decoder. As with the Kalman decoding described in
Jarosiewicz et al. [2015], we used the Kalman filter in conjunction with a bias correcting postprocessing heuristic. We used the feature selection method proposed by Malik et al. [2015] to select D = 60
channels of neural data, and used these same 60 channels for both decoders.
Jarosiewicz et al. [2015] selected the timesteps of data to use for parameter learning by taking the
first two seconds of each trajectory after a 0.3s reaction time. For both decoders, we instead selected
all timesteps in which the cursor was a fixed distance from the cued goal because we found this
alternative method lead to improvements in offline decoding. Both methods for selecting subsets of
the calibration data are designed to compensate for the fact that vectors from cursor to target are not a
reliable estimator for participants? intended aim when the cursor is near the target.
Decoding accuracy. Figure 6 shows that our MSSM decoder had less directional error than the
configuration of the Kalman filter that we compared to. We confirmed the statistical significance
of this result using a Wilcoxon rank sum test. To accommodate the Wilcoxon rank sum test?s
independence assumption, we divided the data into individual trajectories from a starting point
towards a goal, that ended either when the cursor reached the goal or at a timeout (10 seconds). We
then computed the mean squared error of each trajectory, where the squared error is the squared
Euclidean distance between the normalized (unit vector) decoded vectors and the unit vectors from
cursor to target. Within each session, we compared the distributions of these mean squared errors for
trajectories between decoders (p < 10?6 for each session). MSSM also performed better than the
Kalman on metrics from MacKenzie et al. [2001] that measure the smoothness of cursor trajectories
(see Fig. 6).
Figure 7 shows example trajectories as the cursor moves toward its target via the MSSM decoder or
the (bias-corrected) Kalman decoder. Consistent with the quantitative error metrics, the trajectories
produced by the MSSM model were smoother and more direct than those of the Kalman filter,
especially as the cursor approached the goal. The distance ratio (the ratio of the length of the
trajectory to the line from the starting position to the goal) averaged 1.17 for the MSSM decoder and
1.28 for the Kalman decoder, a significant difference (Wilcoxon rank sum test, p < 10?6 ). Some
trajectories for both decoders are shown in Figure 7. Videos of cursor movement under both decoding
algorithms, and additional experimental details, are included in the supplemental material.
Decoding speed. We controlled for speed by configuring both decoders to average the same fast
speed determined in collaboration with clinical research engineers familiar with the participant?s
8
Near Goal
Goal
Figure 7: Examples of realtime decoding trajectories. Left: 20 randomly selected trajectories for the Kalman
decoder, and 20 trajectories for the MSSM decoder. The trajectories are aligned so that the starting position is at
the origin and rotated so the goal position is on the positive, horizontal axis. The MSSM decoder exhibits fewer
abrupt direction changes. Right: The empirical probability of instantaneous angle of movement, after rotating all
trajectories from the realtime data (24 minutes of iBCI use with each decoder). The MSSM distribution (shown
as translucent cyan) is more peaked at zero degrees, corresponding to direct motion towards the goal.
preferred cursor speed. For each decoder, we collected a block of data in which the participant used
that decoder to control the cursor. For each of these blocks, we computed the trimmed mean of
the speed, and then linearly extrapolated the speed gain needed for the desired speed. Although
such an extrapolation is approximate, the average times to acquire a target with each decoder at the
extrapolated speed gains were within 6% of each other: 2.6s for the Kalman decoder versus 2.7s for
the MSSM decoder. This speed discrepancy is dominated by the relative performance improvement
of MSSM over Kalman: the Kalman had a 30.7% greater trajectory mean squared error, 249% more
orthogonal direction changes, and 224% more movement direction changes.
This approach to evaluating decoder performance differs from that suggested by Willett et al. [2016],
which discusses the possibility of optimizing the speed gain and other decoder parameters to minimize
target acquisition time. In contrast, we matched the speed of both decoders and evaluated decoding
error and smoothness. We did not extensively tune the dynamics parameters for either decoder,
instead relying on the Kalman parameters in everyday use by T10. For MSSM we tried two values of
?, which controls the sampling of goal states (6), and chose the remaining parameters offline.
6
Conclusion
We introduce a flexible likelihood model and multiscale semi-Markov (MSSM) dynamics for cursor
control in intracortical brain-computer interfaces. The flexible tuning likelihood model extends the
cosine tuning model to allow for multimodal tuning curves and narrower peaks. The MSSM dynamic
Bayesian network explicitly models the relationship between the goal position, the cursor position,
and the angle of intended movement. Because the goal position changes much less frequently than
the angle of intended movement, a decoder?s past knowledge of the goal position stays relevant for
longer, and the MSSM model can use longer histories of neural activity to infer the direction of
desired movement.
To create a realtime decoding algorithm, we derive an online variant of the junction tree algorithm
with provable accuracy guarantees. We demonstrate a significant improvement over the Kalman
filter in offline experiments with neural recordings, and demonstrate promising preliminary results in
clinical trial tests. Future work will further evaluate the suitability of this method for clinical use.
We hope that the MSSM graphical model will also enable further advances in iBCI decoding, for
example by encoding the structure of a known user interface in the set of latent goals.
Author contributions DJM, JLP, and EBS created the flexible tuning likelihood and the multiscale
semi-Markov dynamics. DJM derived the inference (decoder), wrote software implementations of
these methods, and performed data analyses. DJM, JLP, and EBS designed offline experiments. DJM,
BJ, and JDS designed clinical research sessions. LRH is the sponsor-investigator of the BrainGate2
pilot clinical trial. DJM, JLP, and EBS wrote the manuscript with input from all authors.
9
Acknowledgments
The authors thank Participants T9 and T10 and their families, Brian Franco, Tommy Hosman, Jessica
Kelemen, Dave Rosler, Jad Saab, and Beth Travers for their contributions to this research. Support
for this study was provided by the Office of Research and Development, Rehabilitation R&D Service,
Department of Veterans Affairs (B4853C, B6453R, and N9228C), the National Institute on Deafness
and Other Communication Disorders of National Institutes of Health (NIDCD-NIH: R01DC009899),
MGH-Deane Institute, and The Executive Committee on Research (ECOR) of Massachusetts General
Hospital. The content is solely the responsibility of the authors and does not necessarily represent the
official views of the National Institutes of Health, or the Department of Veterans Affairs or the United
States Government. CAUTION: Investigational Device. Limited by Federal Law to Investigational
Use.
Disclosure: Dr. Hochberg has a financial interest in Synchron Med, Inc., a company developing
a minimally invasive implantable brain device that could help paralyzed patients achieve direct
brain control of assistive technologies. Dr. Hochberg?s interests were reviewed and are managed by
Massachusetts General Hospital, Partners HealthCare, and Brown University in accordance with their
conflict of interest policies.
References
A. B. Ajiboye, F. R. Willett, D. R. Young, W. D. Memberg, B. A. Murphy, J. P. Miller, B. L. Walter,
J. A. Sweet, H. A. Hoyen, M. W. Keith, et al. Restoration of reaching and grasping movements
through brain-controlled muscle stimulation in a person with tetraplegia. The Lancet, 389(10081):
1821?1830, 2017.
B. Amirikian and A. P. Georgopulos. Directional tuning profiles of motor cortical cells. J. Neurosci.
Res., 36(1):73?79, 2000.
D. Bacher, B. Jarosiewicz, N. Y. Masse, S. D. Stavisky, J. D. Simeral, K. Newell, E. M. Oakley, S. S.
Cash, G. Friehs, and L. R. Hochberg. Neural point-and-click communication by a person with
incomplete locked-in syndrome. Neurorehabil Neural Repair, 29(5):462?471, 2015.
X. Boyen and D. Koller. Tractable inference for complex stochastic processes. In Proceedings of
UAI, pages 33?42. Morgan Kaufmann Publishers Inc., 1998.
J. L. Collinger, B. Wodlinger, J. E. Downey, W. Wang, E. C. Tyler-Kabara, D. J. Weber, A. J.
McMorland, M. Velliste, M. L. Boninger, and A. B. Schwartz. High-performance neuroprosthetic
control by an individual with tetraplegia. The Lancet, 381(9866):557?564, 2013.
J. A. Donoghue, J. H. Bagley, V. Zerris, and G. M. Friehs. Youman?s neurological surgery: Chapter
94. Elsevier/Saunders, 2011.
G. W. Fraser, S. M. Chase, A. Whitford, and A. B. Schwartz. Control of a brain?computer interface
without spike sorting. J. Neuroeng, 6(5):055004, 2009.
A. P. Georgopoulos, J. F. Kalaska, R. Caminiti, and J. T. Massey. On the relations between the
direction of two-dimensional arm movements and cell discharge in primate motor cortex. J.
Neurosci., 2(11):1527?1537, 1982.
V. Gilja, C. Pandarinath, C. H. Blabe, P. Nuyujukian, J. D. Simeral, A. A. Sarma, B. L. Sorice, J. A.
Perge, B. Jarosiewicz, L. R. Hochberg, et al. Clinical translation of a high performance neural
prosthesis. Nature Medicine, 21(10):1142, 2015.
L. R. Hochberg, M. D. Serruya, G. M. Friehs, J. A. Mukand, M. Saleh, A. H. Caplan, A. Branner,
D. Chen, R. D. Penn, and J. P. Donoghue. Neuronal ensemble control of prosthetic devices by a
human with tetraplegia. Nature, 442(7099):164?171, 2006.
L. R. Hochberg, D. Bacher, B. Jarosiewicz, N. Y. Masse, J. D. Simeral, J. Vogel, S. Haddadin, J. Liu,
S. S. Cash, P. van der Smagt, et al. Reach and grasp by people with tetraplegia using a neurally
controlled robotic arm. Nature, 485(7398):372, 2012.
10
B. Jarosiewicz, A. A. Sarma, D. Bacher, N. Y. Masse, J. D. Simeral, B. Sorice, E. M. Oakley, C. Blabe,
C. Pandarinath, V. Gilja, et al. Virtual typing by people with tetraplegia using a self-calibrating
intracortical brain-computer interface. Sci. Transl. Med., 7(313):313ra179?313ra179, 2015.
S. P. Kim, J. D. Simeral, L. R. Hochberg, J. P. Donoghue, and M. J. Black. Neural control of computer
cursor velocity by decoding motor cortical spiking activity in humans with tetraplegia. J. Neuroeng,
5(4):455, 2008.
S. Koyama, S. M. Chase, A. S. Whitford, M. Velliste, A. B. Schwartz, and R. E. Kass. Comparison
of brain?computer interface decoding algorithms in open-loop and closed-loop control. J Comput
Neurosci, 29(1-2):73?87, 2010.
P. Levy. Semi-Markovian processes. Proc: III ICM (Amsterdam), pages 416?426, 1954.
I. S. MacKenzie, T. Kauppinen, and M. Silfverberg. Accuracy measures for evaluating computer
pointing devices. In Proceedings of CHI, pages 9?16. ACM, 2001.
W. Q. Malik, L. R. Hochberg, J. P. Donoghue, and E. N. Brown. Modulation depth estimation and
variable selection in state-space models for neural interfaces. IEEE Trans Biomed Eng, 62(2):
570?581, 2015.
K. Murphy and Y. Weiss. The factored frontier algorithm for approximate inference in DBNs. In
Proceedings of UAI, pages 378?385. Morgan Kaufmann Publishers Inc., 2001.
K. P. Murphy. Dynamic Bayesian networks: Representation, inference and learning. PhD thesis,
University of California, Berkeley, 2002.
C. Pandarinath, P. Nuyujukian, C. H. Blabe, B. L. Sorice, J. Saab, F. R. Willett, L. R. Hochberg, K. V.
Shenoy, and J. M. Henderson. High performance communication by people with paralysis using
an intracortical brain-computer interface. eLife, 6:e18554, 2017.
J. D. Simeral, S. P. Kim, M. Black, J. Donoghue, and L. Hochberg. Neural control of cursor trajectory
and click by a human with tetraplegia 1000 days after implant of an intracortical microelectrode
array. J. Neuroeng, 8(2):025027, 2011.
W. L. Smith. Regenerative stochastic processes. Proceedings of the Royal Society of London A:
Mathematical, Physical and Engineering Sciences, 232(1188):6?31, 1955.
F. R. Willett, C. Pandarinath, B. Jarosiewicz, B. A. Murphy, W. D. Memberg, C. H. Blabe, J. Saab,
B. L. Walter, J. A. Sweet, J. P. Miller, et al. Feedback control policies employed by people using
intracortical brain?computer interfaces. J. Neuroeng, 14(1):016001, 2016.
F. R. Willett, B. A. Murphy, W. D. Memberg, C. H. Blabe, C. Pandarinath, B. L. Walter, J. A. Sweet,
J. P. Miller, J. M. Henderson, K. V. Shenoy, et al. Signal-independent noise in intracortical brain?
computer interfaces causes movement time properties inconsistent with Fitts? law. J. Neuroeng, 14
(2):026010, 2017.
S. Z. Yu. Hidden semi-Markov models. Artificial intelligence, 174(2):215?243, 2010.
11
| 6688 |@word trial:10 open:1 tried:1 covariance:6 eng:1 thereby:1 accommodate:1 initial:1 configuration:7 series:1 liu:1 selecting:1 united:1 daniel:1 tuned:5 bc:2 past:1 existing:1 reaction:2 current:3 ka:14 discretization:1 yet:1 written:1 must:2 john:1 enables:1 motor:10 designed:4 plot:2 update:1 intelligence:2 cue:1 device:5 fewer:2 selected:3 beginning:1 affair:3 smith:2 short:1 record:1 provides:1 node:1 location:6 gx:2 mathematical:2 simeral:9 constructed:1 burst:3 along:1 direct:3 transl:1 tommy:1 introduce:4 acquired:1 alm:1 ra:4 expected:2 rapid:1 frequently:1 mukand:1 brain:14 chi:1 inspired:1 relying:1 company:1 provided:1 moreover:1 underlying:1 panel:3 factorized:2 bounded:1 lowest:1 translucent:1 matched:1 informed:1 supplemental:3 caution:2 finding:2 htj:1 ended:1 guarantee:1 temporal:4 quantitative:1 every:1 berkeley:1 axt:1 demonstrates:1 scaled:1 schwartz:3 control:22 unit:8 medical:2 configuring:1 healthcare:1 penn:1 shenoy:2 positive:1 service:2 engineering:2 accordance:1 limit:1 severely:1 despite:1 encoding:1 jarosiewicz:14 solely:1 modulation:2 black:3 chose:1 mssm:36 eb:3 minimally:1 examined:1 collect:1 co:4 hmms:1 limited:2 factorization:1 range:1 locked:1 averaged:1 directed:2 practical:2 acknowledgment:1 testing:1 block:11 implement:1 differs:1 stavisky:1 empirical:4 significantly:4 t10:4 intention:2 integrating:1 regular:3 downsampled:1 get:1 caminiti:1 ga:1 selection:2 applying:1 py:2 measurable:1 center:4 resembling:1 starting:4 duration:4 simplicity:1 abrupt:1 disorder:1 correcting:1 factored:1 estimator:1 array:6 his:2 enabled:1 financial:1 population:1 discharge:1 target:24 pt:7 tan:1 user:4 engage:1 exact:2 dbns:1 us:3 origin:1 harvard:1 element:1 velocity:1 hj2:1 labeled:2 observed:4 electrical:2 capture:4 wang:1 connected:1 intends:1 grasping:1 movement:21 contemporary:1 highest:2 counter:15 disease:1 asked:2 dynamic:33 depend:1 segment:4 predictive:1 upon:1 efficiency:1 basis:3 multimodal:6 connector:1 chapter:1 assistive:1 walter:3 fast:2 shortcoming:1 effective:2 describe:1 london:1 artificial:2 approached:1 outside:2 saunders:1 whose:1 lag:1 stanford:3 heuristic:2 drawing:1 relax:1 otherwise:3 s:10 objectively:1 cov:1 jointly:1 noisy:2 meandering:1 online:10 timeout:2 chase:2 sequence:2 propose:2 regain:1 product:1 closed:1 remainder:1 frequent:2 uci:1 aligned:1 relevant:1 rapidly:2 loop:2 achieve:1 bed:1 moved:1 validate:1 everyday:1 electrode:6 leave:1 rotated:1 tk:1 cued:4 derive:4 illustrate:1 help:1 school:2 keith:1 eq:5 auxiliary:1 predicted:1 indicate:1 come:1 quantify:1 direction:20 filter:23 stochastic:2 centered:1 human:6 enable:2 transient:1 material:3 virtual:1 bin:1 require:1 government:1 hx:1 preliminary:2 suitability:1 brian:1 frontier:1 correction:3 mm:2 lying:1 around:2 ground:2 normal:1 exp:1 mgh:1 tyler:1 branner:1 bj:4 lm:10 pointing:2 a2:2 resample:1 estimation:1 proc:1 label:1 sensitive:3 create:1 weighted:1 hope:1 federal:2 mit:2 neurosurgery:1 sensor:2 gaussian:7 beth:1 aim:18 pacheco:1 rather:1 reaching:1 avoid:1 hj:3 cash:2 varying:1 office:1 conjunction:1 derived:3 focus:1 improvement:3 consistently:1 rank:3 likelihood:22 contrast:4 kim:4 baseline:5 caplan:1 elsevier:1 inference:13 inaccurate:1 typically:3 bt:15 hidden:2 koller:5 relation:1 smagt:1 transformed:1 atan:1 microelectrode:3 biomed:1 flexible:21 proposes:1 development:1 art:3 constrained:1 special:1 marginal:5 field:1 once:2 construct:1 masse:3 beach:1 sampling:2 represents:2 broad:1 yu:2 investigational:4 peaked:1 discrepancy:1 leigh:1 decremented:2 report:4 future:1 few:1 sweet:3 randomly:1 composed:1 simultaneously:1 national:3 individual:4 implantable:1 murphy:8 familiar:1 intended:13 maintain:2 attempt:3 jessica:1 interest:4 highly:1 kinematic:1 possibility:1 evaluation:3 adjust:1 grasp:1 henderson:2 ztj:1 held:2 accurate:2 edge:1 capable:1 respective:1 orthogonal:2 tree:8 incomplete:1 divide:1 euclidean:1 walk:1 desired:4 circle:1 rotating:1 isolated:2 jlp:3 re:1 prosthesis:1 nuyujukian:2 modeling:1 markovian:1 injury:2 contiguous:2 restoration:1 loopy:1 tractability:1 introducing:1 deviation:3 clipping:1 straighter:1 subset:1 uniform:1 iir:1 providence:5 corrupted:1 calibrated:1 person:3 st:1 peak:3 density:3 stay:1 decoding:34 von:8 squared:11 recorded:11 thesis:1 dr:2 semimarkov:1 potential:1 exclude:1 intracortical:10 gy:2 sec:7 pooled:1 includes:1 inc:3 explicitly:4 depends:1 performed:2 root:1 jason:1 responsibility:1 view:1 apparently:1 analyze:3 red:2 start:2 aggregation:1 participant:21 maintains:1 reached:1 whitford:2 contribution:2 ass:1 square:1 nidcd:1 accuracy:6 minimize:1 kaufmann:2 efficiently:1 miller:3 spaced:2 ensemble:1 directional:5 yellow:1 weak:1 bayesian:6 raw:1 produced:1 trajectory:22 confirmed:1 straight:2 dave:1 history:2 reach:5 acquisition:4 invasive:1 mi:8 irvine:1 pilot:2 sampled:1 gain:5 massachusetts:3 tunable:1 knowledge:2 improves:1 appears:1 manuscript:1 friehs:3 higher:1 dt:2 day:3 improved:2 wei:2 reconsideration:1 evaluated:1 box:2 angular:4 just:1 correlation:1 until:3 hand:2 horizontal:1 multiscale:13 nonlinear:1 propagation:1 o:4 gray:1 usa:11 effect:8 calibrating:1 brown:10 true:1 normalized:3 managed:1 evolution:2 alumnus:1 laboratory:1 white:1 sin:1 during:7 self:1 steady:1 cosine:18 m:2 evident:1 ridge:2 demonstrate:6 deafness:1 motion:9 interface:13 tetraplegia:11 postprocessing:1 image:1 weber:1 instantaneous:2 isolates:1 novel:1 nih:1 common:1 superior:1 stimulation:1 spiking:1 physical:1 relating:1 marginals:2 willett:6 expressing:1 measurement:3 significant:2 cambridge:1 smoothness:3 rd:2 tuning:26 grid:4 velliste:2 hj1:1 session:13 had:2 dot:1 moving:1 calibration:3 longer:4 cortex:7 gt:20 add:1 wilcoxon:3 posterior:6 own:1 recent:1 sarma:2 optimizing:1 certain:1 incapable:1 affiliation:1 travers:1 ht1:1 der:1 muscle:2 responsiveness:2 morgan:2 additional:3 greater:1 employed:2 syndrome:1 aggregated:2 dashed:2 paralyzed:3 semi:19 smoother:3 signal:9 unimodal:3 violate:1 multiple:2 infer:2 exceeds:1 full:1 faster:1 smooth:1 clinical:19 long:3 compensate:1 kalaska:1 divided:1 equally:1 fraser:2 regenerative:1 controlled:4 impact:1 prediction:3 variant:5 regression:5 sponsor:1 implanted:2 patient:1 metric:5 sometimes:2 represent:3 bimodal:1 serruya:1 cell:3 receive:1 addition:3 separately:3 interval:1 sudderth:2 sends:1 crucial:1 appropriately:1 biased:1 operate:1 unlike:1 publisher:2 vogel:1 recording:2 hz:3 isolate:1 med:2 regularly:1 incorporates:1 inconsistent:1 near:4 intermediate:1 iii:1 variety:4 independence:2 fit:11 timesteps:3 isolation:1 opposite:1 click:2 reduce:2 inner:1 simplifies:1 tradeoff:1 donoghue:6 whether:1 six:1 trimmed:1 downey:1 cause:1 detailed:2 tune:1 dark:1 extensively:1 specifies:1 problematic:1 djm:5 neuroscience:1 estimated:2 per:1 blue:1 discrete:7 key:1 four:3 threshold:1 drawn:2 pj:2 utilize:1 volitional:1 timestep:2 massey:1 year:1 sum:3 enforced:1 angle:34 extends:1 family:1 realtime:9 hochberg:13 summarizes:1 comparable:1 capturing:1 cyan:2 resampled:1 ct:17 fl:4 display:1 activity:28 constraint:1 extrapolation:1 constrain:1 georgopoulos:3 ri:5 software:1 encodes:3 prosthetic:1 dominated:2 speed:13 franco:1 elife:1 px:2 department:8 structured:1 according:4 developing:1 across:4 remain:2 smaller:2 cable:1 rehabilitation:2 evolves:1 primate:2 outlier:2 repair:1 equation:1 remains:1 previously:1 discus:1 kinematics:1 committee:1 needed:2 know:1 disclosure:1 tractable:1 junction:8 operation:2 eight:1 limb:2 oakley:2 away:2 robustly:1 responsive:1 permission:1 robustness:1 alternative:1 assumes:5 remaining:3 denotes:1 completed:1 graphical:6 medicine:1 especially:1 veteran:3 society:1 unchanged:4 surgery:1 move:7 malik:2 quantity:1 spike:4 concentration:3 dependence:3 rt:1 diagonal:1 neurally:1 exhibit:1 distance:4 separate:1 thank:1 sci:1 decoder:58 koyama:2 hmm:8 majority:1 evaluate:7 partner:1 collected:1 consensus:1 toward:1 provable:1 assuming:1 erik:1 length:1 kalman:36 modeled:1 relationship:3 prompted:1 providing:1 illustration:1 ratio:2 innovation:2 equivalently:1 acquire:1 susceptible:1 favorably:1 implementation:3 zt:5 policy:2 unknown:4 perform:2 allowing:2 discretize:1 upper:1 observation:7 neuron:4 markov:24 snapshot:1 enabling:2 communication:5 head:1 transcutaneous:1 intermittent:1 smoothed:2 arbitrary:2 inferred:2 introduced:3 pair:1 required:1 z1:5 conflict:1 california:2 learned:1 nip:1 trans:1 suggested:1 dynamical:5 pattern:1 mismatch:1 boyen:5 pioneering:1 including:1 reliable:1 video:2 belief:1 royal:1 power:3 typing:2 residual:1 arm:4 improve:1 technology:1 mackenzie:3 axis:3 created:1 extract:1 health:2 alternated:1 prior:1 review:1 geometric:1 evolve:3 relative:1 law:3 embedded:1 loss:1 proportional:2 versus:2 executive:1 integrate:1 degree:5 consistent:1 imposes:1 lancet:2 collaboration:1 translation:1 row:1 extrapolated:2 heading:1 infeasible:1 offline:12 side:1 bias:4 allow:4 institute:5 taking:4 lrh:1 benefit:2 van:1 curve:8 depth:1 cortical:3 evaluating:4 valid:1 neuroprosthetic:1 feedback:1 author:5 made:2 commonly:2 preprocessing:5 historical:1 cope:1 approximate:5 emphasize:1 preferred:3 implicitly:1 jad:1 keep:2 clique:1 unreliable:1 wrote:2 robotic:1 paralysis:2 uai:2 assumed:2 neurology:2 gilja:4 continuous:1 latent:10 reviewed:2 promising:1 nature:3 channel:5 mj:2 ca:3 imagining:1 mse:1 complex:3 necessarily:1 official:1 did:1 significance:1 timescales:1 linearly:1 decrement:2 neurosci:3 noise:7 hyperparameters:1 profile:1 allowed:1 icm:1 neuronal:1 fig:6 referred:1 htd:1 screen:5 deployed:1 slow:2 position:40 vms:2 deterministically:2 decoded:1 comput:1 lie:1 screenshot:1 ib:4 levy:2 third:1 young:1 down:1 minute:6 xt:17 r2:2 decay:2 t9:3 normalizing:1 hxt:1 modulates:1 supplement:2 phd:1 magnitude:1 implant:2 conditioned:2 cursor:58 horizon:1 sorting:1 chen:1 boston:2 suited:1 amsterdam:1 adjustment:3 partially:1 scalar:1 g2:1 neurological:1 newell:1 truth:2 acm:1 ma:3 pandarinath:7 saleh:1 conditional:1 goal:62 narrower:2 towards:4 man:2 content:1 change:13 included:2 typical:1 specifically:1 corrected:1 determined:1 justify:1 wt:2 engineer:1 called:1 hospital:3 pas:1 experimental:1 select:1 people:6 support:2 inability:1 investigator:1 incorporate:1 dept:1 tested:1 ex:4 |
6,288 | 6,689 | PredRNN: Recurrent Neural Networks for Predictive
Learning using Spatiotemporal LSTMs
Yunbo Wang
School of Software
Tsinghua University
[email protected]
Jianmin Wang
School of Software
Tsinghua University
[email protected]
Mingsheng Long?
School of Software
Tsinghua University
[email protected]
Zhifeng Gao
School of Software
Tsinghua University
[email protected]
Philip S. Yu
School of Software
Tsinghua University
[email protected]
Abstract
The predictive learning of spatiotemporal sequences aims to generate future images
by learning from the historical frames, where spatial appearances and temporal variations are two crucial structures. This paper models these structures by presenting a
predictive recurrent neural network (PredRNN). This architecture is enlightened by
the idea that spatiotemporal predictive learning should memorize both spatial appearances and temporal variations in a unified memory pool. Concretely, memory
states are no longer constrained inside each LSTM unit. Instead, they are allowed
to zigzag in two directions: across stacked RNN layers vertically and through all
RNN states horizontally. The core of this network is a new Spatiotemporal LSTM
(ST-LSTM) unit that extracts and memorizes spatial and temporal representations
simultaneously. PredRNN achieves the state-of-the-art prediction performance on
three video prediction datasets and is a more general framework, that can be easily
extended to other predictive learning tasks by integrating with other architectures.
1
Introduction
As a key application of predictive learning, generating images conditioned on given consecutive
frames has received growing interests in machine learning and computer vision communities. To
learn representations of spatiotemporal sequences, recurrent neural networks (RNN) [17, 27] with
the Long Short-Term Memory (LSTM) [9] have been recently extended from supervised sequence
learning tasks, such as machine translation [22, 2], speech recognition [8], action recognition [28, 5]
and video captioning [5], to this spatiotemporal predictive learning scenario [21, 16, 19, 6, 25, 12].
1.1
Why spatiotemporal memory?
In spatiotemporal predictive learning, there are two crucial aspects: spatial correlations and temporal
dynamics. The performance of a prediction system depends on whether it is able to memorize
relevant structures. However, to the best of our knowledge, the state-of-the-art RNN/LSTM predictive
learning methods [19, 21, 6, 12, 25] focus more on modeling temporal variations (such as the object
moving trajectories), with memory states being updated repeatedly over time inside each LSTM unit.
Admittedly, the stacked LSTM architecture is proved powerful for supervised spatiotemporal learning
?
Corresponding author: Mingsheng Long
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
(such as video action recognition [5, 28]). Two conditions are met in this scenario: (1) Temporal
features are strong enough for classification tasks. In contrast, fine-grained spatial appearances prove
to be less significant; (2) There are no complex visual structures to be modeled in the expected
outputs so that spatial representations can be highly abstracted. However, spatiotemporal predictive
leaning does not satisfy these conditions. Here, spatial deformations and temporal dynamics are
equally significant to generating future frames. A straightforward idea is that if we hope to foretell
the future, we need to memorize as many historical details as possible. When we recall something
happened before, we do not just recall object movements, but also recollect visual appearances from
coarse to fine. Motivated by this, we present a new recurrent architecture called Predictive RNN
(PredRNN), which allows memory states belonging to different LSTMs to interact across layers (in
conventional RNNs, they are mutually independent). As the key component of PredRNN, we design
a novel Spatiotemporal LSTM (ST-LSTM) unit. It models spatial and temporal representations in a
unified memory cell and convey the memory both vertically across layers and horizontally over states.
PredRNN achieves the state-of-the-art prediction results on three video datasets. It is a general and
modular framework for predictive learning and is not limited to video prediction.
1.2
Related work
Recent advances in recurrent neural network models provide some useful insights on how to predict
future visual sequences based on historical observations. Ranzato et al. [16] defined a RNN
architecture inspired from language modeling, predicting the frames in a discrete space of patch
clusters. Srivastava et al. [21] adapted the sequence to sequence LSTM framework. Shi et al. [19]
extended this model to further extract visual representations by exploiting convolutions in both
input-to-state and state-to-state transitions. This Convolutional LSTM (ConvLSTM) model has
become a seminal work in this area. Subsequently, Finn et al. [6] constructed a network based on
ConvLSTMs that predicts transformations on the input pixels for next-frame prediction. Lotter et al.
[12] presented a deep predictive coding network where each ConvLSTM layer outputs a layer-specific
prediction at each time step and produces an error term, which is then propagated laterally and
vertically in the network. However, in their settings, the predicted next frame always bases on the
whole previous ground truth sequence. By contrast, we predict sequence from sequence, which is
obviously more challenging. Patraucean et al. [15] and Villegas et al. [25] brought optical flow
into RNNs to model short-term temporal dynamics, which is inspired by the two-stream CNNs [20]
designed for action recognition. However, the optical flow images are hard to use since they would
bring in high additional computational costs and reduce the prediction efficiency. Kalchbrenner
et al. [10] proposed a Video Pixel Network (VPN) that estimates the discrete joint distribution
of the raw pixel values in a video using the well-established PixelCNNs [24]. But it suffers from
high computational complexity. Besides the above RNN architectures, other deep architectures are
involved to solve the visual predictive learning problem. Oh et al. [14] defined a CNN-based action
conditional autoencoder model to predict next frames in Atari games. Mathieu et al. [13] successfully
employed generative adversarial networks [7, 4] to preserve the sharpness of the predicted frames.
In summary, these existing visual prediction models yield different shortcomings due to different
causes. The RNN-based architectures [21, 16, 19, 6, 25, 12] model temporal structures with LSTMs,
but their predicted images tend to blur due to a loss of fine-grained visual appearances. In contrast,
CNN-based networks [13, 14] predict one frame at a time and generate future images recursively,
which are prone to focus on spatial appearances and relatively weak in capturing long-term motions.
In this paper, we explore a new RNN framework for predictive learning and present a novel LSTM
unit for memorizing spatiotemporal information simultaneously.
2
2.1
Preliminaries
Spatiotemporal predictive learning
Suppose we are monitoring a dynamical system (e.g. a video clip) of P measurements over time,
where each measurement (e.g. a RGB channel) is recorded at all locations in a spatial region
represented by an M ? N grid (e.g. video frames). From the spatial view, the observation of these P
measurements at any time can be represented by a tensor X ? RP ?M ?N . From the temporal view,
the observations over T time steps form a sequence of tensors X1 , X2 , . . . , XT . The spatiotemporal
predictive learning problem is to predict the most probable length-K sequence in the future given the
2
previous length-J sequence including the current observation:
Xbt+1 , . . . , Xbt+K =
arg max p (Xt+1 , . . . , Xt+K |Xt?J+1 , . . . , Xt ) .
(1)
Xt+1 ,...,Xt+K
Spatiotemporal predictive learning is an important problem, which could find crucial and high-impact
applications in various domains: video prediction and surveillance, meteorological and environmental
forecasting, energy and smart grid management, economics and finance prediction, etc. Taking video
prediction as an example, the measurements are the three RGB channels, and the observation at
each time step is a 3D video frame of RGB image. Another example is radar-based precipitation
forecasting, where the measurement is radar echo values and the observation at every time step is a
2D radar echo map that can be visualized as an RGB image.
2.2
Convolutional LSTM
Compared with standard LSTMs, the convolutional LSTM (ConvLSTM) [19] is able to model the
spatiotemporal structures simultaneously by explicitly encoding the spatial information into tensors,
overcoming the limitation of vector-variate representations in standard LSTM where the spatial
information is lost. In ConvLSTM, all the inputs X1 , . . . , Xt , cell outputs C1 , . . . , Ct , hidden state
H1 , . . . , Ht , and gates it , ft , gt , ot are 3D tensors in RP ?M ?N , where the first dimension is either the
number of measurement (for inputs) or the number of feature maps (for intermediate representations),
and the last two dimensions are spatial dimensions (rows M and columns N ). To get a better picture
of the inputs and states, we may imagine them as vectors standing on a spatial grid. ConvLSTM
determines the future state of a certain cell in the M ? N grid by the inputs and past states of its
local neighbors. This can easily be achieved by using convolution operators in the state-to-state and
input-to-state transitions. The key equations of ConvLSTM are shown as follows:
gt = tanh(Wxg ? Xt + Whg ? Ht?1 + bg )
it = ?(Wxi ? Xt + Whi ? Ht?1 + Wci Ct?1 + bi )
ft = ?(Wxf ? Xt + Whf ? Ht?1 + Wcf Ct?1 + bf )
Ct = ft Ct?1 + it gt
ot = ?(Wxo ? Xt + Who ? Ht?1 + Wco Ct + bo )
Ht = ot tanh(Ct ),
(2)
where ? is sigmoid activation function, ? and denote the convolution operator and the Hadamard
product respectively. If the states are viewed as the hidden representations of moving objects, then a
ConvLSTM with a larger transitional kernel should be able to capture faster motions while one with a
smaller kernel can capture slower motions [19]. The use of the input gate it , forget gate ft , output
gate ot , and input-modulation gate gt controls information flow across the memory cell Ct . In this
way, the gradient will be prevented from vanishing quickly by being trapped in the memory.
The ConvLSTM network adopts the encoder-decoder RNN architecture that is proposed in [23] and
extended to video prediction in [21]. For a 4-layer ConvLSTM encoder-decoder network, input
frames are fed into the the first layer and future video sequence is generated at the fourth one. In this
process, spatial representations are encoded layer-by-layer, with hidden states being delivered from
bottom to top. However, the memory cells that belong to these four layers are mutually independent
and updated merely in time domain. Under these circumstances, the bottom layer would totally ignore
what had been memorized by the top layer at the previous time step. Overcoming these drawbacks of
this layer-independent memory mechanism is important to the predictive learning of video sequences.
3
PredRNN
In this section, we give detailed descriptions of the predictive recurrent neural network (PredRNN).
Initially, this architecture is enlightened by the idea that a predictive learning system should memorize
both spatial appearances and temporal variations in a unified memory pool. By doing this, we make
the memory states flow through the whole network along a zigzag direction. Then, we would like
to go a step further to see how to make the spatiotemporal memory interact with the original long
short-term memory. Thus we make explorations on the memory cell, memory gate and memory
fusion mechanisms inside LSTMs/ConvLSTMs. We finally derive a novel Spatiotemporal LSTM
(ST-LSTM) unit for PredRNN, which is able to deliver memory states both vertically and horizontally.
3
3.1
Spatiotemporal memory flow
X? t
X? t+1
l=4
l=4
M t?1
, H t?1
W4
W3
M tl=4 , H tl=4
W4
l=3
l=3
M t?1
, H t?1
X? t+2
l=2
l=2
M t?1
, H t?1
W2
M ,H
l=3
l=3
M t+1
, H t+1
W3
M tl=2 , H tl=2
W2
l=1
t?1
l=1
t?1
l=4
l=4
M t+1
, H t+1
W4
M tl=3 , H tl=3
W3
X? t+1
l=2
l=2
M t+1
, H t+1
W2
l=1
t
M ,H
l=1
t
l=1
t+1
M ,H
W1
W1
W1
Xt?1
Xt
Xt+1
l=1
t+1
l=4
l=4
Ct?1
, H t?1
l=3
l=3
Ct?1
, H t?1
l=2
l=2
Ct?1
, H t?1
l=1
l=1
Ct?1
, H t?1
W4
W3
W2
W1
Ctl=4 , H tl=4
H tl=3 l=3 l=3
Ct , H t
H tl=2 l=2 l=2
Ct , H t
H tl=1 l=1 l=1
Ct , H t
l=4
l=4
M t?2
, H t?2
Xt
Figure 1: Left: The convolutional LSTM network with a spatiotemporal memory flow. Right: The
conventional ConvLSTM architecture. The orange arrows denote the memory flow direction for all
memory cells.
For generating spatiotemporal predictions, PredRNN initially exploits convolutional LSTMs (ConvLSTM) [19] as basic building blocks. Stacked ConvLSTMs extract highly abstract features layer-bylayer and then make predictions by mapping them back to the pixel value space. In the conventional
ConvLSTM architecture, as illustrated in Figure 1 (right), the cell states are constrained inside each
ConvLSTM layer and be updated only horizontally. Information is conveyed upwards only by hidden
states. Such a temporal memory flow is reasonable in supervised learning, because according to the
study of the stacked convolutional layers, the hidden representations can be more and more abstract
and class-specific from the bottom layer upwards. However, we suppose in predictive learning,
detailed information in raw input sequence should be maintained. If we want to see into the future,
we need to learn from representations extracted at different-level convolutional layers. Thus, we
apply a unified spatiotemporal memory pool and alter RNN connections as illustrated in Figure 1
(left). The orange arrows denote the feed-forward directions of LSTM memory cells. In the left
figure, a unified memory is shared by all LSTMs which is updated along a zigzag direction. The key
equations of the convolutional LSTM unit with a spatiotemporal memory flow are shown as follows:
gt = tanh(Wxg ? Xt 1{l=1} + Whg ? Htl?1 + bg )
it = ?(Wxi ? Xt 1{l=1} + Whi ? Htl?1 + Wci ? Ml?1
+ bi )
t
ft = ?(Wxf ? Xt 1{l=1} + Whf ? Htl?1 + Wcf ? Ml?1
+ bf )
t
Mlt = ft Ml?1
+ it gt
t
(3)
ot = ?(Wxo ? Xt 1{l=1} + Who ? Htl?1 + Wco ? Mlt + bo )
Htl = ot tanh(Mlt ).
The input gate, input modulation gate, forget gate and output gate no longer depend on the hidden
states and cell states from the previous time step at the same layer. Instead, as illustrated in Figure
1 (left), they rely on hidden states Htl?1 and cell states Ml?1
t (l ? {1, ..., L}) that are updated by
the previous layer at current time step. Specifically, the bottom LSTM unit (l = 1) receives state
L
values from the top layer at the previous time step: Htl?1 = Ht?1
, Ml?1
= ML
t
t?1 . The four layers
in this figure have different sets of input-to-state and state-to-state convolutional parameters, while
they maintain a spatiotemporal memory cell and update its states separately and repeatedly as the
information flows through the current node. Note that, we replace the notation for memory cell from C
to M to emphasize that it flows in the zigzag direction in PredRNN, instead of the horizontal direction
in standard recurrent networks. Different from ConvLSTM that uses Hadamard product for state
transitions in the gates, we adopt convolution operators ? for finer-grained memory transitions.
4
3.2
Spatiotemporal LSTM
Input Gate
Output Gate
it
?
gt
Ctl
Input Modulation Gate
?
l
Ct?1
H
l
t?1
Xt
Forget Gate
ft
ot
Original Temporal
Memory
Spatiotemporal
Memory
M tl
?
X? t +1
W4
W4
M tl=3
?
H tl
W3
it?
Xt ?1
Ctl=2
W2
H tl=2
H tl=1
W1
l=4
M t?1
W4
W3
H tl=3
W2
W1
M tl?1
Ctl=3
H tl=2
M tl=1
ft?
Ctl=4 , H tl=4
W3
W2
X? t +2
H tl=3
M tl=2
gt?
?
X? t
Ctl=1
H tl=1
Xt
W1
M tl=4
Xt +1
Figure 2: ST-LSTM (left) and PredRNN (right). The orange circles in the ST-LSTM unit denotes the
differences compared with the conventional ConvLSTM. The orange arrows in PredRNN denote the
spatiotemporal memory flow, namely the transition path of spatiotemporal memory Mtl in the left.
In this section, we present the predictive recurrent neural network (PredRNN), by replacing convolutional LSTMs with a novel spatiotemporal long short-term memory (ST-LSTM) unit (see Figure 2).
In the architecture presented in the previous sub-section, the spatiotemporal memory cells are updated
in a zigzag direction, information is delivered first upwards across layers then forwards over time. We
wonder what would happen if memory cell states are passed in these two directions simultaneously.
With the aid of ST-LSTMs, the above PredRNN model with a spatiotemporal memory flow evolves
into our ultimate architecture. The equations of ST-LSTM are shown as follows:
l
gt = tanh(Wxg ? Xt + Whg ? Ht?1
+ bg )
l
it = ?(Wxi ? Xt + Whi ? Ht?1
+ bi )
l
ft = ?(Wxf ? Xt + Whf ? Ht?1
+ bf )
l
Ctl = ft Ct?1
+ it gt
0
gt0 = tanh(Wxg
? Xt + Wmg ? Ml?1
+ b0g )
t
0
i0t = ?(Wxi
? Xt + Wmi ? Ml?1
+ b0i )
t
(4)
0
ft0 = ?(Wxf
? Xt + Wmf ? Ml?1
+ b0f )
t
Mlt = ft0 Ml?1
+ i0t gt0
t
l
ot = ?(Wxo ? Xt + Who ? Ht?1
+ Wco ? Ctl + Wmo ? Mlt + bo )
Htl = ot tanh(W1?1 ? [Ctl , Mlt ]).
Two memory cells are maintained: Ctl is the standard temporal cell that is delivered from the previous
node at t ? 1 to the current time step within each LSTM unit. Mtl is the spatiotemporal memory we
described in the current section, which is conveyed vertically from the l ? 1 layer to the current node
at the same time step. For the bottom ST-LSTM layer where l = 1, Ml?1
= ML
t
t?1 , as described in
the previous subsection. We construct another set of gate structures for Mtl , while maintaining the
original gates for Ctl in standard LSTMs. At last, the final hidden states of this node rely on the fused
spatiotemporal memory. We concatenate these memory derived from different directions together and
then apply a 1 ? 1 convolution layer for dimension reduction, which makes the hidden state Htl of the
same dimensions as the memory cells. Different from simple memory concatenation, the ST-LSTM
unit uses a shared output gate for both memory types to enable seamless memory fusion, which can
effectively model the shape deformations and motion trajectories in the spatiotemporal sequences.
5
4
Experiments
Our model is demonstrated to achieve the state-of-the-art performance on three video prediction
datasets including both synthetic and natural video sequences. Our PredRNN model is optimized
with a L1 + L2 loss (other losses have been tried, but L1 + L2 loss works best). All models are
trained using the ADAM optimizer [11] with a starting learning rate of 10?3 . The training process is
stopped after 80, 000 iterations. Unless otherwise specified, the batch size of each iteration is set to 8.
All experiments are implemented in TensorFlow [1] and conducted on NVIDIA TITAN-X GPUs.
4.1
Moving MNIST dataset
Implementation We generate Moving MNIST sequences with the method described in [21]. Each
sequence consists of 20 consecutive frames, 10 for the input and 10 for the prediction. Each frame
contains two or three handwritten digits bouncing inside a 64 ? 64 grid of image. The digits were
chosen randomly from the MNIST training set and placed initially at random locations. For each
digit, we assign a velocity whose direction is randomly chosen by a uniform distribution on a unit
circle, and whose amplitude is chosen randomly in [3, 5). The digits bounce-off the edges of image
and occlude each other when reaching the same location. These properties make it hard for a model to
give accurate predictions without learning the inner dynamics of the movement. With digits generated
quickly on the fly, we are able to have infinite samples size in the training set. The test set is fixed,
consisting of 5,000 sequences. We sample digits from the MNIST test set, assuring the trained model
has never seen them before. Also, the model trained with two digits is tested on another Moving
MNIST dataset with three digits. Such a test setup is able to measure PredRNN?s generalization and
transfer ability, because no frames containing three digits are given throughout the training period.
As a strong competitor, we include the latest state-of-the-art VPN model [10]. We find it hard to
reproduce VPN?s experimental results on Moving MNIST since it is not open source, thus we adopt
its baseline version that uses CNNs instead of PixelCNNs as its decoder and generate each frame in
one pass. We observe that the total number of hidden states has a strong impact on the final accuracy
of PredRNN. After a number of trials, we present a 4-layer architecture with 128 hidden states in each
layer, which yields a high prediction accuracy using reasonable training time and memory footprint.
Table 1: Results of PredRNN with spatiotemporal memory M, PredRNN with ST-LSTMs, and
state-of-the-art models. We report per-frame MSE and Cross-Entropy (CE) of generated sequences
averaged across the Moving MNIST test sets. Lower MSE or CE denotes better prediction accuracy.
Model
FC-LSTM [21]
ConvLSTM Enc.-Dec. (128 ? 4) [19]
CDNA [6]
DFN [3]
VPN baseline [10]
PredRNN with spatiotemporal memory M
PredRNN + ST-LSTM (128 ? 4)
MNIST-2
(CE/frame)
MNIST-2
(MSE/frame)
MNIST-3
(MSE/frame)
483.2
367.0
346.6
285.2
110.1
118.5
97.0
118.3
103.3
97.4
89.0
70.0
74.0
56.8
162.4
142.1
138.2
130.5
125.2
118.2
93.4
Results As an ablation study, PredRNN only with a zigzag memory flow reduces the per-frame
MSE to 74.0 on the Moving MNIST-2 test set (see Table 1). By replacing convolutional LSTMs
with ST-LSTMs, we further decline the sequence MSE from 74.0 down to 56.8. The corresponding
frame-by-frame quantitative comparisons are presented in Figure 3. Compared with VPN, our model
turns out to be more accurate for long-term predictions, especially on Moving MNIST-3. We also
use per-frame cross-entropy likelihood as another evaluation metric on Moving MNIST-2. PredRNN
with ST-LSTMs significantly outperforms all previous methods, while PredRNN with spatiotemporal
memory M performs comparably with VPN baseline.
A qualitative comparison of predicted video sequences is given in Figure 4. Though VPN?s generated
frames look a bit sharper, its predictions gradually deviate from the correct trajectories, as illustrated
in the first example. Moreover, for those sequences that digits are overlapped and entangled, VPN has
difficulties in separating these digits clearly while maintaining their individual shapes. For example,
6
140
180
120
160
Mean Square Error
Mean Square Error
in the right figure, digit ?8? loses its left-side pixels and is predicted as ?3? after overlapping. Other
baseline models suffer from a severer blur effect, especially for longer future time steps. By contrast,
PredRNN?s results are not only sharp enough but also more accurate for long-term motion predictions.
100
80
60
ConvLSTM Enc.-Dec.
CDNA
DFN
VPN baseline
PredRNN + ST-LSTM
40
20
0
1
2
3
4
5 6 7
time steps
(a) MNIST-2
8
9
140
120
100
80
ConvLSTM Enc.-Dec.
CDNA
DFN
VPN baseline
PredRNN + ST-LSTM
60
40
20
1
10
2
3
4
5 6 7
time steps
(b) MNIST-3
8
9
10
Figure 3: Frame-wise MSE comparisons of different models on the Moving MNIST test sets.
Input frames
Ground truth
PredRNN
VPN
baseline
CDNA
ConvLSTM
Enc.-Dec.
Figure 4: Prediction examples on the Moving MNIST-2 test set.
4.2
KTH action dataset
Implementation The KTH action dataset [18] contains six types of human actions (walking,
jogging, running, boxing, hand waving and hand clapping) performed several times by 25 subjects in
four different scenarios: outdoors, outdoors with scale variations, outdoors with different clothes and
indoors. All video clips were taken over homogeneous backgrounds with a static camera in 25fps
frame rate and have a length of four seconds in average. To make the results comparable, we adopt
the experiment setup in [25] that video frames are resized into 128 ? 128 pixels and all videos are
divided with respect to the subjects into a training set (persons 1-16) and a test set (persons 17-25).
All models, including PredRNN as well as the baselines, are trained on the training set across all six
action categories by generating the subsequent 10 frames from the last 10 observations, while the the
presented prediction results in Figure 5 and Figure 6 are obtained on the test set by predicting 20 time
steps into the future. We sample sub-clips using a 20-frame-wide sliding window with a stride of 1 on
the training set. As for evaluation, we broaden the sliding window to 30-frame-wide and set the stride
to 3 for running and jogging, while 20 for the other categories. Sub-clips for running, jogging, and
walking are manually trimmed to ensure humans are always present in the frame sequences. In the
end, we split the database into a training set of 108,717 sequences and a test set of 4,086 sequences.
Results We use the Peak Signal to Noise Ratio (PSNR) and the Structural Similarity Index Measure (SSIM) [26] as metrics to evaluate the prediction results and provide frame-wise quantitative
comparisons in Figure 5. A higher value denotes a better prediction performance. The value of SSIM
ranges between -1 and 1, and a larger score means a greater similarity between two images. PredRNN
consistently outperforms the comparison models. Specifically, the Predictive Coding Network [12]
always exploits the whole ground truth sequence before the current time step to predict the next
7
frame. Thus, it cannot make sequence predictions. Here, we make it predict the next 20 frames by
feeding the 10 ground truth frames and the recursively generated frames in all previous time steps.
The performance of MCnet [25] deteriorates quickly for long-term predictions. Residual connections
of MCnet convey the CNN features of the last frame to the decoder and ignore the previous frames,
which emphasizes the spatial appearances while weakens temporal variations. By contrast, results of
PredRNN in both metrics remain stable over time, only with a slow and reasonable decline. Figure 6
visualizes a sample video sequence from the KTH test set. The ConvLSTM network [19] generates
blurred future frames, since it fails to memorize the detailed spatial representations. MCnet [25]
produces sharper images but is not able to forecast the movement trajectory accurately. Thanks
to ST-LSTMs, PredRNN memorizes detailed visual appearances as well as long-term motions. It
outperforms all baseline models and shows superior predicting power both spatially and temporally.
1
ConvLSTM Enc.-Dec.
MCnet + Res.
Predictive Coding Network
PredRNN + ST-LSTMs
30
Structural Similarity
Peak Signal Noise Ratio
35
25
ConvLSTM Enc.-Dec.
MCnet + Res.
Predictive Coding Network
PredRNN + ST-LSTMs
0.9
0.8
0.7
20
2
4
6
0.6
8 10 12 14 16 18 20
time steps
(a) Frame-wise PSNR
2
4
6
8 10 12 14 16 18 20
time steps
(b) Frame-wise SSIM
Figure 5: Frame-wise PSNR and SSIM comparisons of different models on the KTH action test set.
A higher score denotes a better prediction accuracy.
Input sequence
t=3
t=6
Ground truth and predictions
t=9
t=12
t=15
t=18
t=21
t=24
t=27
t=30
PredRNN
MCnet + Res.
ConvLSTM Enc.-Dec.
Predictive Coding Network
Figure 6: KTH prediction samples. We predict 20 frames into the future by observing 10 frames.
4.3
Radar echo dataset
Predicting the shape and movement of future radar echoes is a real application of predictive learning
and is the foundation of precipitation nowcasting. It is a more challenging task because radar echoes
are not rigid. Also, their speeds are not as fixed as moving digits, their trajectories are not as periodical
as KTH actions, and their shapes may accumulate, dissipate or change rapidly due to the complex
atmospheric environment. Modeling spatial deformation is significant for the prediction of this data.
8
Implementation We first collect the radar echo dataset by adapting the data handling method
described in [19]. Our dataset consists of 10,000 consecutive radar observations, recorded every 6
minutes in Guangzhou, China. For preprocessing, we first map the radar intensities to pixel values,
and represent them as 100 ? 100 gray-scale images. Then we slice the consecutive images with a
20-frame-wide sliding window. Thus, each sequence consists of 20 frames, 10 for the input, and 10
for forecasting. The total 9,600 sequences are split into a training set of 7,800 samples and a test
set of 1,800 samples. The PredRNN model consists of two ST-LSTM layers with 128 hidden states
each. The convolution filters inside ST-LSTMs are set to 3 ? 3. After prediction, we transform the
resulted echo intensities into colored radar maps, as shown in Figure 7, and then calculate the amount
of precipitation at each grid cell of these radar maps using Z-R relationships. Since it would bring in
an additional systematic error to rainfall prediction and makes final results misleading, we do not take
them into account in this paper, but only compare the predicted echo intensity with the ground truth.
Results Two baseline models are considered. The ConvLSTM network [19] is the first architecture
that models sequential radar maps with convolutional LSTMs, but its predictions tend to blur and
obviously inaccurate (see Figure 7). As a strong competitor, we also include the latest state-of-the-art
VPN model [10]. The PixelCNN-based VPN predicts an image pixel by pixel recursively, which
takes around 15 minutes to generate a radar map. Given that precipitation nowcasting has a high
demand on real-time computing, we trade off both prediction accuracy and computation efficiency
and adopt VPN?s baseline model that uses CNNs as its decoders and generates each frame in one pass.
Table 2 shows that the prediction error of PredRNN is significantly lower than VPN baseline. Though
VPN generates more accurate radar maps for the near future, it suffers from a rapid decay for the
long term. Such a phenomenon results from a lack of strong LSTM layers to model spatiotemporal
variations. Furthermore, PredRNN takes only 1/5 memory space and training time as VPN baseline.
Table 2: Quantitative results of different methods on the radar echo dataset.
Model
ConvLSTM Enc.-Dec. [19]
VPN baseline [10]
PredRNN
MSE/frame
Training time/100 batches
Memory usage
68.0
60.7
44.2
105 s
539 s
117 s
1756 MB
11513 MB
2367 MB
Input frames
Ground truth
PredRNN
ConvLSTM
Enc.-Dec.
VPN
baseline
Figure 7: A prediction example on the radar echo test set.
5
Conclusions
In this paper, we propose a novel end-to-end recurrent network named PredRNN for spatiotemporal
predictive learning that models spatial deformations and temporal variations simultaneously. Memory
states zigzag across stacked LSTM layers vertically and through all time states horizontally. Furthermore, we introduce a new spatiotemporal LSTM (ST-LSTM) unit with a gate-controlled dual
memory structure as the key building block of PredRNN. Our model achieves the state-of-the-art
performance on three video prediction datasets including both synthetic and natural video sequences.
9
Acknowledgments
This work was supported by the National Key R&D Program of China (2016YFB1000701), National
Natural Science Foundation of China (61772299, 61325008, 61502265, 61672313) and TNList Fund.
References
[1] M. Abadi, A. Agarwal, P. Barham, E. Brevdo, Z. Chen, C. Citro, G. S. Corrado, A. Davis, J. Dean,
M. Devin, et al. Tensorflow: Large-scale machine learning on heterogeneous distributed systems. arXiv
preprint arXiv:1603.04467, 2016.
[2] K. Cho, B. Van Merri?nboer, D. Bahdanau, and Y. Bengio. On the properties of neural machine translation:
Encoder-decoder approaches. arXiv preprint arXiv:1409.1259, 2014.
[3] B. De Brabandere, X. Jia, T. Tuytelaars, and L. Van Gool. Dynamic filter networks. In NIPS, 2016.
[4] E. L. Denton, S. Chintala, R. Fergus, et al. Deep generative image models using a laplacian pyramid of
adversarial networks. In NIPS, pages 1486?1494, 2015.
[5] J. Donahue, L. Anne Hendricks, S. Guadarrama, M. Rohrbach, S. Venugopalan, K. Saenko, and T. Darrell.
Long-term recurrent convolutional networks for visual recognition and description. In CVPR, pages
2625?2634, 2015.
[6] C. Finn, I. Goodfellow, and S. Levine. Unsupervised learning for physical interaction through video
prediction. In NIPS, 2016.
[7] I. J. Goodfellow, J. Pougetabadie, M. Mirza, B. Xu, D. Wardefarley, S. Ozair, A. Courville, and Y. Bengio.
Generative adversarial networks. NIPS, 3:2672?2680, 2014.
[8] A. Graves and N. Jaitly. Towards end-to-end speech recognition with recurrent neural networks. In ICML,
pages 1764?1772, 2014.
[9] S. Hochreiter and J. Schmidhuber. Long short-term memory. Neural computation, 9(8):1735?1780, 1997.
[10] N. Kalchbrenner, A. v. d. Oord, K. Simonyan, I. Danihelka, O. Vinyals, A. Graves, and K. Kavukcuoglu.
Video pixel networks. In ICML, 2017.
[11] D. Kingma and J. Ba. Adam: A method for stochastic optimization. In ICLR, 2015.
[12] W. Lotter, G. Kreiman, and D. Cox. Deep predictive coding networks for video prediction and unsupervised
learning. In International Conference on Learning Representations (ICLR), 2017.
[13] M. Mathieu, C. Couprie, and Y. LeCun. Deep multi-scale video prediction beyond mean square error. In
ICLR, 2016.
[14] J. Oh, X. Guo, H. Lee, R. L. Lewis, and S. Singh. Action-conditional video prediction using deep networks
in atari games. In NIPS, pages 2863?2871, 2015.
[15] V. Patraucean, A. Handa, and R. Cipolla. Spatio-temporal video autoencoder with differentiable memory.
In ICLR Workshop, 2016.
[16] M. Ranzato, A. Szlam, J. Bruna, M. Mathieu, R. Collobert, and S. Chopra. Video (language) modeling: a
baseline for generative models of natural videos. arXiv preprint arXiv:1412.6604, 2014.
[17] D. E. Rumelhart, G. E. Hinton, and R. J. Williams. Learning representations by back-propagating errors.
Cognitive modeling, 5(3):1, 1988.
[18] C. Schuldt, I. Laptev, and B. Caputo. Recognizing human actions: a local svm approach. In International
Conference on Pattern Recognition, pages 32?36 Vol.3, 2004.
[19] X. Shi, Z. Chen, H. Wang, D.-Y. Yeung, W.-K. Wong, and W.-c. Woo. Convolutional lstm network: A
machine learning approach for precipitation nowcasting. In NIPS, pages 802?810, 2015.
[20] K. Simonyan and A. Zisserman. Two-stream convolutional networks for action recognition in videos. In
NIPS, pages 568?576, 2014.
[21] N. Srivastava, E. Mansimov, and R. Salakhutdinov. Unsupervised learning of video representations using
lstms. In ICML, 2015.
[22] I. Sutskever, J. Martens, and G. E. Hinton. Generating text with recurrent neural networks. In ICML, pages
1017?1024, 2011.
[23] I. Sutskever, O. Vinyals, and Q. V. Le. Sequence to sequence learning with neural networks. NIPS,
4:3104?3112, 2014.
[24] A. van den Oord, N. Kalchbrenner, L. Espeholt, O. Vinyals, A. Graves, et al. Conditional image generation
with pixelcnn decoders. In NIPS, pages 4790?4798, 2016.
[25] R. Villegas, J. Yang, S. Hong, X. Lin, and H. Lee. Decomposing motion and content for natural video
sequence prediction. In International Conference on Learning Representations (ICLR), 2017.
[26] Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli. Image quality assessment: from error visibility
to structural similarity. TIP, 13(4):600, 2004.
[27] P. J. Werbos. Backpropagation through time: what it does and how to do it. Proceedings of the IEEE,
78(10):1550?1560, 1990.
[28] J. Yue-Hei Ng, M. Hausknecht, S. Vijayanarasimhan, O. Vinyals, R. Monga, and G. Toderici. Beyond
short snippets: Deep networks for video classification. In CVPR, pages 4694?4702, 2015.
10
| 6689 |@word trial:1 cnn:3 version:1 wmf:1 wco:3 cox:1 bf:3 open:1 tried:1 rgb:4 tnlist:1 recursively:3 reduction:1 contains:2 score:2 past:1 existing:1 outperforms:3 current:7 guadarrama:1 anne:1 activation:1 devin:1 concatenate:1 happen:1 blur:3 subsequent:1 shape:4 visibility:1 designed:1 update:1 fund:1 occlude:1 generative:4 vanishing:1 core:1 short:6 colored:1 coarse:1 node:4 location:3 along:2 constructed:1 become:1 fps:1 qualitative:1 prove:1 consists:4 abadi:1 inside:6 introduce:1 expected:1 rapid:1 growing:1 multi:1 inspired:2 salakhutdinov:1 toderici:1 window:3 totally:1 precipitation:5 notation:1 moreover:1 what:3 atari:2 unified:5 clothes:1 transformation:1 temporal:18 quantitative:3 every:2 finance:1 laterally:1 rainfall:1 mansimov:1 control:1 unit:14 szlam:1 dfn:3 danihelka:1 before:3 vertically:6 local:2 tsinghua:9 encoding:1 path:1 modulation:3 rnns:2 china:3 collect:1 challenging:2 limited:1 bi:3 range:1 averaged:1 acknowledgment:1 camera:1 lecun:1 lost:1 pixelcnns:2 block:2 backpropagation:1 digit:13 footprint:1 area:1 rnn:11 w4:7 significantly:2 adapting:1 integrating:1 get:1 cannot:1 operator:3 periodical:1 vijayanarasimhan:1 seminal:1 wong:1 conventional:4 map:8 marten:1 demonstrated:1 shi:2 dean:1 straightforward:1 economics:1 go:1 starting:1 latest:2 williams:1 sharpness:1 insight:1 oh:2 variation:8 merri:1 updated:6 imagine:1 suppose:2 assuring:1 homogeneous:1 us:4 goodfellow:2 jaitly:1 overlapped:1 velocity:1 rumelhart:1 recognition:8 walking:2 werbos:1 predicts:2 database:1 bottom:5 ft:10 levine:1 fly:1 preprint:3 wang:4 capture:2 calculate:1 region:1 ranzato:2 movement:4 transitional:1 trade:1 yfb1000701:1 environment:1 complexity:1 nowcasting:3 dynamic:5 radar:16 trained:4 depend:1 singh:1 smart:1 laptev:1 predictive:30 deliver:1 efficiency:2 gt0:2 easily:2 joint:1 represented:2 various:1 stacked:5 shortcoming:1 kalchbrenner:3 whose:2 modular:1 whi:3 solve:1 larger:2 encoded:1 cvpr:2 otherwise:1 encoder:3 ability:1 simonyan:2 uic:1 tuytelaars:1 transform:1 echo:10 delivered:3 final:3 obviously:2 sequence:37 differentiable:1 propose:1 interaction:1 product:2 mb:3 relevant:1 hadamard:2 enc:9 ablation:1 rapidly:1 achieve:1 description:2 exploiting:1 sutskever:2 cluster:1 darrell:1 captioning:1 generating:5 produce:2 adam:2 object:3 derive:1 recurrent:12 weakens:1 propagating:1 school:5 received:1 strong:5 implemented:1 predicted:6 memorize:5 met:1 direction:11 drawback:1 correct:1 cnns:3 subsequently:1 filter:2 exploration:1 human:3 stochastic:1 enable:1 memorized:1 villegas:2 espeholt:1 feeding:1 assign:1 generalization:1 preliminary:1 probable:1 whf:3 around:1 considered:1 ground:7 mapping:1 predict:8 achieves:3 consecutive:4 adopt:4 optimizer:1 tanh:7 ctl:11 successfully:1 hope:1 brought:1 wmi:1 clearly:1 always:3 aim:1 reaching:1 b0i:1 resized:1 surveillance:1 guangzhou:1 derived:1 focus:2 consistently:1 likelihood:1 contrast:5 adversarial:3 baseline:16 rigid:1 inaccurate:1 initially:3 hidden:12 reproduce:1 pixel:10 arg:1 classification:2 dual:1 jianmin:1 spatial:21 constrained:2 art:8 orange:4 construct:1 never:1 beach:1 ng:1 manually:1 yu:1 look:1 denton:1 unsupervised:3 icml:4 alter:1 future:15 report:1 mirza:1 randomly:3 simultaneously:5 preserve:1 resulted:1 individual:1 national:2 consisting:1 maintain:1 lotter:2 interest:1 highly:2 evaluation:2 zigzag:7 accurate:4 edge:1 hausknecht:1 unless:1 circle:2 re:3 deformation:4 stopped:1 column:1 modeling:5 cost:1 uniform:1 wonder:1 recognizing:1 conducted:1 spatiotemporal:40 synthetic:2 cho:1 st:23 person:2 lstm:39 peak:2 thanks:1 oord:2 international:3 standing:1 seamless:1 off:2 systematic:1 lee:2 pool:3 tip:1 together:1 quickly:3 fused:1 w1:8 recorded:2 management:1 containing:1 cognitive:1 account:1 de:1 stride:2 coding:6 blurred:1 vpn:19 satisfy:1 titan:1 explicitly:1 depends:1 dissipate:1 collobert:1 stream:2 bg:3 view:2 memorizes:2 h1:1 doing:1 performed:1 observing:1 waving:1 jia:1 square:3 accuracy:5 convolutional:15 who:3 yield:2 weak:1 raw:2 handwritten:1 kavukcuoglu:1 accurately:1 comparably:1 emphasizes:1 venugopalan:1 trajectory:5 monitoring:1 finer:1 visualizes:1 suffers:2 competitor:2 energy:1 involved:1 chintala:1 wxo:3 static:1 propagated:1 proved:1 dataset:8 recall:2 knowledge:1 subsection:1 psnr:3 amplitude:1 back:2 feed:1 higher:2 supervised:3 patraucean:2 mtl:3 zisserman:1 though:2 furthermore:2 just:1 correlation:1 schuldt:1 hand:2 receives:1 horizontal:1 lstms:20 replacing:2 overlapping:1 lack:1 meteorological:1 assessment:1 outdoors:3 gray:1 quality:1 mingsheng:3 usa:1 building:2 effect:1 usage:1 spatially:1 mlt:6 illustrated:4 game:2 maintained:2 htl:9 davis:1 hong:1 presenting:1 performs:1 motion:7 bring:2 upwards:3 l1:2 image:17 wise:5 handa:1 novel:5 recently:1 jimwang:1 sigmoid:1 superior:1 physical:1 belong:1 cdna:4 significant:3 measurement:6 accumulate:1 grid:6 language:2 had:1 bruna:1 moving:13 stable:1 pixelcnn:2 longer:3 similarity:4 etc:1 base:1 gt:10 something:1 recent:1 scenario:3 schmidhuber:1 certain:1 nvidia:1 seen:1 additional:2 greater:1 employed:1 period:1 corrado:1 signal:2 sliding:3 simoncelli:1 reduces:1 enlightened:2 faster:1 cross:2 long:14 lin:1 divided:1 equally:1 prevented:1 controlled:1 impact:2 prediction:45 laplacian:1 basic:1 heterogeneous:1 vision:1 circumstance:1 metric:3 arxiv:6 iteration:2 kernel:2 represent:1 yeung:1 pyramid:1 achieved:1 cell:19 c1:1 dec:9 background:1 want:1 fine:3 separately:1 agarwal:1 entangled:1 source:1 crucial:3 bovik:1 ot:9 w2:7 yue:1 subject:2 tend:2 bahdanau:1 flow:14 structural:3 near:1 chopra:1 yang:1 intermediate:1 split:2 enough:2 bengio:2 variate:1 w3:7 architecture:16 reduce:1 idea:3 cn:4 inner:1 decline:2 barham:1 bounce:1 whether:1 motivated:1 six:2 ultimate:1 passed:1 trimmed:1 forecasting:3 suffer:1 speech:2 convlstm:26 cause:1 action:13 repeatedly:2 deep:7 useful:1 detailed:4 indoors:1 amount:1 clip:4 visualized:1 category:2 generate:5 wci:2 happened:1 trapped:1 deteriorates:1 per:3 discrete:2 vol:1 key:6 four:4 ce:3 ht:11 merely:1 powerful:1 fourth:1 bouncing:1 named:1 throughout:1 reasonable:3 patch:1 comparable:1 bit:1 capturing:1 layer:31 ct:17 courville:1 kreiman:1 adapted:1 x2:1 software:5 generates:3 aspect:1 speed:1 nboer:1 optical:2 relatively:1 gpus:1 according:1 belonging:1 wxi:4 across:8 smaller:1 remain:1 sheikh:1 evolves:1 memorizing:1 den:1 gradually:1 taken:1 equation:3 mutually:2 hei:1 turn:1 mechanism:2 i0t:2 finn:2 fed:1 end:5 boxing:1 brevdo:1 decomposing:1 clapping:1 apply:2 observe:1 batch:2 gate:19 rp:2 slower:1 original:3 broaden:1 top:3 denotes:4 include:2 running:3 ensure:1 maintaining:2 exploit:2 especially:2 tensor:4 gradient:1 kth:6 iclr:5 separating:1 concatenation:1 philip:1 decoder:7 mail:2 ozair:1 besides:1 length:3 modeled:1 index:1 relationship:1 ratio:2 setup:2 sharper:2 ba:1 design:1 implementation:3 ssim:4 observation:8 convolution:6 datasets:4 snippet:1 extended:4 hinton:2 frame:51 sharp:1 community:1 intensity:3 overcoming:2 atmospheric:1 namely:1 specified:1 wxf:4 connection:2 optimized:1 wxg:4 tensorflow:2 established:1 kingma:1 nip:10 recollect:1 able:7 wardefarley:1 beyond:2 dynamical:1 hendricks:1 pattern:1 program:1 including:4 memory:59 video:36 max:1 gool:1 power:1 hochreiter:1 natural:5 rely:2 difficulty:1 predicting:4 residual:1 misleading:1 temporally:1 mathieu:3 picture:1 woo:1 extract:3 autoencoder:2 deviate:1 text:1 l2:2 graf:3 loss:4 generation:1 limitation:1 wmg:1 foundation:2 conveyed:2 leaning:1 translation:2 row:1 prone:1 summary:1 placed:1 last:4 supported:1 side:1 neighbor:1 wide:3 taking:1 distributed:1 slice:1 van:3 dimension:5 wcf:2 transition:5 concretely:1 author:1 adopts:1 forward:2 preprocessing:1 historical:3 emphasize:1 ignore:2 abstracted:1 ml:12 spatio:1 fergus:1 why:1 table:4 learn:2 channel:2 xbt:2 ca:1 transfer:1 caputo:1 interact:2 mse:8 complex:2 ft0:2 monga:1 domain:2 arrow:3 whole:3 noise:2 allowed:1 jogging:3 convey:2 x1:2 xu:1 tl:24 slow:1 aid:1 sub:3 fails:1 zhifeng:1 grained:3 donahue:1 down:1 minute:2 specific:2 xt:31 brabandere:1 decay:1 svm:1 fusion:2 workshop:1 mnist:17 mcnet:6 effectively:1 sequential:1 conditioned:1 demand:1 forecast:1 chen:2 entropy:2 forget:3 fc:1 appearance:9 explore:1 gao:1 rohrbach:1 visual:9 horizontally:5 vinyals:4 bo:3 cipolla:1 srivastava:2 truth:7 environmental:1 determines:1 extracted:1 loses:1 lewis:1 conditional:3 viewed:1 towards:1 couprie:1 shared:2 replace:1 content:1 hard:3 change:1 specifically:2 infinite:1 admittedly:1 called:1 total:2 pas:2 experimental:1 saenko:1 citro:1 guo:1 evaluate:1 tested:1 phenomenon:1 handling:1 |
6,289 | 669 | Hidden Markov Model Induction by Bayesian
Model Merging
Andreas Stolcke*'**
*Computer Science Division
University of California
Berkeley, CA 94720
[email protected]
Stephen Omohundro"
**International Computer Science Institute
1947 Center Street, Suite 600
Berkeley, CA 94704
[email protected]
Abstract
This paper describes a technique for learning both the number of states and the
topology of Hidden Markov Models from examples. The induction process starts
with the most specific model consistent with the training data and generalizes
by successively merging states. Both the choice of states to merge and the
stopping criterion are guided by the Bayesian posterior probability. We compare
our algorithm with the Baum-Welch method of estimating fixed-size models, and
find that it can induce minimal HMMs from data in cases where fixed estimation
does not converge or requires redundant parameters to converge.
1 INTRODUCTION AND OVERVIEW
Hidden Markov Models (HMMs) are a well-studied approach to the modelling of sequence
data. HMMs can be viewed as a stochastic generalization of finite-state automata, where
both the transitions between states and the generation of output symbols are governed by
probability distributions. HMMs have been important in speech recognition (Rabiner &
Juang, 1986), cryptography, and more recently in other areas such as protein classification
and alignment (Haussler, Krogh, Mian & SjOlander, 1992; Baldi, Chauvin, Hunkapiller &
McClure, 1993).
Practitioners have typically chosen the HMM topology by hand, so that learning the HMM
from sample data means estimating only a fixed number of model parameters. The standard
approach is to find a maximum likelihood (ML) or maximum a posteriori probability (MAP)
estimate of the HMM parameters. The Baum-Welch algorithm uses dynamic programming
11
12
Stokke and Omohundro
to approximate these estimates (Baum, Petrie, Soules & Weiss, 1970).
A more general problem is to additionally find the best HMM topology. This includes
both the number of states and the connectivity (the non-zero transitions and emissions).
One could exhaustively search the model space using the Baum-Welch algorithm on fully
connected models of varying sizes, picking the model size and topology with the highest
posterior probability. (Maximum likelihood estimation is not useful for this comparison
since larger models usually fit the data better.) This approach is very costly and BaumWelch may get stuck at sub-optimal local maxima. Our comparative results later in the
paper show that this often occurs in practice. The problem can be somewhat alleviated by
sampling from several initial conditions, but at a further increase in computational cost.
The HMM induction method proposed in this paper tackles the structure learning problem
in an incremental way. Rather than estimating a fixed-size model from scratch for various
sizes, the model size is adjusted as new evidence arrives. There are two opposing tendencies
in adjusting the model size and structure. Initially new data adds to the model size, because
the HMM has to be augmented to accommodate the new samples. If enough data of a similar
structure is available, however, the algorithm collapses the shared structure, decreasing the
model size. The merging of structure is also what drives generalization, Le., creates HMMs
that generate data not seen during training.
Beyond being incremental, our algorithm is data-driven, in that the samples themselves
completely determine the initial model shape. Baum-Welch estimation, by comparison,
uses an initially random set of parameters for a given-sized HMM and iteratively updates
them until a point is found at which the sample likelihood is locally maximal. What seems
intuitively troublesome with this approach is that the initial model is completely uninformed
by the data. The sample data directs the model formation process only in an indirect manner
as the model approaches a meaningful shape.
2 HIDDEN MARKOV MODELS
For lack of space we cannot give a full introduction to HMMs here; see Rabiner & Juang
(1986) for details. Briefly, an HMM consists of states and transitions like a Markov chain.
In the discrete version considered here, it generates strings by performing random walks
between an initial and a final state, outputting symbols at every state in between. The
probability P(xlM) that a model M generates a string x is determined by the conditional
probabilities of making a transition from one state to another and the probability of emitting
each symbol from each state. Once these are given, the probability of a particular path
through the model generating the string can be computed as the product of all transition
and emission probabilities along the path. The probability of a string x is the sum of the
probabilities of all paths generating x.
For example, the model M3 in Figure 1 generates the strings ab, abab, ababab, ... with
2 2 2
. I
pro b ab I?l??
lUes 3'
3!' 3!' ... , respective
y.
3 HMM INDUCTION BY STATE MERGING
3.1
MODEL MERGING
Omohundro (1992) has proposed an approach to statistical model inference in which initial
Hidden Markov Model Induction by Bayesian Model Merging
models simply replicate the data and generalize by similarity. As more data is received,
component models are fit from more complex model spaces. This allows the formation of
arbitrarily complex models without overfitting along the way. The elementary step used in
modifying the overall model is a merging of sub-models, collapsing the sample sets for the
corresponding sample regions. The search for sub-models to merge is guided by an attempt
to sacrifice as little of the sample likelihood as possible as a result of the merging process.
This search can be done very efficiently if (a) a greedy search strategy can be used, and (b)
likelihood computations can be done locally for each sub-model and don't require global
recomputation on each model update.
3.2 STATE MERGING IN HMMS
We have applied this general approach to the HMM learning task. We describe the algorithm
here mostly by presenting an example. The details are available in Stolcke & Omohundro
(1993).
To obtain an initial model from the data, we first construct an HMM which produces
exactly the input strings. The start state has as many outgoing transitions as there are
strings and each string is represented by a unique path with one state per sample symbol.
The probability of entering these paths from the start state is uniformly distributed. Within
each path there is a unique transition arc whose probability is 1. The emission probabilities
are 1 for each state to produce the corresponding symbol.
As an example, consider the regular language (abt and two samples drawn from it, the
strings ab and abab. The algorithm constructs the initial model Mo depicted in Figure 1.
This is the most specific model accounting for the observed data. It assigns each sample
a probability equal to its relative frequency, and is therefore a maximum likelihood model
for the data.
Learning from the sample data means generalizing from it. This implies trading off
model likelihood against some sort of bias towards 'simpler' models, expressed by a prior
probability distribution over HMMs. Bayesian analysis provides a formal basis for this
tradeoff. Bayes' rule tells us that the posterior model probability P(Mlx) is proportional
to the product of the model prior P(M) and the likelihood of the data P(xlM). Smaller or
simpler models will have a higher prior and this can outweigh the drop in likelihood as long
as the generalization is conservative and keeps the model close to the data. The choice of
model priors is discussed in the next section.
The fundamental idea exploited here is that the initial model Mo can be gradually transformed into the generating model by repeatedly merging states. The intuition for this
heuristic comes from the fact that if we take the paths that generate the samples in an actual
generating HMM M and 'unroll' them to make them completely disjoint, we obtain Mo.
The iterative merging process, then, is an attempt to undo the unrolling, tracing a search
through the model space back to the generating model.
Merging two states q] and q2 in this context means replacing q] and q2 by a new state r with
a transition distribution that is a weighted mixture of the transition probabilities of q], q2,
and with a similar mixture distribution for the emissions. Transition probabilities into q] or
q2 are added up and redirected to r. The weights used in forming the mixture distributions
are the relative frequencies with which q] and q2 are visited in the current model.
Repeatedly performing such merging operations yields a sequence of models Mo, M J ,
13
14
Stokke and Omohundro
Mo:
a
b
log L(xIMo) =
- 1. 39
log L(xIMJ) = log L(xIMo)
a
b
Figure I: Sequence of models obtained by merging samples {ab, abab}. All transitions without
special annotations have probability 1; Output symbols appear above their respective states and also
carry an implicit probability of 1. For each model the log likelihood is given.
M2 ? ..., along which we can search for the MAP model. To make the search for M efficient,
we use a greedy strategy: given Mi. choose a pair of states for merging that maximizes
P(Mi+llX)?
Continuing with the previous example, we find that states 1 and 3 in Mo can be merged
without penalizing the likelihood. This is because they have identical outputs and the loss
due to merging the outgoing transitions is compensated by the merging of the incoming
transitions. The .5/.5 split is simply transferred to outgoing transitions of the merged state.
The same situation obtains for states 2 and 4 once 1 and 3 are merged. From these two
first merges we get model M. in Figure 1. By convention we reuse the smaller of two state
indices to denote the merged state.
At this point the best merge turns out to be between states 2 and 6, giving model M2.
However, there is a penalty in likelihood, which decreases to about .59 of its previous
value. Under all the reasonable priors we considered (see below), the posterior model
probability still increases due to an increase in the prior. Note that the transition probability
ratio at state 2 is now 2/1, since two samples make use of the first transition, whereas only
one takes the second transition.
Finally, states 1 and 5 can be merged without penalty to give M3, the minimal model that
generates (ab)+. Further merging at this point would reduce the likelihood by three orders
of magnitude. The resulting decrease in the posterior probability tells the algorithm to stop
Hidden Markov Model Induction by Bayesian Model Merging
at this point.
3.3 MODEL PRIORS
As noted previously, the likelihoods P(XIMj ) along the sequence of models considered by
the algorithm is monotonically decreasing. The prior P(M) must account for an overall
increase in posterior probability, and is therefore the driving force behind generalization.
As in the work on Bayesian learning of classification trees by Buntine (1992), we can split
the prior P(M) into a term accounting for the model structure, P(Ms), and a term for the
adjustable parameters in a fixed structure P(MpIMs).
We initially relied on the structural prior only, incorporating an explicit bias towards smaller
models. Size here is some function of the number of states and/or transitions, IMI. Such a
prior can be obtained by making P(Ms ) <X e- 1M1 , and can be viewed as a description length
prior that penalizes models according to their coding length (Rissanen, 1983; Wallace &
Freeman, 1987). The constants in this "MOL" term had to be adjusted by hand from
examples of 'desirable' generalization.
For the parameter prior P(MpIMs), it is standard practice to apply some sort of smoothing
or regularizing prior to avoid overfitting the model parameters. Since both the transition
and the emission probabilities are given by multinomial distributions it is natural to use a
Dirichlet conjugate prior in this case (Berger, 1985). The effect of this prior is equivalent
to having a number of 'virtual' samples for each of the possible transitions and emissions
which are added to the actual samples when it comes to estimating the most likely parameter
settings. In our case, the virtual samples made equal use of all potential transitions and
emissions, adding bias towards uniform transition and emission probabilities.
We found that the Dirichlet priors by themselves produce an implicit bias towards smaller
models, a phenomenon that can be explained as follows. The prior alone results in a model
with uniform, flat distributions. Adding actual samples has the effect of putting bumps into
the posterior distributions, so as to fit the data. The more samples are available, the more
peaked the posteriors will get around the maximum likelihood estimates of the parameters,
increasing the MAP value. In estimating HMM parameters, what counts is not the total
number of samples, but the number of samples per state, since transition and emission
distributions are local to each state. As we merge states, the available evidence gets shared
by fewer states, thus allowing the remaining states to produce a better fit to the data.
This phenomenon is similar, but not identical, to the Bayesian 'Occam factors' that prefer
models with fewer parameter (MacKay, 1992). Occam factors are a result of integrating
the posterior over the parameter space, something which we do not do because of the
computational complications it introduces in HMMs (see below).
3.4 APPROXIMATIONS
At each iteration step, our algorithm evaluates the posterior resulting from every possible
merge in the current HMM. To keep this procedure feasible, a number of approximations
are incorporated in the implementation that don't seem to affect its qualitative properties.
? For the purpose of likelihood computation, we consider only the most likely path
through the model for a given sample string (the Viterbi path). This allows us to
15
16
Stokke and Omohundro
express the likelihood in product form, computable from sufficient statistics for each
transition and emission.
? We assume the Viterbi paths are preserved by the merging operation, that is, the
paths previously passing through the merged states now go through the resulting new
state. This allows us to update the sufficient statistics incrementally, and means only
O(number of states) likelihood terms need to be recomputed.
? The posterior probability of the model structure is approximated by the posterior of the
MAP estimates for the model parameters. Rigorously integrating over all parameter
values is not feasible since varying even a single parameter could change the paths of
all samples through the HMM.
? Finally, it has to be kept in mind that our search procedure along the sequence of
merged models finds only local optima, since we stop as soon as the posterior starts to
decrease. A full search of the space would be much more costly. However, we found
a best-first look-ahead strategy to be sufficient in rare cases where a local maximum
caused a problem. In those cases we continue merging along the best-first path for a
fixed number of steps (typically one) to check whether the posterior has undergone
just a temporary decrease.
4 EXPERIMENTS
We have used various artificial finite-state languages to test our algorithm and compare its
performance to the standard Baum-Welch algorithm.
Table 1 summarizes the results on the two sample languages ac? a u bc? band a+ b+a+ b+.
The first of these contains a contingency between initial and final symbols that can be hard
for learning algorithms to uncover.
We used no explicit model size prior in our experiments after we found that the Dirichlet
prior was very robust in giving just the the right amount of bias toward smaller models.!
Summarizing the results, we found that merging very reliably found the generating model
structure from a very small number of samples. The parameter values are determined by
the sample set statistics.
The Baum-Welch algorithm, much like a backpropagation network, may be sensitive to its
random initial parameter settings. We therefore sampled from a number of initial conditions.
Interestingly, we found that Baum-Welch has a good chance of settling into a suboptimal
HMM structure, especially if the number of states is the minimal number required for the
target language. It proved much easier to estimate correct language models when extra
states were provided. Also, increasing the sample size helped it converge to the target
model.
5 RELATED WORK
Our approach is related to several other approaches in the literature.
The concept of state merging is implicit in the notion of state equivalence classes, which is
fundamental to much of automata theory (Hopcroft & Ullman, 1979) and has been applied
I The
number of 'virtual' samples per transition/emission was held constant at 0.1 throughout.
Hidden Markov Model Induction by Bayesian Model Merging
(a)
(b)
Method
Merging
Merging
Baum-Welch
(10 trials)
Baum-Welch
(l0 trials)
Baum-Welch
Baum-Welch
Sample
8 m.p.
20 random
8 m.p.
Method
Merging
Baum-Welch
(3 trials)
Merging
Baum-We1ch
(3 trials)
Sample
5m.p.
5m.p.
20 random
8 m.p.
20 random
10 random
10 random
Entropy
2.295
2.087
2.087
2.773
2.087
2.775
2.384
2.085
Cross-entropy
2.188 ? .020
2.158 ? .033
2.894 ? .023 (best)
4.291 ? .228 (worst)
2.105 ? .031 (best)
2.825 ? .031 (worst)
3.914 ? .271
2.155 ? .032
Entropy
2.163
3.545
3.287
5.009
5.009
6.109
Cross-entropy
7.678 ? .158
8.963 ? .161 (best)
59.663 ? .007 (worst)
5.623 ? .074
5.688 ? .076 (best)
8.395 ? .137 (worst)
Language
ac'a v bc'b
ac'a v bc'b
(a v b)c'(a v b)
(a v b)c'(a v b)
ac'a v bc'b
(a v b)c'(a v b)
ac'a v bc'b
ac'a v bc'b
Language
a+b+a+b+
(a+b+t
(a+b+t
a+b+a+b+
a+b+a+b+
(a+b+t
n
6
6
6
6
6
6
10
10
n
4
4
4
4
4
4
Table 1: Results for merging and Baum-Welch on two regular languages: (a) ac'a v bc'b and (b)
a+b+a+b+. Samples were either the top most probable (m.p.) ones from the target language, or a
set of randomly generated ones. 'Entropy' is the average negative log probability on the training
set, whereas 'cross-entropy' refers to the empirical cross-entropy between the induced model and
the generating model (the lower, the better generalization). n denotes the final number of model
states for merging, or the fixed model size for Baum-Welch. For Baum-Welch, both best and worst
performance over several initial conditions is listed.
to automata learning as well (Angluin & Smith, 1983).
Tomita (1982) is an example of finite-state model space search guided by a (nonprobabilistic) goodness measure.
Horning (1969) describes a Bayesian grammar induction procedure that searches the model
space exhaustively for the MAP model. The procedure provably finds the globally optimal
grammar in finite time, but is infeasible in practice because of its enumerative character.
The incremental augmentation of the HMM by merging in new samples has some of the
flavor of the algorithm used by Porat & Feldman (1991) to induce a finite-state model from
positive-only, ordered examples.
Haussler et al. (1992) use limited HMM 'surgery' (insertions and deletions in a linear
HMM) to adjust the model size to the data, while keeping the topology unchanged.
6 FURTHER RESEARCH
We are investigating several real-world applications for our method. One task is the
construction of unified multiple-pronunciation word models for speech recognition. This
is currently being carried out in collaboration with Chuck Wooters at ICSI, and it appears
that our merging algorithm is able to produce linguistically adequate phonetic models.
Another direction involves an extension of the model space to stochastic context-free
grammars, for which a standard estimation method analogous to Baum-Welch exists (Lari
17
18
Stokke and Omohundro
& Young, 1990). The notions of sample incorporation and merging carry over to this domain
(with merging now involving the non-terminals of the CFO), but need to be complemented
with a mechanism that adds new non-terminals to create hierarchical structure (which we
call chunking).
Acknowledgements
We would like to thank Peter Cheeseman, Wray Buntine, David Stoutamire, and Jerry
Feldman for helpful discussions of the issues in this paper.
References
Angluin, D. & Smith, C. H. (1983), 'Inductive inference: Theory and methods', ACM Computing
Surveys 15(3), 237-269.
Baldi, P., Chauvin, Y., Hunkapiller, T. & McClure, M. A. (1993), 'Hidden Markov Models in
molecular biology: New algorithms and applications' , this volume.
Baum, L. E., Petrie, T., Soules, G. & Weiss, N. (1970), 'A maximization technique occuring in the
statistical analysis of probabilistic functions in Markov chains', The Annals of Mathematical
Statistics 41( 1), 164-17l.
Berger, J. O. (1985), Statistical Decision Theory and Bayesian Analysis, Springer Verlag, New York.
Buntine, W. (1992), Learning classification trees, in D. J. Hand, ed., 'Artificial Intelligence Frontiers
in Statistics: AI and Statistics III' , Chapman & Hall.
Haussler, D., Krogh, A., Mian, 1. S. & Sjolander, K. (1992), Protein modeling using hidden Markov
models: Analysis of globins, Technical Report UCSC-CRL-92-23, Computer and Information
Sciences, University of California, Santa Cruz, Ca. Revised Sept. 1992.
Hopcroft, J. E. & Ullman, J. D. (1979), Introduction to Automata Theory, Languages, and Computation, Addison-Wesley, Reading, Mass.
Horning, J. J. (1969), A study of grammatical inference, Technical Report CS 139, Computer Science
Department, Stanford University, Stanford, Ca.
Lari, K. & Young, S. J. (1990), 'The estimation of stochastic context-free grammars using the
Inside-Outside algorithm' , Computer Speech and Language 4, 35-56.
MacKay, D. J. C. (1992), 'Bayesian interpolation', Neural Computation 4,415-447.
Omohundro, S. M. (1992), Best-first model merging for dynamic learning and recognition, Technical
Report TR-92-004, International Computer Science Institute, Berkeley, Ca.
Porat, S. & Feldman, 1. A. (1991), 'Learning automata from ordered examples', Machine Learning
7, 109-138.
Rabiner, L. R. & Juang, B. H. (1986), 'An introduction to Hidden Markov Models', IEEE ASSP
Magazine 3(1), 4-16.
Rissanen, J. (1983), 'A universal prior for integers and estimation by minimum description length',
The Annals of Statistics 11(2), 416-431 .
Stolcke, A. & Omohundro, S. (1993), Best-first model merging for Hidden Markov Model induction,
Technical Report TR-93-003, International Computer Science Institute, Berkeley, Ca.
Tomita, M. (1982), Dynamic construction of finite automata from examples using hill-climbing, in
'Proceedings of the 4th Annual Conference of the Cognitive Science Society', Ann Arbor,
Mich., pp. 105-108.
Wallace, C. S. & Freeman, P. R. (1987), 'Estimation and inference by compact coding', Journal of
the Royal Statistical Society, Series B 49(3),240-265 .
| 669 |@word trial:4 version:1 briefly:1 seems:1 replicate:1 accounting:2 tr:2 accommodate:1 carry:2 initial:12 contains:1 series:1 bc:7 interestingly:1 soules:2 current:2 must:1 cruz:1 shape:2 drop:1 cfo:1 update:3 alone:1 greedy:2 fewer:2 intelligence:1 smith:2 provides:1 complication:1 simpler:2 mathematical:1 along:6 ucsc:1 redirected:1 qualitative:1 consists:1 baldi:2 inside:1 manner:1 sacrifice:1 themselves:2 wallace:2 terminal:2 freeman:2 globally:1 decreasing:2 horning:2 little:1 actual:3 unrolling:1 increasing:2 provided:1 estimating:5 maximizes:1 mass:1 what:3 string:10 q2:5 unified:1 suite:1 berkeley:6 every:2 tackle:1 exactly:1 appear:1 positive:1 local:4 troublesome:1 path:13 interpolation:1 merge:5 studied:1 equivalence:1 hmms:9 collapse:1 limited:1 unique:2 practice:3 backpropagation:1 procedure:4 area:1 universal:1 empirical:1 alleviated:1 word:1 induce:2 regular:2 integrating:2 refers:1 protein:2 get:4 cannot:1 close:1 context:3 outweigh:1 map:5 equivalent:1 center:1 baum:19 compensated:1 go:1 automaton:6 survey:1 welch:16 assigns:1 m2:2 rule:1 haussler:3 notion:2 analogous:1 ababab:1 target:3 construction:2 annals:2 magazine:1 programming:1 us:2 recognition:3 approximated:1 observed:1 worst:5 region:1 connected:1 decrease:4 highest:1 icsi:3 intuition:1 insertion:1 rigorously:1 dynamic:3 exhaustively:2 creates:1 division:1 completely:3 basis:1 hopcroft:2 indirect:1 various:2 represented:1 describe:1 artificial:2 tell:2 formation:2 outside:1 pronunciation:1 whose:1 heuristic:1 larger:1 stanford:2 grammar:4 statistic:7 final:3 sequence:5 outputting:1 maximal:1 product:3 description:2 juang:3 optimum:1 produce:5 comparative:1 incremental:3 generating:7 ac:7 uninformed:1 received:1 krogh:2 c:1 involves:1 implies:1 trading:1 come:2 convention:1 direction:1 guided:3 merged:7 correct:1 modifying:1 stochastic:3 virtual:3 require:1 generalization:6 probable:1 elementary:1 adjusted:2 extension:1 frontier:1 around:1 considered:3 hall:1 viterbi:2 mo:6 bump:1 driving:1 purpose:1 estimation:7 linguistically:1 currently:1 visited:1 sensitive:1 create:1 weighted:1 rather:1 avoid:1 varying:2 l0:1 emission:11 directs:1 modelling:1 likelihood:18 check:1 summarizing:1 posteriori:1 inference:4 helpful:1 stopping:1 typically:2 initially:3 hidden:11 transformed:1 provably:1 sjolander:2 classification:3 overall:2 issue:1 smoothing:1 special:1 mackay:2 equal:2 once:2 construct:2 having:1 sampling:1 chapman:1 identical:2 biology:1 look:1 peaked:1 report:4 wooters:1 abt:1 randomly:1 opposing:1 ab:5 attempt:2 adjust:1 alignment:1 introduces:1 mixture:3 arrives:1 behind:1 held:1 chain:2 respective:2 tree:2 continuing:1 walk:1 penalizes:1 minimal:3 modeling:1 goodness:1 maximization:1 cost:1 rare:1 uniform:2 imi:1 buntine:3 international:3 fundamental:2 probabilistic:1 off:1 picking:1 connectivity:1 recomputation:1 augmentation:1 successively:1 choose:1 collapsing:1 cognitive:1 ullman:2 account:1 potential:1 coding:2 includes:1 caused:1 later:1 helped:1 start:4 sort:2 bayes:1 relied:1 annotation:1 om:1 efficiently:1 rabiner:3 yield:1 climbing:1 generalize:1 bayesian:11 wray:1 drive:1 ed:1 against:1 evaluates:1 frequency:2 pp:1 mi:2 stop:2 sampled:1 proved:1 adjusting:1 uncover:1 back:1 appears:1 wesley:1 higher:1 wei:2 done:2 just:2 implicit:3 mian:2 until:1 hand:3 replacing:1 lack:1 incrementally:1 effect:2 concept:1 unroll:1 inductive:1 jerry:1 entering:1 iteratively:1 during:1 noted:1 criterion:1 m:2 presenting:1 hill:1 omohundro:9 occuring:1 pro:1 recently:1 petrie:2 multinomial:1 overview:1 volume:1 discussed:1 m1:1 feldman:3 llx:1 ai:1 lues:1 language:11 had:1 similarity:1 add:2 something:1 posterior:14 driven:1 phonetic:1 verlag:1 arbitrarily:1 continue:1 chuck:1 exploited:1 seen:1 minimum:1 somewhat:1 converge:3 determine:1 redundant:1 monotonically:1 stephen:1 full:2 desirable:1 multiple:1 technical:4 cross:4 mcclure:2 long:1 molecular:1 involving:1 iteration:1 globin:1 preserved:1 whereas:2 extra:1 undo:1 induced:1 seem:1 practitioner:1 call:1 structural:1 stoutamire:1 integer:1 split:2 enough:1 stolcke:4 iii:1 affect:1 fit:4 topology:5 suboptimal:1 andreas:1 idea:1 reduce:1 tradeoff:1 computable:1 whether:1 reuse:1 penalty:2 peter:1 speech:3 passing:1 york:1 repeatedly:2 adequate:1 useful:1 santa:1 listed:1 amount:1 locally:2 band:1 generate:2 angluin:2 disjoint:1 per:3 discrete:1 express:1 putting:1 recomputed:1 rissanen:2 drawn:1 penalizing:1 kept:1 sum:1 baumwelch:1 throughout:1 reasonable:1 decision:1 prefer:1 summarizes:1 annual:1 ahead:1 incorporation:1 flat:1 generates:4 performing:2 transferred:1 department:1 according:1 conjugate:1 describes:2 smaller:5 character:1 making:2 intuitively:1 gradually:1 explained:1 chunking:1 lari:2 previously:2 turn:1 count:1 mechanism:1 mind:1 addison:1 generalizes:1 available:4 operation:2 apply:1 hierarchical:1 top:1 dirichlet:3 remaining:1 denotes:1 tomita:2 giving:2 especially:1 society:2 unchanged:1 surgery:1 added:2 occurs:1 strategy:3 costly:2 abab:3 thank:1 street:1 hmm:19 enumerative:1 chauvin:2 induction:9 toward:1 length:3 index:1 berger:2 ratio:1 mostly:1 negative:1 implementation:1 reliably:1 adjustable:1 allowing:1 revised:1 markov:13 arc:1 finite:6 situation:1 incorporated:1 assp:1 david:1 pair:1 required:1 california:2 merges:1 deletion:1 temporary:1 beyond:1 able:1 usually:1 below:2 reading:1 royal:1 natural:1 force:1 settling:1 cheeseman:1 carried:1 sept:1 prior:21 literature:1 acknowledgement:1 relative:2 mlx:1 fully:1 loss:1 nonprobabilistic:1 generation:1 proportional:1 contingency:1 sufficient:3 consistent:1 undergone:1 occam:2 collaboration:1 soon:1 keeping:1 infeasible:1 free:2 bias:5 formal:1 institute:3 stokke:4 tracing:1 distributed:1 grammatical:1 transition:25 world:1 stuck:1 made:1 emitting:1 approximate:1 obtains:1 compact:1 keep:2 ml:1 global:1 overfitting:2 incoming:1 investigating:1 don:2 search:11 iterative:1 mich:1 porat:2 table:2 additionally:1 scratch:1 robust:1 ca:6 hunkapiller:2 mol:1 complex:2 domain:1 cryptography:1 augmented:1 sub:4 explicit:2 governed:1 young:2 specific:2 symbol:7 evidence:2 incorporating:1 exists:1 merging:36 adding:2 magnitude:1 easier:1 flavor:1 entropy:7 depicted:1 generalizing:1 simply:2 likely:2 forming:1 expressed:1 ordered:2 springer:1 chance:1 complemented:1 acm:1 conditional:1 viewed:2 sized:1 ann:1 towards:4 shared:2 xlm:2 feasible:2 change:1 hard:1 crl:1 determined:2 uniformly:1 conservative:1 total:1 tendency:1 arbor:1 m3:2 meaningful:1 outgoing:3 regularizing:1 phenomenon:2 |
6,290 | 6,690 | Detrended Partial Cross Correlation
for Brain Connectivity Analysis
Jaime S Ide?
Yale University
New Haven, CT 06519
[email protected]
Fabio A Cappabianco
Federal University of Sao Paulo
S.J. dos Campos, 12231, Brazil
[email protected]
Fabio A Faria
Federal University of Sao Paulo
S.J. dos Campos, 12231, Brazil
[email protected]
Chiang-shan R Li
Yale University
New Haven, CT
chiang-shan.li-yale.edu
Abstract
Brain connectivity analysis is a critical component of ongoing human connectome
projects to decipher the healthy and diseased brain. Recent work has highlighted
the power-law (multi-time scale) properties of brain signals; however, there remains
a lack of methods to specifically quantify short- vs. long- time range brain connections. In this paper, using detrended partial cross-correlation analysis (DPCCA),
we propose a novel functional connectivity measure to delineate brain interactions
at multiple time scales, while controlling for covariates. We use a rich simulated
fMRI dataset to validate the proposed method, and apply it to a real fMRI dataset
in a cocaine dependence prediction task. We show that, compared to extant methods, the DPCCA-based approach not only distinguishes short and long memory
functional connectivity but also improves feature extraction and enhances classification accuracy. Together, this paper contributes broadly to new computational
methodologies in understanding neural information processing.
1
Introduction
Brain connectivity is crucial to understanding the healthy and diseased brain states [15, 1]. In recent
years, investigators have pursued the construction of human connectomes and made large datasets
available in the public domain [23, 24]. Functional Magnetic Resonance Imaging (fMRI) has been
widely used to examine complex processes of perception and cognition. In particular, functional
connectivity derived from fMRI signals has proven to be effective in delineating biomarkers for many
neuropsychiatric conditions [15].
One of the challenges encountered in functional connectivity analysis is the precise definition of
nodes and edges of connected brain regions [21]. Functional nodes can be defined based on activation
maps or with the use of functional or anatomical atlases. Once nodes are defined, the next step is to
estimate the weights associated with the edges. Traditionally, these functional connectivity weights
are measured using correlation-based metrics. Previous simulation studies have shown that they can
be quite successful, outperforming higher-order statistics (e.g. linear non-gaussian acyclic causal
models) and lag-based approaches (e.g. Granger causality) [20].
On the other hand, very few studies have investigated the power-law cross-correlation properties
(equivalent to multi-time scale measures) of brain connectivity. Recent research suggested that fMRI
?
Corresponding author: Department of Psychiatry, 34 Park St. S110. New Haven CT 06519.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
signals have power-law properties (e.g. their power-spectrum follows a power law) [8, 3] and that the
deviations from the typical range of power-exponents have been noted in neuropsychiatric disorders
[11]. For instance, in [3], using wavelet-based multivariate methods, authors observed that scale-free
properties are characteristic not only of univariate fMRI signals but also of pairwise cross-temporal
dynamics. Moreover, they found an association between the magnitude of scale-free dynamics and
task performance. We hypothesize that power-law correlation measures may capture additional
dimensions of brain connectivity not available from conventional analyses and thus enhance clinical
prediction.
In this paper, we aim to answer three key open questions: (i) whether and how brain networks are
cross-correlated at different time scales with long-range dependencies (?long-memory? process,
equivalent to power-law in the frequency domain); (ii) how to extract the intrinsic association between
two regions controlling for the influence of other interconnected regions; and (iii) whether multi-time
scale connectivity measures can improve clinical prediction. We address the first two questions
by using the detrended partial cross-correlation analsyis (DPCCA) coefficient [25], a measure that
quantifies correlations on multiple time scales between non-stationary time series, as is typically the
case with task-related fMRI signals. DPCCA is an extension of detrended cross-correlation analysis
[17, 13], and has been successfully applied to analyses of complex systems, including climatological
[26] and financial [18] data. Unlike methods based on filtering particular frequency bands, DPCCA
directly informs correlations across multiple time scales, and unlike wavelet-based approaches (e.g.
cross wavelet transformation and wavelet transform coherence [2]), DPCCA has the advantage of
estimating pairwise correlations controlling for the influence of other regions. This is critical because
brain regions and thus fMRI signals thereof are highly interconnected. To answer the third question,
we use the correlation profiles, generated from DPCCA, as input features for different machine
learning methods in classification tasks and compare the performance of DPCCA-based features with
all other competing features.
In Section 2, we describe the simulated and real data sets used in this study, and show how features
of the classification task are extracted from the fMRI signals. In Section 3, we provide further details
about DPCCA (Section 3.1), and present the proposed multi-time scale functional connectivity measure (Section 3.2). In Section 4, we describe core experiments designed to validate the effectiveness
of DPCCA in brain connectivity analysis and clinical prediction. We demonstrate that DPCCA
(i) detects connectivity at multiple-time scales while controlling for covariates (Sections 4.1 and
4.3), (ii) accurately identifies functional connectivity in well-known gold-standard simulated data
(Section 4.2), and (iii) improves classification accuracy of cocaine dependence with fMRI data of
seventy-five cocaine dependent and eighty-eight healthy control individuals (Section 4.4). In Section
5, we conclude by highlighting the significance of the study as well as the limitations and future
work.
2
2.1
Material and Methods
Simulated dataset: NetSim fMRI data
We use fMRI simulation data - NetSim [20] - previously developed for the evaluation of network
modeling methods. Simulating rich and realistic fMRI time series, NetSim is comprised of twentyeight different brain networks, with different levels of complexity. These signals are generated
using dynamic causal modeling (DCM [6]), a generative network model aimed to quantify neuronal
interactions and neurovascular dynamics, as measured by the fMRI signals. NetSim graphs have 5
to 50 nodes organized with ?small-world? topology, in order to reflect real brain networks. NetSim
signals have 200 time points (mostly) sampled with repetition time (TR) of 3 seconds. For each
network, 50 separate realizations (?subjects?) are generated. Thus, we have a total of 1400 synthetic
dataset for testing. Finally, once the signals are generated, white noise of standard deviation 0.1-1%
is added to reproduce the scan thermal noise.
2.2
Real-world dataset: Cocaine dependence prediction
Seventy-five cocaine dependent (CD) and eighty-eight healthy control (HC) individuals matched in
age and gender participated in this study. CD were recruited from the local, greater New Haven area
in a prospective study and met criteria for current cocaine dependence, as diagnosed by the Structured
Clinical Interview for DSM-IV. They were drug-free while staying in an inpatient treatment unit.
2
The Human Investigation committee at Yale University School of Medicine approved the study, and
all subjects signed an informed consent prior to participation. In the MR scanner, they performed a
simple cognitive control paradigm called stop-signal task [14]. FMRI data were collected with 3T
Siemens Trio scanner. Each scan comprised four 10-min runs of the stop signal task. Functional
blood oxygenation level dependent (BOLD) signals were acquired with a single-shot gradient echo
echo-planar imaging (EPI) sequence, with 32 axial slices parallel to the AC-PC line covering the
whole brain: TR=2000 ms, TE=25 ms, bandwidth=2004 Hz/pixel, flip angle=85? , FOV=220?220
mm2 , matrix=66?64, slice thickness=4 mm and no gap. A high-resolution 3D structural image
(MPRAGE; 1 mm resolution) was also obtained for anatomical co-registration. Three hundred
images were acquired in each session. Functional MRI data was pre-processed with standard pipeline
using Statistical Parametric Mapping 12 (SPM12) (Wellcome Department of Imaging Neuroscience,
University College London, U.K.).
2.2.1
Brain activation
We constructed general linear models and localized brain regions responding to conflict (stop signal)
anticipation (encoded by the probability P(stop)) at the group level [10]. The regions responding to
P(stop) comprised the bilateral parietal cortex, the inferior frontal gyrus (IFG) and the right middle
frontal gyrus (MFG); and regions responding to motor slowing bilateral insula, the left precentral
cortex (L.PC), and the supplementary motor area (SMA) (Fig. 1(a))2 . These regions of interest (ROIs)
were used as masks to extract average activation time courses for functional connectivity analyses.
2.2.2
Functional connectivity
We analyzed the frontoparietal circuit involved in conflict anticipation and response adjustment using
a standard Pearson correlation analysis and multivariate Granger causality analysis or mGCA [19]. In
Fig. 1(b), we illustrate fifteen correlation coefficients derived from the six ROIs for each individual
CD and HC as shown in Fig. 1(a). According to mGCA, connectivities from bilateral parietal to L.PC
and SMA were disrupted in CD (Fig. 1(b)). These findings offer circuit-level evidence of altered
cognitive control in cocaine addiction.
Figure 1: Disrupted frontoparietal circuit
in cocaine addicts. The frontoparietal
circuit included six regions responding
to Bayesian conflict anticipation (?S?)
and regions of motor slowing (?RT?): (a)
CD and HC shared connections (orange
arrows). (b) Connectivity strengths between nodes in the frontoparietal circuit.
We show connectivity strengths between
nodes for each individual subject in CD
(red line) and HC (blue line) groups.
(a)
3
(b)
A Novel Measure of Brain Functional Connectivity
3.1
Detrended partial cross-correlation analysis (DPCCA)
Detrended partial cross-correlation is a novel measure recently proposed by [25]. DPCCA combines
the advantages of detrended cross-correlation analysis (DCCA) [17] and standard partial correlation.
Given two time series {x(a) }, {x(b) } ? Xt , where Xt ? IRm , t = 1, 2, ..., N time points, DPCCA is
given by Equation 1:
?DP CCA (a, b; s) = p
2
?Ca,b (s)
,
Ca,a (s).Cb,b (s)
(1)
Peak MNI coordinates for IFG:[39,53,-1], MFG:[42,23,38],bilateral insula:[-33,17,8] and [30,20,2], L.PC:[36,-13,56], and SMA:[-9,-1,50] in mm.
3
where s is the time scale and each term Ca,b (s) is obtained by inverting the matrix ?(s), e.g.
C(s) =??1 (s). The coefficient ?a,b ? ?(s) is the so called DCCA coefficient [13]. The DCCA
coefficient is an extension of the detrented cross correlation analysis [17] combined with detrended
fluctuation analysis (DFA) [12].
Given two time series {x}, {y} ? Xt (indices omitted for the sake of simplicity) with N time
points and time scale s, DCCA coefficient is given by Equation 2:
2
FDCCA
(s)
?(s) =
FDF A,x (s)FDF A,y (s)
,
(2)
where the numerator and denominator are the average of detrended covariances and variances of the
N ? s + 1 windows (partial sums), respectively, as described in Equations 3-4:
PN ?s+1
2
FDCCA
(s)
j=1
=
(3)
N ?s
PN ?s+1
2
FDF
A,x (s)
2
fDCCA
(s, j)
2
fDF
A,x (s, j)
j=1
=
N ?s
.
(4)
The partial sums (profiles) are obtained with sliding windows across the integrated time series
Pt
Pt
Xt = i=1 xi and Yt = i=1 yi . For each time window j with size s, detrended covariances
and variances are computed according to Equations 5-6:
Pj+s?1
2
fDCCA
(s, j)
=
t=j
d
[
(Xt ? X
t,j )(Yt ? Yt,j )
s?1
Pj+s?1
2
fDF
A,x (s, j)
=
t=j
2
[
(Xt ? X
t,j )
s?1
,
,
(5)
(6)
d
[
where X
t,j and Yt,j are polynomial fits of time trends. We used a linear fit as originally proposed
[13], but higher order fits could also be used [25]. DCCA can be used to measure power-law
cross-correlations. However, we focus on DCCA coefficient as a robust measure to detect pairwise
cross-correlation in multiple time scales, while controlling for covariates. Importantly, DPCCA
quantifies correlations among time series with varying levels of non-stationarity [13].
3.2
DPCCA for functional connectivity analysis
In this section, we propose the use of DPCCA as a novel measure of brain functional connectivity.
First, we show in simulation experiments that the measure satisfies desired connectivity properties.
Further, we define the proposed connectivity measure. Although these properties are expected by
mathematical definition of DPCCA, it is critical to confirm its validity on real fMRI data. Additionally,
it is necessary to establish the statistical significance of the computed measures at the group level.
3.2.1
Desired properties
Given real fMRI signals, the measure should accurately detect the time scale in which the pairwise
connections occur, while controlling for the covariates. To verify this, we create synthetic data by
combining real fMRI signals and sinusoidal waves (Fig. 2). To simplify, we assume additive property
of signals and sinusoidal waves reflecting the time onset of the connections. For each simulation, we
randomly sample 100 sets of time series or ?subjects?.
a) Distinction of short and long memory connections. Given two fMRI signals {xA }, {xB }, we
derive three pairs with known connectivity profiles: short-memory {XA = xA + sin(T1 ) + e},
{XB = xB + sin(T1 ) + e}, long-memory {XA = xA + sin(T2 ) + e}, {XB = xB +
sin(T2 ) + e} and mixed {XA = xA + sin(T1 ) + sin(T2 ) + e}, {XB = xB + sin(T1 ) +
sin(T2 ) + e}, where T1 << T2 and e is a Gaussian signal to simulate measurement noise. We
hypothesize that the two nodes A and B are functionally connected at time scales T1 and T2 .
4
b) Control for covariates. Given three fMRI signals {xA }, {xB }, {xC }, we derive three signals
with known connectivity {XAC = xA +xC +sin(T )+e}, {XBC = xB +xC +sin(T )+e},
{XC = xC + e}, where e is the measurement noise. We hypothesize that the two nodes A and B
are functionally connected mostly at scale T, once the mutual influence of node C is controlled.
Figure 2: Illustration of synthetic fMRI signals generated by combining real fMRI signals and sinusoidal waves. (a) Original fMRI
signals, (b) original signals with sin(T =
10s) and sin(T = 30s) waves added.
(a)
3.2.2
(b)
Statistical significance
Given two nodes and their time series, we assume that they are functionally connected if the
max |?DP CCA |, within a time range srange , is significantly greater than the null distribution.
Empirical null distributions are estimated from the original data by randomly shuffling time series
across different subjects and nodes, as proposed in [20]. In this way, we generate realistic distributions
of connectivity weights occurring by chance. Since we have a multivariate measure, the null dataset
is always generated with the same number of nodes as the tested network. Multiple comparisons are
controlled by estimating the false discovery rate. Importantly, the null distribution is also computed on
max |?DP CCA | within the time range srange . We use a srange from 6 to 18 seconds, assuming
that functional connections transpire in this range. Thus, we allow connections with different
time-scales. We use this binary definition of functional connectivity for the current approach to be
comparable with other methods, but it is also possible to work with the whole temporal profile of
?DP CCA (s), as is done in the classification experiment (Section 4.4). To keep the same statistical
criteria, we also generate null distributions for all the other connectivity measures.
3.2.3
DPCCA + Canonical correlation analysis
As further demonstrated by simulation results (Table 1), DPCCA alone has lower true positive rate
(TPR) compared to other competing methods, likely because of its restrictive statistical thresholds. In
order to increase the sensitivity of DPCCA, we augmented the method by including an additional
canonical correlation analysis (CCA) [7]. CCA was previously used in fMRI in different contexts
to detect brain activations [5], functional connectivity [27], and for multimodal information fusion
[4]. In short, given two sets of multivariate time series {XA (t) ? IRm , t = 1, 2, ..., N } and
{XB (t) ? IRn , t = 1, 2, ..., N }, where m and n are the respective dimensions of the two sets
A and B, and N is the number of time points, CCA seeks the linear transformations u and v so that
the correlation between the linear combinations XA (t)u and XB (t)v is maximized. In this work,
we propose the use of CCA to define the existence of a true connection, in addition to the DPCCA
connectivity results. The proposed method is summarized in Algorithm 1. With CCA (Lines 8-14),
we identify the nodes that are strongly connected after linear transformations. In Line 18, we use
CCA to inform DPCCA in terms of positive connections.
4
4.1
Experiments and Results
Connectivity properties: Controlling time scales and covariates
In Figure 3, we observe that DPCCA successfully captured the time scales of the correlations
between time series {XA }, {XB }, despite the noisy nature of fMRI signals. For instance, it
distinguished between short and long-memory connections, represented using T1 = 10s and
T2 = 30s, respectively (Figs. 3a-c). Importantly, it clearly detected the peak connection at 10s after
controlling for the influence of covariate signal XC (Fig. 3f). Further, unlike DPCCA, the original
DCCA method did not rule out the mutual influence of XC with peak at 30s (Fig. 3e).
5
Algorithm 1 DPCCA+CCA
Input: Time series {Xt ? IRm , t = 1, 2, ..., N }, where m is the number of vectors and N is
the number of time points; time range srange with k values
Output: Connectivity matrix F C : [m ? m] and associated matrices
1: Step: DPCCA(Xt )
. Compute pairwise DPCCA
2:
for pair of vectors {x(a) }, {x(b) } ? Xt do
3:
for s in srange do
4:
Compute the coefficient ?DP CCA (a, b; s)
. Equation(1)
5:
F C[a, b] ? max |?DP CCA | in srange
6:
P [a, b] ? statistical significance of F C[a, b] given the null empirical distribution
7:
return F C and P
. Matrix of connection weights and p-values
8: Step: CCA(Xt )
. Compute CCA connectivity
9:
for x(a) ? Xt do
10:
for x(b) ? Xt , b 6= a do
11:
rCCA [a, b] ? (1? CCA between {x(a) }, {x(c) }, c 6= a, b)
. Effect of
excluding node b
12:
indexcon ? k-means(rCCA [a])
. Split connections into binary groups
13:
CCA[a, indexcon ] ? 1
14:
return CCA
. CCA is a binary connectivity matrix
15: Step: DPCCA+CCA(P,CCA)
. Augment DPCCA with CCA results
16:
for pair of nodes {a, b} do
17:
F C ? [a, b] ? 1, if P [a, b] < 0.05
. DPCCA significant connections
. Fill missing connections
18:
F C ? [a, b] ? max(F C ? [a, b], CCA[a, b])
19:
return F C ? , F C and P
. F C ? is a binary matrix
Figure 3: DPCCA temporal profiles
among the synthetic signals (details
in Section 3.2.1). (a)-(c): DPCCA
with peak at T=10s and T=30s, and
mixed. (d) DPCCA of the original fMRI signals used to generate
the synthetics signals. (e) Temporal
profile obtained with DCCA without partial correlation. (f) DPCCA
peak at T=10s after controlling for
XC . Dashed lines are the 95% confidence interval of DPCCA for the
empirical null distribution.
4.2
Simulated networks: Improved connectivity accuracy
The goal of this experiment is to validate the proposed methods in an extensive dataset designed
to test functional connectivity methods. In this dataset, ground truth networks are known with the
architectures aimed to reflect real brain networks. We use the full NetSim dataset comprised of
28 different brain circuits and 50 subjects. For each sample of time series, we compute the partial
correlation (parCorr) and the regularized inverse covariance (ICOV), reported as the best performers
in [20], as well as the proposed DPCCA and DPCCA+CCA methods. For each measure, we construct
empirical null distributions, as described in Section 3.2.2, and generate the binary connectivity matrix
using threshold ? = 0.05. To evaluate their connectivity accuracy, given the ground truth networks,
we compute the true positive and negative rates (TPR and TNR, respectively) and the balanced
N R)
accuracy BAcc= (T P R+T
.
2
Using NetSim fMRI data as the testing benchmark, we observed that the proposed DPCCA+CCA
method provided more accurate functional connectivity results than the best methods reported in the
original paper [20]. Results are summarized in Table 1. Here we use the balanced accuracy (BAcc)
6
as the evaluation metric, since it is a straightforward way to quantify both true positive and negative
connections.
Table 1: Comparison of functional connectivity methods using NetSim dataset. Mean and standard
deviation of balanced accuracy (BAcc), true positive rate (TPR) and true negative rate (TNR) are
reported. ParCorr: partial correlation, ICOV: regularized inverse covariance, DPCCA: detrended
cross correlation analysis, DPCCA+CCA: DPCCA augmented with CCA. DPCCA+CCA balanced
accuracy is significantly higher than the best competing method ICOV (Wilcoxon signed paired test,
Z=3.35 and p=8.1e-04).
Metrics
Mean
Std
4.3
ParCorr
BAcc
TPR TNR
0.834 0.866 0.804
0.096 0.129 0.188
Functional connectivity measures
ICOV
DPCCA
BAcc TPR
TNR
BAcc TPR
TNR
0.841 0.866 0.817 0.846 0.835 0.855
0.095 0.131 0.181 0.095 0.150 0.177
DPCCA+CCA
BAcc TPR
TNR
0.859 0.893 0.824
0.091 0.081 0.169
Real-world dataset: Learning connectivity temporal profiles
We use unsupervised methods to (i) learn representative temporal profiles of connectivity from
DPCCAF ull , and (ii) perform dimensionality reduction. The use of temporal profiles may capture
additional information (such as short- and long-memory connectivity). However, it increases the
feature set dimensionality, imposing additional challenges on classifier training, particularly with
small dataset. The first natural choice for this task is principal component analysis (PCA), which can
represent original features by their linear combination. Additionally, we use two popular non-linear
dimensionality reduction methods Isomap [22] and autoencoders [9]. With Isomap, we attempt to
learn the intrinsic geometry (manifold) of the temporal profile data. With autoencoders, we seek to
represent the data using restricted Boltzmann machines stacked into layers.
In Figure 4, we show some representative correlation profiles obtained by computing DPPCA
among frontoparietal regions (circuit presented in Fig. 1), and the first three principal components.
Interestingly, PCA seemed to learn some of the characteristic temporal profiles. For instance, as
expected, the first components captured the main trend, while the second components captured some
of the short (task-related) and long (resting-state) memory connectivity trends (Figs.4a-b).
Figure 4: Illustration of some
DPCCA profiles and their principal components. IFG: inferior frontal gyrus, SMA: supplementary motor area, PC:
premotor cortex. Explained
variances of the components
are also reported.
4.4
Real-world dataset: Cocaine dependence prediction
The classification task consists of predicting the class membership, cocaine dependence (CD) and
healthy control (HC), given each individual?s fMRI data. After initial preprocessing (Section 2.2), we
extract average time series within the frontoparietal circuit of 6 regions 3 (Figure 1), and compute
the different cross-correlation measures. These coefficients are used as features to train and test
(leave-one-out cross-validation) a set of popular classifiers available in scikit-learn toolbox [16]
(version 0.18.1), including k-nearest neighbors (kNN), support vector machine (SVM), multilayer
perceptron (MLP), Gaussian processes (GP), naive Bayes (NB) and the ensemble method Adaboost
(Ada). For the DPCCA coefficients, we test both peak values DPCCAmax as well as the rich
temporal profiles DPCCAF ull . Finally, we also include the brain activation maps (Section 2.2.1) as
feature set, thus allowing comparison with popular fMRI classification softwares such as PRONTO
(http://www.mlnl.cs.ucl.ac.uk/pronto/). Features are summarized in Table 2.
3
Although these regions are obtained from the whole-group, no class information is used to avoid inflated
classification rates.
7
Table 2: Features used in the cocaine dependence classification task.
Type
Name
P(stop)
UPE
Corr
ParCorr
ICOV
DPCCAmax
DPCCAF ull
DPCCAIso
DPCCAAutoE
DPCCAP CA
Activation
Connectivity
Size
1042
1042
15
15
15
15
270
135-180
30-45
135-180
Description
Brain regions responding to anticipation of stop signals
Brain regions responding to unsigned prediction error of P(stop)
Pearson cross-correlation among the six frontoparietal regions
Partial cross-correlation among the six frontoparietal regions
Regularized inverse covariance among the six frontoparietal regions
Maximum DPCCA within the range 6-40 seconds
Temporal profile of DPCCA within the range 6-40 seconds
Isomap with 9-12 components and 30 neighbors
Autoencoders with 2-3 hidden layers, 5-20 neurons, batch=100, epoch=1000
PCA with 9-12 components
Classification results are summarized in Table 3 and Figure 5. We used the area under curve (AUC)
as an evaluation metric in order to consider both sensitivity and specificity of the classifiers, as well as
balanced accuracy (BAcc). Here we tested all features described in Table 2, including the DPCCA full
profiles after dimensionality reduction (Isomap, autoencoders and PCA). Activation maps produced
poor classification results (P(stop): 0.525?0.048 and UPE: 0.509?0.032), comparable to the results
obtained with PRONTO software using the same features (accuracy 0.556).
Features
Corr
ParCorr
ICOV
DPCCAmax
DPCCAF ull
DPCCAIso
DPCCAAutoE
DPCCAP CA
Mean AUC
(? std)
0.757 (? 0.041)
0.901 (? 0.034)
0.900 (? 0.030)
0.906 (? 0.019)
0.899 (? 0.028)
0.902 (? 0.030)
0.815 (? 0.149)
0.928 (? 0.035)
Mean BAcc
(? std)
0.674 (? 0.037)
0.848 (? 0.025)
0.838 (? 0.023)
0.831 (? 0.022)
0.820 (? 0.052)
0.827 (? 0.068)
0.813 (? 0.106)
0.844 (? 0.064)
Top classifier
(AUC / BAcc)
GP / NB
GP / Ada
GP / SVM
GP / Ada
GP / GP
GP / MLP
SVM / kNN5
Ada / NB
Accuracy
(AUC / BAcc)
0.794 / 0.710
0.948 / 0.875
0.948 / 0.858
0.929 / 0.857
0.957 / 0.874
0.954 / 0.894
0.939 / 0.863
0.963 / 0.911
Table 3: Comparison of classification results for different
features. The DPCCA features combined with PCA produced
the top classifiers according to both criteria (0.963/0.911).
However, DPCCAP CA is not statistically better than ParCorr
or ICOV (Wilcoxon signed paired test, p>0.05). See Figure 5
for accuracy across different classification methods.
5
Figure 5: Comparison of classification results for different features and methods (described in
Section 4.4).
Conclusions
In summary, as a multi-time scale approach to characterize brain connectivity, the proposed method
(DPCCA+CCA) (i) identified connectivity peak-times (Fig. 3), (ii) produced higher connectivity
accuracy than the best competing method ICOV (Table 1), and (iii) distinguished short/long memory
connections between brain regions involved in cognitive control (IFC&SMA and SMA&PC) (Fig.
4). Second, using the connectivity weights as features, DPCCA measures combined with PCA
produced the highest individual accuracies (Table 3). However, it was not statistically different
from the second best feature (ParCorr) across different classifiers. Further separate test set would be
necessary to identify the best classifiers. We performed extensive experiments with a large simulated
fMRI dataset to validate DPCCA as a promising functional connectivity analytic. On the other
hand, our conclusions on clinical prediction (classification task) are still limited to one case. Finally,
further optimization of Isomap and autoencoders methods could improve the learning of connectivity
temporal profiles produced by DPCCA.
Acknowledgments
Supported by FAPESP (2016/21591-5), CNPq (408919/2016-7), NSF (BCS1309260) and NIH
(AA021449, DA023248).
References
[1] DS Bassett and ET Bullmore. Human Brain Networks in Health and Disease. Current opinion in neurology,
22(4):340?347, 2009.
8
[2] C Chang and GH Glover. Time?Frequency Dynamics of Resting-State Brain Connectivity Measured with
fMRI . NeuroImage, 50(1):81 ? 98, 2010.
[3] P Ciuciu, P Abry, and BJ He. Interplay between Functional Connectivity and Scale-Free Dynamics in
Intrinsic fMRI Networks. Neuroimage, 95:248?63, 2014.
[4] NM Correa et al. Fusion of fMRI, sMRI, and EEG Data using Canonical Correlation Analysis. In 2009
IEEE International Conference on Acoustics, Speech and Signal Processing, pages 385?388, April 2009.
[5] O Friman et al. Detection of neural activity in functional MRI using Canonical Correlation Analysis.
Magnetic Resonance in Medicine, 45(2):323?330, February 2001.
[6] KJ Friston, L Harrison, and W Penny. Dynamic Causal Modelling. NeuroImage, 19(4):1273, 2003.
[7] DR Hardoon, SR Szedmak, and JR Shawe-Taylor. Canonical Correlation Analysis: An Overview with
Application to Learning Methods. Neural Comput., 16(12):2639?2664, December 2004.
[8] B. J. He. Scale-free brain activity: past, present, and future. Trends Cogn Sci, 18(9):480?7, 2014.
[9] GE Hinton and RR Salakhutdinov. Reducing the Dimensionality of Data with Neural Networks. Science,
313(5786):504?507, 2006.
[10] S Hu, JS Ide, S Zhang, and CR Li. Anticipating conflict: Neural correlates of a bayesian belief and its
motor consequence. NeuroImage, 119:286?295, 2015. doi: 10.1016/j.neuroimage.2015.06.032.
[11] JS Ide, S Hu, S Zhang, LR Mujica-Parodi, and CR Li. Power spectrum scale invariance as a neural marker
of cocaine misuse and altered cognitive control. NeuroImage: Clinical, 11:349 ? 356, 2016.
[12] JW Kantelhardt et al. Multifractal Detrended Fluctuation Analysis of Nonstationary Time Series. Physica
A: Statistical Mechanics and its Applications, 316(1?4):87 ? 114, 2002.
[13] L Kristoufek. Measuring Correlations between Non-Stationary Series with DCCA Coefficient. Physica
a-Statistical Mechanics and Its Applications, 402:291?298, 2014.
[14] GD Logan, WB Cowan, and KA Davis. On the Ability to Inhibit Simple and Choice Reaction Time
Responses: a Model and a Method. J Exp Psychol Hum Percept Perform, 10(2):276?91, 1984.
[15] PM Matthews and A Hampshire. Clinical Concepts Emerging from fMRI Functional Connectomics.
Neuron, 91(3):511 ? 528, 2016.
[16] F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer,
R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay.
Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825?2830, 2011.
[17] B Podobnik and HE Stanley. Detrended Cross-Correlation Analysis: a New Method for Analyzing two
Nonstationary Time Series. Phys Rev Lett, 100(8):084102, 2008.
[18] XY Qian et al. Detrended Partial Cross-Correlation Analysis of two Nonstationary Time Series Influenced
by Common External Forces. Phys Rev E Stat Nonlin Soft Matter Phys, 91(6):062816, 2015.
[19] AK. Seth, AB. Barrett, and L Barnett. Granger Causality Analysis in Neuroscience and Neuroimaging.
Journal of Neuroscience, 35(8):3293?3297, 2015.
[20] SM Smith et al. Network Modelling Methods for FMRI. Neuroimage, 54(2):875?91, 2011.
[21] O Sporns. The Human Connectome: Origins and Challenges. NeuroImage, 80:53 ? 61, 2013. Mapping
the Connectome.
[22] JB Tenenbaum, V de Silva, and JC Langford. A Global Geometric Framework for Nonlinear Dimensionality
Reduction. Science, 290(5500):2319?2323, 2000.
[23] DC Van Essen et al. The WU-Minn Human Connectome Project: An overview. NeuroImage, 80:62 ? 79,
2013. Mapping the Connectome.
[24] M Xia and Y He. Functional Connectomics from a ?Big Data? Perspective. NeuroImage, pages ?, 2017.
[25] N Yuan et al. Detrended Partial-Cross-Correlation Analysis: a New Method for Analyzing Correlations in
Complex System. Sci Rep, 5:8143, 2015.
[26] N Yuan et al. A Novel Way to Detect Correlations on Multi-Time Scales, with Temporal Evolution and for
Multi-Variables. Sci Rep, 6:27707, 2016.
[27] D Zhou, WK Thompson, and G Siegle. MATLAB Toolbox for Functional Connectivity. NeuroImage, 47
(4):1590 ? 1607, 2009.
9
| 6690 |@word middle:1 version:1 mri:2 polynomial:1 approved:1 open:1 hu:2 simulation:5 seek:2 covariance:5 fifteen:1 tr:2 shot:1 reduction:4 initial:1 series:18 interestingly:1 dubourg:1 past:1 reaction:1 current:3 ka:1 activation:7 connectomics:2 realistic:2 additive:1 oxygenation:1 analytic:1 motor:5 hypothesize:3 atlas:1 designed:2 v:1 stationary:2 pursued:1 generative:1 alone:1 slowing:2 smith:1 short:9 core:1 chiang:2 lr:1 node:15 zhang:2 five:2 mathematical:1 glover:1 constructed:1 addiction:1 consists:1 yuan:2 combine:1 blondel:1 acquired:2 pairwise:5 mask:1 expected:2 examine:1 mechanic:2 multi:7 brain:32 salakhutdinov:1 detects:1 window:3 hardoon:1 project:2 estimating:2 moreover:1 matched:1 circuit:8 provided:1 investigator:1 null:8 emerging:1 developed:1 informed:1 finding:1 transformation:3 temporal:13 knn5:1 delineating:1 ull:4 classifier:7 uk:1 control:8 unit:1 t1:7 positive:5 local:1 tnr:6 consequence:1 despite:1 ak:1 analyzing:2 fluctuation:2 signed:3 fov:1 co:1 limited:1 range:9 statistically:2 acknowledgment:1 testing:2 cogn:1 area:4 empirical:4 drug:1 significantly:2 pre:1 confidence:1 specificity:1 anticipation:4 nb:3 context:1 influence:5 unsigned:1 www:1 jaime:2 map:3 equivalent:2 conventional:1 yt:4 demonstrated:1 missing:1 straightforward:1 misuse:1 thompson:1 resolution:2 simplicity:1 disorder:1 qian:1 rule:1 importantly:3 fill:1 financial:1 traditionally:1 coordinate:1 brazil:2 controlling:9 construction:1 pt:2 origin:1 trend:4 particularly:1 std:3 observed:2 capture:2 region:20 connected:5 highest:1 inhibit:1 balanced:5 disease:1 complexity:1 covariates:6 dynamic:7 passos:1 multimodal:1 seth:1 represented:1 stacked:1 epi:1 train:1 effective:1 describe:2 london:1 doi:1 detected:1 pearson:2 quite:1 lag:1 widely:1 encoded:1 supplementary:2 premotor:1 bullmore:1 statistic:1 knn:1 ability:1 gp:8 highlighted:1 transform:1 echo:2 noisy:1 interplay:1 advantage:2 sequence:1 rr:1 interview:1 ucl:1 propose:3 interaction:2 interconnected:2 combining:2 realization:1 consent:1 gold:1 description:1 validate:4 diseased:2 leave:1 staying:1 illustrate:1 informs:1 ac:2 stat:1 derive:2 axial:1 measured:3 nearest:1 school:1 c:1 quantify:3 met:1 inflated:1 human:6 opinion:1 public:1 material:1 investigation:1 varoquaux:1 parcorr:7 extension:2 physica:2 scanner:2 mm:3 ground:2 roi:2 exp:1 cb:1 cognition:1 mapping:3 bj:1 matthew:1 sma:6 omitted:1 prettenhofer:1 healthy:5 repetition:1 create:1 successfully:2 federal:2 clearly:1 gaussian:3 always:1 aim:1 pn:2 frontoparietal:9 avoid:1 cr:2 varying:1 zhou:1 derived:2 focus:1 upe:2 modelling:2 grisel:1 psychiatry:1 detect:4 dependent:3 membership:1 typically:1 integrated:1 hidden:1 irn:1 reproduce:1 pixel:1 classification:15 among:6 augment:1 exponent:1 resonance:2 gramfort:1 orange:1 mutual:2 once:3 construct:1 extraction:1 beach:1 barnett:1 mm2:1 park:1 seventy:2 unsupervised:1 fmri:35 future:2 jb:1 t2:7 simplify:1 haven:4 few:1 distinguishes:1 eighty:2 randomly:2 individual:6 geometry:1 attempt:1 ab:1 detection:1 stationarity:1 interest:1 mlp:2 inpatient:1 highly:1 cournapeau:1 essen:1 evaluation:3 analyzed:1 pc:6 xb:12 accurate:1 edge:2 partial:14 necessary:2 xy:1 respective:1 iv:1 taylor:1 irm:3 desired:2 logan:1 causal:3 precentral:1 instance:3 modeling:2 wb:1 soft:1 measuring:1 ada:4 dsm:1 deviation:3 hundred:1 comprised:4 successful:1 characterize:1 reported:4 dependency:1 answer:2 trio:1 thickness:1 synthetic:5 combined:3 gd:1 st:2 disrupted:2 peak:7 sensitivity:2 international:1 connectome:5 enhance:1 together:1 extant:1 connectivity:57 reflect:2 nm:1 dr:1 cocaine:12 cognitive:4 external:1 return:3 michel:1 li:4 insula:2 paulo:2 sinusoidal:3 de:1 bold:1 summarized:4 wk:1 coefficient:11 matter:1 jc:1 onset:1 performed:2 bilateral:4 red:1 wave:4 bayes:1 parallel:1 ifc:1 neuropsychiatric:2 accuracy:14 variance:3 characteristic:2 percept:1 maximized:1 ensemble:1 identify:2 decipher:1 bayesian:2 accurately:2 produced:5 inform:1 phys:3 influenced:1 mprage:1 definition:3 frequency:3 involved:2 thereof:1 associated:2 sampled:1 stop:9 dataset:14 treatment:1 popular:3 improves:2 dimensionality:6 organized:1 stanley:1 anticipating:1 reflecting:1 higher:4 originally:1 methodology:1 planar:1 wei:1 response:2 improved:1 adaboost:1 april:1 done:1 delineate:1 diagnosed:1 strongly:1 jw:1 xa:12 correlation:42 autoencoders:5 hand:2 d:1 langford:1 nonlinear:1 scikit:2 lack:1 marker:1 usa:1 effect:1 validity:1 verify:1 true:6 isomap:5 name:1 concept:1 evolution:1 white:1 sin:12 numerator:1 climatological:1 inferior:2 covering:1 noted:1 auc:4 davis:1 ide:4 criterion:3 m:2 demonstrate:1 correa:1 gh:1 silva:1 image:2 novel:5 recently:1 nih:1 common:1 functional:30 overview:2 association:2 he:4 tpr:7 resting:2 functionally:3 measurement:2 significant:1 imposing:1 shuffling:1 pm:1 session:1 shawe:1 cortex:3 wilcoxon:2 j:2 multivariate:4 recent:3 perspective:1 outperforming:1 binary:5 multifractal:1 rep:2 yi:1 captured:3 additional:4 greater:2 mr:1 performer:1 paradigm:1 signal:33 ii:4 sliding:1 multiple:6 dashed:1 full:2 cross:22 long:11 clinical:7 offer:1 netsim:8 paired:2 controlled:2 prediction:8 denominator:1 multilayer:1 metric:4 represent:2 addition:1 participated:1 campos:2 interval:1 harrison:1 crucial:1 unlike:3 sr:1 subject:6 recruited:1 hz:1 cowan:1 december:1 nonlin:1 effectiveness:1 nonstationary:3 structural:1 iii:3 split:1 fit:3 architecture:1 competing:4 topology:1 bandwidth:1 identified:1 br:2 biomarkers:1 whether:2 six:5 pca:6 speech:1 matlab:1 dfa:1 aimed:2 band:1 tenenbaum:1 processed:1 gyrus:3 generate:4 http:1 canonical:5 nsf:1 neuroscience:3 estimated:1 anatomical:2 broadly:1 blue:1 brucher:1 group:5 key:1 four:1 threshold:2 blood:1 pj:2 registration:1 imaging:3 graph:1 year:1 sum:2 run:1 angle:1 inverse:3 wu:1 coherence:1 comparable:2 cca:30 ct:3 shan:2 layer:2 yale:5 encountered:1 mni:1 activity:2 strength:2 occur:1 software:2 sake:1 simulate:1 min:1 department:2 structured:1 according:3 combination:2 poor:1 jr:1 across:5 rev:2 explained:1 restricted:1 pipeline:1 wellcome:1 equation:5 remains:1 previously:2 granger:3 committee:1 mfg:2 thirion:1 flip:1 ge:1 available:3 apply:1 eight:2 observe:1 magnetic:2 simulating:1 distinguished:2 batch:1 existence:1 original:7 responding:6 top:2 include:1 xc:8 medicine:2 restrictive:1 establish:1 february:1 perrot:1 question:3 added:2 hum:1 parametric:1 dependence:7 rt:1 enhances:1 gradient:1 dp:6 fabio:2 separate:2 simulated:6 sci:3 prospective:1 manifold:1 collected:1 assuming:1 connectomes:1 index:1 minn:1 illustration:2 mostly:2 neuroimaging:1 negative:3 ciuciu:1 boltzmann:1 perform:2 allowing:1 neuron:2 datasets:1 sm:1 benchmark:1 thermal:1 parietal:2 hinton:1 excluding:1 precise:1 dc:1 inverting:1 pair:3 vanderplas:1 toolbox:2 extensive:2 connection:17 conflict:4 acoustic:1 distinction:1 nip:1 address:1 suggested:1 perception:1 challenge:3 including:4 memory:9 max:4 belief:1 sporns:1 detrended:15 critical:3 power:10 natural:1 friston:1 regularized:3 participation:1 predicting:1 force:1 improve:2 altered:2 identifies:1 psychol:1 extract:3 naive:1 health:1 kj:1 szedmak:1 prior:1 understanding:2 discovery:1 epoch:1 python:1 geometric:1 law:7 mixed:2 limitation:1 filtering:1 proven:1 acyclic:1 localized:1 age:1 validation:1 sao:2 cd:7 course:1 summary:1 supported:1 cnpq:1 free:5 allow:1 perceptron:1 neighbor:2 penny:1 van:1 slice:2 curve:1 dimension:2 lett:1 world:4 xia:1 rich:3 seemed:1 author:2 made:1 preprocessing:1 correlate:1 keep:1 confirm:1 global:1 conclude:1 xi:1 neurology:1 spectrum:2 quantifies:2 table:10 additionally:2 promising:1 nature:1 learn:5 robust:1 ca:7 contributes:1 eeg:1 investigated:1 complex:3 hc:5 domain:2 did:1 significance:4 dcca:9 main:1 arrow:1 whole:3 noise:4 bassett:1 profile:17 big:1 neuronal:1 causality:3 fig:12 augmented:2 representative:2 neuroimage:11 duchesnay:1 comput:1 third:1 wavelet:4 ifg:3 xt:12 covariate:1 barrett:1 svm:3 evidence:1 fusion:2 intrinsic:3 false:1 corr:2 magnitude:1 te:1 occurring:1 gap:1 univariate:1 likely:1 highlighting:1 adjustment:1 chang:1 gender:1 truth:2 satisfies:1 chance:1 extracted:1 dcm:1 goal:1 shared:1 included:1 specifically:1 typical:1 reducing:1 principal:3 hampshire:1 total:1 called:2 invariance:1 siemens:1 pedregosa:1 college:1 support:1 scan:2 frontal:3 ongoing:1 evaluate:1 tested:2 correlated:1 |
6,291 | 6,691 | Contrastive Learning for Image Captioning
Bo Dai
Dahua Lin
Department of Information Engineering, The Chinese University of Hong Kong
[email protected]
[email protected]
Abstract
Image captioning, a popular topic in computer vision, has achieved substantial
progress in recent years. However, the distinctiveness of natural descriptions is
often overlooked in previous work. It is closely related to the quality of captions,
as distinctive captions are more likely to describe images with their unique aspects.
In this work, we propose a new learning method, Contrastive Learning (CL), for
image captioning. Specifically, via two constraints formulated on top of a reference
model, the proposed method can encourage distinctiveness, while maintaining the
overall quality of the generated captions. We tested our method on two challenging
datasets, where it improves the baseline model by significant margins. We also
showed in our studies that the proposed method is generic and can be used for
models with various structures.
1
Introduction
Image captioning, a task to generate natural descriptions of images, has been an active research
topic in computer vision and machine learning. Thanks to the advances in deep neural networks,
especially the wide adoption of RNN and LSTM, there has been substantial progress on this topic
in recent years [23, 24, 15, 19]. However, studies [1, 3, 2, 10] have shown that even the captions
generated by state-of-the-art models still leave a lot to be desired. Compared to human descriptions,
machine-generated captions are often quite rigid and tend to favor a ?safe? (i.e. matching parts of
the training captions in a word-by-word manner) but restrictive way. As a consequence, captions
generated for different images, especially those that contain objects of the same categories, are
sometimes very similar [1], despite their differences in other aspects.
We argue that distinctiveness, a property often overlooked in previous work, is significant in natural
language descriptions. To be more specific, when people describe an image, they often mention or
even emphasize the distinctive aspects of an image that distinguish it from others. With a distinctive
description, someone can easily identify the image it is referring to, among a number of similar
images. In this work, we performed a self-retrieval study (see Section 4.1), which reveals the lack of
distinctiveness affects the quality of descriptions.
From a technical standpoint, the lack of distinctiveness is partly related to the way that the captioning
model was learned. A majority of image captioning models are learned by Maximum Likelihood
Estimation (MLE), where the probabilities of training captions conditioned on corresponding images
are maximized. While well grounded in statistics, this approach does not explicitly promote distinctiveness. Specifically, the differences among the captions of different images are not explicitly taken
into account. We found empirically that the resultant captions highly resemble the training set in a
word-by-word manner, but are not distinctive.
In this paper, we propose Contrastive Learning (CL), a new learning method for image captioning,
which explicitly encourages distinctiveness, while maintaining the overall quality of the generated
captions. Specifically, it employs a baseline, e.g. a state-of-the-art model, as a reference. During
learning, in addition to true image-caption pairs, denoted as (I, c), this method also takes as input
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
mismatched pairs, denoted as (I, c/ ), where c/ is a caption describing another image. Then, the
target model is learned to meet two goals, namely (1) giving higher probabilities p(c|I) to positive
pairs, and (2) lower probabilities p(c/ |I) to negative pairs, compared to the reference model. The
former ensures that the overall performance of the target model is not inferior to the reference; while
the latter encourages distinctiveness.
It is noteworthy that the proposed learning method (CL) is generic. While in this paper, we focused on
models based on recurrent neural networks [23, 15], the proposed method can also generalize well to
models based on other formulations, e.g. probabilistic graphical models [4, 9]. Also, by choosing the
state-of-the-art model as the reference model in CL, one can build on top of the latest advancement in
image captioning to obtain improved performances.
2
Related Work
Models for Image Captioning The history of image captioning can date back to decades ago.
Early attempts are mostly based on detections, which first detect visual concepts (e.g. objects and
their attributes) [9, 4] followed by template filling [9] or nearest neighbor retrieving for caption
generation [2, 4]. With the development of neural networks, a more powerful paradigm, encoderand-decoder, was proposed by [23], which then becomes the core of most state-of-the-art image
captioning models. It uses a CNN [20] to represent the input image with a feature vector, and applies
a LSTM net [6] upon the feature to generate words one by one.
Based on the encoder-and-decoder, many variants are proposed, where attention mechanism [24] appears to be the most effective add-on. Specifically, attention mechanism replaces the feature vector
with a set of feature vectors, such as the features from different regions [24] , and those under different
conditions [27]. It also uses the LSTM net to generate words one by one, where the difference is that
at each step, a mixed guiding feature over the whole feature set, will be dynamically computed. In
recent years, there are also approaches combining attention mechanism and detection. Instead of
doing attention on features, they consider the attention on a set of detected visual concepts, such as
attributes [25] and objects [26].
Despite of the specific structure of any image captioning model, it is able to give p(c|I), the probability
of a caption conditioned on an image. Therefore, all image captioning models can be used as the
target or the reference in CL method.
Learning Methods for Image Captioning Many state-of-the-art image captioning models adopt
Maximum Likelihood Estimation (MLE) as their learning method, which maximizes the conditional
log-likelihood of the training samples, as:
X
Ti
X
(t)
(t?1)
ln p(wi |Ii , wi
(1)
, ..., wi , ?),
(1)
(ci ,Ii )?D t=1
(1)
(2)
(T )
where ? is the parameter vector, Ii and ci = (wi , wi , ..., wi i ) are a training image and its
caption. Although effective, some issues, including high resemblance in model-gerenated captions,
are observed [1] on models learned by MLE.
Facing these issues, alternative learning methods are proposed in recent years. Techniques of
reinforcement learning (RL) have been introduced in image captioning by [19] and [14]. RL sees the
procedure of caption generation as a procedure of sequentially sampling actions (words) in a policy
space (vocabulary). The rewards in RL are defined to be evaluation scores of sampled captions. Note
that distinctiveness has not been considered in both approaches, RL and MLE.
Prior to this work, some relevant ideas have been explored [21, 16, 1]. Specifically, [21, 16] proposed
an introspective learning (IL) approach that learns the target model by comparing its outputs on (I, c)
and (I/ , c). Note that IL uses the target model itself as a reference. On the contrary, the reference
model in CL provides more independent and stable indications about distinctiveness. In addition,
(I/ , c) in IL is pre-defined and fixed across the learning procedure, while the negative sample in
CL, i.e. (I, c/ ), is dynamically sampled, making it more diverse and random. Recently, Generative
Adversarial Networks (GAN) was also adopted for image captioning [1], which involves an evaluator
that may help promote the distinctiveness. However, this evaluator is learned to directly measure the
2
A man performing stunt
in the air at skate park
A man doing a trick
on a skateboard
Self Retrieval
Self Retrieval
(a) Nondistinctive Caption
(b) Distinctive Caption
Figure 1: This figure illustrates respectively a nondistinctive and distinctive captions of an image, where the
nondistinctive one fails to retrieve back the original image in self retrieval task.
Self Retrieval Top-K Recall
Method
Neuraltalk2 [8]
AdaptiveAttention [15]
AdaptiveAttention + CL
Captioning
1
5
50
500
ROUGE_L
CIDEr
0.02
0.10
0.32
0.32
0.96
1.18
3.02
11.76
11.84
27.50
78.46
80.96
0.652
0.689
0.695
0.827
1.004
1.029
Table 1: This table lists results of self retrieval and captioning of different models. The results are reported on
standard MSCOCO test set. See sec 4.1 for more details.
distinctiveness as a parameterized approximation, and the approximation accuracy is not ensured in
GAN. In CL, the fixed reference provides stable bounds about the distinctiveness, and the bounds are
supported by the model?s performance on image captioning. Besides that, [1] is specifically designed
for models that generate captions word-by-word, while CL is more generic.
3
Background
Our formulation is partly inspired by Noise Contrastive Estimation (NCE) [5]. NCE is originally
introduced for estimating probability distributions, where the partition functions can be difficult or
even infeasible to compute. To estimate a parametric distribution pm (.; ?), which we refer to as the
target distribution, NCE employs not only the observed samples X = (x1 , x2 , ..., xTm ), but also
the samples drawn from a reference distribution pn , denoted as Y = (y1 , y2 , ..., yTn ). Instead of
estimating pm (.; ?) directly, NCE estimates the density ratio pm /pn by training a classifier based on
logistic regression.
Specifically, let U = (u1 , ..., uTm +Tn ) be the union of X and Y . A binary class label Ct is assigned
to each ut , where Ct = 1 if ut ? X and Ct = 0 if ut ? Y . The posterior probabilities for the class
labels are therefore
P (C = 1|u, ?) =
pm (u; ?)
,
pm (u; ?) + ?pn (u)
P (C = 0|u, ?) =
?pn (u)
,
pm (u; ?) + ?pn (u)
(2)
where ? = Tn /Tm . Let G(u; ?) = ln pm (u; ?) ? ln pn (u) and h(u, ?) = P (C = 1|u, ?), then we
can write
1
h(u; ?) = r? (G(u; ?)), with r? (z) =
.
(3)
1 + ? exp(?z)
The objective function of NCE is the joint conditional log-probabilities of Ct given the samples U ,
which can be written as
L(?; X, Y ) =
Tm
X
ln[h(xt ; ?)] +
t=1
Tn
X
ln[1 ? h(yt ; ?)].
(4)
t=1
Maximizing this objective with respect to ? leads to an estimation of G(?; ?), the logarithm of the
density ratio pm /pn . As pn is a known distribution, pm (: |?) can be readily derived.
4
Contrastive Learning for Image Captioning
Learning a model by characterizing desired properties relative to a strong baseline is a convenient
and often quite effective way in situations where it is hard to describe these properties directly.
Specifically, in image captioning, it is difficult to characterize the distinctiveness of natural image
descriptions via a set of rules, without running into the risk that some subtle but significant points are
3
missed. Our idea in this work is to introduce a baseline model as a reference, and try to enhance the
distinctiveness on top, while maintaining the overall quality of the generated captions.
In the following we will first present an empirical study on the correlation between distinctiveness of
its generated captions and the overall performance of a captioning model. Subsequently, we introduce
the main framework of Contrastive Learning in detail.
4.1
Empirical Study: Self Retrieval
In most of the existing learning methods of image captioning, models are asked to generate a caption
that best describes the semantics of a given image. In the meantime, distinctiveness of the caption,
which, on the other hand, requires the image to be the best matching among all images for the caption,
has not been explored. However, distinctiveness is crucial for high-quality captions. A study by Jas
[7] showed that specificity is common in human descriptions, which implies that image descriptions
often involve distinctive aspects. Intuitively, a caption satisfying this property is very likely to contain
key and unique content of the image, so that the original image could easily be retrieved when the
caption is presented.
To verify this intuition, we conducted an empirical study which we refer to as self retrieval. In this
experiment, we try to retrieve the original image given its model-generated caption and investigate topk recalls, as illustrated in Figure 1. Specifically, we randomly sampled 5, 000 images (I1 , I2 , ..., I5000 )
from standard MSCOCO [13] test set as the experiment benchmark. For an image captioning model
pm (:, ?), we first ran it on the benchmark to get corresponding captions (c1 , c2 , ..., c5000 ) for the
images. After that, using each caption ct as a query, we computed the conditional probabilities
(pm (ct |I1 ), pm (ct |I2 ), ..., pm (ct |I5000 )), which were used to get a ranked list of images, denoted by
rt . Based on all ranked lists, we can compute top-k recalls, which is the fraction of images within
top-k positions of their corresponding ranked lists. The top-k recalls are good indicators of how well
a model captures the distinctiveness of descriptions.
In this experiment, we compared three different models, including Neuraltalk2 [8] and AdaptiveAttention [15] that are learned by MLE, as well as AdaptiveAttention learned by our method. The
top-k recalls are listed in Table 1, along with overall performances of these models in terms of
Rouge [12] and Cider [22]. These results clearly show that the recalls of self retrieval are positively
correlated to the performances of image captioning models in classical captioning metrics. Although
most of the models are not explicitly learned to promote distinctiveness, the one with better recalls
of self retrieval, which means the generated-captions are more distinctive, performs better in the
image captioning evaluation. Such positive correlation clearly demonstrates the significance of
distinctiveness to captioning performance.
4.2
Contrastive Learning
In Contrastive Learning (CL), we learn a target image captioning model pm (:; ?) with parameter ?
by constraining its behaviors relative to a reference model pn (:; ?) with parameter ?. The learning
procedure requires two sets of data: (1) the observed data X, which is a set of ground-truth imagecaption pairs ((c1 , I1 ), (c2 , I2 ), ..., (cTm , ITm )), and is readily available in any image captioning
dataset, (2) the noise set Y , which contains mismatched pairs ((c/1 , I1 ), (c/2 , I2 ), ..., (c/Tn , ITn )),
and can be generated by randomly sampling c/t ? C/It for each image It , where C/It is the set of
all ground-truth captions except captions of image It . We refer to X as positive pairs while Y as
negative pairs.
For any pair (c, I), the target model and the reference model will respectively give their estimated
conditional probabilities pm (c|I, ?) and pn (c|I, ?). We wish that pm (ct |It , ?) is greater than
pn (ct |It , ?) for any positive pair (ct , It ), and vice versa for any negative pair (c/t , It ). Following this intuition, our initial attempt was to define D((c, I); ?, ?), the difference between pm (c|I, ?)
and pn (c|I, ?), as
D((c, I); ?, ?) = pm (c|I, ?) ? pn (c|I, ?),
(5)
and set the loss function to be:
L0 (?; X, Y, ?) =
Tm
X
D((ct , It ); ?, ?) ?
t=1
Tn
X
t=1
4
D((c/t , It ); ?, ?).
(6)
In practice, this formulation would meet with several difficulties. First, pm (c|I, ?) and pn (c|I, ?) are
very small (? 1e-8), which may result in numerical problems. Second, Eq (6) treats easy samples,
hard samples, and mistaken samples equally. This, however, is not the most effective way. For
example, when D((ct , It ); ?, ?) 0 for some positive pair, further increasing D((ct , It ); ?, ?)
is probably not as effective as updating D((ct0 , It0 ); ?, ?) for another positive pair, for which
D((ct0 , It0 ); ?, ?) is much smaller.
To resolve these issues, we adopted an alternative formulation inspired by NCE (sec 3), where we
replace the difference function D((c, I); ?, ?) with a log-ratio function G((c, I); ?, ?):
G((c, I); ?, ?) = ln pm (c|I, ?) ? ln pn (c|I, ?),
(7)
and further use a logistic function r? (Eq(3)) after G((c, I); ?, ?) to saturate the influence of easy
samples. Following the notations in NCE, we let ? = Tn /Tm , and turn D((c, I); ?, ?) into:
h((c, I); ?, ?) = r? (G((c, I); ?, ?))).
(8)
Note that h((c, I); ?, ?) ? (0, 1). Then, we define our updated loss function as:
L(?; X, Y, ?) =
Tm
X
ln[h((ct , It ); ?, ?)] +
t=1
Tn
X
ln[1 ? h((c/t , It ); ?, ?)].
(9)
t=1
For the setting of ? = Tn /Tm , we choose ? = 1, i.e. Tn = Tm , to ensure balanced influences from
both positive and negative pairs. This setting consistently yields good performance in our experiments.
Furthermore, we copy X for K times and sample K different Y s, in order to involve more diverse
negative pairs without overfitted to them. In practice we found K = 5 is sufficient to make the
learning stable. Finally, our objective function is defined to be
K
1 1 X
J(?) =
L(?; X, Yk , ?).
(10)
K Tm
k=1
Note that J(?) attains its upper bound 0 if positive and negative pairs can be perfectly distinguished,
namely, for all t, h((ct , It ); ?, ?) = 1 and h((c/t , It ); ?, ?) = 0. In this case, G((ct , It ); ?, ?) ? ?
and G((c/t , It ); ?, ?) ? ??, which indicates the target model will give higher probability p(ct |It )
and lower probability p(c/t |It ), compared to the reference model. Towards this goal, the learning
process would encourage distinctiveness by suppressing negative pairs, while maintaining the overall
performance by maximizing the probability values on positive pairs.
4.3
Discussion
Maximum Likelihood Estimation (MLE) is a popular learning method in the area of image captioning
[23, 24, 15]. The objective of MLE is to maximize only the probabilities of ground-truth imagecaption pairs, which may lead to some issues [1], including high resemblance in generated captions.
While in CL, the probabilities of ground-truth pairs are indirectly ensured by the positive constraint
(the first term in Eq(9)), and the negative constraint (the second term in Eq(9)) suppresses the
probabilities of mismatched pairs, forcing the target model to also learn from distinctiveness.
Generative Adversarial Network (GAN) [1] is a similar learning method that involves an auxiliary
model. However, in GAN the auxiliary model and the target model follow two opposite goals, while
in CL the auxiliary model and the target model are models in the same track. Moreover, in CL the
auxiliary model is stable across the learning procedure, while itself needs careful learning in GAN.
It?s worth noting that although our CL method bears certain level of resemblance with Noise Contrastive Estimation (NCE) [5]. The motivation and the actual technical formulation of CL and NCE
are essentially different. For example, in NCE the logistic function is a result of computing posterior
probabilities, while in CL it is explicitly introduced to saturate the influence of easy samples.
As CL requires only pm (c|I) and pn (c|I), the choices of the target model and the reference model
can range from models based on LSTMs [6] to models in other formats, such as MRFs [4] and
memory-networks [18]. On the other hand, although in CL, the reference model is usually fixed
across the learning procedure, one can replace the reference model with the latest target model
periodically. The reasons are (1) ?J(?) 6= 0 when the target model and the reference model are
identical, (2) latest target model is usually stronger than the reference model, (3) and a stronger
reference model can provide stronger bounds and lead to a stronger target model.
5
COCO Online Testing Server C5
Method
B-1
B-2
B-3
B-4
METEOR
ROUGE_L
CIDEr
Google NIC [23]
Hard-Attention[24]
AdaptiveAttention [15]
AdpativeAttention + CL (Ours)
PG-BCMR [14]
0.713
0.705
0.735
0.742
0.754
0.542
0.528
0.569
0.577
0.591
0.407
0.383
0.429
0.436
0.445
0.309
0.277
0.323
0.326
0.332
0.254
0.241
0.258
0.260
0.257
0.530
0.516
0.541
0.544
0.550
0.943
0.865
1.001
1.010
1.013
ATT-FCN? [26]
MSM? [25]
AdaptiveAttention? [15]
Att2in? [19]
0.731
0.739
0.746
-
0.565
0.575
0.582
-
0.424
0.436
0.443
-
0.316
0.330
0.335
0.344
0.250
0.256
0.264
0.268
0.535
0.542
0.550
0.559
0.943
0.984
1.037
1.123
COCO Online Testing Server C40
Method
B-1
B-2
B-3
B-4
METEOR
ROUGE_L
CIDEr
Google NIC [23]
Hard-Attention [24]
AdaptiveAttention [15]
AdaptiveAttention + CL (Ours)
PG-BCMR [14]
0.895
0.881
0.906
0.910
-
0.802
0.779
0.823
0.831
-
0.694
0.658
0.717
0.728
-
0.587
0.537
0.607
0.617
-
0.346
0.322
0.347
0.350
-
0.682
0.654
0.689
0.695
-
0.946
0.893
1.004
1.029
-
ATT-FCN? [26]
MSM? [25]
AdaptiveAttention? [15]
Att2in? [19]
0.900
0.919
0.918
-
0.815
0.842
0.842
-
0.709
0.740
0.740
-
0.599
0.632
0.633
-
0.335
0.350
0.359
-
0.682
0.700
0.706
-
0.958
1.003
1.051
-
Table 2: This table lists published results of state-of-the-art image captioning models on the online COCO
testing server. ? indicates ensemble model. "-" indicates not reported. In this table, CL improves the base model
(AdaptiveAttention [15]) to gain the best results among all single models on C40.
5
5.1
Experiment
Datasets
We use two large scale datasets to test our contrastive learning method. The first dataset is MSCOCO
[13], which contains 122, 585 images for training and validation. Each image in MSCOCO has 5
human annotated captions. Following splits in [15], we reserved 2, 000 images for validation. A
more challenging dataset, InstaPIC-1.1M [18], is used as the second dataset, which contains 648, 761
images for training, and 5, 000 images for testing. The images and their ground-truth captions
are acquired from Instagram, where people post images with related descriptions. Each image in
InstaPIC-1.1M is paired with 1 caption. This dataset is challenging, as its captions are natural posts
with varying formats. In practice, we reserved 2, 000 images from the training set for validation.
On both datasets, non-alphabet characters except emojis are removed, and alphabet characters are
converted to lowercases. Words and emojis that appeared less than 5 times are replaced with UNK.
And all captions are truncated to have at most 18 words and emojis. As a result, we obtained a
vocabulary of size 9, 567 on MSCOCO, and a vocabulary of size 22, 886 on InstaPIC-1.1M.
5.2
Settings
To study the generalization ability of proposed CL method, we tested it on two different image
captioning models, namely Neuraltalk2 [8] and AdaptiveAttention [15]. Both models are based
on encoder-and-decoder [23], where no attention mechanism is used in the former, and an adaptive
attention component is used in the latter.
For both models, we have pretrained them by MLE, and use the pretrain checkpoints as initializations.
In all experiments except for the experiment on model choices, we choose the same model and use
the same initialization for target model and reference model. In all our experiments, we fixed the
learning rate to be 1e-6 for all components, and used Adam optimizer. Seven evaluation metrics have
been selected to compare the performances of different models, including Bleu-1,2,3,4 [17], Meteor
[11], Rouge [12] and Cider [22]. All experiments for ablation studies are conducted on the validation
set of MSCOCO.
6
Two people on a tennis court
playing tennis
Two tennis players shaking
AA + CL
hands on a tennis court
A fighter jet flying through a
blue sky
A fighter jet flying over a lush
green field
A row of boats on a river near
a river
A row of boats docked in a
river
A bathroom with a toilet and a
sink
A bathroom with a red toilet
and red walls
Three clocks are mounted to
the side of a building
Three three clocks with three
AA + CL
different time zones
Two people on a yellow
yellow and yellow motorcycle
Two people riding a yellow
motorcycle in a forest
A baseball player pitching a
ball on top of a field
A baseball game in progress
with pitcher throwing the ball
A bunch of lights hanging
from a ceiling
A bunch of baseballs bats
hanging from a ceiling
AA
AA
Figure 2: This figure illustrates several images with captions generated by different models, where AA represents
AdaptiveAttention [15] learned by MLE, and AA + CL represents the same model learned by CL. Compared to
AA, AA + CL generated more distinctive captions for these images.
Method
Google NIC [23]
Hard-Attention [24]
CSMN [18]
AdaptiveAttention [15]
AdaptiveAttention + CL (Ours)
B-1
B-2
B-3
B-4
METEOR
ROUGE_L
CIDEr
0.055
0.106
0.079
0.065
0.072
0.019
0.015
0.032
0.026
0.028
0.007
0.000
0.015
0.011
0.013
0.003
0.000
0.008
0.005
0.006
0.038
0.026
0.037
0.029
0.032
0.081
0.140
0.120
0.093
0.101
0.004
0.049
0.133
0.126
0.144
Table 3: This table lists results of different models on the test split of InstaPIC-1.1M [18], where CL improves
the base model (AdaptiveAttention [15]) by significant margins, achieving the best result on Cider.
5.3
Results
Overall Results We compared our best model (AdaptiveAttention [15] learned by CL) with stateof-the-art models on two datasets. On MSCOCO, we submitted the results to the online COCO
testing server. The results along with other published results are listed in Table 2. Compared to
MLE-learned AdaptiveAttention, CL improves the performace of it by significant margins across
all metrics. While most of state-of-the-art results are achieved by ensembling multiple models, our
improved AdaptiveAttention gains competitive results as a single model. Specifically, on Cider, CL
improves AdaptiveAttention from 1.003 to 1.029, which is the best single-model result on C40 among
all published ones. In terms of Cider, if we use MLE, we need to combine 5 models to get 4.5% boost
on C40 for AdaptiveAttention. Using CL, we improve the performance by 2.5% with just a single
model. On InstaPIC-1.1M, CL improves the performance of AdaptiveAttention by 14% in terms of
Cider, which is the state-of-the-art. Some qualitative results are shown in Figure 2. It?s worth noting
that the proposed learning method can be used with stronger base models to obtain better results
without any modification.
Compare Learning Methods Using AdaptiveAttention learned by MLE as base model and initialization, we compared our CL with similar learning methods, including CL(P) and CL(N) that
Method
AdaptiveAttention [15] (Base)
Base + IL [21]
Base + GAN [1]
Base + CL(P)
Base + CL(N)
Base + CL(Full)
B-1
B-2
B-3
B-4
METEOR
ROUGE_L
CIDEr
0.733
0.706
0.629
0.735
0.539
0.755
0.572
0.544
0.437
0.573
0.411
0.598
0.433
0.408
0.290
0.437
0.299
0.460
0.327
0.307
0.190
0.334
0.212
0.353
0.260
0.253
0.212
0.262
0.246
0.271
0.540
0.530
0.458
0.545
0.479
0.559
1.042
1.004
0.700
1.059
0.603
1.142
Table 4: This table lists results of a model learned by different methods. The best result is obtained by the one
learned with full CL, containing both the positive constraint and negative constraint.
7
Target Model
Reference Model
B-1
B-2
B-3
B-4
METEOR
ROUGE_L
CIDEr
NT
NT
NT
NT
AA
0.697
0.708
0.716
0.525
0.536
0.547
0.389
0.399
0.411
0.291
0.300
0.311
0.238
0.242
0.249
0.516
0.524
0.533
0.882
0.905
0.956
AA
AA
AA
0.733
0.755
0.572
0.598
0.433
0.460
0.327
0.353
0.260
0.271
0.540
0.559
1.042
1.142
Table 5: This table lists results of different model choices on MSCOCO. In this table, NT represents Neuraltalk2
[8], and AA represents AdaptiveAttention [15]. "-" indicates the target model is learned using MLE.
Run
B-1
B-2
B-3
B-4
METEOR
ROUGE_L
CIDEr
0
1
2
0.733
0.755
0.756
0.572
0.598
0.598
0.433
0.460
0.460
0.327
0.353
0.353
0.260
0.271
0.272
0.540
0.559
0.559
1.042
1.142
1.142
Table 6: This table lists results of periodical replacement of the reference in CL. In run 0, the model is learned
by MLE, which are used as both the target and the reference in run 1. In run 2, the reference is replaced with the
best target in run 1.
respectively contains only the positive constraint and the negative constraint in CL. We also compared
with IL [21], and GAN [1]. The results on MSCOCO are listed in Table 4, where (1) among IL, CL
and GAN, CL improves performance of the base model, while both IL and GAN decrease the results.
This indicates the trade-off between learning distinctiveness and maintaining overall performance is
not well settled in IL and GAN. (2) comparing models learned by CL(P), CL(N) and CL, we found
using the positive constraint or the negative constraint alone is not sufficient, as only one source of
guidance is provided. While CL(P) gives the base model lower improvement than full CL, CL(N)
downgrades the base model, indicating overfits on distinctiveness. Combining CL(P) and CL(N),
CL is able to encourage distinctiveness while also emphasizing on overall performance, resulting in
largest improvements on all metrics.
Compare Model Choices To study the generalization ability of CL, AdaptiveAttention and Neuraltalk2 are respectively chosen as both the target and the reference in CL. In addition, AdaptiveAttention learned by MLE, as a better model, is chosen to be the reference, for Neuraltalk2. The
results are listed in Table 5, where compared to models learned by MLE, both AdaptiveAttention
and Neuraltalk2 are improved after learning using CL. For example, on Cider, AdaptiveAttention
improves from 1.042 to 1.142, and Neuraltalk2 improves from 0.882 to 0.905. Moreover, by using
a stronger model, AdaptiveAttention, as the reference, Neuraltalk2 improves further from 0.905 to
0.956, which indicates stronger references empirically provide tighter bounds on both the positive
constraint and the negative constraint.
Reference Replacement As discussed in sec 4.3, one can periodically replace the reference with
latest best target model, to further improve the performance. In our study, using AdaptiveAttention
learned by MLE as a start, each run we fix the reference model util the target saturates its performance
on the validation set, then we replace the reference with latest best target model and rerun the learning.
As listed in Table 6, in second run, the relative improvements of the target model is incremental,
compared to its improvement in the first run. Therefore, when learning a model using CL, with a
sufficiently strong reference, the improvement is usually saturated in the first run, and there is no
need, in terms of overall performance, to replace the reference multiple times.
6
Conclusion
In this paper, we propose Contrastive Learning, a new learning method for image captioning. By
employing a state-of-the-art model as a reference, the proposed method is able to maintain the
optimality of the target model, while encouraging it to learn from distinctiveness, which is an
important property of high quality captions. On two challenging datasets, namely MSCOCO and
InstaPIC-1.1M, the proposed method improves the target model by significant margins, and gains
state-of-the-art results across multiple metrics. On comparative studies, the proposed method extends
well to models with different structures, which clearly shows its generalization ability.
8
Acknowledgment This work is partially supported by the Big Data Collaboration Research grant
from SenseTime Group (CUHK Agreement No.TS1610626), the General Research Fund (GRF) of
Hong Kong (No.14236516) and the Early Career Scheme (ECS) of Hong Kong (No.24204215).
References
[1] Bo Dai, Sanja Fidler, Raquel Urtasun, and Dahua Lin. Towards diverse and natural image
descriptions via a conditional gan. In Proceedings of the IEEE International Conference on
Computer Vision, 2017.
[2] Jacob Devlin, Saurabh Gupta, Ross Girshick, Margaret Mitchell, and C Lawrence Zitnick.
Exploring nearest neighbor approaches for image captioning. arXiv preprint arXiv:1505.04467,
2015.
[3] Hao Fang, Saurabh Gupta, Forrest Iandola, Rupesh K Srivastava, Li Deng, Piotr Doll?r, Jianfeng
Gao, Xiaodong He, Margaret Mitchell, John C Platt, et al. From captions to visual concepts and
back. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,
pages 1473?1482, 2015.
[4] Ali Farhadi, Mohsen Hejrati, Mohammad Amin Sadeghi, Peter Young, Cyrus Rashtchian, Julia
Hockenmaier, and David Forsyth. Every picture tells a story: Generating sentences from images.
In European conference on computer vision, pages 15?29. Springer, 2010.
[5] Michael U Gutmann and Aapo Hyv?rinen. Noise-contrastive estimation of unnormalized
statistical models, with applications to natural image statistics. Journal of Machine Learning
Research, 13(Feb):307?361, 2012.
[6] Sepp Hochreiter and J?rgen Schmidhuber. Long short-term memory. Neural computation,
9(8):1735?1780, 1997.
[7] Mainak Jas and Devi Parikh. Image specificity. In Proceedings of the IEEE Conference on
Computer Vision and Pattern Recognition, pages 2727?2736, 2015.
[8] Andrej Karpathy and Li Fei-Fei. Deep visual-semantic alignments for generating image descriptions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,
pages 3128?3137, 2015.
[9] Girish Kulkarni, Visruth Premraj, Vicente Ordonez, Sagnik Dhar, Siming Li, Yejin Choi,
Alexander C Berg, and Tamara L Berg. Babytalk: Understanding and generating simple image
descriptions. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(12):2891?
2903, 2013.
[10] Polina Kuznetsova, Vicente Ordonez, Tamara L Berg, and Yejin Choi. Treetalk: Composition
and compression of trees for image descriptions. TACL, 2(10):351?362, 2014.
[11] Michael Denkowski Alon Lavie. Meteor universal: Language specific translation evaluation for
any target language. ACL 2014, page 376, 2014.
[12] Chin-Yew Lin. Rouge: A package for automatic evaluation of summaries. In Text summarization
branches out: Proceedings of the ACL-04 workshop, volume 8. Barcelona, Spain, 2004.
[13] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr
Doll?r, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In European
Conference on Computer Vision, pages 740?755. Springer, 2014.
[14] Siqi Liu, Zhenhai Zhu, Ning Ye, Sergio Guadarrama, and Kevin Murphy. Optimization of image
description metrics using policy gradient methods. arXiv preprint arXiv:1612.00370, 2016.
[15] Jiasen Lu, Caiming Xiong, Devi Parikh, and Richard Socher. Knowing when to look: Adaptive
attention via a visual sentinel for image captioning. arXiv preprint arXiv:1612.01887, 2016.
[16] Junhua Mao, Jonathan Huang, Alexander Toshev, Oana Camburu, Alan L Yuille, and Kevin
Murphy. Generation and comprehension of unambiguous object descriptions. In Proceedings
of the IEEE Conference on Computer Vision and Pattern Recognition, pages 11?20, 2016.
9
[17] Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. Bleu: a method for automatic
evaluation of machine translation. In Proceedings of the 40th annual meeting on association for
computational linguistics, pages 311?318. Association for Computational Linguistics, 2002.
[18] Cesc Chunseong Park, Byeongchang Kim, and Gunhee Kim. Attend to you: Personalized image
captioning with context sequence memory networks. In CVPR, 2017.
[19] Steven J Rennie, Etienne Marcheret, Youssef Mroueh, Jarret Ross, and Vaibhava Goel. Selfcritical sequence training for image captioning. arXiv preprint arXiv:1612.00563, 2016.
[20] Karen Simonyan and Andrew Zisserman. Very deep convolutional networks for large-scale
image recognition. arXiv preprint arXiv:1409.1556, 2014.
[21] Ramakrishna Vedantam, Samy Bengio, Kevin Murphy, Devi Parikh, and Gal Chechik. Contextaware captions from context-agnostic supervision. arXiv preprint arXiv:1701.02870, 2017.
[22] Ramakrishna Vedantam, C Lawrence Zitnick, and Devi Parikh. Cider: Consensus-based image
description evaluation. In Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, pages 4566?4575, 2015.
[23] Oriol Vinyals, Alexander Toshev, Samy Bengio, and Dumitru Erhan. Show and tell: A neural
image caption generator. In Proceedings of the IEEE Conference on Computer Vision and
Pattern Recognition, pages 3156?3164, 2015.
[24] Kelvin Xu, Jimmy Ba, Ryan Kiros, Kyunghyun Cho, Aaron C Courville, Ruslan Salakhutdinov,
Richard S Zemel, and Yoshua Bengio. Show, attend and tell: Neural image caption generation
with visual attention. In ICML, volume 14, pages 77?81, 2015.
[25] Ting Yao, Yingwei Pan, Yehao Li, Zhaofan Qiu, and Tao Mei. Boosting image captioning with
attributes. arXiv preprint arXiv:1611.01646, 2016.
[26] Quanzeng You, Hailin Jin, Zhaowen Wang, Chen Fang, and Jiebo Luo. Image captioning with
semantic attention. In Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, pages 4651?4659, 2016.
[27] Luowei Zhou, Chenliang Xu, Parker Koch, and Jason J Corso. Image caption generation with
text-conditional semantic attention. arXiv preprint arXiv:1606.04621, 2016.
10
| 6691 |@word kong:3 cnn:1 compression:1 stronger:7 hyv:1 jacob:1 contrastive:12 pg:2 mention:1 ytn:1 initial:1 liu:1 contains:4 score:1 att:2 ours:3 suppressing:1 existing:1 guadarrama:1 comparing:2 nt:5 luo:1 written:1 readily:2 john:1 neuraltalk2:9 numerical:1 partition:1 periodically:2 designed:1 fund:1 alone:1 generative:2 selected:1 advancement:1 intelligence:1 core:1 short:1 provides:2 boosting:1 evaluator:2 along:2 c2:2 retrieving:1 qualitative:1 combine:1 introduce:2 manner:2 acquired:1 behavior:1 kiros:1 inspired:2 salakhutdinov:1 resolve:1 actual:1 encouraging:1 farhadi:1 increasing:1 becomes:1 provided:1 estimating:2 notation:1 moreover:2 maximizes:1 agnostic:1 spain:1 suppresses:1 gal:1 hejrati:1 sky:1 every:1 ti:1 ensured:2 classifier:1 demonstrates:1 platt:1 ramanan:1 grant:1 kelvin:1 positive:14 engineering:1 attend:2 treat:1 todd:1 consequence:1 rouge:3 despite:2 meet:2 noteworthy:1 acl:2 initialization:3 dynamically:2 challenging:4 someone:1 pitching:1 gunhee:1 range:1 adoption:1 bat:1 unique:2 acknowledgment:1 testing:5 union:1 practice:3 procedure:6 maire:1 mei:1 area:1 rnn:1 empirical:3 universal:1 matching:2 convenient:1 word:11 pre:1 performace:1 specificity:2 sensetime:1 chechik:1 get:3 andrej:1 periodical:1 salim:1 risk:1 influence:3 context:3 yt:1 maximizing:2 latest:5 attention:14 pitcher:1 sepp:1 jimmy:1 focused:1 rule:1 fang:2 retrieve:2 updated:1 target:30 rinen:1 caption:51 us:3 premraj:1 samy:2 agreement:1 trick:1 satisfying:1 recognition:8 updating:1 observed:3 steven:1 preprint:8 wang:1 capture:1 region:1 ensures:1 gutmann:1 decrease:1 removed:1 overfitted:1 trade:1 ran:1 substantial:2 intuition:2 balanced:1 yk:1 dhlin:1 reward:1 asked:1 mohsen:1 deva:1 ali:1 flying:2 distinctive:9 upon:1 baseball:3 toilet:2 yuille:1 sink:1 easily:2 joint:1 various:1 alphabet:2 describe:3 effective:5 detected:1 query:1 tell:3 youssef:1 jianfeng:1 choosing:1 kevin:3 zemel:1 quite:2 cvpr:1 rennie:1 encoder:2 favor:1 statistic:2 ability:3 ward:1 simonyan:1 itself:2 online:4 sequence:2 indication:1 net:2 propose:3 relevant:1 combining:2 ablation:1 date:1 motorcycle:2 shaking:1 papineni:1 margaret:2 grf:1 description:18 amin:1 rashtchian:1 jing:1 captioning:41 comparative:1 adam:1 leave:1 incremental:1 object:5 help:1 generating:3 recurrent:1 alon:1 andrew:1 nearest:2 progress:3 eq:4 strong:2 auxiliary:4 resemble:1 involves:2 implies:1 safe:1 ning:1 closely:1 meteor:8 attribute:3 annotated:1 subsequently:1 human:3 vaibhava:1 fix:1 generalization:3 wall:1 tighter:1 ryan:1 comprehension:1 exploring:1 sufficiently:1 considered:1 ground:5 koch:1 exp:1 lawrence:3 rgen:1 optimizer:1 early:2 adopt:1 ctm:1 estimation:7 ruslan:1 label:2 ross:2 largest:1 vice:1 clearly:3 cider:15 pn:16 zhou:1 varying:1 derived:1 l0:1 improvement:5 consistently:1 likelihood:4 indicates:6 hk:2 pretrain:1 adversarial:2 attains:1 baseline:4 detect:1 kim:2 mrfs:1 rigid:1 lowercase:1 xtm:1 rupesh:1 perona:1 i1:4 semantics:1 tao:1 rerun:1 overall:11 among:6 issue:4 unk:1 denoted:4 stateof:1 development:1 art:11 field:2 saurabh:2 beach:1 sampling:2 piotr:2 identical:1 represents:4 park:2 look:1 icml:1 filling:1 promote:3 fcn:2 siqi:1 others:1 yoshua:1 richard:2 employ:2 randomly:2 murphy:3 replaced:2 replacement:2 maintain:1 microsoft:1 attempt:2 detection:2 highly:1 investigate:1 evaluation:7 saturated:1 alignment:1 light:1 encourage:3 itm:1 tree:1 logarithm:1 desired:2 guidance:1 girshick:1 babytalk:1 conducted:2 characterize:1 reported:2 cho:1 referring:1 thanks:1 st:1 lstm:3 density:2 lush:1 ie:2 river:3 international:1 probabilistic:1 off:1 enhance:1 michael:3 yao:1 topk:1 settled:1 containing:1 choose:2 huang:1 emojis:3 li:4 account:1 converted:1 sec:3 forsyth:1 explicitly:5 tsung:1 performed:1 try:2 lot:1 jason:1 doing:2 overfits:1 red:2 competitive:1 start:1 il:8 air:1 accuracy:1 convolutional:1 reserved:2 maximized:1 yield:1 identify:1 ensemble:1 yew:1 yellow:4 serge:1 generalize:1 lu:1 bunch:2 worth:2 published:3 ago:1 history:1 submitted:1 corso:1 tamara:2 james:1 resultant:1 sampled:3 gain:3 dataset:5 popular:2 mitchell:2 recall:7 ut:3 improves:11 subtle:1 back:3 appears:1 higher:2 originally:1 follow:1 zisserman:1 improved:3 wei:1 formulation:5 furthermore:1 just:1 correlation:2 clock:2 hand:3 tacl:1 lstms:1 lack:2 google:3 logistic:3 quality:7 ordonez:2 resemblance:3 riding:1 building:1 xiaodong:1 ye:1 usa:1 contain:2 true:1 concept:3 y2:1 former:2 kyunghyun:1 assigned:1 verify:1 fidler:1 i2:4 illustrated:1 semantic:3 during:1 self:10 encourages:2 inferior:1 game:1 unambiguous:1 unnormalized:1 hong:3 chin:1 mohammad:1 julia:1 tn:9 performs:1 image:89 recently:1 parikh:4 common:2 empirically:2 rl:4 volume:2 discussed:1 he:1 association:2 dahua:2 significant:6 refer:3 composition:1 versa:1 mistaken:1 automatic:2 mroueh:1 pm:21 language:3 sanja:1 stable:4 tennis:4 supervision:1 ct0:2 add:1 base:13 feb:1 sergio:1 posterior:2 recent:4 showed:2 retrieved:1 forcing:1 coco:5 skateboard:1 certain:1 server:4 schmidhuber:1 hay:1 binary:1 meeting:1 yi:1 dai:2 greater:1 bathroom:2 deng:1 goel:1 paradigm:1 cuhk:3 itn:1 maximize:1 ii:3 branch:1 multiple:3 full:3 alan:1 technical:2 jet:2 long:2 lin:4 retrieval:10 c40:4 fighter:2 post:2 mle:17 equally:1 paired:1 variant:1 regression:1 aapo:1 vision:12 metric:6 essentially:1 arxiv:16 sometimes:1 grounded:1 represent:1 girish:1 achieved:2 hochreiter:1 c1:2 addition:3 background:1 source:1 standpoint:1 crucial:1 probably:1 tend:1 contrary:1 near:1 noting:2 constraining:1 split:2 easy:3 bengio:3 affect:1 perfectly:1 opposite:1 idea:2 tm:8 devlin:1 court:2 knowing:1 peter:1 karen:1 action:1 deep:3 polina:1 involve:2 listed:5 karpathy:1 nic:3 category:1 generate:5 estimated:1 track:1 blue:1 diverse:3 write:1 group:1 key:1 achieving:1 drawn:1 pietro:1 dhar:1 fraction:1 year:4 nce:10 run:9 package:1 parameterized:1 powerful:1 you:2 raquel:1 extends:1 chunseong:1 forrest:1 missed:1 bound:5 ct:18 followed:1 distinguish:1 courville:1 replaces:1 annual:1 constraint:11 throwing:1 fei:2 x2:1 personalized:1 toshev:2 aspect:4 u1:1 optimality:1 performing:1 format:2 department:1 zhaowen:1 hanging:2 ball:2 lavie:1 across:5 describes:1 smaller:1 character:2 pan:1 wi:6 hockenmaier:1 making:1 modification:1 jas:2 intuitively:1 taken:1 ceiling:2 ln:9 jiasen:1 describing:1 turn:1 mechanism:4 adopted:2 available:1 doll:2 generic:3 indirectly:1 caiming:1 distinguished:1 xiong:1 alternative:2 original:3 top:9 running:1 ensure:1 linguistics:2 gan:11 graphical:1 maintaining:5 etienne:1 giving:1 restrictive:1 ting:1 chinese:1 especially:2 build:1 classical:1 objective:4 parametric:1 rt:1 gradient:1 majority:1 decoder:3 topic:3 seven:1 argue:1 consensus:1 urtasun:1 reason:1 bleu:2 besides:1 ratio:3 difficult:2 mostly:1 skate:1 hao:1 negative:13 ba:1 policy:2 summarization:1 upper:1 datasets:6 benchmark:2 jin:1 introspective:1 truncated:1 situation:1 saturates:1 y1:1 jiebo:1 overlooked:2 introduced:3 david:1 pair:21 namely:4 sentence:1 learned:21 boost:1 barcelona:1 nip:1 able:3 usually:3 pattern:8 appeared:1 including:5 memory:3 green:1 natural:7 ranked:3 difficulty:1 meantime:1 indicator:1 boat:2 zhu:2 scheme:1 improve:2 sadeghi:1 picture:1 text:2 prior:1 understanding:1 relative:3 loss:2 cesc:1 bear:1 mixed:1 generation:5 msm:2 mounted:1 facing:1 ramakrishna:2 generator:1 validation:5 kuznetsova:1 sufficient:2 story:1 playing:1 roukos:1 collaboration:1 translation:2 row:2 summary:1 supported:2 copy:1 infeasible:1 side:1 mismatched:3 wide:1 distinctiveness:27 template:1 neighbor:2 characterizing:1 vocabulary:3 c5:1 reinforcement:1 adaptive:2 employing:1 ec:1 kishore:1 transaction:1 erhan:1 emphasize:1 active:1 reveals:1 sequentially:1 cyrus:1 it0:2 belongie:1 sentinel:1 vedantam:2 decade:1 table:19 learn:3 ca:1 career:1 forest:1 cl:59 european:2 zitnick:3 significance:1 main:1 whole:1 noise:4 motivation:1 big:1 junhua:1 qiu:1 x1:1 positively:1 ensembling:1 xu:2 parker:1 mscoco:10 fails:1 position:1 guiding:1 wish:1 mao:1 learns:1 young:1 saturate:2 emphasizing:1 choi:2 dumitru:1 specific:3 xt:1 explored:2 list:9 instagram:1 gupta:2 workshop:1 socher:1 ci:2 zhenhai:1 conditioned:2 illustrates:2 margin:4 chen:1 likely:2 gao:1 devi:4 visual:6 vinyals:1 iandola:1 partially:1 bo:2 pretrained:1 applies:1 srivastava:1 aa:13 springer:2 truth:5 conditional:6 goal:3 formulated:1 careful:1 towards:2 replace:5 man:2 content:1 hard:5 vicente:2 checkpoint:1 specifically:10 except:3 partly:2 player:2 zone:1 indicating:1 aaron:1 berg:3 people:5 latter:2 jonathan:1 alexander:3 kulkarni:1 oriol:1 tested:2 correlated:1 |
6,292 | 6,692 | Safe Model-based Reinforcement Learning with
Stability Guarantees
Felix Berkenkamp
Department of Computer Science
ETH Zurich
[email protected]
Matteo Turchetta
Department of Computer Science,
ETH Zurich
[email protected]
Angela P. Schoellig
Institute for Aerospace Studies
University of Toronto
[email protected]
Andreas Krause
Department of Computer Science
ETH Zurich
[email protected]
Abstract
Reinforcement learning is a powerful paradigm for learning optimal policies from
experimental data. However, to find optimal policies, most reinforcement learning
algorithms explore all possible actions, which may be harmful for real-world systems. As a consequence, learning algorithms are rarely applied on safety-critical
systems in the real world. In this paper, we present a learning algorithm that
explicitly considers safety, defined in terms of stability guarantees. Specifically,
we extend control-theoretic results on Lyapunov stability verification and show
how to use statistical models of the dynamics to obtain high-performance control
policies with provable stability certificates. Moreover, under additional regularity
assumptions in terms of a Gaussian process prior, we prove that one can effectively
and safely collect data in order to learn about the dynamics and thus both improve
control performance and expand the safe region of the state space. In our experiments, we show how the resulting algorithm can safely optimize a neural network
policy on a simulated inverted pendulum, without the pendulum ever falling down.
1
Introduction
While reinforcement learning (RL, [1]) algorithms have achieved impressive results in games, for
example on the Atari platform [2], they are rarely applied to real-world physical systems (e.g., robots)
outside of academia. The main reason is that RL algorithms provide optimal policies only in the
long-term, so that intermediate policies may be unsafe, break the system, or harm their environment.
This is especially true in safety-critical systems that can affect human lives. Despite this, safety in RL
has remained largely an open problem [3].
Consider, for example, a self-driving car. While it is desirable for the algorithm that drives the
car to improve over time (e.g., by adapting to driver preferences and changing environments), any
policy applied to the system has to guarantee safe driving. Thus, it is not possible to learn about the
system through random exploratory actions, which almost certainly lead to a crash. In order to avoid
this problem, the learning algorithm needs to consider its ability to safely recover from exploratory
actions. In particular, we want the car to be able to recover to a safe state, for example, driving at a
reasonable speed in the middle of the lane. This ability to recover is known as asymptotic stability
in control theory [4]. Specifically, we care about the region of attraction of the closed-loop system
under a policy. This is a subset of the state space that is forward invariant so that any state trajectory
that starts within this set stays within it for all times and converges to a goal state eventually.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
In this paper, we present a RL algorithm for continuous state-action spaces that provides these kind
of high-probability safety guarantees for policies. In particular, we show how, starting from an initial,
safe policy we can expand our estimate of the region of attraction by collecting data inside the safe
region and adapt the policy to both increase the region of attraction and improve control performance.
Related work Safety is an active research topic in RL and different definitions of safety exist [5, 6].
Discrete Markov decision processes (MDPs) are one class of tractable models that have been analyzed.
In risk-sensitive RL, one specifies risk-aversion in the reward [7]. For example, [8] define risk as
the probability of driving the agent to a set of known, undesirable states. Similarly, robust MDPs
maximize rewards when transition probabilities are uncertain [9, 10]. Both [11] and [12] introduce
algorithms to safely explore MDPs so that the agent never gets stuck without safe actions. All these
methods require an accurate probabilistic model of the system.
In continuous state-action spaces, model-free policy search algorithms have been successful. These
update policies without a system model by repeatedly executing the same task [13]. In this setting, [14] introduces safety guarantees in terms of constraint satisfaction that hold in expectation.
High-probability worst-case safety guarantees are available for methods based on Bayesian optimization [15] together with Gaussian process models (GP, [16]) of the cost function. The algorithms
in [17] and [18] provide high-probability safety guarantees for any parameter that is evaluated on
the real system. These methods are used in [19] to safely optimize a parametric control policy on a
quadrotor. However, resulting policies are task-specific and require the system to be reset.
In the model-based RL setting, research has focused on safety in terms of state constraints. In
[20, 21], a priori known, safe global backup policies are used, while [22] learns to switch between
several safe policies. However, it is not clear how one may find these policies in the first place.
Other approaches use model predictive control with constraints, a model-based technique where the
control actions are optimized online. For example, [23] models uncertain environmental constraints,
while [24] uses approximate uncertainty propagation of GP dynamics along trajectories. In this
setting, robust feasability and constraint satisfaction can be guaranteed for a learned model with
bounded errors using robust model predictive control [25]. The method in [26] uses reachability
analysis to construct safe regions in the state space. The theoretical guarantees depend on the solution
to a partial differential equation, which is approximated.
Theoretical guarantees for the stability exist for the more tractable stability analysis and verification
under a fixed control policy. In control, stability of a known system can be verified using a Lyapunov
function [27]. A similar approach is used by [28] for deterministic, but unknown dynamics that are
modeled as a GP, which allows for provably safe learning of regions of attraction for fixed policies.
Similar results are shown in [29] for stochastic systems that are modeled as a GP. They use Bayesian
quadrature to compute provably accurate estimates of the region of attraction. These approaches do
not update the policy.
Our contributions We introduce a novel algorithm that can safely optimize policies in continuous
state-action spaces while providing high-probability safety guarantees in terms of stability. Moreover,
we show that it is possible to exploit the regularity properties of the system in order to safely learn
about the dynamics and thus improve the policy and increase the estimated safe region of attraction
without ever leaving it. Specifically, starting from a policy that is known to stabilize the system
locally, we gather data at informative, safe points and improve the policy safely based on the improved
model of the system and prove that any exploration algorithm that gathers data at these points reaches
a natural notion of full exploration. We show how the theoretical results transfer to a practical
algorithm with safety guarantees and apply it to a simulated inverted pendulum stabilization task.
2
Background and Assumptions
We consider a deterministic, discrete-time dynamic system
xt+1 = f (xt , ut ) = h(xt , ut ) + g(xt , ut ),
(1)
with states x ? X ? Rq and control actions u ? U ? Rp and a discrete time index t ? N. The true
dynamics f : X ? U ? X consist of two parts: h(xt , ut ) is a known, prior model that can be
obtained from first principles, while g(xt , ut ) represents a priori unknown model errors. While the
model errors are unknown, we can obtain noisy measurements of f (x, u) by driving the system to
the state x and taking action u. We want this system to behave in a certain way, e.g., the car driving
2
on the road. To this end, we need to specify a control policy ? : X ? U that, given the current state,
determines the appropriate control action that drives the system to some goal state, which we set as
the origin without loss of generality [4]. We encode the performance requirements of how to drive
the system to the origin through a positive cost r(x, u) that is associated with states and actions and
has r(0, 0) = 0. The policy aims to minimize the cumulative, discounted costs for each starting state.
The goal is to safely learn about the dynamics from measurements and adapt the policy for performance, without encountering system failures. Specifically, we define the safety constraint on the
state divergence that occurs when leaving the region of attraction. This means that adapting the
policy is not allowed to decrease the region of attraction and exploratory actions to learn about the
dynamics f (?) are not allowed to drive the system outside the region of attraction. The region of
attraction is not known a priori, but is implicitly defined through the system dynamics and the choice
of policy. Thus, the policy not only defines performance as in typical RL, but also determines safety
and where we can obtain measurements.
Model assumptions In general, this kind of safe learning is impossible without further assumptions.
For example, in a discontinuous system even a slight change in the control policy can lead to drastically
different behavior. Moreover, to expand the safe set we need to generalize learned knowledge about
the dynamics to (potentially unsafe) states that we have not visited. To this end, we restrict ourselves
to the general and practically relevant class of models that are Lipschitz continuous. This is a typical
assumption in the control community [4]. Additionally, to ensure that the closed-loop system remains
Lipschitz continuous when the control policy is applied, we restrict policies to the rich class of
L? -Lipschitz continuous functions ?L , which also contains certain types of neural networks [30].
Assumption 1 (continuity). The dynamics h(?) and g(?) in (1) are Lh - and Lg Lipschitz continuous
with respect to the 1-norm. The considered control policies ? lie in a set ?L of functions that
are L? -Lipschitz continuous with respect to the 1-norm.
To enable safe learning, we require a reliable statistical model. While we commit to GPs for the
exploration analysis, for safety any suitable, well-calibrated model is applicable.
Assumption 2 (well-calibrated model). Let ?n (?) and ?n (?) denote the posterior mean and covariance matrix functions of the statistical model of the dynamics (1) conditioned on n noisy measurements.
1/2
With ?n (?) = trace(?n (?)), there exists a ?n > 0 such that with probability at least (1 ? ?) it holds
for all n ? 0, x ? X , and u ? U that kf (x, u) ? ?n (x, u)k1 ? ?n ?n (x, u).
This assumption ensures that we can build confidence intervals on the dynamics that, when scaled by
an appropriate constant ?n , cover the true function with high probability. We introduce a specific
statistical model that fulfills both assumptions under certain regularity assumptions in Sec. 3.
Lyapunov function To satisfy the specified safety constraints for safe learning, we require a tool
to determine whether individual states and actions are safe. In control theory, this safety is defined
through the region of attraction, which can be computed for a fixed policy using Lyapunov functions [4]. Lyapunov functions are continuously differentiable functions v : X ? R?0 with v(0) = 0
and v(x) > 0 for all x ? X \ {0}. The key idea behind using Lyapunov functions to show stability
of the system (1) is similar to that of gradient descent on strictly quasiconvex functions: if one can
show that, given a policy ?, applying the dynamics f on the state maps it to strictly smaller values
on the Lyapunov function (?going downhill?), then the state eventually converges to the equilibrium
point at the origin (minimum). In particular, the assumptions in Theorem 1 below imply that v is
strictly quasiconvex within the region of attraction if the dynamics are Lipschitz continuous. As a
result, the one step decrease property for all states within a level set guarantees eventual convergence
to the origin.
Theorem 1 ([4]). Let v be a Lyapunov function, f Lipschitz continuous dynamics, and ? a policy. If
v(f (x, ?(x))) < v(x) for all x within the level set V(c) = {x ? X \ {0} | v(x) ? c}, c > 0, then
V(c) is a region of attraction, so that x0 ? V(c) implies xt ? V(c) for all t > 0 and limt?? xt = 0.
It is convenient to characterize the region of attraction through a level set of the Lyapunov function,
since it replaces the challenging test for convergence with a one-step decrease condition on the
Lyapunov function. For the theoretical analysis in this paper, we assume that a Lyapunov function is
given to determine the region of attraction. For ease of notation, we also assume ?v(x)/?x 6= 0 for
all x ? X \ 0, which ensures that level sets V(c) are connected if c > 0. Since Lyapunov functions
are continuously differentiable, they are Lv -Lipschitz continuous over the compact set X .
3
In general, it is not easy to find suitable Lyapunov functions. However, for physical models, like the
prior model h in (1), the energy of the system (e.g., kinetic and potential for mechanical systems) is a
good candidate Lyapunov function. Moreover, it has recently been shown that it is possible to compute
suitable Lyapunov functions [31, 32]. In our experiments, we exploit the fact that value functions in
RL are Lyapunov functions if the costs are strictly positive away from the origin. This follows directly
from the definition of the value function, where v(x) = r(x, ?(x)) + v(f (x, ?(x)) ? v(f (x, ?(x))).
Thus, we can obtain Lyapunov candidates as a by-product of approximate dynamic programming.
Initial safe policy Lastly, we need to ensure that there exists a safe starting point for the learning
process. Thus, we assume that we have an initial policy ?0 that renders the origin of the system in (1)
asymptotically stable within some small set of states S0x . For example, this policy may be designed
using the prior model h in (1), since most models are locally accurate but deteriorate in quality as
state magnitude increases. This policy is explicitly not safe to use throughout the state space X \ S0x .
3
Theory
In this section, we use these assumptions for safe reinforcement learning. We start by computing the
region of attraction for a fixed policy under the statistical model. Next, we optimize the policy in order
to expand the region of attraction. Lastly, we show that it is possible to safely learn about the dynamics
and, under additional assumptions about the model and the system?s reachability properties, that this
approach expands the estimated region of attraction safely. We consider an idealized algorithm that is
amenable to analysis, which we convert to a practical variant in Sec. 4. See Fig. 1 for an illustrative
run of the algorithm and examples of the sets defined below.
Region of attraction We start by computing the region of attraction for a fixed policy. This is an
extension of the method in [28] to discrete-time systems. We want to use the Lyapunov decrease condition in Theorem 1 to guarantee safety for the statistical model of the dynamics. However, the posterior
uncertainty in the statistical model of the dynamics means that one step predictions about v(f (?)) are
uncertain too. We account for this by constructing high-probability confidence intervals on v(f (x, u)):
Qn (x, u) := [v(?n?1 (x, u)) ? Lv ?n ?n?1 (x, u)]. From Assumption 2 together with the Lipschitz
property of v, we know that v(f (x, u)) is contained in Qn (x, u) with probability at least (1 ? ?). For
our exploration analysis, we need to ensure that safe state-actions cannot become unsafe; that is, an
initial set of safe set S0 remains safe (defined later). To this end, we intersect the confidence intervals:
Cn (x, u) := Cn?1 ? Qn (x, u), where the set C is initialized to C0 (x, u) = (??, v(x) ? L?v ? )
when (x, u) ? S0 and C0 (x, u) = R otherwise. Note that v(f (x, u)) is contained in Cn (x, u) with
the same (1 ? ?) probability as in Assumption 2. The upper and lower bounds on v(f (?)) are defined
as un (x, u) := max Cn (x, u) and ln (x, u) := min Cn (x, u).
Given these high-probability confidence intervals, the system is stable according to Theorem 1 if
v(f (x, u)) ? un (x) < v(x) for all x ? V(c). However, it is intractable to verify this condition
directly on the continuous domain without additional, restrictive assumptions about the model.
Instead, we consider a discretization of the state space X? ? X into cells, so that kx ? [x]? k1 ? ?
holds for all x ? X . Here, [x]? denotes the point in X? with the smallest l1 distance to x. Given this
discretization, we bound the decrease variation on the Lyapunov function for states in X? and use the
Lipschitz continuity to generalize to the continuous state space X .
Theorem 2. Under Assumptions 1 and 2 with L?v := Lv Lf (L? + 1) + Lv , let X? be a discretization of X such that kx ? [x]? k1 ? ? for all x ? X . If, for all x ? V(c) ? X? with c > 0, u = ?(x),
and for some n ? 0 it holds that un (x, u) < v(x) ? L?v ?, then v(f (x, ?(x))) < v(x) holds for all
x ? V(c) with probability at least (1 ? ?) and V(c) is a region of attraction for (1) under policy ?.
The proof is given in Appendix A.1. Theorem 2 states that, given confidence intervals on the statistical
model of the dynamics, it is sufficient to check the stricter decrease condition in Theorem 2 on the
discretized domain X? to guarantee the requirements for the region of attraction in the continuous
domain in Theorem 1. The bound in Theorem 2 becomes tight as the discretization constant ?
and |v(f (?)) ? un (?)| go to zero. Thus, the discretization constant trades off computation costs for
accuracy, while un approaches v(f
? (?)) as we obtain more measurement data and the posterior model
uncertainty about the dynamics, ?n ?n decreases. The confidence intervals on v(f (x, ?(x)) ? v(x)
and the corresponding estimated region of attraction (red line) can be seen in the bottom half of Fig. 1.
Policy optimization So far, we have focused on estimating the region of attraction for a fixed policy.
Safety is a property of states under a fixed policy. This means that the policy directly determines
4
?30
?15
policy ?0
? (x, u)
S30
V(c15 )
u0
?L?v ?
l0
state x
(a) Initial safe set (in red).
state x
(b) Exploration: 15 data points.
state x
?v(x, ?(x))
v(x)
action u
D30
(c) Final policy after 30 evaluations.
Figure 1: Example application of Algorithm 1. Due to input constraints, the system becomes unstable
for large states. We start from an initial, local policy ?0 that has a small, safe region of attraction (red
lines) in Fig. 1(a). The algorithm selects safe, informative state-action pairs within Sn (top, white
shaded), which can be evaluated without leaving the region of attraction V(cn ) (red lines) of the
current policy ?n . As we gather more data (blue crosses), the uncertainty in the model decreases
(top, background) and we use (3) to update the policy so that it lies within Dn (top, red shaded) and
fulfills the Lyapunov decrease condition. The algorithm converges to the largest safe set in Fig. 1(c).
It improves the policy without evaluating unsafe state-action pairs and thereby without system failure.
which states are safe. Specifically, to form a region of attraction all states in the discretizaton X?
within a level set of the Lyapunov function need to fulfill the decrease condition in Theorem 2 that
depends on the policy choice. The set of all state-action pairs that fulfill this decrease condition is
given by
Dn = (x, u) ? X? ? U | un (x, u) ? v(x) < ?L?v ? ,
(2)
see Fig. 1(c) (top, red shaded). In order to estimate the region of attraction based on this set, we
need to commit to a policy. Specifically, we want to pick the policy that leads to the largest possible
region of attraction according to Theorem 2. This requires that for each discrete state in X? the
corresponding state-action pair under the policy must be in the set Dn . Thus, we optimize the policy
according to
?n , cn =
argmax c,
???L ,c?R>0
such that for all x ? V(c) ? X? : (x, ?(x)) ? Dn .
(3)
The region of attraction that corresponds to the optimized policy ?n according to (3) is given
by V(cn ), see Fig. 1(b). It is the largest level set of the Lyapunov function for which all state-action
pairs (x, ?n (x)) that correspond to discrete states within V(cn ) ? X? are contained in Dn . This
means that these state-action pairs fulfill the requirements of Theorem 2 and V(cn ) is a region of
attraction of the true system under policy ?n . The following theorem is thus a direct consequence
of Theorem 2 and (3).
Theorem 3. Let R?n be the true region of attraction of (1) under the policy ?n . For any ? ? (0, 1),
we have with probability at least (1 ? ?) that V(cn ) ? R?n for all n > 0.
Thus, when we optimize the policy subject to the constraint in (3) the estimated region of attraction is
always an inner approximation of the true region of attraction. However, solving the optimization
problem in (3) is intractable in general. We approximate the policy update step in Sec. 4.
Collecting measurements Given these stability guarantees, it is natural to ask how one might obtain
data points in order to improve the model of g(?) and thus efficiently increase the region of attraction.
This question is difficult to answer in general, since it depends on the property of the statistical model.
In particular, for general statistical models it is often not clear whether the confidence intervals
contract sufficiently quickly. In the following, we make additional assumptions about the model and
reachability within V(cn ) in order to provide exploration guarantees. These assumptions allow us to
highlight fundamental requirements for safe data acquisition and that safe exploration is possible.
5
We assume that the unknown model errors g(?) have bounded norm in a reproducing kernel Hilbert
space (RKHS, [33]) corresponding to a differentiable
kernel k, kg(?)kk ? Bg . These are a class of
P?
well-behaved functions of the form g(z) = i=0 ?i k(zi , z) defined through representer points zi
and weights ?i that decay sufficiently fast with i. This assumption ensures
p that g satisfies the
Lipschitz property in Assumption 1, see [28]. Moreover, with ?n = Bg + 4? ?n + 1 + ln(1/?) we
can use GP models for the dynamics that fulfill Assumption 2 if the state if fully observable and the
measurement noise is ?-sub-Gaussian (e.g., bounded in [??, ?]), see [34]. Here ?n is the information
capacity. It corresponds to the amount of mutual information that can be obtained about g from nq
measurements, a measure of the size of the function class encoded by the model. The information
capacity has a sublinear dependence on n for common kernels and upper bounds can be computed
efficiently [35]. More details about this model are given in Appendix A.2.
In order to quantify the exploration properties of our algorithm, we consider a discrete action
space U? ? U. We define exploration as the number of state-action pairs in X? ? U? that we can
safely learn about without leaving the true region of attraction. Note that despite this discretization,
the policy takes values on the continuous domain. Moreover, instead of using the confidence intervals
directly as in (3), we consider an algorithm that uses the Lipschitz constants to slowly expand the safe
set. We use this in our analysis to quantify the ability to generalize beyond the current safe set. In
practice, nearby states are sufficiently correlated under the model to enable generalization using (2).
Suppose we are given a set S0 of state-action pairs about which we can learn safely. Specifically, this
means that we have a policy such that, for any state-action pair (x, u) in S0 , if we apply action u in
state x and then apply actions according to the policy, the state converges to the origin. Such a set
can be constructed using the initial policy ?0 from Sec. 2 as S0 = {(x, ?0 (x)) | x ? S0x }. Starting
from this set, we want to update the policy to expand the region of attraction according to Theorem 2.
To this end, we use the confidence intervals on v(f (?)) for states inside S0 to determine state-action
pairs that fulfill the decrease condition. We thus redefine Dn for the exploration analysis to
[
0
Dn =
z ? X? ? U? | un (x, u) ? v(x) + L?v kz0 ? (x, u)k1 < ?L?v ? . (4)
(x,u)?Sn?1
This formulation is equivalent to (2), except that it uses the Lipschitz constant to generalize safety.
Given Dn , we can again find a region of attraction V(cn ) by committing to a policy according to (3).
In order to expand this region of attraction effectively we need to decrease the posterior model
uncertainty about the dynamics of the GP by collecting measurements. However, to ensure safety
as outlined in Sec. 2, we are not only restricted to states within V(cn ), but also need to ensure that
the state after taking an action is safe; that is, the dynamics map the state back into the region of
attraction V(cn ). We again use the Lipschitz constant in order to determine this set,
[
Sn =
z0 ? V(cn ) ? X? ? U? | un (z) + Lv Lf kz ? z0 k1 ? cn }.
(5)
z?Sn?1
The set Sn contains state-action pairs that we can safely evaluate under the current policy ?n without
leaving the region of attraction, see Fig. 1 (top, white shaded).
What remains is to define a strategy for collecting data points within Sn to effectively decrease model
uncertainty. We specifically focus on the high-level requirements for any exploration scheme without
committing to a specific method. In practice, any (model-based) exploration strategy that aims to
decrease model uncertainty by driving the system to specific states may be used. Safety can be
ensured by picking actions according to ?n whenever the exploration strategy reaches the boundary
of the safe region V(cn ); that is, when un (x, u) > cn . This way, we can use ?n as a backup policy
for exploration.
The high-level goal of the exploration strategy is to shrink the confidence intervals at state-action
pairs Sn in order to expand the safe region. Specifically, the exploration strategy should aim to visit
state-action pairs in Sn at which we are the most uncertain about the dynamics; that is, where the
confidence interval is the largest:
(xn , un ) = argmax un (x, u) ? ln (x, u).
(6)
(x,u)?Sn
As we keep collecting data points according to (6), we decrease the uncertainty about the dynamics
for different actions throughout the region of attraction and adapt the policy, until eventually we
6
Algorithm 1 S AFE LYAPUNOV L EARNING
1: Input: Initial safe policy ?0 , dynamics model GP(?(z), k(z, z0 ))
2: for all n = 1, . . . do
3:
Compute policy ?n via SGD on (7)
4:
cn = argmaxc c, such that ?x ? V(cn ) ? X? : un (x, ?n (x)) ? v(x) < ?L?v ?
5:
Sn = {(x, u) ? V(cn ) ? U? | un (x, u) ? cn }
6:
Select (xn , un ) within Sn using (6) and drive system there with backup policy ?n
7:
Update GP with measurements f (xn , un ) + n
have gathered enough information in order to expand it. While (6) implicitly assumes that any
state within V(cn ) can be reached by the exploration policy, it achieves the high-level goal of any
exploration algorithm that aims to reduce model uncertainty. In practice, any safe exploration scheme
is limited by unreachable parts of the state space.
We compare the active learning scheme in (6) to an oracle baseline that starts from the same initial
safe set S0 and knows v(f (x, u)) up to accuracy within the safe set. The oracle also uses knowledge
about the Lipschitz constants and the optimal policy in ?L at each iteration. We denote the set that this
baseline manages to determine as safe with R (S0 ) and provide a detailed definition in Appendix A.3.
Theorem 4. Assume ?-sub-Gaussian measurement noise and that the model error g(?)
in (1) has RKHSpnorm smaller than Bg .
Under the assumptions of Theorem 2,
with ?n = Bg + 4? ?n + 1 + ln(1/?), and with measurements collected according to (6), let
?
0 )|+1)
n? be the smallest positive integer so that ? 2 n? ? ? Cq(|R(S
where C = 8/ log(1 + ? ?2 ).
L2v 2
n? n
Let R? be the true region of attraction of (1) under a policy ?. For any > 0, and ? ? (0, 1), the
following holds jointly with probability at least (1 ? ?) for all n > 0:
(i) V(cn ) ? R?n
(ii) f (x, u) ? R?n ?(x, u) ? Sn .
(iii) R (S0 ) ? Sn ? R0 (S0 ).
Theorem 4 states that, when selecting data points according to (6), the estimated region of attraction V(cn ) is (i) contained in the true region of attraction under the current policy and (ii) selected
data points do not cause the system to leave the region of attraction. This means that any exploration
method that considers the safety constraint (5) is able to safely learn about the system without leaving
the region of attraction. The last part of Theorem 4, (iii), states that after a finite number of data
points n? we achieve at least the exploration performance of the oracle baseline, while we do not
classify unsafe state-action pairs as safe. This means that the algorithm explores the largest region
of attraction possible for a given Lyapunov function with residual uncertaint about v(f (?)) smaller
than . Details of the comparison baseline are given in the appendix. In practice, this means that any
exploration method that manages to reduce the maximal uncertainty about the dynamics within Sn is
able to expand the region of attraction.
An example run of repeatedly evaluating (6) for a one-dimensional state-space is shown in Fig. 1. It
can be seen that, by only selecting data points within the current estimate of the region of attraction,
the algorithm can efficiently optimize the policy and expand the safe region over time.
4
Practical Implementation and Experiments
In the previous section, we have given strong theoretical results on safety and exploration for an
idealized algorithm that can solve (3). In this section, we provide a practical variant of the theoretical
algorithm in the previous section. In particular, while we retain safety guarantees, we sacrifice
exploration guarantees to obtain a more practical algorithm. This is summarized in Algorithm 1.
The policy optimization problem in (3) is intractable to solve and only considers safety, rather
than a performance metric. We propose to use an approximate policy update that that maximizes
approximate performance while providing stability guarantees. It proceeds by optimizing the policy
first and then computes the region of attraction V(cn ) for the new, fixed policy. This does not
impact safety, since data is still only collected inside the region of attraction. Moreover, should the
optimization fail and the region of attraction decrease, one can always revert to the previous policy,
which is guaranteed to be safe.
7
V(c50 )
angle [rad]
angular velocity [rad/s]
0.30
5
V(c0 )
0
unsafe region
?5
safely optimized policy
initial policy
0.25
0.20
0.15
?0
0.10
?50
0.05
0.00
?1.0
?0.5
0.0
0.5
1.0
0.0
angle [rad]
0.5
1.0
1.5
time [s]
(a) Estimated safe set.
(b) State trajectory (lower is better).
Figure 2: Optimization results for an inverted pendulum. Fig. 2(a) shows the initial safe set (yellow)
under the policy ?0 , while the green region represents the estimated region of attraction under the
optimized neural network policy. It is contained within the true region of attraction (white). Fig. 2(b)
shows the improved performance of the safely learned policy over the policy for the prior model.
In our experiments, we use approximate dynamic programming [36] to capture the performance of the
policy. Given a policy ?? with parameters ?, we compute an estimate of the cost-to-go J?? (?) for the
mean dynamics ?n based on the cost r(x, u) ? 0. At each state, J?? (x) is the sum of ?-discounted
rewards encountered when following the policy ?? . The goal is to adapt the parameters of the policy
for minimum cost as measured by J?? , while ensuring that the safety constraint on the worst-case
decrease on the Lyapunov function in Theorem 2 is not violated. A Lagrangian formulation to this
constrained optimization problem is
X
?n = argmin
r(x, ?? (x)) + ?J?? (?n?1 (x, ?? (x)) + ? un (x, ?? (x)) ? v(x) + L?v ? , (7)
?? ??L
x?X?
where the first term measures long-term cost to go and ? ? 0 is a Lagrange multiplier for the safety
constraint from Theorem 2. In our experiments, we use the value function as a Lyapunov function
candidate, v = J with r(?, ?) ? 0, and set ? = 1. In this case, (7) corresponds to an high-probability
upper bound on the cost-to-go given the uncertainty in the dynamics. This is similar to worst-case
performance formulations found in robust MDPs [9, 10], which consider worst-case value functions
given parametric uncertainty in MDP transition model. Moreover, since L?v depends on the Lipschitz
constant of the policy, this simultaneously serves as a regularizer on the parameters ?.
To verify safety, we use the GP confidence intervals ln and un directly, as in (2). We also use
confidence to compute Sn for the active learning scheme, see Algorithm 1, Line 5. In practice, we
do not need to compute the entire set Sn to solve (3), but can use a global optimization method or
even a random sampling scheme within V(cn ) to find suitable state-actions. Moreover, measurements
for actions that are far away from the current policy are unlikely to expand V(cn ), see Fig. 1(c). As
we optimize (7) via gradient descent, the policy changes only locally. Thus, we can achieve better
data-efficiency by restricting the exploratory actions u with (x, u) ? Sn to be close to ?n , u ?
[?n (x) ? u
?, ?n (x) + u
?] for some constant u
?.
Computing the region of attraction by verifying the stability condition on a discretized domain suffers
from the curse of dimensionality. However, it is not necessary to update policies in real time. In
particular, since any policy that is returned by the algorithm is provably safe within some level set,
any of these policies can be used safely for an arbitrary number of time steps. To scale this method to
higher-dimensional system, one would have to consider an adaptive discretization for the verification
as in [27].
Experiments A Python implementation of Algorithm 1 and the experiments based on TensorFlow [37] and GPflow [38] is available at https://github.com/befelix/safe_learning.
We verify our approach on an inverted pendulum benchmark problem. The true, continuous-time
dynamics are given by ml2 ?? = gml sin(?) ? ??? + u, where ? is the angle, m the mass, g the
gravitational constant, and u the torque applied to the pendulum. The control torque is limited, so that
the pendulum necessarily falls down beyond a certain angle. We use a GP model for the discrete-time
dynamics, where the mean dynamics are given by a linearized and discretized model of the true
8
dynamics that considers a wrong, lower mass and neglects friction. As a result, the optimal policy for
the mean dynamics does not perform well and has a small region of attraction as it underactuates the
system. We use a combination of linear and Mat?rn kernels in order to capture the model errors that
result from parameter and integration errors.
For the policy, we use a neural network with two hidden layers and 32 neurons with ReLU activations
each. We compute a conservative estimate of the Lipschitz constant as in [30]. We use standard
approximate dynamic programming with a quadratic, normalized cost r(x, u) = xT Qx + uT Ru,
where Q and R are positive-definite, to compute the cost-to-go J?? . Specifically, we use a piecewiselinear triangulation of the state-space as to approximate J?? , see [39]. This allows us to quickly
verify the assumptions that we made about the Lyapunov function in Sec. 2 using a graph search. In
practice, one may use other function approximators. We optimize the policy via stochastic gradient
descent on (7).
The theoretical confidence intervals for the GP model are conservative. To enable more data-efficient
learning, we fix ?n = 2. This corresponds to a high-probability decrease condition per-state, rather
than jointly over the state space. Moreover, we use local Lipschitz constants of the Lyapunov function
rather than the global one. While this does not affect guarantees, it greatly speeds up exploration.
For the initial policy, we use approximate dynamic programming to compute the optimal policy for
the prior mean dynamics. This policy is unstable for large deviations from the initial state and has poor
performance, as shown in Fig. 2(b). Under this initial, suboptimal policy, the system is stable within
a small region of the state-space Fig. 2(a). Starting from this initial safe set, the algorithm proceeds to
collect safe data points and improve the policy. As the uncertainty about the dynamics decreases, the
policy improves and the estimated region of attraction increases. The region of attraction after 50 data
points is shown in Fig. 2(a). The resulting set V(cn ) is contained within the true safe region of the
optimized policy ?n . At the same time, the control performance improves drastically relative to the
initial policy, as can be seen in Fig. 2(b). Overall, the approach enables safe learning about dynamic
systems, as all data points collected during learning are safely collected under the current policy.
5
Conclusion
We have shown how classical reinforcement learning can be combined with safety constraints in terms
of stability. Specifically, we showed how to safely optimize policies and give stability certificates
based on statistical models of the dynamics. Moreover, we provided theoretical safety and exploration
guarantees for an algorithm that can drive the system to desired state-action pairs during learning. We
believe that our results present an important first step towards safe reinforcement learning algorithms
that are applicable to real-world problems.
Acknowledgments
This research was supported by SNSF grant 200020_159557, the Max Planck ETH Center for
Learning Systems, NSERC grant RGPIN-2014-04634, and the Ontario Early Researcher Award.
References
[1] Richard S. Sutton and Andrew G. Barto. Reinforcement learning: an introduction. MIT press,
1998.
[2] Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Andrei A. Rusu, Joel Veness, Marc G.
Bellemare, Alex Graves, Martin Riedmiller, Andreas K. Fidjeland, Georg Ostrovski, Stig Petersen, Charles Beattie, Amir Sadik, Ioannis Antonoglou, Helen King, Dharshan Kumaran, Daan
Wierstra, Shane Legg, and Demis Hassabis. Human-level control through deep reinforcement
learning. Nature, 518(7540):529?533, 2015.
[3] Dario Amodei, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Man?.
Concrete problems in AI safety. arXiv:1606.06565 [cs], 2016.
[4] Hassan K. Khalil and J. W. Grizzle. Nonlinear systems, volume 3. Prentice Hall, 1996.
9
[5] Martin Pecka and Tomas Svoboda. Safe exploration techniques for reinforcement learning ?
an overview. In Modelling and Simulation for Autonomous Systems, pages 357?375. Springer,
2014.
[6] Javier Garc?a and Fernando Fern?ndez. A comprehensive survey on safe reinforcement learning.
Journal of Machine Learning Research (JMLR), 16:1437?1480, 2015.
[7] Stefano P. Coraluppi and Steven I. Marcus. Risk-sensitive and minimax control of discrete-time,
finite-state Markov decision processes. Automatica, 35(2):301?309, 1999.
[8] Peter Geibel and Fritz Wysotzki. Risk-sensitive reinforcement learning applied to control under
constraints. J. Artif. Intell. Res.(JAIR), 24:81?108, 2005.
[9] Aviv Tamar, Shie Mannor, and Huan Xu. Scaling Up Robust MDPs by Reinforcement Learning.
In Proc. of the International Conference on Machine Learning (ICML), 2014.
[10] Wolfram Wiesemann, Daniel Kuhn, and Ber? Rustem. Robust Markov Decision Processes.
Mathematics of Operations Research, 38(1):153?183, 2012.
[11] Teodor Mihai Moldovan and Pieter Abbeel. Safe exploration in Markov decision processes. In
Proc. of the International Conference on Machine Learning (ICML), pages 1711?1718, 2012.
[12] Matteo Turchetta, Felix Berkenkamp, and Andreas Krause. Safe exploration in finite markov
decision processes with gaussian processes. pages 4305?4313, 2016.
[13] Jan Peters and Stefan Schaal. Policy gradient methods for robotics. In Proc. of the IEEE/RSJ
International Conference on Intelligent Robots and Systems, pages 2219?2225, 2006.
[14] Joshua Achiam, David Held, Aviv Tamar, and Pieter Abbeel. Constrained policy optimization.
In Proc. of the International Conference on Machine Learning (ICML), 2017.
[15] Jonas Mockus. Bayesian approach to global optimization, volume 37 of Mathematics and Its
Applications. Springer, Dordrecht, 1989.
[16] Carl Edward Rasmussen and Christopher K.I Williams. Gaussian processes for machine
learning. MIT Press, Cambridge MA, 2006.
[17] Jens Schreiter, Duy Nguyen-Tuong, Mona Eberts, Bastian Bischoff, Heiner Markert, and Marc
Toussaint. Safe exploration for active learning with Gaussian processes. In Machine Learning
and Knowledge Discovery in Databases, number 9286, pages 133?149. Springer International
Publishing, 2015.
[18] Yanan Sui, Alkis Gotovos, Joel W. Burdick, and Andreas Krause. Safe exploration for optimization with Gaussian processes. In Proc. of the International Conference on Machine Learning
(ICML), pages 997?1005, 2015.
[19] Felix Berkenkamp, Angela P. Schoellig, and Andreas Krause. Safe controller optimization for
quadrotors with Gaussian processes. In Proc. of the IEEE International Conference on Robotics
and Automation (ICRA), pages 493?496, 2016.
[20] J. Garcia and F. Fernandez. Safe exploration of state and action spaces in reinforcement learning.
Journal of Artificial Intelligence Research, pages 515?564, 2012.
[21] Alexander Hans, Daniel Schneega?, Anton Maximilian Sch?fer, and Steffen Udluft. Safe
exploration for reinforcement learning. In Proc. of the European Symposium on Artificial
Neural Networks (ESANN), pages 143?148, 2008.
[22] Theodore J. Perkins and Andrew G. Barto. Lyapunov design for safe reinforcement learning.
The Journal of Machine Learning Research, 3:803?832, 2003.
[23] Dorsa Sadigh and Ashish Kapoor. Safe control under uncertainty with Probabilistic Signal
Temporal Logic. In Proc. of Robotics: Science and Systems, 2016.
10
[24] Chris J. Ostafew, Angela P. Schoellig, and Timothy D. Barfoot. Robust constrained learningbased NMPC enabling reliable mobile robot path tracking. The International Journal of Robotics
Research (IJRR), 35(13):1547?1536, 2016.
[25] Anil Aswani, Humberto Gonzalez, S. Shankar Sastry, and Claire Tomlin. Provably safe and
robust learning-based model predictive control. Automatica, 49(5):1216?1226, 2013.
[26] Anayo K. Akametalu, Shahab Kaynama, Jaime F. Fisac, Melanie N. Zeilinger, Jeremy H.
Gillula, and Claire J. Tomlin. Reachability-based safe learning with Gaussian processes. In
Proc. of the IEEE Conference on Decision and Control (CDC), pages 1424?1431, 2014.
[27] Ruxandra Bobiti and Mircea Lazar. A sampling approach to finding Lyapunov functions for
nonlinear discrete-time systems. In Proc. of the European Control Conference (ECC), pages
561?566, 2016.
[28] Felix Berkenkamp, Riccardo Moriconi, Angela P. Schoellig, and Andreas Krause. Safe learning
of regions of attraction in nonlinear systems with Gaussian processes. In Proc. of the Conference
on Decision and Control (CDC), pages 4661?4666, 2016.
[29] Julia Vinogradska, Bastian Bischoff, Duy Nguyen-Tuong, Henner Schmidt, Anne Romer, and
Jan Peters. Stability of controllers for Gaussian process forward models. In Proceedings of the
International Conference on Machine Learning (ICML), pages 545?554, 2016.
[30] Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, and Rob Fergus. Intriguing properties of neural networks. In Proc. of the International
Conference on Learning Representations (ICLR), 2014.
[31] Huijuan Li and Lars Gr?ne. Computation of local ISS Lyapunov functions for discrete-time systems via linear programming. Journal of Mathematical Analysis and Applications, 438(2):701?
719, 2016.
[32] Peter Giesl and Sigurdur Hafstein. Review on computational methods for Lyapunov functions.
Discrete and Continuous Dynamical Systems, Series B, 20(8):2291?2337, 2015.
[33] Bernhard Sch?lkopf. Learning with kernels: support vector machines, regularization, optimization, and beyond. Adaptive computation and machine learning. MIT Press, Cambridge, Mass,
2002.
[34] Sayak Ray Chowdhury and Aditya Gopalan. On kernelized multi-armed bandits. arXiv preprint
arXiv:1704.00445, 2017.
[35] Niranjan Srinivas, Andreas Krause, Sham M. Kakade, and Matthias Seeger. Gaussian Process
Optimization in the Bandit Setting: No Regret and Experimental Design. IEEE Transactions on
Information Theory, 58(5):3250?3265, 2012.
[36] Warren B. Powell. Approximate dynamic programming: solving the curses of dimensionality.
John Wiley & Sons, 2007.
[37] Mart?n Abadi, Ashish Agarwal, Paul Barham, Eugene Brevdo, Zhifeng Chen, Craig Citro,
Greg S. Corrado, Andy Davis, Jeffrey Dean, Matthieu Devin, Sanjay Ghemawat, Ian Goodfellow,
Andrew Harp, Geoffrey Irving, Michael Isard, Yangqing Jia, Rafal Jozefowicz, Lukasz Kaiser,
Manjunath Kudlur, Josh Levenberg, Dan Mane, Rajat Monga, Sherry Moore, Derek Murray,
Chris Olah, Mike Schuster, Jonathon Shlens, Benoit Steiner, Ilya Sutskever, Kunal Talwar, Paul
Tucker, Vincent Vanhoucke, Vijay Vasudevan, Fernanda Viegas, Oriol Vinyals, Pete Warden,
Martin Wattenberg, Martin Wicke, Yuan Yu, and Xiaoqiang Zheng. TensorFlow: Large-Scale
Machine Learning on Heterogeneous Distributed Systems. arXiv:1603.04467 [cs], 2016.
[38] Alexander G. de G. Matthews, Mark van der Wilk, Tom Nickson, Keisuke Fujii, Alexis
Boukouvalas, Pablo Le?n-Villagr?, Zoubin Ghahramani, and James Hensman. GPflow: a
Gaussian process library using TensorFlow. Journal of Machine Learning Research, 18(40):1?6,
2017.
11
[39] Scott Davies. Multidimensional triangulation and interpolation for reinforcement learning. In
Proc. of the Conference on Neural Information Processing Systems (NIPS), pages 1005?1011,
1996.
[40] Andreas Christmann and Ingo Steinwart. Support Vector Machines. Information Science and
Statistics. Springer, New York, NY, 2008.
12
| 6692 |@word middle:1 norm:3 mockus:1 c0:3 open:1 pieter:2 simulation:1 linearized:1 covariance:1 jacob:1 schoellig:5 evaluating:2 pick:1 sgd:1 thereby:1 initial:16 ndez:1 series:1 contains:2 selecting:2 daniel:2 rkhs:1 steiner:1 current:8 discretization:7 com:1 anne:1 activation:1 intriguing:1 must:1 john:2 devin:1 academia:1 informative:2 christian:1 burdick:1 enables:1 designed:1 update:8 intelligence:1 selected:1 half:1 nq:1 amir:1 isard:1 keisuke:1 wolfram:1 provides:1 certificate:2 mannor:1 toronto:1 preference:1 fujii:1 wierstra:1 along:1 olah:2 dn:8 become:1 symposium:1 differential:1 jonas:1 abadi:1 yuan:1 prove:2 constructed:1 direct:1 driver:1 ray:1 redefine:1 inside:3 dan:2 introduce:3 x0:1 deteriorate:1 sacrifice:1 villagr:1 behavior:1 multi:1 steffen:1 discretized:3 torque:2 discounted:2 curse:2 armed:1 becomes:2 provided:1 estimating:1 bounded:3 notation:1 maximizes:1 mass:3 moreover:11 what:1 kg:1 atari:1 kind:2 dharshan:1 argmin:1 finding:1 guarantee:21 safely:20 temporal:1 wiesemann:1 multidimensional:1 expands:1 collecting:5 rustem:1 stricter:1 zaremba:1 ensured:1 wrong:1 scaled:1 control:29 grant:2 planck:1 safety:34 felix:4 ecc:1 local:3 positive:4 consequence:2 despite:2 sutton:1 path:1 matteo:2 interpolation:1 might:1 theodore:1 collect:2 shaded:4 challenging:1 ease:1 limited:2 acknowledgment:1 practical:5 practice:6 regret:1 definite:1 lf:2 learningbased:1 jan:2 demis:1 powell:1 intersect:1 riedmiller:1 eth:4 adapting:2 convenient:1 davy:1 confidence:14 road:1 petersen:1 zoubin:1 get:1 cannot:1 close:1 tuong:2 undesirable:1 shankar:1 prentice:1 risk:5 impossible:1 applying:1 bellemare:1 transition:2 optimize:10 jaime:1 dean:1 equivalent:1 deterministic:2 map:2 lagrangian:1 williams:1 go:5 starting:6 center:1 helen:1 focused:2 survey:1 tomas:1 matthieu:1 attraction:60 shlens:1 stability:16 exploratory:4 notion:1 autonomous:1 variation:1 suppose:1 svoboda:1 programming:6 gps:1 us:5 carl:1 alexis:1 goodfellow:2 origin:7 kunal:1 velocity:1 approximated:1 barfoot:1 s0x:3 database:1 mike:1 bottom:1 steven:1 preprint:1 capture:2 verifying:1 worst:4 region:70 ensures:3 connected:1 l2v:1 trade:1 decrease:20 rq:1 environment:2 bischoff:2 reward:3 dynamic:45 depend:1 tight:1 solving:2 predictive:3 duy:2 efficiency:1 regularizer:1 revert:1 committing:2 fast:1 artificial:2 gotovos:1 outside:2 lazar:1 dordrecht:1 encoded:1 solve:3 otherwise:1 ability:3 statistic:1 rgpin:1 tomlin:2 gp:11 jointly:2 noisy:2 commit:2 final:1 online:1 schneega:1 differentiable:3 matthias:1 propose:1 product:1 maximal:1 reset:1 fer:1 relevant:1 loop:2 kapoor:1 achieve:2 ontario:1 khalil:1 sutskever:2 convergence:2 regularity:3 requirement:5 silver:1 leave:1 converges:4 executing:1 andrew:3 measured:1 strong:1 edward:1 esann:1 c:2 reachability:4 implies:1 christmann:1 quantify:2 lyapunov:32 ml2:1 safe:67 kuhn:1 discontinuous:1 stochastic:2 lars:1 exploration:33 human:2 stabilization:1 enable:3 jonathon:1 hassan:1 garc:1 require:4 abbeel:2 generalization:1 fix:1 strictly:4 extension:1 hold:6 gravitational:1 practically:1 hall:1 considered:1 sufficiently:3 equilibrium:1 matthew:1 driving:7 achieves:1 early:1 smallest:2 schreiter:1 proc:13 applicable:2 visited:1 sensitive:3 largest:5 tool:1 stefan:1 mit:3 snsf:1 gaussian:14 always:2 aim:4 fulfill:5 rather:3 avoid:1 rusu:1 mobile:1 barto:2 encode:1 l0:1 focus:1 schaal:1 legg:1 mane:1 modelling:1 check:1 greatly:1 seeger:1 huijuan:1 baseline:4 entire:1 unlikely:1 kernelized:1 bandit:2 hidden:1 expand:12 going:1 selects:1 provably:4 overall:1 unreachable:1 akametalu:1 priori:3 platform:1 integration:1 constrained:3 mutual:1 construct:1 never:1 feasability:1 beach:1 sampling:2 veness:1 koray:1 represents:2 yu:1 icml:5 representer:1 intelligent:1 richard:1 simultaneously:1 divergence:1 intell:1 comprehensive:1 individual:1 argmax:2 ourselves:1 jeffrey:1 eberts:1 ostrovski:1 mnih:1 zheng:1 chowdhury:1 evaluation:1 certainly:1 joel:2 benoit:1 introduces:1 analyzed:1 behind:1 held:1 amenable:1 accurate:3 andy:1 partial:1 necessary:1 afe:1 lh:1 huan:1 harmful:1 initialized:1 desired:1 re:1 theoretical:8 uncertain:4 alkis:1 classify:1 cover:1 cost:12 deviation:1 subset:1 nickson:1 successful:1 gr:1 too:1 characterize:1 answer:1 kudlur:1 calibrated:2 combined:1 st:1 fritz:1 international:10 explores:1 fundamental:1 stay:1 retain:1 probabilistic:2 off:1 contract:1 picking:1 michael:1 together:2 quickly:2 ashish:2 continuously:2 concrete:1 ilya:2 again:2 rafal:1 slowly:1 lukasz:1 wojciech:1 li:1 szegedy:1 volodymyr:1 potential:1 jeremy:1 de:1 account:1 sec:6 stabilize:1 ioannis:1 automation:1 summarized:1 satisfy:1 explicitly:2 fernandez:1 idealized:2 bg:4 depends:3 later:1 break:1 closed:2 pendulum:7 red:6 start:5 s30:1 reached:1 recover:3 jia:1 contribution:1 minimize:1 greg:1 accuracy:2 largely:1 efficiently:3 correspond:1 gathered:1 yellow:1 generalize:4 lkopf:1 bayesian:3 vincent:1 anton:1 kavukcuoglu:1 manages:2 craig:1 fern:1 trajectory:3 researcher:1 drive:6 reach:2 suffers:1 whenever:1 definition:3 failure:2 coraluppi:1 energy:1 acquisition:1 c15:1 tucker:1 james:1 derek:1 proof:1 associated:1 argmaxc:1 ask:1 knowledge:3 car:4 dimensionality:2 ut:6 hilbert:1 improves:3 javier:1 mathematical:1 wicke:1 back:1 higher:1 jair:1 tom:1 specify:1 improved:2 formulation:3 evaluated:2 shrink:1 generality:1 angular:1 lastly:2 until:1 steinwart:1 christopher:1 nonlinear:3 propagation:1 continuity:2 defines:1 quality:1 behaved:1 believe:1 aviv:2 mdp:1 artif:1 dario:1 usa:1 verify:4 true:13 normalized:1 multiplier:1 vasudevan:1 regularization:1 moore:1 white:3 sin:1 piecewiselinear:1 self:1 game:1 irving:1 davis:1 illustrative:1 levenberg:1 during:2 gpflow:2 theoretic:1 julia:1 l1:1 stefano:1 novel:1 recently:1 charles:1 d30:1 common:1 rl:9 physical:2 overview:1 volume:2 extend:1 slight:1 jozefowicz:1 mihai:1 measurement:13 cambridge:2 ai:1 outlined:1 mathematics:2 sastry:1 similarly:1 bruna:1 stable:3 robot:3 impressive:1 han:1 encountering:1 pete:1 posterior:4 grizzle:1 showed:1 triangulation:2 optimizing:1 inf:2 wattenberg:1 sherry:1 certain:4 fernanda:1 approximators:1 der:1 joshua:1 life:1 jens:1 seen:3 inverted:4 additional:4 care:1 minimum:2 r0:1 determine:5 maximize:1 fernando:1 corrado:1 paradigm:1 u0:1 ii:2 christiano:1 desirable:1 sham:1 full:1 signal:1 adapt:4 cross:1 long:3 niranjan:1 visit:1 award:1 impact:1 ensuring:1 prediction:1 variant:2 geibel:1 heterogeneous:1 controller:2 metric:1 expectation:1 arxiv:4 iteration:1 kernel:5 limt:1 monga:1 agarwal:1 robotics:4 cell:1 achieved:1 background:2 crash:1 krause:6 want:5 interval:13 leaving:6 sch:2 warden:1 shane:1 subject:1 shie:1 integer:1 intermediate:1 iii:2 easy:1 enough:1 switch:1 affect:2 relu:1 zi:2 restrict:2 suboptimal:1 inner:1 idea:1 reduce:2 barham:1 andreas:8 cn:30 tamar:2 whether:2 turchetta:2 manjunath:1 render:1 peter:4 returned:1 york:1 cause:1 action:41 repeatedly:2 deep:1 detailed:1 clear:2 gopalan:1 teodor:1 amount:1 locally:3 http:1 specifies:1 exist:2 estimated:8 per:1 blue:1 discrete:12 mat:1 georg:1 harp:1 key:1 quadrotor:1 yangqing:1 falling:1 changing:1 verified:1 graph:1 asymptotically:1 convert:1 sum:1 run:2 angle:4 talwar:1 uncertainty:14 powerful:1 place:1 throughout:2 reasonable:1 almost:1 earning:1 gonzalez:1 decision:7 appendix:4 scaling:1 bound:5 layer:1 guaranteed:2 bastian:2 quadratic:1 replaces:1 encountered:1 oracle:3 constraint:14 perkins:1 alex:1 lane:1 nearby:1 speed:2 friction:1 min:1 martin:4 department:3 according:11 amodei:1 combination:1 poor:1 smaller:3 wysotzki:1 son:1 kakade:1 rob:1 restricted:1 invariant:1 ln:5 equation:1 zurich:3 remains:3 c50:1 eventually:3 fail:1 heiner:1 know:2 tractable:2 antonoglou:1 serf:1 end:4 available:2 operation:1 brevdo:1 moldovan:1 apply:3 away:2 appropriate:2 stig:1 romer:1 schmidt:1 hassabis:1 rp:1 denotes:1 angela:4 ensure:5 top:5 publishing:1 assumes:1 neglect:1 exploit:2 restrictive:1 ghahramani:1 especially:1 murray:1 build:1 classical:1 icra:1 k1:5 rsj:1 question:1 occurs:1 kaiser:1 parametric:2 strategy:5 dependence:1 gradient:4 iclr:1 distance:1 simulated:2 capacity:2 fidjeland:1 chris:3 topic:1 collected:4 considers:4 unstable:2 reason:1 provable:1 marcus:1 ru:1 modeled:2 index:1 kk:1 providing:2 cq:1 riccardo:1 lg:1 difficult:1 potentially:1 trace:1 design:2 implementation:2 policy:119 unknown:4 perform:1 upper:3 neuron:1 kumaran:1 markov:5 benchmark:1 ingo:1 wilk:1 descent:3 behave:1 finite:3 enabling:1 daan:1 ever:2 rn:1 reproducing:1 mona:1 arbitrary:1 community:1 pablo:1 david:2 pair:15 mechanical:1 specified:1 optimized:5 rad:3 aerospace:1 xiaoqiang:1 learned:3 tensorflow:3 unsafe:6 nip:2 boukouvalas:1 beyond:3 able:3 proceeds:2 below:2 dynamical:1 sanjay:1 scott:1 achiam:1 max:2 reliable:2 green:1 suitable:4 satisfaction:2 critical:2 natural:2 residual:1 melanie:1 minimax:1 scheme:5 improve:7 github:1 mdps:5 imply:1 ijrr:1 library:1 ne:1 sn:17 udluft:1 joan:1 eugene:1 prior:6 review:1 schulman:1 discovery:1 kf:1 python:1 graf:1 relative:1 asymptotic:1 fully:1 loss:1 cdc:2 highlight:1 sublinear:1 geoffrey:1 lv:5 toussaint:1 aversion:1 krausea:1 vanhoucke:1 gather:3 sufficient:1 verification:3 s0:10 agent:2 principle:1 claire:2 supported:1 last:1 free:1 rasmussen:1 drastically:2 warren:1 allow:1 ber:1 institute:1 fall:1 srinivas:1 taking:2 distributed:1 van:1 boundary:1 hensman:1 xn:3 world:4 cumulative:1 rich:1 qn:3 kz:1 forward:2 made:1 reinforcement:17 stuck:1 adaptive:2 computes:1 nguyen:2 erhan:1 far:2 transaction:1 qx:1 approximate:10 compact:1 observable:1 implicitly:2 bernhard:1 keep:1 logic:1 global:4 active:4 harm:1 automatica:2 fergus:1 un:17 continuous:17 search:2 additionally:1 nature:1 transfer:1 learn:9 ca:2 robust:8 necessarily:1 european:2 constructing:1 domain:5 marc:2 main:1 backup:3 noise:2 paul:3 allowed:2 quadrature:1 befelix:2 xu:1 fig:15 i:1 andrei:1 ny:1 wiley:1 quasiconvex:2 sub:2 downhill:1 lie:2 candidate:3 jmlr:1 learns:1 zhifeng:1 ian:2 anil:1 theorem:22 remained:1 dumitru:1 sui:1 specific:4 xt:9 utoronto:1 down:2 z0:3 ghemawat:1 decay:1 intractable:3 consist:1 exists:2 restricting:1 effectively:3 magnitude:1 conditioned:1 maximilian:1 kx:2 chen:1 vijay:1 garcia:1 timothy:1 explore:2 steinhardt:1 josh:1 lagrange:1 vinyals:1 nserc:1 contained:6 tracking:1 aditya:1 gml:1 sadik:1 springer:4 ch:3 corresponds:4 determines:3 satisfies:1 environmental:1 ma:1 mart:1 kinetic:1 kz0:1 goal:6 king:1 towards:1 eventual:1 lipschitz:18 man:1 change:2 aswani:1 specifically:11 except:1 typical:2 beattie:1 conservative:2 experimental:2 zeilinger:1 berkenkamp:4 citro:1 rarely:2 select:1 mark:1 support:2 fulfills:2 alexander:2 ethz:3 rajat:1 oriol:1 violated:1 evaluate:1 schuster:1 correlated:1 |
6,293 | 6,693 | Online Multiclass Boosting
Young Hun Jung
Jack Goetz
Department of Statistics
University of Michigan
Ann Arbor, MI 48109
{yhjung, jrgoetz, tewaria}@umich.edu
Ambuj Tewari
Abstract
Recent work has extended the theoretical analysis of boosting algorithms to multiclass problems and to online settings. However, the multiclass extension is in the
batch setting and the online extensions only consider binary classification. We fill
this gap in the literature by defining, and justifying, a weak learning condition for
online multiclass boosting. This condition leads to an optimal boosting algorithm
that requires the minimal number of weak learners to achieve a certain accuracy.
Additionally, we propose an adaptive algorithm which is near optimal and enjoys
an excellent performance on real data due to its adaptive property.
1
Introduction
Boosting methods are a ensemble learning methods that aggregate several (not necessarily) weak
learners to build a stronger learner. When used to aggregate reasonably strong learners, boosting has
been shown to produce results competitive with other state-of-the-art methods (e.g., Korytkowski
et al. [1], Zhang and Wang [2]). Until recently theoretical development in this area has been focused
on batch binary settings where the learner can observe the entire training set at once, and the labels
are restricted to be binary (cf. Schapire and Freund [3]). In the past few years, progress has been
made to extend the theory and algorithms to more general settings.
Dealing with multiclass classification turned out to be more subtle than initially expected. Mukherjee
and Schapire [4] unify several different proposals made earlier in the literature and provide a general
framework for multiclass boosting. They state their weak learning conditions in terms of cost matrices
that have to satisfy certain restrictions: for example, labeling with the ground truth should have less
cost than labeling with some other labels. A weak learning condition, just like the binary condition,
states that the performance of a learner, now judged using a cost matrix, should be better than a
random guessing baseline. One particular condition they call the edge-over-random condition, proves
to be sufficient for boostability. The edge-over-random condition will also figure prominently in this
paper. They also consider a necessary and sufficient condition for boostability but it turns out to be
computationally intractable to be used in practice.
A recent trend in modern machine learning is to train learners in an online setting where the instances
come sequentially and the learner has to make predictions instantly. Oza [5] initially proposed an
online boosting algorithm that has accuracy comparable with the batch version, but it took several
years to design an algorithm with theoretical justification (Chen et al. [6]). Beygelzimer et al. [7]
achieved a breakthrough by proposing an optimal algorithm in online binary settings and an adaptive
algorithm that works quite well in practice. These theories in online binary boosting have led to
several extensions. For example, Chen et al. [8] combine one vs all method with binary boosting
algorithms to tackle online multiclass problems with bandit feedback, and Hu et al. [9] build a theory
of boosting in regression setting.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
In this paper, we combine the insights and techniques of Mukherjee and Schapire [4] and Beygelzimer
et al. [7] to provide a framework for online multiclass boosting. The cost matrix framework from the
former work is adopted to propose an online weak learning condition that defines how well a learner
can perform over a random guess (Definition 1). We show this condition is naturally derived from its
batch setting counterpart. From this weak learning condition, a boosting algorithm (Algorithm 1) is
proposed which is theoretically optimal in that it requires the minimal number of learners and sample
complexity to attain a specified level of accuracy. We also develop an adaptive algorithm (Algorithm
2) which allows learners to have variable strengths. This algorithm is theoretically less efficient than
the optimal one, but the experimental results show that it is quite comparable and sometimes even
better due to its adaptive property. Both algorithms not only possess theoretical proofs of mistake
bounds, but also demonstrate superior performance over preexisting methods.
2
Preliminaries
We first describe the basic setup for online boosting. While in the batch setting, an additional weak
learner is trained at every iteration, in the online setting, the algorithm starts with a fixed count
of N weak learners and a booster which manages the weak learners. There are k possible labels
[k] := {1, ? ? ? , k} and k is known to the learners. At each iteration t = 1, ? ? ? , T , an adversary picks
a labeled example (xt , yt ) ? X ? [k], where X is some domain, and reveals xt to the booster. Once
the booster observes the unlabeled data xt , it gathers the weak learners? predictions and makes a final
prediction. Throughout this paper, index i takes values from 1 to N ; t from 1 to T ; and l from 1 to k.
We utilize the cost matrix framework, first proposed by Mukherjee and Schapire [4], to develop
multiclass boosting algorithms. This is a key ingredient in the multiclass extension as it enables
different penalization for each pair of correct label and prediction, and we further develop this
framework to suit the online setting. The booster sequentially computes cost matrices {Cit ?
Rk?k | i = 1, ? ? ? , N }, sends (xt , Cit ) to the ith weak learner W Li , and gets its prediction lti ? [k].
Here the cost matrix Cit plays a role of loss function in that W Li tries to minimize the cumulative
P
cost t Cit [yt , lti ]. As the booster wants each learner to predict the correct label, it wants to set the
diagonal entries of Cit to be minimal among its row. At this stage, the true label yt is not revealed yet,
but the previous weak learners? predictions can affect the computation of the cost matrix for the next
learner. Given a matrix C, the (i, j)th entry will be denoted by C[i, j], and ith row vector by C[i].
Once all the learners make predictions, the booster makes the final prediction y?t by majority votes.
The booster can either take simple majority votes or weighted ones. In fact for the adaptive algorithm,
we will allow weighted votes so that the booster can assign more weights on well-performing learners.
The weight for W Li at iteration t will be denoted by ?ti . After observing the booster?s final decision,
the adversary reveals the true label yt , and the booster suffers 0-1 loss 1(?
yt 6= yt ). The booster also
shares the true label to the weak learners so that they can train on this data point.
Two main issues have to be resolved to design a good boosting algorithm. First, we need to design
the booster?s strategy for producing cost matrices. Second, we need to quantify weak learner?s
PT
ability to reduce the cumulative cost t=1 Cit [yt , lti ]. The first issue will be resolved by introducing
potential functions, which will be thoroughly discussed in Section 3.1. For the second issue, we
introduce our online weak learning condition, a generalization of the weak learning assumption in
Beygelzimer et al. [7], stating that for any adaptively given sequence of cost matrices, weak learners
can produce predictions whose cumulative cost is less than that incurred by random guessing. The
online weak learning condition will be discussed in the following section. For the analysis of the
adaptive algorithm, we use empirical edges instead of the online weak learning condition.
2.1
Online weak learning condition
In this section, we propose an online weak learning condition that states the weak learners are better
than a random guess. We first define a baseline condition that is better than a random guess. Let
?[k] denote a family of distributions over [k] and ul? ? ?[k] be a uniform distribution that puts ?
1??
1??
more weight on the label l. For example, u1? = ( 1??
k + ?, k , ? ? ? , k ). For a given sequence of
examples {(xt , yt ) | t = 1, ? ? ? , T }, U? ? RT ?k consists of rows uy?t . Then we restrict the booster?s
2
choice of cost matrices to
C1eor := {C ? Rk?k | ?l, r ? [k], C[l, l] = 0, C[l, r] ? 0, and ||C[l]||1 = 1}.
Note that diagonal entries are minimal among the row, and C1eor also has a normalization constraint.
A broader choice of cost matrices is allowed if one can assign importance weights on observations,
which is possible for various learners. Even if the learner does not take the importance weight as an
input, we can achieve a similar effect by sending to the learner an instance with probability that is
proportional to its weight. Interested readers can refer Beygelzimer et al. [7, Lemma 1]. From now
on, we will assume that our weak learners can take weight wt as an input.
We are ready to present our online weak learning condition. This condition is in fact naturally derived
from the batch setting counterpart that is well studied by Mukherjee and Schapire [4]. The link is
thoroughly discussed in Appendix A. For the scaling issue, we assume the weights wt lie in [0, 1].
Definition 1. (Online multiclass weak learning condition) For parameters ?, ? ? (0, 1), and
S > 0, a pair of online learner and an adversary is said to satisfy online weak learning condition
with parameters ?, ?, and S if for any sample length T , any adaptive sequence of labeled examples,
and for any adaptively chosen series of pairs of weight and cost matrix {(wt , Ct ) ? [0, 1] ? C1eor | t =
1, ? ? ? , T }, the learner can generate predictions y?t such that with probability at least 1 ? ?,
T
X
wt Ct [yt , y?t ] ? C ? U0? + S =
t=1
1??
||w||1 + S,
k
(1)
where C ? RT ?k consists of rows of wt Ct [yt ] and A ? B0 denotes the Frobenius inner product
Tr(AB0 ). w = (w1 , ? ? ? , wT ) and the last equality holds due to the normalized condition on C1eor . ?
is called an edge, and S an excess loss.
Remark. Notice that this condition is imposed on a pair of learner and adversary instead of solely
on a learner. This is because no learner can satisfy this condition if the adversary draws samples
in a completely adaptive manner. The probabilistic statement is necessary because many online
algorithms? predictions are not deterministic. The excess loss requirement is needed since an online
learner cannot produce meaningful predictions before observing a sufficient number of examples.
3
An optimal algorithm
In this section, we describe the booster?s optimal strategy for designing cost matrices. We first
introduce a general theory without specifying the loss, and later investigate the asymptotic behavior
of cumulative loss suffered by our algorithm under the specific 0-1 loss. We adopt the potential
function framework from Mukherjee and Schapire [4] and extend it to the online setting. Potential
functions help both in designing cost matrices and in proving the mistake bound of the algorithm.
3.1
A general online multiclass boost-by-majority (OnlineMBBM) algorithm
We will keep track of the weighted cumulative votes of the first i weak learners for the sample xt by
Pi
sit := j=1 ?tj elj , where ?ti is the weight of W Li , lti is its prediction and ej is the j th standard basis
t
vector. For the optimal algorithm, we assume that ?ti = 1, ?i, t. In other words, the booster makes
the final decision by simple majority votes. Given a cumulative vote s ? Rk , suppose we have a loss
function Lr (s) where r denotes the correct label. We call a loss function proper, if it is a decreasing
function of s[r] and an increasing function of other coordinates (we alert the reader that ?proper loss?
has at least one other meaning in the literature). From now on, we will assume that our loss function
is proper. A good example of proper loss is multiclass 0-1 loss:
Lr (s) := 1(max s[l] ? s[r]).
l6=r
(2)
The purpose of the potential function ?ri (s) is to estimate the booster?s loss when there remain i
learners until the final decision and the current cumulative vote is s. More precisely, we want potential
functions to satisfy the following conditions:
?r0 (s)
r
?i+1 (s)
= Lr (s),
= El?ur? ?ri (s + el ).
3
(3)
Algorithm 1 Online Multiclass Boost-by-Majority (OnlineMBBM)
1: for t = 1, ? ? ? , T do
2:
Receive example xt
3:
Set s0t = 0 ? Rk
4:
for i = 1, ? ? ? , N do
5:
Set the normalized cost matrix Dit according to (5) and pass it to W Li
6:
Get weak predictions lti = W Li (xt ) and update sit = sti?1 + elti
7:
end for
8:
Predict y?t := argmaxl sN
t [l] and receive true label yt
9:
for i = 1, ? ? ? , N do
Pk
10:
Set wi [t] = l=1 [?yNt?i (si?1
+ el ) ? ?yNt?i (sti?1 + eyt )]
t
11:
Pass training example with weight (xt , yt , wi [t]) to W Li
12:
end for
13: end for
Readers should note that ?ri (s) also inherits the proper property of the loss function, which can be
shown by induction. The condition (3) can be loosened by replacing both equalities by inequalities
???, but in practice we usually use equalities.
Now we describe the booster?s strategy for designing cost matrices. After observing xt , the booster
sequentially sets a cost matrix Cit for W Li , gets the weak learner?s prediction lti and uses this in the
computation of the next cost matrix Ci+1
t . Ultimately, booster wants to set
Cit [r, l] = ?rN ?i (si?1
+ el ).
t
(4)
C1eor ,
However, this cost matrix does not satisfy the condition of
and thus should be modified in order
to utilize the weak learning condition. First to make the cost for the true label equal to 0, we subtract
Cit [r, r] from every element of Cit [r]. Since the potential function is proper, our new cost matrix still
has non-negative elements after the subtraction. We then normalize the row so that each row has `1
norm equal to 1. In other words, we get new normalized cost matrix
Dit [r, l] =
?rN ?i (si?1
+ el ) ? ?rN ?i (si?1
+ er )
t
t
,
i
w [t]
(5)
Pk
where wi [t] := l=1 ?rN ?i (si?1
+ el ) ? ?rN ?i (si?1
+ er ) plays the role of weight. It is still possible
t
t
i
that a row vector Ct [r] is a zero vector so that normalization is impossible. In this case, we just leave
it as a zero vector. Our weak learning condition (1) still works with cost matrices some of whose row
vectors are zeros because however the learner predicts, it incurs no cost.
After defining cost matrices, the rest of the algorithm is straightforward except we have to estimate
||wi ||? to normalize the weight. This is necessary because the weak learning condition assumes
the weights lying in [0, 1]. We cannot compute the exact value of ||wi ||? until the last instance is
revealed, which is fine as we need this value only in proving the mistake bound. The estimate wi? for
||wi ||? requires to specify the loss, and we postpone the technical parts to Appendix B.2. Interested
readers may directly refer Lemma 10 before proceeding. Once the learners generate predictions after
observing cost matrices, the final decision is made by simple majority votes. After the true label
is revealed, the booster updates the weight and sends the labeled instance with weight to the weak
learners. The pseudocode for the entire algorithm is depicted in Algorithm 1. The algorithm is named
after Beygelzimer et al. [7, OnlineBBM], which is in fact OnlineMBBM with binary labels.
We present our first main result regarding the mistake bound of general OnlineMBBM. The proof
appears in Appendix B.1 where the main idea is adopted from Beygelzimer et al. [7, Lemma 3].
Theorem 2. (Cumulative loss bound for OnlineMBBM) Suppose weak learners and an adversary
satisfy the online weak learning condition (1) with parameters ?, ?, and S. For any T and N satisfying
? N1 , and any adaptive sequence of labeled examples generated by the adversary, the final loss
suffered by OnlineMBBM satisfies the following inequality with probability 1 ? N ?:
T
X
yt
L
(sN
t )
?
?1N (0)T
t=1
+S
N
X
i=1
4
wi? .
(6)
Here ?1N (0) plays a role of asymptotic error rate and the second term determines the sample complexity. We will investigate the behavior of those terms under the 0-1 loss in the following section.
3.2
Mistake bound under 0-1 loss and its optimality
From now on, we will specify the loss to be multiclass 0-1 loss defined in (2), which might be the
most relevant measure in multiclass problems. To present a specific mistake bound, two terms in
the RHS of (6) should be bounded. This requires an approximation of potentials, which is technical
and postponed to Appendix B.2. Lemma 9 and 10 provide the bounds for those terms. We also
mention another bound for the weight in the remark after Lemma 10 so that one can use whichever
tighter. Combining the above lemmas with Theorem 2 gives the following corollary. The additional
constraint on ? comes from Lemma 10.
Corollary 3. (0-1 loss bound of OnlineMBBM) Suppose weak learners and an adversary satisfy
the online weak learning condition (1) with parameters ?, ?, and S, where ? < 12 . For any T and
N satisfying ? N1 and any adaptive sequence of labeled examples generated by the adversary,
OnlineMBBM can generate predictions y?t that satisfy the following inequality with probability 1?N ?:
T
X
1(yt 6= y?t ) ? (k ? 1)e?
?2 N
2
?
? 5/2 N S).
T + O(k
(7)
t=1
Therefore in order to achieve error rate , it suffices to use N = ?( ?12 ln k ) weak learners, which
? k5/2 S).
gives an excess loss bound of ?(
?
5/2
? k S). If
Remark. Note that the above excess loss bound gives a sample complexity bound of ?(
?
we use alternative weight bound to get kN S as an upper bound for the second term in (6), we end up
?
? k2 S).
having O(kN
S). This will give an excess loss bound of ?(
?
We now provide lower bounds on the number of learners and sample complexity for arbitrary online
boosting algorithms to evaluate the optimality of OnlineMBBM under 0-1 loss. In particular, we
construct weak learners that satisfy the online weak learning condition (1) and have almost matching
asymptotic error rate and excess loss compared to those of OnlineMBBM as in (7). Indeed we
can prove that the number of learners and sample complexity of OnlineMBBM is optimal up to
logarithmic factors, ignoring the influence of the number of classes k. Our bounds are possibly
suboptimal up to polynomial factors in k, and the problem to fill the gap remains open. The detailed
proof and a discussion of the gap can be found in Appendix B.3. Our lower bound is a multiclass
version of Beygelzimer et al. [7, Theorem 3].
k ln( 1 )
Theorem 4. (Lower bounds for N and T ) For any ? ? (0, 14 ), ?, ? (0, 1), and S ? ? ? , there
exists an adversary with a family of learners satisfying the online weak learning condition (1) with
parameters ?, ?, and S, such that to achieve asymptotic error rate , an online boosting algorithm
k
requires at least ?( k21? 2 ln 1 ) learners and a sample complexity of ?( ?
S).
4
An adaptive algorithm
The online weak learning condition imposes minimal assumptions on the asymptotic accuracy of
learners, and obviously it leads to a solid theory of online boosting. However, it has two main practical
limitations. The first is the difficulty of estimating the edge ?. Given a learner and an adversary, it
is by no means a simple task to find the maximum edge that satisfies (1). The second issue is that
different learners may have different edges. Some learners may in fact be quite strong with significant
edges, while others are just slightly better than a random guess. In this case, OnlineMBBM has to
pick the minimum edge as it assumes common ? for all weak learners. It is obviously inefficient in
that the booster underestimates the strong learners? accuracy.
Our adaptive algorithm will discard the online weak learning condition to provide a more practical
method. Empirical edges ?1 , ? ? ? , ?N (see Section 4.2 for the definition) are measured for the weak
learners and are used to bound the number of mistakes made by the boosting algorithm.
5
4.1
Choice of loss function
Adaboost, proposed by Freund et al. [10], is arguably the most popular boosting algorithm in practice.
It aims to minimize the exponential loss, and has many variants which use some other surrogate
loss. The main reason of using a surrogate loss is ease of optimization; while 0-1 loss is not even
continuous, most surrogate losses are convex. We adopt the use of a surrogate loss for the same reason,
and throughout this section will discuss our choice of surrogate loss for the adaptive algorithm.
Exponential loss is a very strong candidate in that it provides a closed form for computing potential
functions, which are used to design cost matrices (cf. Mukherjee and Schapire [4, Theorem 13]).
One property of online setting, however, makes it unfavorable. Like OnlineMBBM, each data point
will have a different weight depending on weak learners? performance, and if the algorithm uses
exponential loss, this weight will be an exponential function of difference in weighted cumulative
votes. With this exponentially varying weights among samples, the algorithm might end up depending
on very small portion of observed samples. This is undesirable because it is easier for the adversary
to manipulate the sample sequence to perturb the learner.
To overcome exponentially varying weights, Beygelzimer et al. [7] use logistic loss in their adaptive
algorithm. Logistic loss is more desirable in that its derivative is bounded and thus weights will be
relatively smooth. For this reason, we will also use multiclass version of logistic loss:
X
Lr (s) =:
log(1 + exp(s[r] ? s[r])).
(8)
l6=r
We still need to compute potential functions from logistic loss in order to calculate cost matrices.
Unfortunately, Mukherjee and Schapire [4] use a unique property of exponential loss to get a closed
form for potential functions, which cannot be adopted to logistic loss. However, the optimal cost
matrix induced from exponential loss has a very close connection with the gradient of the loss (cf.
Mukherjee and Schapire [4, Lemma 22]). From this, we will design our cost matrices as following:
(
1
, if l 6= r
i?1
i?1
i
P t [r]?st [l])
(9)
Ct [r, l] := 1+exp(s
1
? j6=r 1+exp(si?1 [r]?si?1 [j]) , if l = r.
t
t
Cit [r]
Readers should note that the row vector
is simply the gradient of Lr (si?1
). Also note that this
t
matrix does not belong to C1eor , but it does guarantee that the correct prediction gets the minimal cost.
The choice of logistic loss over exponential loss is somewhat subjective. The undesirable property
of exponential loss does not necessarily mean that we cannot build an adaptive algorithm using this
loss. In fact, we can slightly modify Algorithm 2 to develop algorithms using different surrogates
(exponential loss and square hinge loss). However, their theoretical bounds are inferior to the one with
logistic loss. Interested readers can refer Appendix D, but it assumes understanding of Algorithm 2.
4.2
Adaboost.OLM
Our work is a generalization of Adaboost.OL by Beygelzimer et al. [7], from which the name
Adaboost.OLM comes with M standing for multiclass. We introduce a new concept of an expert.
From N weak learners, we can produce N experts where expert i makes its prediction by weighted
majority votes among the first i learners. Unlike OnlineMBBM, we allow
weights ?ti over
P varying
yt i
the learners. As we are working with logistic loss, we want to minimize t L (st ) for each i, where
the loss is given in (8). We want to alert the readers to note that even though the algorithm tries to
minimize the cumulative surrogate loss, its performance is still evaluated by 0-1 loss. The surrogate
loss only plays a role of a bridge that makes the algorithm adaptive.
We do not impose the online weak
learning condition on weak learners, but instead just measure the
P i
i
i
t Ct [yt ,lt ]
P
performance of W L by ?i :=
. This empirical edge will be used to bound the number of
i
t Ct [yt ,yt ]
mistakes made by Adaboost.OLM. By definition of cost matrix, we can check
Cit [yt , yt ] ? Cit [yt , l] ? ?Cit [yt , yt ], ?l ? [k],
from which we can prove ?1 ? ?i ? 1, ?i. If the online weak learning condition is met with edge ?,
then one can show that ?i ? ? with high probability when the sample size is sufficiently large.
6
Algorithm 2 Adaboost.OLM
1: Initialize: ?i, v1i = 1, ?1i = 0
2: for t = 1, ? ? ? , T do
3:
Receive example xt
4:
Set s0t = 0 ? Rk
5:
for i = 1, ? ? ? , N do
6:
Compute Cit according to (9) and pass it to W Li
7:
Set lti = W Li (xt ) and sit = si?1
+ ?ti elti
t
i
i
8:
Set y?t = argmaxl st [l], the prediction of expert i
9:
end for
10:
Randomly draw it with P(it = i) ? vti
11:
Predict y?t = y?tit and receive the true label yt
12:
for i = 1, ? ? ? , N do
0
i
13:
Set ?t+1
= ?(?ti ? ?t fti (?ti )) using (10) and ?t =
?
2 2?
(k?1) t
Ci [y ,y ]
t t
Set wi [t] = ? tk?1
and pass (xt , yt , wi [t]) to W Li
i
15:
Set vt+1
= vti ? exp(?1(yt 6= y?ti ))
16:
end for
17: end for
14:
Unlike the optimal algorithm, we cannot show the last expert that utilizes all the learners has the
best accuracy. However, we can show at least one expert has a good predicting power. Therefore
we will use classical Hedge algorithm (Littlestone and Warmuth [11] and Freund and Schapire [12])
to randomly choose an expert at each iteration with adaptive probability weight depending on each
expert?s prediction history.
Finally we need to address how to set the weight ?ti for each weak learner.
tries to
P As our algorithm
i
i
minimize the cumulative logistic loss, we want to set ?ti to minimize t Lyt (si?1
+
?
e
).
This
t
t lt
is again a classical topic in online learning, and we will use online gradient descent, proposed
by Zinkevich [13]. By letting, fti (?) := Lyt (si?1
+ ?elti ), we need an online algorithm ensuring
t
P i i
P i
i
f
(?
)
?
min
f
(?)
+
R
(T
)
where
F
is a feasible set to be specified later, and Ri (T )
??F
t
t t
t t
is a regret that is sublinear in T . To apply Zinkevich [13, Theorem 1], we need fti to be convex
and F to be compact. The first assumption is met by our choice of logistic loss, and for the second
assumption, we will set F = [?2, 2]. There is no harm to restrict the choice of ?ti by F because we
can always scale the weights without affecting the result of weighted majority votes.
By taking derivatives, we get
0
fti (?)
(
=
1
i?1 i
1+exp(si?1
P t [yt ]?st [lt ]??)
1
? j6=yt 1+exp(si?1 [j]+??s
i?1
[yt ])
t
t
, if lti 6= yt
, if lti = yt .
(10)
0
This provides |fti (?)| ? k ? 1. Now let ?(?) represent a projection onto ?
F : ?(?) :=
2 2?
i
i
i0
i
max{?2, min{2, ?}}. By setting ?t+1 = ?(?t ? ?t ft (?t )) where ?t = (k?1) t , we get
?
?
Ri (T ) ? 4 2(k ? 1) T . Readers should note that any learning rate of the form ?t = ?ct would
work, but our choice is optimized to ensure the minimal regret.
The pseudocode for Adaboost.OLM is presented in Algorithm 2. In fact, if we put k = 2, Adaboost.OLM has the same structure with Adaboost.OL. As in OnlineMBBM, the booster also needs
to pass the weight along with labeled instance. According to (9), it can be inferred that the weight is
proportional to ?Cit [yt , yt ].
4.3
Mistake bound and comparison to the optimal algorithm
Now we present our second main result that provides a mistake bound of Adaboost.OLM. The main
structure of the proof is adopted from Beygelzimer et al. [7, Theorem 4] but in a generalized cost
matrix framework. The proof appears in Appendix C.
7
Theorem 5. (Mistake bound of Adaboost.OLM) For any T and N , with probability 1 ? ?, the
number of mistakes made by Adaboost.OLM satisfies the following inequality:
T
X
t=1
2
? PkN
T + O(
),
N
2
2
i=1 ?i
i=1 ?i
8(k ? 1)
1(yt 6= y?t ) ? PN
? notation suppresses dependence on log 1 .
where O
?
Remark. Note that this theorem naturally implies Beygelzimer et al. [7, Theorem 4]. The difference
in coefficients is due to different scaling of ?i . In fact, their ?i ranges from [? 12 , 12 ].
Now that we have established a mistake bound, it is worthwhile to compare the bound with the
optimal boosting algorithm. Suppose the weak learners satisfy the weak learning
condition (1)
P i
i
t Ct [yt ,lt ]
P
??
with edge ?. For simplicity, we will ignore the excess loss S. As we have ?i =
i
t Ct [yt ,yt ]
8(k?1)
kN
?
with high probability, the mistake bound becomes 2 T + O( 2 ). In order to achieve error
? N
?
8(k?1)
? 2k2 4 ) sample size. Note that
learners
and
T
=
?(
2
?
?
? k2 )}. Adaboost.OLM is
? k5/2 ), ?(
?( ?12 ln k ) and T = min{?(
?
?
rate , Adaboost.OLM requires N ?
OnlineMBBM requires N =
obviously suboptimal, but due to its adaptive feature, its performance on real data is quite comparable
to that by OnlineMBBM.
5
Experiments
We compare the new algorithms to existing ones for online boosting on several UCI data sets, each
with k classes1 . Table 1 contains some highlights, with additional results and experimental details in
the Appendix E. Here we show both the average accuracy on the final 20% of each data set, as well as
the average run time for each algorithm. Best decision tree gives the performance of the best of 100
online decision trees fit using the VFDT algorithm in Domingos and Hulten [14], which were used as
the weak learners in all other algorithms, and Online Boosting is an algorithm taken from Oza [5].
Both provide a baseline for comparison with the new Adaboost.OLM and OnlineMBBM algorithms.
Best MBBM takes the best result from running the OnlineMBBM with five different values of the
edge parameter ?.
Despite being theoretically weaker, Adaboost.OLM often demonstrates similar accuracy and sometimes outperforms Best MBBM, which exemplifies the power of adaptivity in practice. This power
comes from the ability to use diverse learners efficiently, instead of being limited by the strength of
the weakest learner. OnlineMBBM suffers from high computational cost, as well as the difficulty of
choosing the correct value of ?, which in general is unknown, but when the correct value of ? is used
it peforms very well. Finally in all cases Adaboost.OLM and OnlineMBBM algorithms outperform
both the best tree and the preexisting Online Boosting algorithm, while also enjoying theoretical
accuracy bounds.
Table 1: Comparison of algorithm accuracy on final 20% of data set and run time in seconds. Best
accuracy on a data set reported in bold.
1
Data sets
k
Best decision tree
Online Boosting
Adaboost.OLM
Best MBBM
Balance
Mice
Cars
Mushroom
Nursery
ISOLET
Movement
3
8
4
2
4
26
5
0.768
0.608
0.924
0.999
0.953
0.515
0.915
0.772
0.399
0.914
1.000
0.941
0.149
0.870
0.754
0.561
0.930
1.000
0.966
0.521
0.962
0.821
42
0.695 2173
0.914
56
1.000
325
0.969 1510
0.635 64707
0.988 18676
8
105
39
241
526
470
1960
19
263
27
169
302
1497
3437
20
416
59
355
735
2422
5072
Codes are available at https://github.com/yhjung88/OnlineBoostingWithVFDT
8
Acknowledgments
We acknowledge the support of NSF under grants CAREER IIS-1452099 and CIF-1422157.
References
[1] Marcin Korytkowski, Leszek Rutkowski, and Rafa? Scherer. Fast image classification by boosting fuzzy
classifiers. Information Sciences, 327:175?182, 2016.
[2] Xiao-Lei Zhang and DeLiang Wang. Boosted deep neural networks and multi-resolution cochleagram
features for voice activity detection. In INTERSPEECH, pages 1534?1538, 2014.
[3] Robert E Schapire and Yoav Freund. Boosting: Foundations and algorithms. MIT press, 2012.
[4] Indraneel Mukherjee and Robert E Schapire. A theory of multiclass boosting. Journal of Machine Learning
Research, 14(Feb):437?497, 2013.
[5] Nikunj C Oza. Online bagging and boosting. In 2005 IEEE international conference on systems, man and
cybernetics, volume 3, pages 2340?2345. IEEE, 2005.
[6] Shang-Tse Chen, Hsuan-Tien Lin, and Chi-Jen Lu. An online boosting algorithm with theoretical
justifications. ICML, 2012.
[7] Alina Beygelzimer, Satyen Kale, and Haipeng Luo. Optimal and adaptive algorithms for online boosting.
ICML, 2015.
[8] Shang-Tse Chen, Hsuan-Tien Lin, and Chi-Jen Lu. Boosting with online binary learners for the multiclass
bandit problem. In Proceedings of The 31st ICML, pages 342?350, 2014.
[9] Hanzhang Hu, Wen Sun, Arun Venkatraman, Martial Hebert, and Andrew Bagnell. Gradient boosting on
stochastic data streams. In Artificial Intelligence and Statistics, pages 595?603, 2017.
[10] Yoav Freund, Robert Schapire, and N Abe. A short introduction to boosting. Journal-Japanese Society For
Artificial Intelligence, 14(771-780):1612, 1999.
[11] Nick Littlestone and Manfred K Warmuth. The weighted majority algorithm. In Foundations of Computer
Science, 1989., 30th Annual Symposium on, pages 256?261. IEEE, 1989.
[12] Yoav Freund and Robert E Schapire. A desicion-theoretic generalization of on-line learning and an
application to boosting. In European conference on computational learning theory, pages 23?37. Springer,
1995.
[13] Martin Zinkevich. Online convex programming and generalized infinitesimal gradient ascent. In Proceedings of 20th ICML, 2003.
[14] Pedro Domingos and Geoff Hulten. Mining high-speed data streams. In Proceedings of the sixth ACM
SIGKDD international conference on Knowledge discovery and data mining, pages 71?80. ACM, 2000.
[15] Amit Daniely, Sivan Sabato, Shai Ben-David, and Shai Shalev-Shwartz. Multiclass learnability and the
erm principle. In COLT, pages 207?232, 2011.
[16] Nick Littlestone. Learning quickly when irrelevant attributes abound: A new linear-threshold algorithm.
Machine learning, 2(4):285?318, 1988.
[17] Volodimir G Vovk. Aggregating strategies. In Proc. Third Workshop on Computational Learning Theory,
pages 371?383. Morgan Kaufmann, 1990.
[18] Nicolo Cesa-Bianchi and G?bor Lugosi. Prediction, learning, and games. Cambridge university press,
2006.
R in Optimization,
[19] Elad Hazan et al. Introduction to online convex optimization. Foundations and Trends
2(3-4):157?325, 2016.
[20] Robert E Schapire. Drifting games. Machine Learning, 43(3):265?291, 2001.
[21] Eric V Slud. Distribution inequalities for the binomial law. The Annals of Probability, pages 404?412,
1977.
[22] C.L. Blake and C.J. Merz. UCI machine learning repository, 1998. URL http://archive.ics.uci.
edu/ml.
9
[23] Cios KJ Higuera C, Gardiner KJ. Self-organizing feature maps identify proteins critical to learning in a
mouse model of down syndrome. PLoS ONE, 2015.
[24] Wallace Ugulino, D?bora Cardador, Katia Vega, Eduardo Velloso, Ruy Milidi?, and Hugo Fuks. Wearable
computing: Accelerometers? data classification of body postures and movements. In Advances in Artificial
Intelligence-SBIA 2012, pages 52?61. Springer, 2012.
10
| 6693 |@word repository:1 version:3 polynomial:1 stronger:1 norm:1 cochleagram:1 open:1 hu:2 pick:2 incurs:1 mention:1 tr:1 solid:1 series:1 contains:1 past:1 subjective:1 existing:1 current:1 outperforms:1 com:1 beygelzimer:12 luo:1 si:14 yet:1 mushroom:1 enables:1 update:2 v:1 intelligence:3 guess:4 warmuth:2 ith:2 short:1 lr:5 manfred:1 provides:3 boosting:36 zhang:2 five:1 alert:2 along:1 symposium:1 consists:2 prove:2 combine:2 manner:1 introduce:3 theoretically:3 indeed:1 expected:1 sbia:1 behavior:2 wallace:1 multi:1 ol:2 chi:2 decreasing:1 increasing:1 becomes:1 fti:5 estimating:1 bounded:2 notation:1 abound:1 suppresses:1 fuzzy:1 proposing:1 eduardo:1 guarantee:1 tewaria:1 every:2 ti:11 tackle:1 k2:3 demonstrates:1 classifier:1 grant:1 producing:1 arguably:1 before:2 aggregating:1 modify:1 mistake:14 despite:1 solely:1 lugosi:1 might:2 studied:1 specifying:1 ease:1 limited:1 range:1 uy:1 practical:2 unique:1 acknowledgment:1 practice:5 regret:2 postpone:1 volodimir:1 area:1 empirical:3 attain:1 matching:1 projection:1 word:2 protein:1 get:9 cannot:5 unlabeled:1 undesirable:2 close:1 judged:1 put:2 onto:1 impossible:1 influence:1 restriction:1 zinkevich:3 imposed:1 deterministic:1 yt:37 map:1 straightforward:1 kale:1 convex:4 focused:1 resolution:1 unify:1 simplicity:1 hsuan:2 insight:1 isolet:1 fill:2 proving:2 coordinate:1 justification:2 annals:1 pt:1 play:4 suppose:4 exact:1 programming:1 us:2 designing:3 domingo:2 trend:2 element:2 satisfying:3 nikunj:1 mukherjee:9 predicts:1 boostability:2 labeled:6 observed:1 role:4 ft:1 wang:2 oza:3 calculate:1 sun:1 plo:1 movement:2 observes:1 complexity:6 ultimately:1 trained:1 argmaxl:2 tit:1 eric:1 learner:71 completely:1 basis:1 resolved:2 geoff:1 various:1 train:2 fast:1 describe:3 preexisting:2 artificial:3 labeling:2 aggregate:2 choosing:1 shalev:1 quite:4 whose:2 elad:1 ability:2 statistic:2 satyen:1 final:9 online:54 obviously:3 sequence:6 took:1 propose:3 product:1 turned:1 relevant:1 combining:1 uci:3 organizing:1 achieve:5 frobenius:1 haipeng:1 normalize:2 requirement:1 produce:4 leave:1 ben:1 tk:1 help:1 depending:3 develop:4 stating:1 andrew:1 measured:1 b0:1 progress:1 strong:4 come:4 implies:1 quantify:1 met:2 correct:6 attribute:1 stochastic:1 assign:2 suffices:1 generalization:3 preliminary:1 tighter:1 indraneel:1 extension:4 hold:1 lying:1 sufficiently:1 ground:1 blake:1 exp:6 ic:1 predict:3 pkn:1 adopt:2 purpose:1 proc:1 label:15 bridge:1 v1i:1 arun:1 weighted:7 mit:1 always:1 aim:1 modified:1 pn:1 ej:1 boosted:1 varying:3 hun:1 broader:1 hulten:2 corollary:2 derived:2 inherits:1 exemplifies:1 check:1 sigkdd:1 baseline:3 el:6 i0:1 entire:2 initially:2 bandit:2 marcin:1 interested:3 issue:5 classification:4 among:4 colt:1 denoted:2 development:1 art:1 breakthrough:1 initialize:1 equal:2 once:4 construct:1 having:1 beach:1 icml:4 venkatraman:1 others:1 few:1 wen:1 modern:1 randomly:2 eyt:1 n1:2 suit:1 detection:1 investigate:2 mining:2 tj:1 edge:14 necessary:3 tree:4 enjoying:1 littlestone:3 theoretical:7 minimal:7 instance:5 tse:2 earlier:1 yoav:3 cost:38 introducing:1 entry:3 daniely:1 uniform:1 learnability:1 reported:1 kn:3 thoroughly:2 st:6 adaptively:2 international:2 standing:1 probabilistic:1 mouse:2 quickly:1 w1:1 again:1 cesa:1 choose:1 possibly:1 booster:22 expert:8 inefficient:1 derivative:2 li:11 potential:10 accelerometer:1 bold:1 coefficient:1 satisfy:10 stream:2 later:2 try:3 closed:2 observing:4 hazan:1 portion:1 competitive:1 start:1 shai:2 minimize:6 square:1 cio:1 accuracy:11 kaufmann:1 ynt:2 efficiently:1 ensemble:1 identify:1 weak:55 bor:1 manages:1 lu:2 cybernetics:1 j6:2 history:1 suffers:2 definition:4 infinitesimal:1 sixth:1 underestimate:1 naturally:3 proof:5 mi:1 wearable:1 popular:1 knowledge:1 car:1 subtle:1 appears:2 adaboost:18 specify:2 evaluated:1 though:1 just:4 stage:1 until:3 working:1 replacing:1 defines:1 logistic:10 lei:1 usa:1 effect:1 name:1 normalized:3 true:7 concept:1 counterpart:2 former:1 equality:3 game:2 interspeech:1 self:1 inferior:1 higuera:1 generalized:2 theoretic:1 demonstrate:1 loosened:1 meaning:1 image:1 jack:1 vega:1 recently:1 superior:1 common:1 pseudocode:2 hugo:1 exponentially:2 volume:1 extend:2 discussed:3 belong:1 refer:3 significant:1 cambridge:1 feb:1 deliang:1 nicolo:1 recent:2 irrelevant:1 discard:1 certain:2 inequality:5 binary:9 vt:1 postponed:1 tien:2 morgan:1 minimum:1 additional:3 somewhat:1 impose:1 syndrome:1 r0:1 subtraction:1 u0:1 ii:1 desirable:1 smooth:1 technical:2 long:1 lin:2 goetz:1 justifying:1 manipulate:1 ensuring:1 prediction:22 variant:1 regression:1 basic:1 iteration:4 sometimes:2 normalization:2 represent:1 achieved:1 proposal:1 receive:4 want:7 fine:1 affecting:1 desicion:1 sends:2 suffered:2 sabato:1 rest:1 unlike:2 posse:1 archive:1 ascent:1 induced:1 call:2 near:1 revealed:3 affect:1 fit:1 restrict:2 suboptimal:2 reduce:1 inner:1 regarding:1 idea:1 multiclass:22 url:1 ul:1 cif:1 remark:4 deep:1 tewari:1 detailed:1 cit:16 dit:2 schapire:15 generate:3 outperform:1 http:2 nsf:1 notice:1 track:1 instantly:1 diverse:1 nursery:1 key:1 sivan:1 threshold:1 alina:1 lti:9 utilize:2 olm:15 year:2 sti:2 run:2 named:1 throughout:2 family:2 reader:8 almost:1 utilizes:1 draw:2 decision:7 appendix:8 scaling:2 comparable:3 bound:30 ct:10 annual:1 activity:1 strength:2 gardiner:1 constraint:2 precisely:1 ri:5 u1:1 speed:1 optimality:2 min:3 performing:1 relatively:1 martin:1 department:1 according:3 remain:1 slightly:2 ur:1 wi:10 restricted:1 erm:1 taken:1 computationally:1 ln:4 remains:1 turn:1 count:1 discus:1 needed:1 letting:1 lyt:2 whichever:1 end:8 umich:1 sending:1 adopted:4 available:1 k5:2 apply:1 observe:1 worthwhile:1 batch:6 alternative:1 voice:1 drifting:1 bagging:1 denotes:2 assumes:3 cf:3 ensure:1 running:1 binomial:1 hinge:1 l6:2 perturb:1 build:3 prof:1 amit:1 classical:2 society:1 posture:1 strategy:4 rt:2 dependence:1 diagonal:2 guessing:2 said:1 surrogate:8 gradient:5 bagnell:1 link:1 majority:9 topic:1 reason:3 induction:1 length:1 code:1 index:1 balance:1 setup:1 unfortunately:1 robert:5 statement:1 negative:1 design:5 proper:6 unknown:1 perform:1 bianchi:1 upper:1 observation:1 acknowledge:1 descent:1 defining:2 extended:1 rn:5 arbitrary:1 abe:1 inferred:1 david:1 pair:4 specified:2 connection:1 optimized:1 nick:2 established:1 boost:2 nip:1 address:1 adversary:12 usually:1 ambuj:1 max:2 power:3 critical:1 difficulty:2 predicting:1 github:1 martial:1 ready:1 sn:2 kj:2 literature:3 understanding:1 discovery:1 asymptotic:5 law:1 freund:6 loss:61 katia:1 highlight:1 sublinear:1 adaptivity:1 limitation:1 proportional:2 scherer:1 ingredient:1 penalization:1 foundation:3 incurred:1 gather:1 sufficient:3 imposes:1 xiao:1 principle:1 share:1 pi:1 row:10 jung:1 last:3 hebert:1 enjoys:1 allow:2 weaker:1 taking:1 feedback:1 overcome:1 cumulative:11 computes:1 made:6 adaptive:20 excess:7 compact:1 ignore:1 keep:1 dealing:1 ml:1 sequentially:3 reveals:2 vti:2 harm:1 shwartz:1 continuous:1 table:2 additionally:1 reasonably:1 ca:1 career:1 ignoring:1 excellent:1 necessarily:2 japanese:1 european:1 domain:1 pk:2 main:7 rh:1 allowed:1 body:1 s0t:2 exponential:9 prominently:1 lie:1 candidate:1 third:1 young:1 rk:5 theorem:10 down:1 xt:13 specific:2 jen:2 er:2 k21:1 weakest:1 sit:3 intractable:1 exists:1 workshop:1 importance:2 ci:2 gap:3 easier:1 chen:4 subtract:1 depicted:1 michigan:1 led:1 logarithmic:1 simply:1 elj:1 lt:4 springer:2 pedro:1 truth:1 satisfies:3 determines:1 acm:2 hedge:1 ann:1 man:1 feasible:1 except:1 vovk:1 wt:6 lemma:8 shang:2 called:1 pas:5 arbor:1 experimental:2 unfavorable:1 vote:11 meaningful:1 merz:1 support:1 evaluate:1 |
6,294 | 6,694 | Matching on Balanced Nonlinear Representations for
Treatment Effects Estimation
Yun Fu
Northeastern University
Boston, MA
[email protected]
Sheng Li
Adobe Research
San Jose, CA
[email protected]
Abstract
Estimating treatment effects from observational data is challenging due to the
missing counterfactuals. Matching is an effective strategy to tackle this problem.
The widely used matching estimators such as nearest neighbor matching (NNM)
pair the treated units with the most similar control units in terms of covariates,
and then estimate treatment effects accordingly. However, the existing matching
estimators have poor performance when the distributions of control and treatment
groups are unbalanced. Moreover, theoretical analysis suggests that the bias of
causal effect estimation would increase with the dimension of covariates. In this
paper, we aim to address these problems by learning low-dimensional balanced and
nonlinear representations (BNR) for observational data. In particular, we convert
counterfactual prediction as a classification problem, develop a kernel learning
model with domain adaptation constraint, and design a novel matching estimator.
The dimension of covariates will be significantly reduced after projecting data
to a low-dimensional subspace. Experiments on several synthetic and real-world
datasets demonstrate the effectiveness of our approach.
1
Introduction
Causal questions exist in many areas, such as health care [24, 12], economics [14], political science [17], education [36], digital marketing [6, 43, 5, 15, 44], etc. In the field of health care, it is
critical to understand if a new medicine could cure a certain illness and perform better than the old
ones. In political science, it is of great importance to evaluate whether the government should fund a
job training program, by assessing if the program is the true factor that leads to the success of job
hunting. All of these causal questions can be addressed by the causal inference technique. Formally,
causal inference estimates the treatment effect on some units after interventions [33, 20]. In the
above example of heath care, the units could be patients, and the intervention would be taking new
medicines. Due to the wide applications of causal questions, effective causal inference techniques are
highly desired to address these problems.
Generally, the causal inference problems can be tackled by either experimental study or observational
study. Experimental study is popular in traditional causal inference problems, but it is time-consuming
and sometimes impractical. As an alternative strategy, observational study has attracted increasing
attention in the past decades, which extracts causal knowledge only from the observed data. Two
major paradigms for observational study have been developed in computer science and statistics,
including the causal graphical model [29] and the potential outcome framework [27, 33]. The former
builds directed acyclic graphs (DAG) from covariates, treatment and outcome, and uses probabilistic
inference to determine causal relationships; while the latter estimates counterfactuals for each treated
unit, and gives a precise definition of causal effect. The equivalence of two paradigms has been
discussed in [11]. In this paper, we mainly focus on the potential outcome framework.
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
A missing data problem needs to be dealt with in the potential outcome framework. As each unit is
either treated or not treated, it is impossible to observe its outcomes in both scenarios. In other words,
one has to predict the missing counterfactuals. A widely used solution to estimating counterfactuals
is matching. According to the (binary) treatment assignments, a set of units can be divided into a
treatment group and a control group. For each treated unit, matching methods select its counterpart in
the control group based on certain criteria, and treat the selected unit as a counterfactual. Then the
treatment effect can be estimated by comparing the outcomes of treated units and the corresponding
counterfactuals. Some popular matching estimators include nearest neighbor matching (NNM) [32],
propensity score matching [31], coarsened exact matching (CEM) [17], genetic matching [9], etc.
Existing matching methods have three major drawbacks. First, they either perform matching in the
original covariate space (e.g., NNM, CEM) or in the one-dimensional propensity score space (e.g.,
PSM). The potential of using intermediate representations has not been extensively studied before.
Second, existing methods work well for data with a moderate number of covariates, but may fail
for data with a large number of covariates, as theoretical analysis suggests that the bias of treatment
effect estimation would increase with the dimension of covariates [1]. Third, most matching methods
do not take into account whether the distributions of two groups are balanced or not. The matching
process would make no sense if the distributions of two groups have little overlap.
To address the above problems, we propose to learn balanced and nonlinear representations (BNR)
from observational data, and design a novel matching estimator named BNR-NNM. First, the
counterfactual prediction problem is converted to a multi-class classification problem, by categorizing
the outcomes to ordinal labels. Then, we propose a novel criterion named ordinal scatter discrepancy
(OSD) for supervised kernel learning on data with ordinal labels, and extract low-dimensional
nonlinear representations from covariates. Further, to achieve balanced distributions in the lowdimensional space, a maximum mean discrepancy (MMD) criterion [4] is incorporated to the model.
Finally, matching strategy is performed on the extracted balanced representations, in order to provide
a robust estimation of causal effect. In summary, the main contributions of our work include:
? We propose a novel matching estimator, BNR-NNM, which learns low-dimensional balanced
and nonlinear representations via kernel learning.
? We convert the counterfactual prediction problem into a multi-class classification problem,
and design an OSD criterion for nonlinear kernel learning with ordinal labels.
? We incorporate a domain adaptation constraint to feature learning by using the maximum
mean discrepancy criterion, which leads to balanced representations.
? We evaluate the proposed estimator on both synthetic datasets and real-world datasets, and
demonstrate its superiority over the state-of-the-art methods.
2
Background
Potential Outcome Framework. The potential outcome framework is proposed by Neyman and
Rubin [27, 33]. Considering binary treatments for a set of units, there are two possible outcomes for
each unit. Formally, for unit k, the outcome is defined as Yk (1) if it received treatment, and Yk (0) if
it did not. Then, the individual-level treatment effect is defined as ?k = Yk (1) ? Yk (0). Clearly, each
unit only belongs to one of the two groups, and therefore, we can only observe one of the two possible
outcomes. This is the well-known missing data problem in causal inference. In particular, if unit k
received treatment, Yk (1) is the observed outcome, and Yk (0) is missing data, i.e., counterfactual.
The potential outcome framework usually makes the following assumptions [19].
Assumption 1. Stable Unit Treatment Value Assumption (SUTVA): The potential outcomes for
any units do not vary with the treatments assigned to other units, and for each unit there are no
differences forms or versions of each treatment level, which lead to different potential outcomes.
|=
Assumption 2. Strongly Ignorable Treatment Assignment (SITA): Conditional on covariates xk ,
treatment Tk is independent of potential outcomes.
(Yk (1), Yk (0))
Tk |xk . (Unconfoundedness)
(1)
0 < Pr(Tk = 1|xk ) < 1. (Overlap)
These assumptions enable the modeling of treatment of one unit with respect to covariates, independent of outcomes and other units.
Matching Estimators. To address the aforementioned missing data problem, a simple yet effective
strategy has been developed, which is matching [32, 33, 14, 40]. The idea of matching is to estimate
2
the counterfactual for a treated unit by seeking its most similar counterpart in the control group.
Existing matching methods can be roughly divided into three categories: nearest neighbor matching
(NNM), weighting, and subclassification. We mainly focus on NNM in this paper.
Let XC ? Rd?NC and XT ? Rd?NT denote the covariates of a control group and a treatment group,
respectively, where d is the number of covariates, NC and NT are the group sizes. T is a binary
vector indicating if the units received treatments (i.e., Tk = 1) or not (i.e., Tk = 0). Y is an outcome
vector. For each treated unit k, NNM finds its nearest neighbor in the control group in terms of the
covariates. The outcome of the selected control unit is considered as an estimation of counterfactual.
Then, the average treatment effect on treated (ATT) is defined as:
1 X
AT T =
Yk (1) ? Y?k (0) ,
(2)
NT
k:Tk =1
where Y?k (0) is the counterfactual estimated from unit k?s nearest neighbor in the control group.
NNM can be implemented in various ways, such as using different distance metrics, or choosing
different number of neighbors. Euclidean distance and Mahalanobis distance are two widely-used
distance metrics for NNM. They work well when there are a few covariates with normal distributions [34]. Another important matching estimator is propensity score matching (PSM) [31]. PSM
estimates the propensity score (i..e., the probability of receiving treatment) for each unit via logistic
regression, and pairs the units from two groups with similar scores [35, 8, 30]. Most recently, a
covariate balancing propensity score (CBPS) method is developed to balance the distributions of two
groups by weighting the covariates, and has shown promising performance [18].
The key differences between the proposed BNR-NNM estimator and the traditional matching estimators are two-fold. First, BNR-NNM performs matching in an intermediate low-dimensional
subspace that could guarantee a low estimation bias, while the traditional estimators adopt either the
original covariate space or the one-dimensional space. Second, BNR-NNM explicitly considers the
balanced distributions across treatment and control groups, while the traditional estimators usually
fail to achieve such a property.
Machine Learning for Causal Inference. In recent years, researchers have been exploring the
relationships between causal inference and machine learning [39, 10, 38]. A number of predictive
models have been designed to estimate the causal effects, such as causal trees [3] and causal
forests [42]. Balancing the distributions of two groups is considered as a key issue in observational
study, which is closely related to covariate shift and in general domain adaptation [2]. Meanwhile,
causal inference has also been incorporated to improve the performance of domain adaptation [46, 45].
Most recently, the idea of representation learning is introduced to learn new features from covariates
through random projections [25], informative subspace learning [7], and deep neural networks [21,
37].
3
Learning Balanced and Nonlinear Representations (BNR)
In this section, we first define the notations that will be used throughout this paper. Then we introduce
how to convert the counterfactual prediction problem into a multi-class classification problem, and
justify the rationality of this strategy. We will also present the details of how to learn nonlinear and
balanced representations, and derive the closed-form solutions to the model.
Notations. Let X = [XC , XT ] ? Rd?N denote the covariates of all units, where XC ? Rd?NC is
the control group with NC units, and XT ? Rd?NT is the treatment group with NT units. N is the
total number of units, and d is the number of covariates for each unit. ? : x ? Rd ? ?(x) ? F is
a nonlinear mapping function from sample space R to an implicit feature space F. T ? RN ?1 is a
binary vector to indicate if the units received treatments or not. Y ? RN ?1 is an outcome vector. The
elements in Y could be either discrete or continuous values.
3.1
From Counterfactual Prediction to Multi-Class Classification
When estimating the treatment effects as shown in Eq.(2), we only have the observed outcome Yk (1),
but need to estimate the counterfactual Y?k (0). Ideally, we would train a model Y?k (0) = Fcf (xk ) that
can predict the counterfactual for any units, given the covariate vector xk . One strategy is to build a
predictive model (e.g., regression) that maps each unit xi to its output Yi , which has been extensively
3
studied before. Alternatively, we can convert the counterfactual prediction problem into a multi-class
classification problem.
Given a set of units X and the corresponding outcome vector Y , we aim to learn a predictive
model Fcf (xk ) that maps from the covariate space to the outcome space. In particular, we propose
to seek an intermediate representation space in which the units close to each other should have
very similar outcomes. The outcome vector Y usually contains continuous values. We categorize
outcomes in Y into multiple levels on the basis of the magnitude of outcome value, and consider
them as (pseudo) class labels. Clustering or kernel density estimation can be used for discretizing
Y . Finally, Y is converted to a (pseudo) class label vector Yc with c categories. For example,
Y = [0.3, 0.5, 1.1, 1.2, 2.4] could be categorized as Y3 = [1, 1, 2, 2, 3]. As a result, we could use Yc
and X to train a classifier.
Note that the Yc actually contains ordinal labels, as the discretized labels carry additional information.
In particular, the labels [1, 2, 3] are not totally independent. We actually assume that Class 1 should
be more close to Class 2 than Class 3, since the outcome values in Class 1 are closer to those in Class
2. We will make use of such ordinal label information when designing the classification model.
3.2
Learning Nonlinear Representations via Ordinal Scatter Discrepancy
To obtain effective representations from X, we propose to train a nonlinear classifier in a reproducing
kernel Hilbert space (RKHS). The reasons of employing the RKHS based nonlinear models are as
follows. First, compared to linear models, nonlinear models are usually more capable of dealing
with complicated data distributions. It is well known that the treatment and control groups might
have diverse distributions, and the nonlinear models would be able to tightly couple them in a shared
low-dimensional subspace. Second, the RKHS based nonlinear models usually have closed-form
solutions because of the kernel trick, which is beneficial for handling large-scale data.
Let ?(xi ) denote the mapped counterpart of xi in kernel space, and then ?(X) =
[?(x1 ), ?(x2 ), ? ? ? , ?(xN )]. In light of the maximum scatter difference criterion [26], we take
into account the ordinal label information, and propose a novel criterion named Ordinal Scatter Discrepancy (OSD) to achieve the desired data distribution after projecting ?(X) to a low-dimensional
subspace. In particular, OSD minimizes the within-class scatter, and meanwhile maximize the
noncontiguous-class scatter matrix. Let P denote a transformation matrix, OSD maps samples onto a
subspace by maximizing the differences of noncontiguous-class scatter and within-class scatter. We
perform OSD in kernel space to learn nonlinear representations, and have the following objective
function:
arg max F (P, ?(X), Yc ) = tr(P > (KI ? ?KW )P ),
P
(3)
s.t.
P > P = I,
where ? is a non-negative trade-off parameter, tr(?) is the trace operator for matrix, and I is an identity
matrix. The orthogonal constraint P > P = I is introduced to reduce the redundant information in
projection.
In Eq.(3), KI and KW are the noncontiguous-class scatter matrix and within-class scatter matrix in
kernel space, respectively. The detailed definitions are:
KI?
=
?
KW
c(c?1)
2
=
1
N
c
c
P
P
e(j?i) (mi ? mj )(mi ? mj )>
(4)
i=1 j=i+1
ni
c P
P
(?(xij ) ? m)(?(x
?
?i )>
ij ) ? m
(5)
i=1 j=1
where ?(xij ) = [k(x1 , xij ), k(x2 , xij ), ? ? ? , k(xN , xij )]> , mi is the mean vector of ?(xij ) that
belongs to the i-th class, m
? is the mean vector of all ?(xij ), and ni is the number of units in the
i-th class. k(xi , xj ) = h?(xi ), ?(xj )i is a kernel function, which is utilized to avoid calculating the
explicit form of function ? (i.e., the kernel trick).
Eq. (4) characterizes the scatter of a set of classes with (pseudo) ordinal labels. It measures the scatter
of every pair of classes. The factor e(j?i) is used to penalize the classes that are noncontiguous. The
intuition is that, for ordinal labels, we may expect the contiguous classes will be close to each other
after projection, while the noncontiguous classes should be pushed away. Therefore, we put larger
4
weights for the noncontiguous classes. For example, e(2?1) < e(3?1) , since Class 1 should be more
close to Class 2 than Class 3, as we explained in Section 3.1.
Eq. (5) measures the within-class scatter. We expect that the units having the same (pseudo) class
labels will be very close to each other in the feature space, and therefore they will have similar feature
representations after projection.
The differences between the proposed OSD criterion and other discriminative criteria (e.g., Fisher
criterion, maximum scatter difference criterion) are two-fold. (1) OSD criterion learns nonlinear
projection and feature representations in the RKHS space; (2) OSD explicitly makes use of the
ordinal label information that are usually ignored by existing criteria. Moreover, the maximum scatter
difference criterion is a special case of OSD.
3.3
Learning Balanced Representations via Maximum Mean Discrepancy
Balanced distributions of control and treatment groups, in terms of covariates, would greatly facilitate
the causal inference methods such as NNM. To this end, we adopt the idea of maximum mean
discrepancy (MMD) [4] when learning the transformation P , and finally obtain balanced nonlinear
representations. The MMD criterion has been successfully applied to some problems like domain
adaptation [28].
Assume that the control group XC and treatment group XT are random variable sets with distributions
P and Q, MMD implies the empirical estimation of the distance between P and Q. In particular,
MMD estimates the distance between nonlinear feature sets ?(XC ) and ?(XT ), which can be
formulated as:
n
nT
C
P
P
Dist(?(XC ), ?(XT )) = k N1C
?(XCi ) ? N1T
?(XT i )k2F ,
(6)
i=1
i=1
where F denotes a kernel space.
By utilizing the kernel trick, Dist(?(XC ), ?(XT )) in the original kernel space can be equivalently
converted to:
Dist(?(XC ), ?(XT )) = tr(KL),
(7)
KCC KCT
where K =
is a kernel matrix, KCC , KT T , and KT C are kernel matrices defined
KT C KT T
on control group, treatment group, and cross groups, respectively. L is a constant matrix. If
xi , xj ? XC , Lij = N12 ; if xi , xj ? XT , Lij = N12 ; otherwise, Lij = ? NC1NT .
C
T
As all the units are projected into a new space via projection P , we need to measure the MMD for
new representations ?(XC ) = P > ?(XC ) and ?(XT ) = P > ?(XT ), and rewrite Eq.(7) into the
following form after some derivations:
Dist(?(XC ), ?(XT )) = tr(P > KLKP ).
3.4
(8)
BNR Model and Solutions
The representation learning objectives described in Section 3.2 and Section 3.3 are actually performed
on the same data set with different partitions. For nonlinear representation learning, we merge the
control group and treatment group, assign a (pseudo) ordinal label for each unit, and then learn
discriminative nonlinear features accordingly. For balanced representation learning, we aim to
mitigate the distribution discrepancy between control group and treatment group. Two learning
objectives are motivated from different perspectives, and therefore they are complementary to each
other. By combing the objectives for nonlinear and balanced representations in Eq.(3) and Eq.(8), we
can extract effective representations for the purpose of treatment effect estimation.
The objective function of BNR is formulated as follows:
arg max F (P, ?(X), Yc ) ? ?Dist(?(XC ), ?(XT ))
P
s.t.
= tr(P > (KI ? ?KW )P ) ? ?tr(P > KLKP ),
P > P = I,
(9)
where ? is a trade-off parameter to balance the effects of two terms. A negative sign is added before
?Dist(?(XC ), ?(XT )) in order to adapt it into this maximization problem.
5
The problem Eq.(9) can be efficiently solved by using a closed-form solution described in Proposition
1. The proof is provided in the supplementary document due to space limit.
Proposition 1. The optimal solution of P in problem Eq.(9) is the eigenvectors of matrix (KI ?
?KW ? ?KLK), which correspond to the m leading eigenvalues.
4
BNR for Nearest Neighbor Matching
Leveraging on the balanced nonlinear representations extracted from observational data, we propose
a novel nearest neighbor matching estimator named BNR-NNM.
After obtaining the transformation P in kernel space, we could generate nonlinear and balanced
representations for control and treated units as: X?C = P > KC , X?T = P > KT , where KC and KT
are kernel matrices defined in control and treatment groups, respectively. Then we follow the basic
idea of nearest neighbor matching. On the new representations X?C and X?T , we calculate the distance
between each treated unit and control unit, and choose the one with the smallest distance. The
outcome of the selected control unit serves as the estimation of counterfactual. Finally, the average
treatment effect on treated (ATT) can be calculated, as defined in Eq.(2). The complete procedures of
BNR-NNM are summarized in Algorithm 1.
Algorithm 1. BNR-NNM
The estimated ATT is dependent on the transforInput: Treatment group XT ? Rd?Nt
mation matrix P . Although P is optimal for the
Control group XC ? Rd?Nc
representation learning model Eq.(9), it might not
Outcome vectors YT and YC
be optimal for the whole causal inference process,
Total sample size N
for three reasons. First, the model Eq.(9) contains
Kernel function k
two major hyperparameters, ? and ?. Different ?opParameters ?, ?, c
timal? transformations P would be obtained with
1: Convert outcomes to (pseudo) ordinal labels
different parameter settings. Second, the ground2: Construct KI and KW using Eqs.(4) and (5)
3: Construct kernel matrix K using Eq.(7)
truth label information required by supervised learn4: Learn the transformation P using Eq.(9)
ing are unknown. Recall that we categorize the
5: Construct kernel matrix KC and KT
outcome vector as pseudo labels, which introduces
6: Project KC and KT using P
considerable uncertainty. Third, the ground-truth
X?C = P > KC , X?T = P > KT .
information of causal effect is unknown in observa7: Perform NNM between X?C and X?T
tional studies with real-world data. Therefore, it is
8: Estimate the ATT A from Eq.(2)
impossible to use the faithful supervision informaOutput: Return A
tion of causal effect to guide the learning process.
These uncertainties from three perspectives might result in an unreliable estimation of ATT.
Thus, we present two strategies to tackle the above issue. (1) Median causal effect from multiple
estimations. Following the randomized NNM estimator [25], we implement multiple settings of
BNR-NNM with different parameters ?, ? and c, calculate multiple ATT values, and finally choose
the median value as the final estimation. In this way, a robust estimation of causal effect can be
obtained. (2) Model selection by cross-validation. Alternatively, the cross-validation strategy can be
employed to select proper values for ? and ?, by equally dividing the data and pseudo labels into k
subsets. Although the multiple runs in the above strategies would increase the computational cost,
our method is still efficient for three reasons. First, the dimension of covariates will be significantly
reduced, which enables a faster matching process. Second, owing to the closed-form solution to P
introduced in Proposition 1, the representation learning procedure is efficient. Third, these settings
are independent from each other, and therefore they can be executed in parallel.
5
Experiments and Analysis
Synthetic Dataset. Data Generation. We generate a synthetic dataset by following the protocols
described in [41, 25]. In particular, the sample size N is set to 1000, and the number of covariates d is
set to 100. The following basis functions are adopted in the data generation process: g1 (x) = x ? 0.5,
g2 (x) = (x ? 0.5)2 + 2, g3 (x) = x2 ? 1/3, g4 (x) = ?2 sin(2x), g5 (x) = e?x ? e?1 ? 1,
g6 (x) = e?x , g7 (x) = x2 , g8 (x) = x, g9 (x) = Ix>0 , and g10 (x) = cos(x). For each unit, the
covariates x1 , x2 , ? ? ? , xd are drawn independently from the standard normal distribution N (0, 1).
We only consider binary treatment in this paper, and define the treatment vector T as T |x = 1 if
P5
vector T , the
k=1 gk (xk ) > 0 and T |x = 0 otherwise. Given covariate vector x and the treatment
P5
outcome variables in Y are generated from the following model: Y |x, T ? N ( j=1 gj+5 (xj ) +
6
T, 1). It is obvious that Y contains continuous values. The first five covariates are correlated to the
treatments in T and the outcomes in Y , simulating a confounding effect, while the rest are noisy
components. By definition, the true causal effect (i.e., the ground truth of ATT) in this dataset is 1.
Baselines and Settings. We compare our matching estimator BNR-NNM with the following baseline
methods: Euclidean distance based NNM (Eud-NNM), Mahalanobis distance based NNM (MahNNM) [34], PSM [31], principal component analysis based NNM (PCA-NNM), locality preserving
projections based NNM (LPP-NNM), and randomized NNM (RNNM) [25].
Eud?NNM
PSM
Mah?NNM
PCA?NNM
LPP?NNM
RNNM
BNR?NNM (Ours)
0
10
Mean Square Error
PSM is a classical causal inference approach,
which estimates the propensity scores for each
control or treated unit using logistic regression,
and then perform matching on these scores. As
our approach learns new representations via
transformations, we also implement two matching estimators based on the popular subspace
learning methods PCA [22] and LPP [13]. The
nearest neighbor matching is performed on the
low-dimensional feature space learned by PCA
and LPP, respectively. RNNM is the state-ofthe-art matching estimator, especially for highdimensional data. It projects units to multiple
random subspaces, performs matching in each
of them, and finally selects the median value of
estimations. In RNNM, the number of random
projections is set to 20. The proposed BNRNNM and RNNM share a similar idea on projecting data to low-dimensional subspaces, but
they have different motivations and learn different data representations.
?1
10
?2
10
2
5
10
20
30
40
50
Dimension
60
70
80
90
100
Figure 1: MSE of different estimators on the synthetic dataset. Note that Eud-NNM and Mah-NNM
only involve matching in the original 100 dimensional data space.
The major parameters in BNR-NNM include ?, ?, and c. In the experiments, ? is empirically set to
1. ? is chosen from {10?3 , 10?1 , 1, 10, 103 }. The number of categories c is chosen from {2, 4, 6, 8}.
As described in Section 4, the median ATT of multiple estimations is used as the final result. We use
the Gaussian kernel function k(xi , xj ) = exp(?kxi ? xj k2 /2? 2 ), in which the bandwidth parameter
? is empirically set to 5. In the experiments we observe that our approach allows flexible settings
for these parameters, and intuitively selecting parameters from a wider range would lead to a robust
estimation of ATT.
Results and Discussions. To ensure a robust estimation of the performance of each matching
estimator, we repeat the data generation process 500 times, calculate the ATT for each estimator
in every replication, and compute the mean square error (MSE) with standard error (SD) for each
estimator over all of the replications. Eud-NNM and Mah-NNM perform matching in the original
covariate space, and PSM maps each unit to a single score. Thus we only have a single point
estimation for each of them. For PCA-NNM, LPP-NNM, RNNM and our method, we can choose the
dimension of feature space where the matching is conducted. Specifically, we increase the dimension
from 2 to 100, and calculate MSE and SD in each case. Figure 1 shows the MSE and SD (shown
as error bars) of each estimator when varying the dimensions. We observe from Figure 1 that the
proposed estimator BNR-NNM obtains lower MSE than all other methods in every case. The lowest
MSE is achieved when the dimension is 5. In addition, we have analyzed the sensitivity of parameter
settings. The detailed results are provided in the supplementary document.
IHDP Dataset with Simulated Outcomes. IHDP data [16] is an experimental dataset collected by
the Infant Health and Development Program. In particular, a randomized experiment was conducted,
where intensive high-quality care were provided to the low-birth-weight and premature infants. By
using the original data, an observation study can be conducted by removing a nonrandom subset of
the treatment group: all children with non-white mothers. After this preprocessing step, there are in
total 24 pretreatment covariates (excluding race) and 747 units, including 608 control units and 139
treatment units. The outcomes are simulated by using the pretreatment covariates and the treatment
assignment information, in order to hold the unconfoundedness assumption.
7
Due to the space limit, the outcome simulation procedures Table 1: Results on IHDP dataset.
are provided in the supplementary document. We repeat such
Method
?AT T
procedures for 200 times and generate 200 sets of simulated
Eu-NNM
0.18?0.06
outcomes, in order to conduct extensive evaluations. For each
Mah-NNM 0.31?0.12
set of simulated outcomes, we run our method and the basePSM
0.26?0.08
lines introduced above, and report the results in Table 1. We
PCA-NNM 0.19?0.11
use the error in average treatment effect on treated (ATT),
LPP-NNM
0.25?0.13
?AT T , as the evaluation metric. It is defined as the absolute
RNNM
0.16?0.07
[
difference between true ATT and estimated ATT (AT
T ), i.e.,
BNR-NNM 0.16?0.06
[
?AT T = |AT T ? AT
T |. Table 1 shows that the proposed
BNR-NNM estimator outperforms most baselines, which further validates the effectiveness of the
balanced and nonlinear representations.
LaLonde Dataset with Real Outcomes. The LaLonde dataset is a widely used benchmark for
observational studies [23]. It consists of a treatment group and a control group. The treatment group
contains 297 units from a randomized study of a job training program (the ?National Supported
Work Demonstration?), where an unbiased estimate of the average treatment effect is available. The
original LaLonde dataset contains 425 control units that are collected from the Current Population
Survey. Recently, Imai et al. augmented the data by including 2,490 units from the Panel Study
of Income Dynamics [18]. Thus, the sample size of control group is increased to 2,915. For each
sample, the covariates include age, education, race (black, white, or Hispanic), marriage status, high
school degree, earnings in 1974, and earnings in 1975. The outcome variable is earnings in 1978. In
this benchmark dataset, the unbiased estimation of ATT is $886 with a standard error of $448.
We compare our estimator with the baselines Table 2: Results on LaLonde dataset. BIAS (%) is
used in the previous experiments. In addi- the bias in percentage of the true effect.
tion, we also compare with a recently proMethod
ATT
SD
BIAS (%)
posed matching estimator, covariate balancing
Ground
Truth
886
488
N/A
propensity score (CBPS) [18] and a deep neural
Eu-NNM
-565.9 592.8
164%
network (DNN) method [37]. CBPS aims to
Mah-NNM
-67.9
526.1
108%
achieve balanced distributions between control
PSM
-947.6 567.9
201%
and treatment groups by adjusting the weights
PCA-NNM
-499.8
592.5
156%
for covariates. The DNN method utilizes a deep
LPP-NNM
-457.1 581.2
152%
neural network architecture for counterfactual
RNNM
-557.6 584.9
163%
regression, which is the state-of-the-art method
CBPS
423.3 1295.2
52%
on representation learning based counterfactual
DNN
742.0
N/A
16%
inference. For BNR-NNM, we use the same set783.6
546.3
12%
tings for ? and c as in the previous experiments. BNR-NNM
Table 2 shows the ground truth of ATT, and the estimations of different methods. We can observe
from Table 2 that CBPS and DNN obtain better results than other baselines, as both of them consider
the balanced property across treatment and control groups. Moreover, our BNR-NNM estimator
achieves the best result, due to the fully exploitation of balanced and nonlinear feature representations.
The evaluations on runtime behavior of each compared method are provided in the supplementary
document due to space limit.
6
Conclusions
In this paper, we propose a novel matching estimator based on balanced and nonlinear representations
for treatment effect estimation. Our method leverages on the predictive power of machine learning
models to estimate counterfactuals, and achieves balanced distributions in an intermediate feature
space. In particular, an ordinal scatter discrepancy criterion is designed to extract discriminative
features from observational data with ordinal pseudo labels, while a maximum mean discrepancy
criterion is incorporated to achieve balanced distributions. Extensive experimental results on three
synthetic and real-world datasets show that our approach provides more accurate estimation of causal
effects than the state-of-the-art matching estimators and representation learning methods. In future
work, we will extend the balanced representation learning model to other causal inference strategies
such as weighting and regression, and design estimators for multiple levels of treatments.
Acknowledgement. This research is supported in part by the NSF IIS award 1651902, ONR Young
Investigator Award N00014-14-1-0484, and U.S. Army Research Office Award W911NF-17-1-0367.
8
References
[1] Alberto Abadie and Guido W Imbens. Large sample properties of matching estimators for average
treatment effects. Econometrica, 74(1):235?267, 2006.
[2] Deepak Agarwal, Lihong Li, and Alexander J Smola. Linear-time estimators for propensity scores. In
Proceedings of the International Conference on Artificial Intelligence and Statistics, pages 93?100, 2011.
[3] Susan Athey and Guido Imbens. Recursive partitioning for heterogeneous causal effects. Proceedings of
the National Academy of Sciences, 113(27):7353?7360, 2016.
[4] Karsten M Borgwardt, Arthur Gretton, Malte J Rasch, Hans-Peter Kriegel, Bernhard Sch?lkopf, and Alex J
Smola. Integrating structured biological data by kernel maximum mean discrepancy. Bioinformatics,
22(14):e49?e57, 2006.
[5] Kay H Brodersen, Fabian Gallusser, Jim Koehler, Nicolas Remy, and Steven L Scott. Inferring causal
impact using bayesian structural time-series models. The Annals of Applied Statistics, 9(1):247?274, 2015.
[6] David Chan, Rong Ge, Ori Gershony, Tim Hesterberg, and Diane Lambert. Evaluating online ad campaigns
in a pipeline: causal models at scale. In Proceedings of the 16th ACM SIGKDD International Conference
on Knowledge Discovery and Data Mining, pages 7?16. ACM, 2010.
[7] Yale Chang and Jennifer G Dy. Informative subspace learning for counterfactual inference. In Proceedings
of the Thirty-First AAAI Conference on Artificial Intelligence, pages 1770?1776, 2017.
[8] Rajeev H Dehejia and Sadek Wahba. Propensity score-matching methods for nonexperimental causal
studies. Review of Economics and Statistics, 84(1):151?161, 2002.
[9] Alexis Diamond and Jasjeet S Sekhon. Genetic matching for estimating causal effects: A general
multivariate matching method for achieving balance in observational studies. Review of Economics and
Statistics, 95(3):932?945, 2013.
[10] Doris Entner, Patrik Hoyer, and Peter Spirtes. Data-driven covariate selection for nonparametric estimation
of causal effects. In Proceedings of the Sixteenth International Conference on Artificial Intelligence and
Statistics, pages 256?264, 2013.
[11] David Galles and Judea Pearl. An axiomatic characterization of causal counterfactuals. Foundations of
Science, 3(1):151?182, 1998.
[12] Thomas A Glass, Steven N Goodman, Miguel A Hern?n, and Jonathan M Samet. Causal inference in
public health. Annual Review of Public Health, 34:61?75, 2013.
[13] Xiaofei He and Partha Niyogi. Locality preserving projections. In Advances in Neural Information
Processing Systems, pages 153?160, 2004.
[14] James J Heckman, Hidehiko Ichimura, and Petra Todd. Matching as an econometric evaluation estimator.
The Review of Economic Studies, 65(2):261?294, 1998.
[15] Daniel N Hill, Robert Moakler, Alan E Hubbard, Vadim Tsemekhman, Foster Provost, and Kiril Tsemekhman. Measuring causal impact of online actions via natural experiments: application to display
advertising. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery
and Data Mining, pages 1839?1847. ACM, 2015.
[16] Jennifer L Hill. Bayesian nonparametric modeling for causal inference. Journal of Computational and
Graphical Statistics, 20(1):217?240, 2012.
[17] Stefano M Iacus, Gary King, and Giuseppe Porro. Causal inference without balance checking: Coarsened
exact matching. Political Analysis, 20(1):1?24, 2011.
[18] Kosuke Imai and Marc Ratkovic. Covariate balancing propensity score. Journal of the Royal Statistical
Society: Series B (Statistical Methodology), 76(1):243?263, 2014.
[19] Guido Imbens and Donald B. Rubin. Causal Inference for Statistics, Social, and Biomedical Sciences: An
Introduction. Cambridge University Press, 2015.
[20] Hui Jin and Donald B Rubin. Principal stratification for causal inference with extended partial compliance.
Journal of the American Statistical Association, 103(481):101?111, 2008.
[21] Fredrik D. Johansson, Uri Shalit, and David Sontag. Learning representations for counterfactual inference.
In Proceedings of the 33nd International Conference on Machine Learning, pages 3020?3029, 2016.
9
[22] Ian Jolliffe. Principal Component Analysis. John Wiley and Sons, 2002.
[23] Robert J LaLonde. Evaluating the econometric evaluations of training programs with experimental data.
The American Economic Review, pages 604?620, 1986.
[24] Brian K Lee, Justin Lessler, and Elizabeth A Stuart. Improving propensity score weighting using machine
learning. Statistics in Medicine, 29(3):337?346, 2010.
[25] Sheng Li, Nikos Vlassis, Jaya Kawale, and Yun Fu. Matching via dimensionality reduction for estimation
of treatment effects in digital marketing campaigns. In Proceedings of the Twenty-Fifth International Joint
Conference on Artificial Intelligence, pages 3768?3774, 2016.
[26] Qingshan Liu, Xiaoou Tang, Hanqing Lu, Songde Ma, et al. Face recognition using kernel scatterdifference-based discriminant analysis. IEEE Transactions on Neural Networks, 17(4):1081?1085, 2006.
[27] Jerzy Neyman. On the application of probability theory to agricultural experiments. Essay on principles.
Section 9. Statistical Science, 5(4):465?480, 1923.
[28] Sinno Jialin Pan, James T Kwok, and Qiang Yang. Transfer learning via dimensionality reduction. In
Proceedings of the Twenty-Third AAAI Conference on Artificial Intelligence, volume 8, pages 677?682,
2008.
[29] Judea Pearl. Causality. Cambridge University Press, 2009.
[30] Deborah N Peikes, Lorenzo Moreno, and Sean Michael Orzol. Propensity score matching: A note of
caution for evaluators of social programs. The American Statistician, 62(3):222?231, 2008.
[31] Paul R Rosenbaum and Donald B Rubin. The central role of the propensity score in observational studies
for causal effects. Biometrika, 70(1):41?55, 1983.
[32] Donald B Rubin. Matching to remove bias in observational studies. Biometrics, pages 159?183, 1973.
[33] Donald B Rubin. Estimating causal effects of treatments in randomized and nonrandomized studies.
Journal of Educational Psychology, 66(5):688?701, 1974.
[34] Donald B Rubin. Using multivariate matched sampling and regression adjustment to control bias in
observational studies. Journal of the American Statistical Association, 74(366):318?328, 1979.
[35] Donald B Rubin and Neal Thomas. Combining propensity score matching with additional adjustments for
prognostic covariates. Journal of the American Statistical Association, 95(450):573?585, 2000.
[36] Adam C Sales, Asa Wilks, and John F Pane. Student usage predicts treatment effect heterogeneity in the
cognitive tutor algebra i program. In Proceedings of the International Conference on Educational Data
Mining, pages 207?214, 2016.
[37] Uri Shalit, Fredrik Johansson, and David Sontag. Bounding and minimizing counterfactual error. arXiv
preprint arXiv:1606.03976, 2016.
[38] Ricardo Silva and Robin Evans. Causal inference through a witness protection program. In Advances in
Neural Information Processing Systems, pages 298?306, 2014.
[39] Peter Spirtes. Introduction to causal inference. Journal of Machine Learning Research, 11(May):1643?
1662, 2010.
[40] Elizabeth A Stuart. Matching methods for causal inference: A review and a look forward. Statistical
science: a review journal of the Institute of Mathematical Statistics, 25(1):1?21, 2010.
[41] Wei Sun, Pengyuan Wang, Dawei Yin, Jian Yang, and Yi Chang. Causal inference via sparse additive
models with application to online advertising. In Proceedings of Twenty-Ninth AAAI Conference on
Artificial Intelligence, pages 297?303, 2015.
[42] Stefan Wager and Susan Athey. Estimation and inference of heterogeneous treatment effects using random
forests. arXiv preprint arXiv:1510.04342, 2015.
[43] Pengyuan Wang, Wei Sun, Dawei Yin, Jian Yang, and Yi Chang. Robust tree-based causal inference for
complex ad effectiveness analysis. In Proceedings of the Eighth ACM International Conference on Web
Search and Data Mining, pages 67?76. ACM, 2015.
10
[44] Pengyuan Wang, Dawei Yin, Jian Yang, Yi Chang, and Marsha Meytlis. Rethink targeting: detect
?smart cheating? in online advertising through causal inference. In Proceedings of the 24th International
Conference on World Wide Web Companion, pages 133?134, 2015.
[45] Kun Zhang, Mingming Gong, and Bernhard Sch?lkopf. Multi-source domain adaptation: A causal view.
In Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, pages 3150?3157, 2015.
[46] Kun Zhang, Bernhard Sch?lkopf, Krikamol Muandet, and Zhikun Wang. Domain adaptation under target
and conditional shift. In Proceedings of the International Conference on Machine Learning (3), pages
819?827, 2013.
11
| 6694 |@word exploitation:1 version:1 prognostic:1 nd:1 johansson:2 essay:1 seek:1 simulation:1 lpp:7 tr:6 klk:1 carry:1 reduction:2 hunting:1 liu:1 contains:6 score:17 att:16 selecting:1 series:2 genetic:2 rkhs:4 document:4 ours:1 daniel:1 past:1 existing:5 outperforms:1 current:1 com:1 comparing:1 nt:7 protection:1 scatter:16 yet:1 attracted:1 john:2 evans:1 additive:1 partition:1 informative:2 enables:1 krikamol:1 moreno:1 designed:2 remove:1 fund:1 infant:2 intelligence:7 selected:3 accordingly:2 xk:7 provides:1 characterization:1 earnings:3 evaluator:1 five:1 zhang:2 mathematical:1 replication:2 consists:1 introduce:1 g4:1 qingshan:1 behavior:1 karsten:1 dist:6 roughly:1 multi:6 discretized:1 little:1 considering:1 increasing:1 totally:1 provided:5 estimating:5 moreover:3 notation:2 project:2 panel:1 sinno:1 lowest:1 matched:1 minimizes:1 developed:3 caution:1 transformation:6 impractical:1 remy:1 guarantee:1 pseudo:9 mitigate:1 y3:1 every:3 nonrandom:1 tackle:2 xd:1 runtime:1 biometrika:1 classifier:2 k2:1 control:30 unit:56 partitioning:1 intervention:2 superiority:1 sale:1 before:3 treat:1 sd:4 limit:3 todd:1 merge:1 might:3 black:1 studied:2 equivalence:1 suggests:2 challenging:1 co:1 hidehiko:1 campaign:2 g7:1 range:1 directed:1 faithful:1 thirty:1 recursive:1 implement:2 procedure:4 area:1 empirical:1 significantly:2 matching:57 projection:9 word:1 integrating:1 donald:7 onto:1 close:5 selection:2 operator:1 targeting:1 put:1 impossible:2 map:4 xci:1 missing:6 maximizing:1 yt:1 educational:2 economics:3 attention:1 independently:1 survey:1 estimator:34 utilizing:1 kay:1 population:1 n12:2 annals:1 target:1 rationality:1 exact:2 guido:3 us:1 designing:1 alexis:1 trick:3 element:1 recognition:1 utilized:1 ignorable:1 predicts:1 agricultural:1 observed:3 coarsened:2 steven:2 p5:2 role:1 solved:1 preprint:2 wang:4 calculate:4 susan:2 subclassification:1 sun:2 eu:2 trade:2 kawale:1 yk:10 balanced:26 intuition:1 deborah:1 covariates:28 ideally:1 econometrica:1 dynamic:1 n1t:1 rewrite:1 algebra:1 smart:1 predictive:4 asa:1 basis:2 joint:1 xiaoou:1 various:1 derivation:1 train:3 effective:5 artificial:7 outcome:42 choosing:1 birth:1 widely:4 larger:1 supplementary:4 posed:1 otherwise:2 statistic:10 addi:1 g1:1 niyogi:1 noisy:1 validates:1 final:2 online:4 mah:5 eigenvalue:1 propose:8 lowdimensional:1 adaptation:7 combining:1 achieve:5 academy:1 sixteenth:1 g9:1 assessing:1 adam:1 tk:6 wider:1 derive:1 develop:1 tim:1 gong:1 miguel:1 ij:1 nearest:9 school:1 received:4 eq:16 job:3 dividing:1 implemented:1 fredrik:2 indicate:1 implies:1 rosenbaum:1 rasch:1 petra:1 drawback:1 closely:1 owing:1 enable:1 observational:14 public:2 education:2 government:1 assign:1 samet:1 proposition:3 biological:1 brian:1 exploring:1 rong:1 jerzy:1 hold:1 sekhon:1 marriage:1 considered:2 ground:4 normal:2 exp:1 great:1 mapping:1 predict:2 g8:1 major:4 hispanic:1 achieves:2 smallest:1 vary:1 adopt:2 purpose:1 estimation:26 axiomatic:1 label:20 propensity:14 hubbard:1 successfully:1 stefan:1 clearly:1 gaussian:1 mation:1 aim:4 brodersen:1 avoid:1 varying:1 office:1 categorizing:1 focus:2 jaya:1 mainly:2 greatly:1 political:3 sigkdd:2 baseline:5 detect:1 sense:1 nonexperimental:1 inference:29 tional:1 dependent:1 glass:1 hesterberg:1 kc:5 dnn:4 selects:1 issue:2 classification:7 aforementioned:1 flexible:1 arg:2 development:1 art:4 special:1 field:1 construct:3 having:1 beach:1 sampling:1 stratification:1 qiang:1 kw:6 stuart:2 look:1 k2f:1 athey:2 discrepancy:11 future:1 report:1 few:1 tightly:1 national:2 individual:1 statistician:1 highly:1 mining:4 evaluation:5 introduces:1 analyzed:1 light:1 jialin:1 wager:1 kt:9 accurate:1 fu:2 closer:1 capable:1 arthur:1 partial:1 orthogonal:1 biometrics:1 tree:2 conduct:1 old:1 euclidean:2 desired:2 shalit:2 causal:53 theoretical:2 increased:1 modeling:2 contiguous:1 w911nf:1 measuring:1 assignment:3 maximization:1 e49:1 cost:1 subset:2 conducted:3 kxi:1 synthetic:6 muandet:1 st:1 density:1 international:10 randomized:5 sensitivity:1 borgwardt:1 probabilistic:1 off:2 receiving:1 lee:1 michael:1 kcc:2 aaai:4 central:1 choose:3 cognitive:1 american:5 leading:1 return:1 ricardo:1 li:3 combing:1 account:2 potential:10 converted:3 summarized:1 student:1 pengyuan:3 explicitly:2 race:2 ad:2 performed:3 tion:2 ori:1 closed:4 view:1 counterfactuals:7 characterizes:1 complicated:1 parallel:1 contribution:1 partha:1 square:2 ni:2 efficiently:1 doris:1 correspond:1 ofthe:1 dealt:1 lkopf:3 bayesian:2 lambert:1 lu:1 advertising:3 researcher:1 neu:1 definition:3 kct:1 james:2 obvious:1 proof:1 mi:3 couple:1 judea:2 dataset:12 treatment:58 popular:3 adjusting:1 counterfactual:19 recall:1 knowledge:3 nnm:55 dimensionality:2 hilbert:1 sean:1 actually:3 supervised:2 follow:1 ichimura:1 methodology:1 wei:2 strongly:1 marketing:2 implicit:1 smola:2 biomedical:1 sheng:2 web:2 nonlinear:27 rajeev:1 logistic:2 quality:1 usage:1 usa:1 effect:36 facilitate:1 true:4 unbiased:2 counterpart:3 former:1 assigned:1 spirtes:2 neal:1 white:2 mahalanobis:2 sin:1 wilks:1 criterion:17 yun:2 hill:2 complete:1 demonstrate:2 performs:2 stefano:1 silva:1 dawei:3 novel:7 recently:4 empirically:2 volume:1 discussed:1 extend:1 illness:1 lalonde:5 he:1 association:3 cambridge:2 dag:1 mother:1 rd:8 lihong:1 stable:1 han:1 supervision:1 gj:1 etc:2 multivariate:2 recent:1 confounding:1 perspective:2 rnnm:8 moderate:1 belongs:2 chan:1 scenario:1 driven:1 certain:2 n00014:1 binary:5 success:1 discretizing:1 onr:1 yi:4 preserving:2 additional:2 care:4 nikos:1 employed:1 determine:1 paradigm:2 maximize:1 redundant:1 imai:2 ii:1 multiple:8 gretton:1 ing:1 alan:1 faster:1 adapt:1 cross:3 long:1 divided:2 alberto:1 equally:1 award:3 adobe:2 prediction:6 impact:2 regression:6 basic:1 heterogeneous:2 patient:1 metric:3 arxiv:4 kernel:25 sometimes:1 mmd:6 agarwal:1 achieved:1 penalize:1 background:1 addition:1 addressed:1 median:4 jian:3 source:1 sch:3 goodman:1 rest:1 vadim:1 heath:1 zhikun:1 compliance:1 leveraging:1 effectiveness:3 structural:1 leverage:1 yang:4 intermediate:4 xj:7 psychology:1 architecture:1 bandwidth:1 wahba:1 reduce:1 idea:5 economic:2 intensive:1 shift:2 whether:2 motivated:1 pca:7 unconfoundedness:2 peter:3 sontag:2 action:1 deep:3 ignored:1 generally:1 detailed:2 eigenvectors:1 involve:1 giuseppe:1 nonparametric:2 extensively:2 category:3 reduced:2 generate:3 exist:1 xij:7 percentage:1 nsf:1 kosuke:1 sign:1 estimated:4 diverse:1 discrete:1 n1c:1 group:40 key:2 achieving:1 drawn:1 econometric:2 graph:1 convert:5 year:1 run:2 jose:1 psm:8 uncertainty:2 named:4 throughout:1 utilizes:1 dy:1 pushed:1 ki:6 tackled:1 yale:1 fold:2 display:1 annual:1 constraint:3 alex:1 x2:5 noncontiguous:6 pane:1 structured:1 according:1 poor:1 across:2 beneficial:1 son:1 elizabeth:2 pan:1 g3:1 imbens:3 projecting:3 explained:1 pr:1 intuitively:1 handling:1 g6:1 pipeline:1 neyman:2 jennifer:2 hern:1 fail:2 jolliffe:1 ordinal:16 ge:1 end:1 serf:1 adopted:1 available:1 observe:5 kwok:1 away:1 simulating:1 alternative:1 original:7 thomas:2 denotes:1 clustering:1 include:4 ensure:1 graphical:2 e57:1 xc:15 medicine:3 calculating:1 ting:1 build:2 especially:1 classical:1 society:1 tutor:1 seeking:1 objective:5 question:3 added:1 koehler:1 strategy:10 traditional:4 g5:1 hoyer:1 heckman:1 subspace:10 distance:10 mapped:1 simulated:4 considers:1 collected:2 discriminant:1 reason:3 relationship:2 balance:4 demonstration:1 minimizing:1 nc:5 equivalently:1 executed:1 kun:2 robert:2 gk:1 trace:1 negative:2 design:4 proper:1 unknown:2 perform:6 diamond:1 twenty:4 observation:1 ihdp:3 datasets:4 benchmark:2 fabian:1 xiaofei:1 jin:1 heterogeneity:1 witness:1 extended:1 incorporated:3 precise:1 excluding:1 jim:1 rn:2 vlassis:1 reproducing:1 provost:1 ninth:2 introduced:4 david:4 pair:3 timal:1 kl:1 required:1 extensive:2 cheating:1 learned:1 pearl:2 nip:1 address:4 able:1 bar:1 kriegel:1 usually:6 justin:1 scott:1 yc:6 eud:4 eighth:1 program:8 including:3 max:2 royal:1 power:1 critical:1 overlap:2 treated:14 malte:1 natural:1 improve:1 lorenzo:1 fcf:2 extract:4 health:5 lij:3 patrik:1 review:7 acknowledgement:1 discovery:2 checking:1 fully:1 expect:2 generation:3 acyclic:1 age:1 digital:2 validation:2 foundation:1 degree:1 rubin:8 foster:1 principle:1 share:1 balancing:4 summary:1 repeat:2 supported:2 bias:8 guide:1 understand:1 institute:1 neighbor:10 wide:2 taking:1 face:1 deepak:1 absolute:1 fifth:1 sparse:1 dimension:9 xn:2 world:5 cure:1 calculated:1 evaluating:2 forward:1 san:1 projected:1 osd:10 premature:1 employing:1 preprocessing:1 income:1 social:2 transaction:1 obtains:1 status:1 bernhard:3 unreliable:1 dealing:1 cem:2 consuming:1 xi:8 discriminative:3 alternatively:2 continuous:3 search:1 decade:1 table:6 robin:1 promising:1 learn:8 mj:2 robust:5 nicolas:1 ca:2 transfer:1 obtaining:1 improving:1 forest:2 mse:6 diane:1 complex:1 meanwhile:2 domain:7 protocol:1 marc:1 did:1 main:1 whole:1 motivation:1 hyperparameters:1 paul:1 bounding:1 child:1 complementary:1 categorized:1 x1:3 augmented:1 causality:1 wiley:1 inferring:1 explicit:1 third:4 weighting:4 learns:3 ix:1 young:1 northeastern:1 removing:1 ian:1 tang:1 companion:1 xt:16 covariate:11 abadie:1 g10:1 importance:1 hui:1 magnitude:1 uri:2 boston:1 locality:2 yin:3 army:1 adjustment:2 g2:1 chang:4 gary:1 truth:5 extracted:2 ma:2 acm:6 conditional:2 identity:1 formulated:2 king:1 shared:1 fisher:1 considerable:1 specifically:1 justify:1 principal:3 total:3 ece:1 experimental:5 entner:1 indicating:1 formally:2 select:2 highdimensional:1 latter:1 unbalanced:1 categorize:2 alexander:1 bioinformatics:1 jonathan:1 investigator:1 incorporate:1 evaluate:2 correlated:1 |
6,295 | 6,695 | Learning Overcomplete HMMs
Vatsal Sharan
Stanford University
[email protected]
Sham Kakade
University of Washington
[email protected]
Percy Liang
Stanford University
[email protected]
Gregory Valiant
Stanford University
[email protected]
Abstract
We study the problem of learning overcomplete HMMs?those that have many
hidden states but a small output alphabet. Despite having significant practical
importance, such HMMs are poorly understood with no known positive or negative
results for efficient learning. In this paper, we present several new results?both
positive and negative?which help define the boundaries between the tractable and
intractable settings. Specifically, we show positive results for a large subclass of
HMMs whose transition matrices are sparse, well-conditioned, and have small
probability mass on short cycles. On the other hand, we show that learning is
impossible given only a polynomial number of samples for HMMs with a small
output alphabet and whose transition matrices are random regular graphs with large
degree. We also discuss these results in the context of learning HMMs which can
capture long-term dependencies.
1
Introduction
Hidden Markov Models (HMMs) are commonly used for data with natural sequential structure (e.g.,
speech, language, video). This paper focuses on overcomplete HMMs, where the number of output
symbols m is much smaller than the number of hidden states n. As an example, for an HMM that
outputs natural language documents one character at a time, the number of characters m is quite
small, but the number of hidden states n would need to be very large to encode the rich syntactic,
semantic, and discourse structure of the document.
Most algorithms for learning HMMs with provable guarantees assume the transition T ? Rn?n and
observation O ? Rm?n matrices are full rank [2, 3, 20] and hence do not apply to the overcomplete
regime. A notable exception is the recent work of Huang et al. [14] who studied this setting where
m n and showed that generic HMMs can be learned in polynomial time given exact moments
of the output process (which requires infinite data). Though understanding properties of generic
HMMs is an important first step, in reality, HMMs with a large number of hidden states typically have
structured, non-generic transition matrices?e.g., consider sparse transition matrices or transition
matrices of factorial HMMs [12]. Huang et al. [14] also assume access to exact moments, which
leaves open the question of when learning is possible with efficient sample complexity. Summarizing,
we are interested in the following questions:
1. What are the fundamental limitations for learning overcomplete HMMs?
2. What properties of HMMs make learning possible with polynomial samples?
3. Are there structured HMMs which can be learned in the overcomplete regime?
Our contributions. We make progress on all three questions in this work, sharpening our understanding of the boundary between tractable and intractable learning. We begin by stating a negative
result, which perhaps explains some of the difficulty of obtaining strong learning guarantees in the
overcomplete setting.
Theorem 1. The parameters of HMMs where i) the transition matrix encodes a random walk on a
regular graph on n nodes with degree polynomial in n, ii) the output alphabet m = polylog(n) and,
iii) the output distribution for each hidden state is chosen uniformly and independently at random,
cannot be learned (even approximately) using polynomially many samples over any window length
polynomial in n, with high probability over the choice of the observation matrix.
Theorem 1 is somewhat surprising, as parameters of HMMs with such transition matrices can be
easily learned in the non-overcomplete (m ? n) regime. This is because such transition matrices
are full-rank and their condition numbers are polynomial in n; hence spectral techniques such
as Anandkumar et al. [3] can be applied. Theorem 1 is also fundamentally of a different nature
as compared to lower bounds based on parity with noise reductions for HMMs [20], as ours is
information-theoretic.1 Also, it seems far more damning as the hard cases are seemingly innocuous
classes such as random walks on dense graphs. The lower bound also shows that analyzing generic or
random HMMs might not be the right framework to consider in the overcomplete regime as these
might not be learnable with polynomial samples even though they are identifiable. This further
motivates the need for understanding HMMs with structured transition matrices. We provide a proof
of Theorem 1 with more explicitly stated conditions in Appendix D.
For our positive results we focus on understanding properties of structured transition matrices
which make learning tractable. To disentangle additional complications due to the choice of the
observation matrix, we will assume that the observation matrix is drawn at random throughout the
paper. Long-standing open problems on learning aliased HMMs (HMMs where multiple hidden states
have identical output distributions) [7, 15, 23] hint that understanding learnability with respect to
properties of the observation matrix is a daunting task in itself, and is perhaps best studied separately
from understanding how properties of the transition matrix affect learning.
Our positive result on learnability (Theorem 2) depends on two natural graph-theoretic properties of
the transition matrix. We consider transition matrices which are i) sparse (hidden states have constant
degree) and ii) have small probability mass on cycles shorter than 10 logm n states?and show that
these HMMs can be learned efficiently using tensor decomposition and the method of moments,
given random observation matrices. The condition prohibiting short cycles might seem mysterious.
Intuitively, we need this condition to ensure that the Markov Chain visits a sufficient large portion of
the state space in a short interval of time, and in fact the condition stems from information-theoretic
considerations. We discuss these further in Sections 2.4 and 3.1. We also discuss how our results
relate to learning HMMs which capture long-term dependencies in their outputs, and introduce a new
notion of how well an HMM captures long-term dependencies. These are discussed in Section 5.
We also show new identifiability results for sparse HMMs. These results provide a finer picture of
identifiability than Huang et al. [14], as ours hold for sparse transition matrices which are not generic.
Technical contribution. To prove Theorem 2 we show that the Khatri-Rao product of dependent
random vectors is well-conditioned under certain conditions. Previously, Bhaskara et al. [6] showed
that the Khatri-Rao product of independent random vectors is well-conditioned to perform a smoothed
analysis of tensor decomposition, their techniques however do not extend to the dependent case. For
the dependent case, we show a similar result using a novel Markov chain coupling based argument
which relates the condition number to the best coupling of output distributions of two random walks
with disjoint starting distributions. The technique is outlined in Section 2.2.
Related work. Spectral methods for learning HMMs have been studied in Anandkumar et al.
[3], Bhaskara et al. [5], Allman et al. [1], Hsu et al. [13], but these results require m ? n. In Allman
et al. [1], the authors show that that HMMs are identifiable given moments of continuous observations
over a time interval of length N = 2? + 1 for some ? such that ? +m?1
? n. When m n
m?1
1/m
this requires ? = O(n
). Bhaskara et al. [5] give another bound on window size which requires
? = O(n/m). However, with a output alphabet of size m, specifying all moments in a N length
continuous time interval requires mN time and samples, and therefore all of these approaches lead to
exponential runtimes when m is constant with respect to n. Also relevant is the work by Anandkumar
et al. [4] on guarantees for learning certain latent variable models such as Gaussian mixtures in the
overcomplete setting through tensor decomposition. As mentioned earlier, the work closest to ours is
Huang et al. [14] who showed that generic HMMs are identifiable with ? = O(logm n), which gives
the first polynomial runtimes for the case when m is constant.
1
Parity with noise is information theoretically easy given observations over a window of length at least the
number of inputs to the parity. This is linear in the number of hidden states of the parity with noise HMM,
whereas Theorem 1 says that the sample complexity must be super polynomial for any polynomial sized window.
2
Outline. Section 2 introduces the notation and setup. It also provides examples and a high-level
overview of our proof approach. Section 3 states the learnability result, discusses our assumptions
and HMMs which satisfy these assumptions. Section 4 contains our identifiability results for sparse
HMMs. Section 5 discusses natural measures of long-term dependencies in HMMs. We conclude in
Section 6. Proof details are deferred to the Appendix.
2
Setup and preliminaries
In this section we first introduce the required notation, and then outline the method of moments
approach for parameter recovery. We also go over some examples to provide a better understanding
of the classes of HMMs we aim to learn, and give a high level proof strategy.
2.1
Notation and preliminaries
We will denote the output at time t by yt and the hidden state at time t by ht . Let the number of hidden
states be n and the number of observations be m. Assume that the output alphabet is {0, . . . , m ? 1}
without loss of generality. Let T be the transition matrix and O be the observation matrix of the
HMM, both of these are defined so that the columns add up to one. For any matrix A, we refer to
the ith column of A as Ai . T 0 is defined as the transition matrix of the time-reversed Markov chain,
but we do not assume reversibility and hence T may not equal T 0 . Let yij = yi , . . . , yj denote the
sequence of outputs from time i to time j. Let lij = li , . . . , lj refer to a string of length i + j ? 1
over the output alphabet, denoting a particular output sequence from time i to j. Define a bijective
mapping L which maps an output sequence l1? ? {0, . . . , m?1}? into an index L(l1? ) ? {1, . . . , m? }
and the associated inverse mapping L?1 .
Throughout the paper, we assume that the transition matrix T is ergodic, and hence has a stationary
distribution. We also assume that every hidden state has stationary probability at least 1/poly(n).
This is a necessary condition, as otherwise we might not even visit all states in poly(n) samples. We
also assume that the output process of the HMM is stationary. A stochastic process is stationary
if the distribution of any subset of random variables is invariant with respect to shifts in the time
? +T
?
?
?
?
index?that is, P[y??
= l??
] = P[y??
+T = l?? ] for any ?, T and string l?? . This is true if the
initial hidden state is chosen according to the stationary distribution.
Our results depend on the conditioning of the matrix T with respect to the `1 norm. We define
(1)
?min (T ) as the minimum `1 gain of the transition matrix T over all vectors x having unit `1 norm
(not just non-negative vectors x, for which the ratio would always be 1):
(1)
?min (T ) = minn
x?R
(1)
kT xk1
kxk1
(1)
?min (T ) is also a natural parameter to measure the long-term dependence of the HMM?if ?min (T )
is large then T preserves significant information about the distribution of hidden states at time 0 at a
future time t, for all initial distributions at time 0. We discuss this further in Section 5.
2.2
Method of moments for learning HMMs
Our algorithm for learning HMMs follows the method of moments based approach, outlined for
example in Anandkumar et al. [2] and Huang et al. [14]. In contrast to the more popular ExpectationMaximization (EM) approach which can suffer from slow convergence and local optima [21], the
method of moments approach ensures guaranteed recovery of the parameters under mild conditions.
More details about tensor decomposition and the method of moments approach to learning HMMs
can be found in Appendix A.
The method of moments approach to learning HMMs has two high-level steps. In the first step, we
write down a tensor of empirical moments of the data, such that the factors of the tensor correspond to
parameters of the underlying model. In the second step, we perform tensor decomposition to recover
the factors of the tensor?and then recover the parameters of the model from the factors. The key fact
that enables the second step is that tensors have a unique decomposition under mild conditions on
their factors, for example tensors have a unique decomposition if all the factors are full rank. The
uniqueness of tensor decomposition permits unique recovery of the parameters of the model.
?
We will learn the HMM using the moments of observation sequences y??
from time ?? to ? .
Since the output process is assumed to be stationary, the distribution of outputs is the same for
3
any contiguous time interval of the same length, and we use the interval ?? to ? in our setup for
convenience. We call the length of the observation sequences used for learning the window length
N = 2? + 1. Since the number of samples required to estimate moments over a window of length N
is mN , it is desirable to keep N small. Note that to ensure polynomial runtime and sample complexity
for the method of moments approach, the window length N must be O(logm n).
We will now define our moment tensor. Given moments over a window of length N = 2? + 1, we
?
?
can construct the third-order moment tensor M ? Rm ?m ?m using the mapping L from strings of
outputs to indices in the tensor:
?
?
M(L(l? ),L(l?? ),l0 ) = P[y??
= l??
].
1
?1
M is simply the tensor of the moments of the HMM over a window length N , and can be estimated
directly from data. We can write M as an outer product because of the Markov property:
M =A?B?C
?
where A ? Rm
state at time 0):
?n
, B ? Rm
?
?n
, C ? Rm?n are defined as follows (here h0 denotes the hidden
AL(l1? ),i = P[y1? = l1? | h0 = i]
??
??
BL(l?? ),i = P[y?1
= l?1
| h0 = i]
?1
Cl0 ,i = P[y0 = l, h0 = i]
T and O can be related in a simple manner to A, B and C. If we can decompose the tensor M into the
factors A, B and C, we can recover T and O from A, B and C. Kruskal?s condition [18] guarantees
that tensors have a unique decomposition whenever A and B are full rank and no two column of C
are the same. We refer the reader to Appendix A for more details, specifically Algorithm 1.
2.3
High-level proof strategy
As the transition and observation matrices can be recovered from the factors of the tensors, our goal
is to analyze the conditions under which the tensor decomposition step works provably. Note that the
factor matrix A is the likelihood of observing each sequence of observations conditioned on starting
at a given hidden state. We?ll refer to A as the likelihood matrix for this reason. B is the equivalent
matrix for the time-reversed Markov chain. If we show that A, B are full rank and no two columns of
C are the same, then the HMM can be learned provided the exact moments using the simultaneous
diagonalization algorithm, also known as Jennrich?s algorithm (see Algorithm 1). We show this
property for our identifiability results. For our learnability results, we show that the matrices A and
B are well-conditioned (have condition numbers polynomial in n), which implies learnability from
polynomial samples. This is the main technical contribution of the paper, and requires analyzing
the condition number of the Khatri-Rao product of dependent random vectors. Before sketching the
argument, we first introduce some notation. We can define A(t) as the likelihood matrix over t steps:
(t)
AL(lt ),i = P[y1t = l1t | h0 = i].
1
A(t) can be recursively written down as follows:
A(0) = OT, A(t) = (O A(t?1) )T
(1)
where A B, denotes the Khatri-Rao product of the matrices A and B. If A and B are two matrices
of size m1 ? r and m2 ? r then the Khatri-Rao product is a m1 m2 ? r matrix whose ith column is
the outer product Ai ? Bi flattened into a vector. Note that A(? ) is the same as A. We now sketch
our argument for showing that A(? ) is well-conditioned under appropriate conditions.
Coupling random walks to analyze the Khatri-Rao product. As mentioned in the introduction,
in this paper we are interested in the setting where the transition matrix is fixed but the observation
matrix is drawn at random. If we could draw fresh random matrices O at each time step of the
recursion in Eq. 1, then A would be well-conditioned by the smoothed analysis of the Khatri-Rao
product due to Bhaskara et al. [6]. However, our setting is significantly more difficult, as we do not
have access to fresh randomness at each time step, so the techniques of Bhaskara et al. [6] cannot be
applied here. As pointed out earlier, the condition number of A in this scenario depends crucially on
the transition matrix T , as A is not even full rank if T = I.
4
(a) Transition matrix is a cycle, or a
permutation on the hidden states.
(b) Transition matrix is a random walk on a
graph with small degree and no short cycles.
Figure 1: Examples of transition matrices which we can learn, refer to Section 2.4 and Section 3.2.
Instead, we analyze A by a coupling argument. To get some intuition for this, note that if A does not
have full rank, then there are two disjoint sets of columns of A whose linear combinations are equal,
and these combination weights can be used to setup the initial states of two random walks defined by
the transition matrix T which have the same output distribution for ? time steps. More generally, if
A is ill-conditioned then there are two random walks with disjoint starting states which have very
similar output distributions. We show that if two random walks have very similar output distributions
over ? time steps for a randomly chosen observation matrix O, then most of the probability mass in
(1)
these random walks can be coupled. On the other hand, if (?min (T ))? is sufficiently large, the total
variational distance between random walks starting at two different starting states must be at least
(1)
(?min (T ))? after ? time steps, and so there cannot be a good coupling, and A is well-conditioned.
We provide a sketch of the argument for a simple case in Appendix 1 before we prove Theorem 2.
2.4
Illustrative examples
We now provide a few simple examples which will illustrate some classes of HMMs we can and
cannot learn. We first provide an example of a class of simple HMMs which can be handled by our
results, but has non-generic transition matrices and hence does not fit into the framework of Huang
et al. [14]. Consider an HMM where the transition matrix is a permutation or cyclic shift on the
hidden states (see Fig. 1a). Our results imply that such HMMs are learnable in polynomial time from
polynomial samples if the output distributions of the hidden states are chosen at random. We will
try to provide some intuition about why an HMM with the transition matrix as in Fig. 1a should be
efficiently learnable. Let us consider the simple case when the the outputs are binary (so m = 2) and
each hidden state deterministically outputs a 0 or a 1, and is labeled by a 0 or a 1 accordingly. If
the labels are assigned at random, then with high probability the string of labels of any continuous
sequence of 2 log2 n hidden states in the cycle in Fig. 1a will be unique. This means that the output
distribution in a 2 log2 n time window is unique for every initial hidden state, and it can be shown
that this ensures that the moment tensor has a unique factorization. By showing that the output
distribution in a 2 log2 n time window is very different for different initial hidden states?in addition
to being unique?we can show that the factors of the moment tensor are well-conditioned, which
allows recovery with efficient sample complexity. As another slightly more complex example of
an HMM we can learn, Fig. 1b depicts an HMM whose transition matrix is a random walk on a
graph with small degree and no short cycles. Our learnability result can handle such HMMs having
structured transition matrices.
As an example of an HMM which cannot be learned in our framework, consider an HMM with
transition matrix T = I and binary observations (m = 2), see Fig. 2a. In this case, the probability of
an output sequence only depends on the total number of zeros or ones in the sequence. Therefore,
we only get t independent measurements from windows of length t, hence windows of length O(n)
instead of O(log2 n) are necessary for identifiability (also refer to Blischke [8] for more discussions
on this case). More generally, we prove in Proposition 1 that for small m a transition matrix composed
only of cycles of constant length (see Fig. 2b) requires the window length to be polynomial in n to
become identifiable.
Proposition 1. Consider an HMM on n hidden states and m observations with the transition
matrix
c
being a permutation composed of cycles of length c. Then windows of length O(n1/m ) are necessary
for the model to be identifiable, which is polynomial in n for constant c and m.
The root cause of the difficulty in learning HMMs having short cycles is that they do not visit a large
enough portion of the state space in O(logm n) steps, and hence moments over a O(logm n) time
5
(a) Transition matrix is the identity on
8 hidden states.
(b) Transition matrix is a union of 4 cycles,
each on 5 hidden states.
Figure 2: Examples of transition matrices which do not fit in our framework. Proposition 1 shows
that such HMMs where the transition matrix is composed of a union of cycles of constant length are
not even identifiable from short windows of length O(logm n)
window do not carry sufficient information for learning. Our results cannot handle such classes of
transition matrices, also see Section 3.1 for more discussion.
3
Learnability results for overcomplete HMMs
In this section, we state our learnability result, discuss the assumptions and provide examples of
HMMs which satisfy these assumptions. Our learnability results hold under the following conditions:
Assumptions: For fixed constants c1 , c2 , c3 > 1, the HMM satisfies the following properties for
some c > 0:
1. Transition matrix is well-conditioned: Both T and the transition matrix T 0 of the time
(1)
(1)
reversed Markov Chain are well-conditioned in the `1 -norm: ?min (T ), ?min (T 0 ) ? 1/mc/c1
2. Transition matrix does not have short cycles: For both T and T 0 , every state visits at least
10 logm n states in 15 logm n time except with probability ?1 ? 1/nc .
3. All hidden states have small ?degree?: There exists ?2 such that for every hidden state i, the
transition distributions Ti and Ti0 have cumulative mass at most ?2 on all but d states, with
d ? m1/c2 and ?2 ? 1/nc . Hence this is a soft ?degree? requirement.
4. Output distributions are random and have small support : There exists ?3 such that for every
hidden state i the output distribution Oi has cumulative mass at most ?3 on all but k outputs,
with k ? m1/c3 and ?3 ? 1/nc . Also, the output distribution Oi is drawn uniformly on
these k outputs.
The constants c1 , c2 , c3 are can be made explicit, for example, c1 = 20, c2 = 16 and c3 = 10 works.
Under these conditions, we show that HMMs can be learned using polynomially many samples:
Theorem 2. If an HMM satisfies the above conditions, then with high probability over the choice
of O, the parameters of the HMM are learnable to within additive error with observations over
windows of length 2? + 1, ? = 15 logm n, with the sample complexity poly(n, 1/).
Appendix C also states a corollary of Theorem 2 in terms of the minimum singular value ?min (T ) of
(1)
the matrix T , instead of ?min (T ). We discuss the conditions for Theorem 2 next, and subsequently
provide examples of HMMs which satisfy these conditions.
3.1
Discussion of the assumptions
1. Transition matrix is well-conditioned: Note that singular transition matrices might not even be
identifiable. Moreover, Mossel and Roch [20] showed that learning HMMs with singular transition
matrices is as hard as learning parity with noise, which is widely conjectured to be computationally
hard. Hence, it is necessary to exclude at least some classes of ill-conditioned transition matrices.
2. Transition matrix does not have short cycles: Due to Proposition 1, we know that a HMM might
not even be identifiable from short windows if it is composed of a union of short cycles, hence we
expect a similar condition for learning the HMM with polynomial samples; though there is a gap
between the upper and lower bounds in terms of the probability mass which is allowed on the short
cycles. We performed some simulations to understand how the length of cycles in the transition
matrix and the probability mass assigned to short cycles affects the condition number of the likelihood
matrix A; recall that the condition number of A determines the stability of the method of moments
6
Condition number of matrix A
Condition number of matrix A
100
80
60
40
Cycle length 2
Cycle length 4
Cycle length 8
20
0
0.1
0.2
0.3
Epsilon
0.4
0.5
(a) The conditioning becomes worse
when cycles are smaller or when more
probability mass is put on short cycles.
200
150
Degree 2
Degree 4
Degree 8
100
50
0
0.01
0.02
0.03
Epsilon
0.04
0.05
(b) The conditioning becomes worse as the
degree increases, and when more probabiltiy
mass is put on the dense part of T .
Figure 3: Experiments to study the effect of sparsity and short cycles on the learnability of HMMs.
The condition number of the likelihood matrix A determines the stability or sample complexity of the
method of moments approach. The condition numbers are averaged over 10 trials.
approach. We take the number of hidden states n = 128, and let P128 be a cycle on the n hidden
states (as in Fig. 1a). Let Pc be a union of short cycles of length c on the n states (refer to Fig. 2b
for an example). We take the transition matrix to be T = Pc + (1 ? )P128 for different values of
c and . Fig. 3a shows that the condition number of A becomes worse and hence learning requires
more samples if the cycles are shorter in length, and if more probability mass is assigned to the short
cycles, hinting that our conditions are perhaps not be too stringent.
3. All hidden states have a small degree: Condition 3 in Theorem 2 can be reinterpreted as saying
that the transition probabilities out of any hidden state must have mass at most 1/n1+c on any hidden
state except a set of d hidden states, for any c > 0. While this soft constraint is weaker than a hard
constraint on the degree, it natural to ask whether any sparsity is necessary to learn HMMs. As above,
we carry out simulations to understand how the degree affects the condition number of the likelihood
matrix A. We consider transition matrices on n = 128 hidden states which are a combination of a
dense part and a cycle. Define P128 to be a cycle as before. Define Gd as the adjacency matrix of a
directed regular graph with degree d. We take the transition matrix T = Gd + (1 ? d)P128 . Hence
the transition distribution of every hidden state has mass on a set of d neighbors, and the residual
probability mass is assigned to the permutation P128 . Fig. 3b shows that the condition number of
A becomes worse as the degree d becomes larger, and as more probability mass is assigned to
the dense part Gd of the transition matrix T , providing some weak evidence for the necessity of
Condition 3. Also, recall that Theorem 1 shows that HMMs where the transition matrix is a random
walk on an undirected regular graph with large degree (degree polynomial in n) cannot be learned
using polynomially many samples if m is a constant with
? respect to n. However, such graphs have
all eigenvalues except the first one to be less than O(1/ d), hence it is not clear if the hardness of
learning depends on the large degree itself or is only due to T being ill-conditioned. More concretely,
we pose the following open question:
Open question: Consider an HMM with a transition matrix T = (1 ? )P + U , where P is the
cyclic permutation on n hidden states (such as in Fig. 1a) and U is a random walk on a undirected,
regular graph with large degree (polynomial in n) and > 0 is a constant. Can this HMM be learned
using polynomial samples when m is small (constant) with respect to n? This example approximately
preserves ?min (T ) by the addition of the permutation, and hence the difficulty is only due to the
transition matrix having large degree.
4. Output distributions are random and have small support: As discussed in the introduction, if we
do not assume that the observation matrices are random, then even simple HMMs with a cycle or
permutation as the transition matrix might require long windows even to become identifiable, see Fig.
4. Hence some assumptions on the output distribution do seem necessary for learning the model from
short time windows, though our assumptions are probably not tight. For instance, the assumption that
the output distributions have a small support makes learning easier as it leads to the outputs being
more discriminative of the hidden states, but it is not clear that this is a necessary assumption. Ideally,
we would like to prove our learnability results under a smoothed model for O, where an adversary is
allowed to see the transition matrix T and pick any worst-case O, but random noise is then added to
7
the output distributions, which limits the power of the adversary. We believe our results should hold
under such a smoothed setting, but set this aside for future work.
Figure 4: Consider two HMMs with transition matrices being cycles on n = 16 states with binary
outputs, and outputs conditioned on the hidden states are deterministic. The states labeled as 0 always
emit a 0 and the states labeled as 1 always emit a 1. The two HMMs are not distinguishable from
windows of length less than 8. Hence with worst case O even simple HMMs like the cycle could
require long windows to even become identifiable.
3.2
Examples of transition matrices which satisfy our assumptions
We revisit the examples from Fig. 1a and Fig. 1b, showing that they satisfy our assumptions.
1. Transition matrices where the Markov Chain is a permutation: If the Markov chain is a permutation
with all cycles longer than 10 logm n then the transition matrix obeys all the conditions for Theorem
2. This is because all the singular values of a permutation are 1, the degree is 1 and all hidden states
visit 10 logm n different states in 15 logm n time steps.
2. Transition matrices which are random walks on graphs with small degree and large girth:
For directed graphs, Condition 2 can be equivalently stated as that the graph representation of the
transition matrix has a large girth (girth of a graph is defined as the length of its shortest cycle).
3. Transition matrices of factorial HMMs: Factorial HMMs [12] factor the latent state at any time into
D dimensions, each of which independently evolves according to a Markov process. For D = 2, this
is equivalent to saying that the hidden states are indexed by two labels (i, j) and if T1 and T2 represent
the transition matrices for the two dimensions, then P[(i1 , j1 ) ? (i2 , j2 )] = T1 (i2 , i1 )T2 (j2 , j1 ).
This naturally models settings where there are multiple latent concepts which evolve independently.
The following properties are easy to show:
1. If either of T1 or T2 visit N different states in 15 logm n time steps with probability (1 ? ?),
then T visits N different states in 15 logm n time steps with probability (1 ? ?).
2. ?min (T ) = ?min (T1 )?min (T2 )
3. If all hidden states in T1 and T2 have mass at most ? on all but d1 states and d2 states
respectively, then T has mass at most 2? on all but d1 d2 states.
Therefore, factorial HMMs are learnable with random O if the underlying processes obey conditions
similar to the assumptions for Theorem 2. If both T1 and T2 are well-conditioned and at least one of
them does not have short cycles, and either has small degree, then T is learnable with random O.
4
Identifiability of HMMs from short windows
As it is not obvious that some of the requirements for Theorem 2 are necessary, it is natural to attempt
to derive stronger results for just identifiability of HMMs having structured transition matrices. In this
section, we state our results for identifiability of HMMs from windows of size O(logm n). Huang
et al. [14] showed that all HMMs except those belonging to a measure zero set become identifiable
from windows of length 2? + 1 with ? = 8dlogm ne. However, the measure zero set itself might
possibly contain interesting classes of HMMs (see Fig. 1), for example sparse HMMs also belong to
a measure zero set. We refine the identifiability results in this section, and show that a natural sparsity
condition on the transition matrix guarantees identifiability from short windows. Given any transition
matrix T , we regard T as being supported by a set of indices S if the non-zero entries of T all lie in
S. We now state our result for identifiability of sparse HMMs.
Theorem 3. Let S be a set of indices which supports a permutation where all cycles have at least
2dlogm ne hidden states. Then the set T of all transition matrices with support S is identifiable from
windows of length 4dlogm ne + 1 for all observation matrices O except for a measure zero set of
transition matrices in T and observation matrices O.
8
We hypothesize that excluding a measure zero set of transition matrices in Theorem 3 should not be
necessary as long as the transition matrix is full rank, but are unable to show this. Note that our result
on identifiability is more flexible in allowing short cycles in transition matrices than Theorem 2, and
is closer to the lower bound on identifiability in Proposition 1.
We also strengthen the result of Huang et al. [14] for identifiability of generic HMMs. Huang
et al. [14] conjectured that windows of length 2dlogm ne + 1 are sufficient for generic HMMs to be
identifiable. The constant 2 is the information theoretic bound as an HMM on n hidden states and
m outputs has O(n2 + nm) independent parameters, and hence needs observations over a window
of size 2dlogm ne + 1 to be uniquely identifiable. Proposition 2 settles this conjecture, proving
the optimal window length requirement for generic HMMs to be identifiable. As the number of
possible outputs over a window of length t is mt , the size of the moment tensor in Section 2.2 is itself
exponential in the window length. Therefore even a factor of 2 improvement in the window length
requirement leads to a quadratic improvement in the sample and time complexity.
Proposition 2. The set of all HMMs is identifiable from observations over windows of length
2dlogm ne + 1 except for a measure zero set of transition matrices T and observation matrices O.
5
Discussion on long-term dependencies in HMMs
In this section, we discuss long-term dependencies in HMMs, and show how our results on overcomplete HMMs improve the understanding of how HMMs can capture long-term dependencies, both
(1)
with respect to the Markov chain and the outputs. Recall the definition of ?min (T ):
kT xk1
(1)
?min (T ) = minn
x?R
kxk1
(1)
We claim that if ?min (T ) is large, then the transition matrix preserves significant information about
the distribution of hidden states at time 0 at a future time t, for all initial distributions at time 0.
Consider any two distributions p0 and q0 at time 0. Let pt and qt be the distributions of the hidden
states at time t given that the distribution at time 0 is p0 and q0 respectively. Then the `1 distance
(1)
between pt and qt is kpt ? qt k1 ? (?min (T ))t kp0 ? q0 k1 , verifying our claim. It is interesting
to compare this notion with the mixing time of the transition matrix. Defining mixing time as
the time until the `1 distance between any two starting distributions is at most 1/2, it follows that
(1)
(1)
the mixing time ?mix ? 1/ log(1/?min (T )), therefore if ?min (T )) is large then the chain is slowly
(1)
mixing. However, the converse is not true??min (T ) might be small even if the chain never mixes,
for example if the graph is disconnected but the connected components mix very quickly. Therefore,
(1)
?min (T ) is possibly a better notion of the long-term dependence of the transition matrix, as it requires
that information is preserved about the past state ?in all directions?.
Another reasonable notion of the long-term dependence of the HMM is the long-term dependence in
the output process instead of in the hidden Markov chain, which is the utility of past observations
when making predictions about the distant future (given outputs y?? , . . . , y1 , y2 , . . . , yt , at time t
how far back do we need to remember about the past to make a good prediction about yt ?). This does
not depend in a simple way on the T and O matrices, but we do note that if the Markov chain is fast
mixing then the output process can certainly not have long-term dependencies. We also note that with
respect to long-term dependencies in the output process, the setting m n seems to be much more
interesting than when m is comparable to n. The reason is that in the small output alphabet setting
we only receive a small amount of information about the true hidden state at each step, and hence
longer windows are necessary to infer the hidden state and make a good prediction. We also refer the
reader to Kakade et al. [16] for related discussions on the memory of output processes of HMMs.
6
Conclusion and Future Work
The setting where the output alphabet m is much smaller than the number of hidden states n is
well-motivated in practice and seems to have several interesting theoretical questions about new
lower bounds and algorithms. Though some of our results are obtained in more restrictive conditions
than seems necessary, we hope the ideas and techniques pave the way for much sharper results in
this setting. Some open problems which we think might be particularly useful for improving our
understanding is relaxing the condition on the observation matrix being random to some structural
constraint on the observation matrix (such as on its Kruskal rank), and more thoroughly investigating
the requirement for the transition matrix being sparse and not having short cycles.
9
References
[1] E. S. Allman, C. Matias, and J. A. Rhodes. Identifiability of parameters in latent structure
models with many observed variables. Annals of Statistics, 37:3099?3132, 2009.
[2] A. Anandkumar, D. J. Hsu, and S. M. Kakade. A method of moments for mixture models and
hidden markov models. In COLT, volume 1, page 4, 2012.
[3] A. Anandkumar, R. Ge, D. Hsu, S. M. Kakade, and M. Telgarsky. Tensor decompositions for
learning latent variable models. arXiv, 2013.
[4] A. Anandkumar, R. Ge, and M. Janzamin. Learning overcomplete latent variable models
through tensor methods. In COLT, pages 36?112, 2015.
[5] A. Bhaskara, M. Charikar, and A. Vijayaraghavan. Uniqueness of tensor decompositions with
applications to polynomial identifiability. CoRR, abs/1304.8087, 2013.
[6] A. Bhaskara, M. Charikar, A. Moitra, and A. Vijayaraghavan. Smoothed analysis of tensor
decompositions. In Proceedings of the 46th Annual ACM Symposium on Theory of Computing,
pages 594?603. ACM, 2014.
[7] D. Blackwell and L. Koopmans. On the identifiability problem for functions of finite Markov
chains. Annals of Mathematical Statistics, 28:1011?1015, 1957.
[8] W. Blischke. Estimating the parameters of mixtures of binomial distributions. Journal of the
American Statistical Association, 59(306):510?528, 1964.
[9] J. T. Chang. Full reconstruction of markov models on evolutionary trees: identifiability and
consistency. Mathematical biosciences, 137(1):51?73, 1996.
[10] A. Flaxman, A. W. Harrow, and G. B. Sorkin. Strings with maximally many distinct subsequences and substrings. Electron. J. Combin, 11(1):R8, 2004.
[11] J. Friedman. A proof of Alon?s second eigenvalue conjecture. In Proceedings of the thirty-fifth
Annual ACM Symposium on Theory of Computing, pages 720?724. ACM, 2003.
[12] Z. Ghahramani and M. Jordan. Factorial hidden markov models. Machine Learning, 1:31, 1997.
[13] D. Hsu, S. M. Kakade, and T. Zhang. A spectral algorithm for learning hidden markov models.
Journal of Computer and System Sciences, 78(5):1460?1480, 2012.
[14] Q. Huang, R. Ge, S. Kakade, and M. Dahleh. Minimal realization problems for hidden markov
models. IEEE Transactions on Signal Processing, 64(7):1896?1904, 2016.
[15] H. Ito, S.-I. Amari, and K. Kobayashi. Identifiability of hidden markov information sources and
their minimum degrees of freedom. IEEE transactions on information theory, 38(2):324?333,
1992.
[16] S. Kakade, P. Liang, V. Sharan, and G. Valiant. Prediction with a short memory. arXiv preprint
arXiv:1612.02526, 2016.
[17] M. Krivelevich, B. Sudakov, V. H. Vu, and N. C. Wormald. Random regular graphs of high
degree. Random Structures & Algorithms, 18(4):346?363, 2001.
[18] J. B. Kruskal. Three-way arrays: rank and uniqueness of trilinear decompositions, with
application to arithmetic complexity and statistics. Linear Algebra and its Applications, 18(2),
1977.
[19] S. Leurgans, R. Ross, and R. Abel. A decomposition for three-way arrays. SIAM Journal on
Matrix Analysis and Applications, 14(4):1064?1083, 1993.
[20] E. Mossel and S. Roch. Learning nonsingular phylogenies and hidden markov models. In
Proceedings of the thirty-seventh Annual ACM Symposium on Theory of Computing, pages
366?375. ACM, 2005.
[21] R. A. Redner and H. F. Walker. Mixture densities, maximum likelihood and the em algorithm.
SIAM review, 26(2):195?239, 1984.
[22] E. Shamir and E. Upfal. Large regular factors in random graphs. North-Holland Mathematics
Studies, 87:271?282, 1984.
[23] R. Weiss and B. Nadler. Learning parametric-output hmms with two aliased states. In ICML,
pages 635?644, 2015.
10
| 6695 |@word mild:2 trial:1 koopmans:1 polynomial:22 seems:4 norm:3 stronger:1 open:5 d2:2 simulation:2 crucially:1 decomposition:15 p0:2 pick:1 recursively:1 carry:2 reduction:1 moment:27 cyclic:2 contains:1 initial:6 necessity:1 denoting:1 document:2 ours:3 past:3 recovered:1 surprising:1 must:4 written:1 additive:1 distant:1 j1:2 enables:1 hypothesize:1 aside:1 stationary:6 leaf:1 accordingly:1 ith:2 short:24 provides:1 node:1 complication:1 zhang:1 mathematical:2 c2:4 become:4 symposium:3 prove:4 manner:1 introduce:3 theoretically:1 hardness:1 kp0:1 window:36 becomes:5 begin:1 provided:1 notation:4 underlying:2 moreover:1 estimating:1 mass:16 aliased:2 what:2 string:5 sudakov:1 sharpening:1 guarantee:5 remember:1 every:6 subclass:1 ti:1 runtime:1 rm:5 unit:1 converse:1 positive:5 before:3 understood:1 local:1 t1:6 kobayashi:1 limit:1 despite:1 analyzing:2 approximately:2 might:10 wormald:1 studied:3 specifying:1 innocuous:1 relaxing:1 hmms:75 factorization:1 bi:1 averaged:1 obeys:1 directed:2 practical:1 unique:8 thirty:2 yj:1 vu:1 union:4 practice:1 dahleh:1 kpt:1 empirical:1 significantly:1 regular:7 get:2 cannot:7 convenience:1 put:2 context:1 impossible:1 equivalent:2 map:1 deterministic:1 yt:3 go:1 starting:6 independently:3 ergodic:1 recovery:4 m2:2 array:2 stability:2 handle:2 notion:4 proving:1 annals:2 pt:2 shamir:1 strengthen:1 exact:3 particularly:1 labeled:3 kxk1:2 observed:1 preprint:1 capture:4 worst:2 verifying:1 ensures:2 cycle:39 connected:1 mentioned:2 intuition:2 complexity:8 abel:1 ti0:1 ideally:1 depend:2 tight:1 algebra:1 easily:1 alphabet:8 distinct:1 fast:1 h0:5 whose:5 quite:1 stanford:6 y1t:1 widely:1 say:1 larger:1 otherwise:1 amari:1 statistic:3 syntactic:1 think:1 itself:4 seemingly:1 sequence:9 eigenvalue:2 reconstruction:1 product:9 j2:2 relevant:1 realization:1 mixing:5 poorly:1 dlogm:6 convergence:1 optimum:1 requirement:5 telgarsky:1 help:1 polylog:1 alon:1 coupling:5 pose:1 stating:1 illustrate:1 derive:1 qt:3 expectationmaximization:1 progress:1 eq:1 strong:1 c:2 implies:1 direction:1 stochastic:1 subsequently:1 stringent:1 settle:1 adjacency:1 explains:1 require:3 preliminary:2 decompose:1 proposition:7 yij:1 hold:3 sufficiently:1 nadler:1 mapping:3 claim:2 electron:1 kruskal:3 uniqueness:3 rhodes:1 label:3 ross:1 hope:1 gaussian:1 always:3 super:1 aim:1 corollary:1 encode:1 l0:1 focus:2 improvement:2 rank:10 likelihood:7 contrast:1 sharan:2 summarizing:1 dependent:4 typically:1 lj:1 hidden:57 i1:2 jennrich:1 provably:1 interested:2 ill:3 flexible:1 colt:2 equal:2 construct:1 never:1 having:7 washington:2 reversibility:1 runtimes:2 identical:1 icml:1 future:5 t2:6 fundamentally:1 hint:1 few:1 randomly:1 composed:4 preserve:3 logm:15 n1:2 attempt:1 ab:1 friedman:1 freedom:1 reinterpreted:1 certainly:1 deferred:1 introduces:1 mixture:4 pc:2 chain:13 kt:2 emit:2 closer:1 necessary:11 janzamin:1 shorter:2 indexed:1 tree:1 walk:14 overcomplete:13 combin:1 theoretical:1 minimal:1 instance:1 column:6 earlier:2 soft:2 rao:7 contiguous:1 subset:1 entry:1 seventh:1 too:1 learnability:11 dependency:9 gregory:1 gd:3 thoroughly:1 density:1 fundamental:1 siam:2 standing:1 sketching:1 quickly:1 nm:1 moitra:1 huang:10 possibly:2 slowly:1 worse:4 american:1 li:1 exclude:1 north:1 satisfy:5 notable:1 explicitly:1 depends:4 performed:1 try:1 root:1 analyze:3 observing:1 portion:2 recover:3 identifiability:19 contribution:3 oi:2 who:2 efficiently:2 correspond:1 nonsingular:1 trilinear:1 weak:1 mc:1 substring:1 finer:1 randomness:1 simultaneous:1 khatri:7 whenever:1 definition:1 matias:1 mysterious:1 obvious:1 naturally:1 proof:6 associated:1 bioscience:1 hsu:4 gain:1 popular:1 ask:1 recall:3 redner:1 back:1 maximally:1 daunting:1 wei:1 though:5 generality:1 just:2 xk1:2 until:1 hand:2 sketch:2 perhaps:3 believe:1 effect:1 concept:1 true:3 contain:1 y2:1 hence:18 assigned:5 q0:3 semantic:1 i2:2 ll:1 uniquely:1 illustrative:1 bijective:1 outline:2 theoretic:4 percy:1 l1:4 variational:1 consideration:1 novel:1 mt:1 overview:1 conditioning:3 volume:1 discussed:2 extend:1 m1:4 belong:1 association:1 significant:3 refer:8 measurement:1 ai:2 leurgans:1 outlined:2 consistency:1 mathematics:1 pointed:1 language:2 access:2 longer:2 add:1 disentangle:1 closest:1 recent:1 showed:5 conjectured:2 scenario:1 certain:2 binary:3 yi:1 minimum:3 additional:1 somewhat:1 shortest:1 signal:1 arithmetic:1 ii:2 relates:1 full:9 multiple:2 sham:2 stem:1 desirable:1 mix:3 technical:2 infer:1 long:17 visit:7 prediction:4 arxiv:3 represent:1 c1:4 preserved:1 whereas:1 addition:2 separately:1 receive:1 interval:5 singular:4 source:1 walker:1 ot:1 probably:1 undirected:2 vijayaraghavan:2 seem:2 jordan:1 anandkumar:7 call:1 allman:3 structural:1 iii:1 easy:2 enough:1 affect:3 fit:2 sorkin:1 l1t:1 idea:1 shift:2 whether:1 motivated:1 handled:1 utility:1 suffer:1 speech:1 cause:1 krivelevich:1 generally:2 useful:1 clear:2 factorial:5 amount:1 revisit:1 estimated:1 disjoint:3 write:2 key:1 drawn:3 ht:1 graph:17 inverse:1 throughout:2 reader:2 saying:2 reasonable:1 draw:1 appendix:6 comparable:1 bound:7 guaranteed:1 quadratic:1 refine:1 identifiable:16 annual:3 constraint:3 prohibiting:1 encodes:1 argument:5 min:22 conjecture:2 structured:6 charikar:2 according:2 combination:3 disconnected:1 belonging:1 smaller:3 slightly:1 em:2 character:2 y0:1 kakade:7 evolves:1 making:1 intuitively:1 invariant:1 computationally:1 previously:1 discus:9 know:1 ge:3 tractable:3 permit:1 apply:1 obey:1 generic:10 spectral:3 appropriate:1 denotes:2 binomial:1 ensure:2 log2:4 restrictive:1 epsilon:2 k1:2 ghahramani:1 bl:1 tensor:26 question:6 added:1 strategy:2 parametric:1 dependence:4 pave:1 evolutionary:1 reversed:3 distance:3 unable:1 hmm:25 outer:2 reason:2 provable:1 fresh:2 length:37 index:5 minn:2 ratio:1 providing:1 liang:2 setup:4 difficult:1 nc:3 equivalently:1 sharper:1 relate:1 negative:4 stated:2 motivates:1 cl0:1 pliang:1 perform:2 upper:1 allowing:1 observation:28 markov:21 finite:1 defining:1 excluding:1 y1:2 rn:1 smoothed:5 required:2 blackwell:1 c3:4 learned:10 roch:2 adversary:2 regime:4 sparsity:3 memory:2 video:1 power:1 natural:8 difficulty:3 recursion:1 residual:1 mn:2 improve:1 mossel:2 imply:1 ne:6 picture:1 coupled:1 flaxman:1 lij:1 review:1 understanding:9 evolve:1 loss:1 expect:1 permutation:11 interesting:4 limitation:1 upfal:1 degree:26 sufficient:3 supported:1 parity:5 weaker:1 understand:2 neighbor:1 fifth:1 sparse:9 regard:1 boundary:2 dimension:2 transition:81 cumulative:2 rich:1 author:1 commonly:1 made:1 concretely:1 far:2 polynomially:3 transaction:2 keep:1 investigating:1 conclude:1 assumed:1 discriminative:1 subsequence:1 continuous:3 latent:6 why:1 reality:1 nature:1 learn:6 obtaining:1 improving:1 poly:3 complex:1 dense:4 main:1 noise:5 n2:1 allowed:2 fig:15 depicts:1 slow:1 deterministically:1 explicit:1 exponential:2 lie:1 third:1 ito:1 bhaskara:7 theorem:19 down:2 showing:3 symbol:1 learnable:6 hinting:1 r8:1 evidence:1 intractable:2 exists:2 sequential:1 valiant:3 importance:1 flattened:1 corr:1 diagonalization:1 conditioned:17 gap:1 easier:1 lt:1 distinguishable:1 simply:1 girth:3 chang:1 holland:1 satisfies:2 discourse:1 determines:2 acm:6 sized:1 goal:1 identity:1 hard:4 specifically:2 infinite:1 uniformly:2 except:6 total:2 exception:1 phylogeny:1 support:5 d1:2 |
6,296 | 6,696 | GP CaKe: Effective brain connectivity with causal
kernels
Luca Ambrogioni
Radboud University
[email protected]
Max Hinne
Radboud University
[email protected]
Marcel A. J. van Gerven
Radboud University
[email protected]
Eric Maris
Radboud University
[email protected]
Abstract
A fundamental goal in network neuroscience is to understand how activity in one
brain region drives activity elsewhere, a process referred to as effective connectivity.
Here we propose to model this causal interaction using integro-differential equations and causal kernels that allow for a rich analysis of effective connectivity. The
approach combines the tractability and flexibility of autoregressive modeling with
the biophysical interpretability of dynamic causal modeling. The causal kernels are
learned nonparametrically using Gaussian process regression, yielding an efficient
framework for causal inference. We construct a novel class of causal covariance
functions that enforce the desired properties of the causal kernels, an approach
which we call GP CaKe. By construction, the model and its hyperparameters have
biophysical meaning and are therefore easily interpretable. We demonstrate the
efficacy of GP CaKe on a number of simulations and give an example of a realistic
application on magnetoencephalography (MEG) data.
1
Introduction
In recent years, substantial effort was dedicated to the study of the network properties of neural
systems, ranging from individual neurons to macroscopic brain areas. It has become commonplace to
describe the brain as a network that may be further understood by considering either its anatomical
(static) scaffolding, the functional dynamics that reside on top of that or the causal influence that
the network nodes exert on one another [1?3]. The latter is known as effective connectivity and has
inspired a surge of data analysis methods that can be used to estimate the information flow between
neural sources from their electrical or haemodynamic activity[2, 4]. In electrophysiology, the most
popular connectivity methods are variations on the autoregressive (AR) framework [5]. Specifically,
Granger causality (GC) and related methods, such as partial directed coherence and directed transfer
function, have been successfully applied to many kinds of neuroscientific data [6, 7]. These methods
can be either parametric or non-parametric, but are not based on a specific biophysical model [8, 9].
Consequently, the connectivity estimates obtained from these methods are only statistical in nature
and cannot be directly interpreted in terms of biophysical interactions [10]. This contrasts with the
framework of dynamic causal modeling (DCM), which allows for Bayesian inference (using Bayes
factors) with respect to biophysical models of interacting neuronal populations [11]. These models
are usually formulated in terms of either deterministic or stochastic differential equations, in which
the effective connectivity between neuronal populations depends on a series of scalar parameters
that specify the strength of the interactions and the conduction delays [12]. DCMs are usually
less flexible than AR models since they depend on an appropriate parametrization of the effective
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
connectivity kernel, which in turn depends on detailed prior biophysical knowledge or Bayesian
model comparison.
In this paper, we introduce a new method that is aimed to bridge the gap between biophysically
inspired models, such as DCM, and statistical models, such as AR, using the powerful tools of
Bayesian nonparametrics [13]. We model the interacting neuronal populations with a system of
stochastic integro-differential equations. In particular, the intrinsic dynamic of each population is
modeled using a linear differential operator while the effective connectivity between populations
is modeled using causal integral operators. The differential operators can account for a wide range
of dynamic behaviors, such as stochastic relaxation and stochastic oscillations. While this class of
models cannot account for non-linearities, it has the advantage of being analytically tractable. Using
the framework of Gaussian process (GP) regression, we can obtain the posterior distribution of the
effective connectivity kernel without specifying a predetermined parametric form. We call this new
effective connectivity method Gaussian process Causal Kernels (GP CaKe). The GP CaKe method
can be seen as a nonparametric extension of linear DCM for which the exact posterior distribution can
be obtained in closed-form without resorting to variational approximations. In this way, the method
combines the flexibility and statistical simplicity of AR modeling with the biophysical interpretability
of a linear DCM.
The paper is structured as follows. In Section 2 we describe the model for the activity of neuronal
populations and their driving interactions. In Section 3 we construct a Bayesian hierarchical model
that allows us to learn the causal interaction functions. Next, in Subsection 3.2, we show that these
causal kernels may be learned analytically using Gaussian process regression. Subsequently in
Section 4, we validate GP CaKe using a number of simulations and demonstrate its usefulness on
MEG data in Section 5. Finally, we discuss the wide array of possible extensions and applications of
the model in Section 6.
2
Neuronal dynamics
We model the activity of a neuronal population xj (t) using the stochastic differential equation
Dj xj (t) = Ij (t) + wj (t) ,
(1)
where Ij (t) is the total synaptic input coming from other neuronal populations and wj (t) is Gaussian
PP
dp
white noise with mean 0 and variance ? 2 . The differential operator Dj = ?0 + p=1 ?p dt
p specifies
the internal dynamic of the neuronal population. For example, oscillatory dynamic can be modeled
d
d2
2
using the damped harmonic operator DjH = dt
, where ?0 is the (undamped) peak
2 + ? dt + ?0
angular frequency and ? is the damping coefficient.
In Eq. 1, the term Ij (t) accounts for the effective connectivity between neuronal populations. Assuming that the interactions are linear and stationary over time, the most general form for Ij (t) is given
by a sum of convolutions:
N
X
Ij (t) =
ci?j ? xi (t) ,
(2)
i=1
where the function ci?j (t) is the causal kernel, modeling the effective connectivity from population i
to population j, and ? indicates the convolution operator. The causal kernel ci?j (t) gives a complete
characterization of the linear effective connectivity between the two neuronal populations, accounting
for the excitatory or inhibitory nature of the connection, the time delay, and the strength of the
interaction. Importantly, in order to preserve the causality of the system, we assume that ci?j (t) is
identically equal to zero for negative lags (t < 0).
Inserting Eq. 2 into Eq. 1, we obtain the following system of stochastic integro-differential equations:
Dj xj (t) =
N
X
ci?j ? xi (t) + wj (t),
j = 1...N ,
(3)
i=1
which fully characterizes the stochastic dynamic of a functional network consisting of N neuronal
populations.
2
3
The Bayesian model
We can frame the estimation of the effective connectivity between neuronal populations as a nonparametric Bayesian regression problem. In order to do this, we assign a GP prior distribution to the kernel
functions ci?j (t) for every presynaptic population i and postsynaptic population j. A stochastic function f (t) is said to follow a GP distribution when all its marginal distributions p(f (t1 ), . . . , f (tn )) are
distributed as a multivariate Gaussian [14]. Since these marginals are determined by their mean vector
and covariance matrix, the GP is fully specified by a mean and a covariance function, respectively
mf (t) = hf (t)i and Kf (t1 , t2 ) = h(f (t1 ) ? mf (t1 ))(f (t2 ) ? mf (t2 ))i. Using the results of the
previous subsection we can summarize the problem of Bayesian nonparametric effective connectivity
estimation in the following way:
ci?j (t) ? GP (0, K(t1 , t2 ))
wj (t) ? N (0, ? 2 )
Dj xj (t) =
N
X
(4)
(ci?j ? xi ) (t) + wj (t) ,
i=1
where expressions such as f (t) ? GP m(t), K(t1 , t2 ) mean that the stochastic process f (t) follows
a GP distribution with mean function m(t) and covariance function K(t1 , t2 ).
Our aim is to obtain the posterior distributions of the effective connectivity kernels given a set of
samples from all the neuronal processes. As a consequence of the time shift invariance, the system
of integro-differential equations becomes a system of decoupled linear algebraic equations in the
frequency domain. It is therefore convenient to rewrite the regression problem in the frequency
domain:
ci?j (?) ? CGP 0, K(?1 , ?2 )
wj (?) ? CN (0, ? 2 )
Pj (?)xj (?) =
N
X
(5)
xi (?)ci?j (?) + wj (?) ,
i=1
PP
where Pj (?) = p=0 ?p (?i?)p is a complex-valued polynomial since the application of a differential operator in the time domain is equivalent to multiplication with a polynomial in the frequency
domain. In the previous expression, CN (?, ?) denotes a circularly-symmetric complex normal
distribution with mean ? and variance ?, while CGP (m(t), K(?)) denotes a circularly-symmetric
complex valued GP with mean function m(?) and Hermitian covariance function K(?1 , ?2 ) [15].
Importantly, the complex valued Hermitian covariance function K(?1 , ?2 ) can be obtained from
K(t1 , t2 ) by taking the Fourier transform of both its arguments:
Z +? Z +?
(6)
K(?1 , ?2 ) =
e?i?1 t1 ?i?2 t2 K(t1 , t2 )dt1 dt2 .
??
3.1
??
Causal covariance functions
In order to be applicable for causal inference, the prior covariance function K(t1 , t2 ) must reflect
three basic assumptions about the connectivity kernel: I) temporal localization, II) causality and
III) smoothness. Since we perform the GP analysis in the frequency domain, we will work with
K(?1 , ?2 ), i.e. the double Fourier transform of the covariance function.
First, the connectivity kernel should be localized in time, as the range of plausible delays in axonal
communication between neuronal populations is bounded. In order to enforce this constraint, we
need a covariance function K(t1 , t2 ) that vanishes when either t1 or t2 becomes much larger than a
time constant ?. In the frequency domain, this temporal localization can be implemented by inducing
correlations between the Fourier coefficients of neighboring frequencies. In fact, local correlations in
the time domain are associated with a Fourier transform that vanishes for high values of ?. From
Fourier duality, this implies that local correlations in the frequency domain are associated with a
function that vanishes for high values of t. We model these spectral correlations using a squared
exponential covariance function:
KSE (?1 , ?2 ) = e??
(?2 ??1 )2
2
+its (?2 ??1 )
3
= e??
?2
2
+its ?
,
(7)
where ? = ?2 ? ?1 . Since we expect the connectivity to be highest after a minimal conduction delay
ts , we introduced a time shift factor its ? in the exponent that translates the peak of the variance from
0 to ts , which follows from the Fourier shift theorem. As this covariance function depends solely on
the difference between frequencies ?, it can be written (with a slight abuse of notation) as KSE (?).
Second, we want the connectivity kernel to be causal, meaning that information cannot propagate
back from the future. In order to enforce causality, we introduce a new family of covariance functions
that vanish when the lag t2 ? t1 is negative. In the frequency domain, a causal covariance function
can be obtained by adding an imaginary part to Eq. 7 that is equal to its Hilbert transform H [16].
Causal covariance functions are the Fourier dual of quadrature covariance functions, which define GP
distributions over the space of analytic functions, i.e. functions whose Fourier coefficients are zero
for all negative frequencies [15]. The causal covariance function is given by the following formula:
KC (?) = KSE (?) + iHKSE (?) .
(8)
Finally, as communication between neuronal populations is mediated by smooth biological processes
such as synaptic release of neurotransmitters and dendritic propagation of potentials, we want the
connectivity kernel to be a smooth function of the time lag. Smoothness in the time domain can be
imposed by discounting high frequencies. Here, we use the following discounting function:
f (?1 , ?2 ) = e??
2 +? 2
?1
2
2
.
(9)
This discounting function induces a process that is smooth (infinitely differentiable) and with time
scale equal to ? [14]. Our final covariance function is given by
K(?1 , ?2 ) = f (?1 , ?2 ) (KSE (?) + iHKSE (?)) .
(10)
Unfortunately, the temporal smoothing breaks the strict causality of the covariance function because
it introduces leakage from the positive lags to the negative lags. Nevertheless, the covariance function
closely approximates a causal covariance function when ? is not much bigger than ts .
3.2
Gaussian process regression
In order to explain how to obtain the posterior distribution of the causal kernel, we need to review some
basic results of nonparametric Bayesian regression and GP regression in particular. Nonparametric
Bayesian statistics deals with inference problems where the prior distribution has infinitely many
degrees of freedom [13]. We focus on the following nonparametric regression problem, where the aim
is to reconstruct a series of real-valued functions from a finite number of noisy mixed observations:
X
yt =
?i (t)fi (t) + wt ,
(11)
i
where yt is the t-th entry of the data vector y, fi (t) is an unknown latent function and wt is a
random variable that models the observation noise with diagonal covariance matrix D. The mixing
functions ?i (t) are assumed to be known and determine how the latent functions generate the data.
In nonparametric Bayesian regression, we specify prior probability distributions over the whole
(infinitely dimensional) space of functions fi (t). Specifically, in the GP regression framework this
distribution is chosen to be a zero-mean GP. In order to infer the value of the function f (t) at an
?
arbitrary set of target points T ? = {t?
1 , ..., tm }, we organize these values in the vector f with entries
?
fl = f (tl ). The posterior expected value of f , that we will denote as mfj |y , is given by
X
?1
mfj |y = Kf?j ?j
?i Kfi ?i + D
y,
(12)
i
where the covariance matrix Kf is defined by the entries [Kf ]uv = Kf (tu , tv ) and the crosscovariance matrix K?? is defined by the entries [Kf? ]uv = Kf (t?
u , tv ) [14]. The matrices ?i are
square and diagonal, with the entries [?i ]uu given by ?i (tu ).
It is easy to see that the problem defined by Eq. 5 has the exact same form as the generalized
regression problem given by Eq. 11, with ? as dependent variable. In particular, the weight functions
wj (?)
?2
?i (?) are given by Pxij(?)
(?) and the noise term Pj (?) has variance |Pj (?)|2 . Therefore, the expectation
of the posterior distributions p(ci?j (?)|{x1 (?h )}, . . . , {xN (?h )}) can be obtained in closed from
from Eq. 12.
4
4
Effective connectivity simulation study
We performed a simulation study to assess the performance of the GP CaKe approach in recovering
the connectivity kernel from a network of simulated sources. The neuronal time series xj (t) are
generated by discretizing a system of integro-differential equations, as expressed in Eq. 3. Time series
data was then generated for each of the sources using the Ornstein-Uhlenbeck process dynamic, i.e.
d
D(1) =
+?,
(13)
dt
where the positive parameter ? is the relaxation coefficient of the process. The bigger ? is, the faster
the process reverts to its mean (i.e. zero) after a perturbation. The discretization of this dynamic is
equivalent to a first order autoregressive process. As ground truth effective connectivity, we used
functions of the form
?
(14)
ci?j (? ) = ai?j ? e? s ,
where ? is a (non-negative) time lag, ai?j is the connectivity strength from i to j and s is the
connectivity time scale.
In order to recover the connectivity kernels ci?j (t) we first need to estimate the differential operator
D(1) . For simplicity, we estimated the parameters of the differential operator by maximizing the
univariate marginal likelihood of each individual source. This procedure requires that the variance
of the structured input from the other neuronal populations is smaller than the variance of the
unstructured white noise input so that the estimation of the intrinsic dynamic is not too much affected
by the coupling.
Since most commonly used effective connectivity measures (e.g. Granger causality, partial directed
coherence, directed transfer function) are obtained from fitted vector autoregression (VAR) coefficients, we use VAR as a comparison method. Since the least-squares solution for the VAR coefficients
is not regularized, we also compare with a ridge regularized VAR model, whose penalty term is
learned using cross-validation on separately generated training data. This comparison is particularly
natural since our connectivity kernel is the continuous-time equivalent of the lagged AR coefficients
between two time series.
4.1
Recovery of the effective connectivity kernels
We explore the effects of different parameter values to demonstrate the intuitiveness of the kernel
parameters. Whenever a parameter is not specifically adjusted, we use the following default values:
noise level ? = 0.05, temporal smoothing ? = 0.15 and temporal localization ? = ?. Furthermore,
we set ts = 0.05 throughout.
Figure 1 illustrates connectivity kernels recovered by GP CaKe. These kernels have a connection
strength of ai?j = 5.0 if i feeds into j and ai?j = 0 otherwise. This applies to both the two node
and the three node network. As these kernels show, our method recovers the desired shape as well
as the magnitude of the effective connectivity for both connected and disconnected edges. At the
same time, Fig. 1B demonstrates that the indirect pathway through two connections does not lead to a
non-zero estimated kernel. Note furthermore that the kernels become non-zero after the zero-lag mark
(indicated by the dashed lines), demonstrating that there is no significant anti-causal information
leakage.
The effects of the different kernel parameter settings are shown in Fig. 2A, where again the method is
estimating connectivity for a two node network with one active connection, with ai?j = 5.0. We
show the mean squared error (MSE) as well as the correlation between the ground truth effective
connectivity and the estimates obtained using our method. We do this for different values of the
temporal smoothing, the noise level and the temporal localization parameters. Figure 2B shows
the estimated kernels that correspond to these settings. As to be expected, underestimating the
temporal smoothness results in increased variance due to the lack of regularization. On the other hand,
overestimating the smoothness results in a highly biased estimate as well as anti-causal information
leakage. Overestimating the noise level does not induce anti-causal information leakage but leads to
substantial bias. Finally, overestimating the temporal localization leads to an underestimation of the
duration of the causal influence.
Figure 3 shows a quantitative comparison between GP CaKe and the (regularized and unregularized)
VAR model for the networks shown in Fig. 1A and Fig. 1B. The connection strength ai?j was
5
A
B
1
2
1
2
Ground truth
GP CaKe
3
0 lag
Figure 1: Example of estimated connectivity. A. The estimated connectivity kernels for two connections: one present (2 ? 1) and one absent (1 ? 2). B. A three-node network in which node 1 feeds
into node 2 and node 2 feeds into node 3. The disconnected edge from 1 to 3 is correctly estimated,
as the estimated kernel is approximately zero. For visual clarity, estimated connectivity kernels for
other absent connections (2 ? 1, 3 ? 2 and 3 ?1) are omitted in the second panel. The shaded
areas indicate the 95% posterior density interval over 200 trials.
varied to study its effect on the kernel estimation. It is clear that GP CaKe greatly outperforms both
VAR models and that ridge regularization is beneficial for the VAR approach. Note that, when the
connection strength is low, the MSE is actually smallest for the fully disconnected model. Conversely,
both GP CaKe and VAR always outperform the disconnected estimate with respect to the correlation
measure.
5
Brain connectivity
In this section we investigate the effective connectivity structure of a network of cortical sources. In
particular, we focus on sources characterized by alpha oscillations (8?12Hz), the dominant rhythm in
MEG recordings. The participant was asked to watch one-minute long video clips selected from an
American television series. During these blocks the participant was instructed to fixate on a cross
in the center of the screen. At the onset of each block a visually presented message instructed the
participant to pay attention to either the auditory or the visual stream. The experiment also included a
so-called ?resting state? condition in which the participant was instructed to fixate on a cross in the
center of a black screen. Brain activity was recorded using a 275 channels axial MEG system.
The GP CaKe method can be applied to a set of signals whose intrinsic dynamic can be characterized
by stochastic differential equations. Raw MEG measurements can be seen as a mixture of dynamical
signals, each characterized by a different intrinsic dynamic. Therefore, in order to apply the method
on MEG data, we need to isolate a set of dynamic components. We extracted a series of unmixed
neural sources by applying independent component analysis (ICA) on the sensor recordings. These
components were chosen to have a clear dipolar pattern, the signature of a localized cortical source.
These local sources have a dynamic that can be well approximated with a linear mixture of linear
stochastic differential equations [17]. We used the recently introduced temporal GP decomposition
in order to decompose the components? time series into a series of dynamic components [17]. In
particular, for each ICA source we independently extracted the alpha oscillation component, which
d2
d
2
we modeled with a damped harmonic oscillator: DjH = dt
2 + ? dt + ?0 . Note that the temporal GP
decomposition automatically estimates the parameters ? and ?0 through a non-linear least-squares
procedure [17].
We computed the effective connectivity between the sources that corresponded to occipital, parietal
and left- and right auditory cortices (see Fig. 4A) using GP CaKe with the following parameter
settings: temporal smoothing ? = 0.01, temporal shift ts = 0.004, temporal localization ? = 8?
and noise level ? = 0.05. To estimate the causal structure of the network, we performed a z-test
on the maximum values of the kernels for each of the three conditions. The results were corrected
6
Temporal smoothing
1.0
? = 0.10
? = 1.00
? = 10.00
1.0
3
2
0.2
0 lag
0.4
MSE
Correlation
5
4
0.0
1
10
1.0
Noise level
-1
?
0
10
1
10
0.0
0.8
Correlation
MSE
0.8
0
0.6
0.4
0.4
0.0
-2
10
-1
10
?
0
10
1
10
Temporal localization
1.0
0.0
1.0
2.0
0.0
1.0
2.0
0.0
1.0
? = 0.01
? = 0.10
? = 1.00
? = 10.00
?=?
? = 2?
? = 3?
? = 4?
2.0
1.0
0.2
0.2
2.0
MSE
0.6
1.0
GP CaKe
0.0
-2
10
0.0
0
0.30
0.8
Correlation
1.0
0.20
MSE
0.6
0.4
0.0
0.10
0.2
0.0
? = 0.01
6
0.6
Correlation
B
7
0.8
Ground truth
A
?
2?
?
3?
0.0
4?
1.0
2.0
Time lag (s)
0.0
1.0
Time lag (s)
2.0
0.0
1.0
Time lag (s)
2.0
0.0
1.0
Time lag (s)
2.0
Figure 2: The effect of the the temporal localization, smoothness and noise level parameters on a
present connection. A. The correlation and mean squared error between the ground truth connectivity
kernel and the estimation by GP CaKe. B. The shapes of the estimated kernels as determined by the
indicated parameter. Default values for the parameters that remain fixed are ? = 0.05, ? = 0.15 and
? = ?. The dashed line indicates the zero-lag moment at which point the causal effect deviates from
zero. The shaded areas indicate the 95% posterior density interval over 200 trials.
B
Two-node network
0.2
0.0
?0.2
1.0
5.0
10.0
Connection weight
100
10-1
10-2
108
0.2
101
Correlation
Mean squared error
Correlation
0.4
Three-node network
0.3
102
Mean squared error
A
0.1
0.0
?0.1
1.0
5.0
?0.2
10.0
Connection weight
GP CaKe
1.0
2.5
5.0
Connection weight
VAR, Ridge
VAR
106
104
102
100
10-2
1.0
2.5
5.0
Connection weight
Baseline
Figure 3: The performance of the recovery of the effective connectivity kernels in terms of the
correlation and mean squared error between the actual and the recovered kernel. Left column: results
for the two node graph shown in Fig. 1A. Right column: results for the three node graph shown in
Fig. 1B. The dashed line indicates the baseline that estimates all node pairs as disconnected.
for multiple comparisons using FDR correction with ? = 0.05. The resulting structure is shown
in Fig. 4A, with the corresponding causal kernels in Fig. 4B. The three conditions are clearly
distinguishable from their estimated connectivity structure. For example, during the auditory attention
condition, alpha band causal influence from parietal to occipital cortex is suppressed relative to the
other conditions. Furthermore, a number of connections (i.e. right to left auditory cortex, as well as
both auditory cortices to occipital cortex) are only present during the resting state.
7
B
R, V, A
0
?1
?2
?3
R.A.
audio
2
1
0
?1
?2
?3
1e?18
2
1
0
?1
?2
?3
video
2
1
0
?1
?2
?3
1e?18
rest
Par. cortex
A
R, V, A
,A
,V
,A
,V
R
R
V
R
R
,A
,V
R
L.A.
V
L. auditory cortex R. auditory cortex
1
1e?18
R
Parietal cortex
1e?18
2
onset
Par.
R, V
R. aud. cortex L. aud. cortex
Occ.
Occipital cortex
Occ. cortex
A
0.00
0.04
0.08
0.12
Time lag (s)
0.00
0.04
0.08
0.12
Time lag (s)
0.00
0.04
0.08
0.12
Time lag (s)
Time lag (s)
Figure 4: Effective connectivity using MEG for three conditions: I. resting state (R), II. attention to
video stream (V) and III. attention to audio stream (A). Shown are the connections between occipital
cortex, parietal cortex and left and right auditory cortices. A. The binary network for each of the
three conditions. B. The kernels for each of the connections. Note that the magnitude of the kernels
depends on the noise level ?, and as the true strength is unknown, this is in arbitrary units.
6
Discussion
We introduced a new effective connectivity method based on GP regression and integro-differential
dynamical systems, referred to as GP CaKe. GP CaKe can be seen as a nonparametric extension of
DCM [11] where the posterior distribution over the effective connectivity kernel can be obtained in
closed form. In order to regularize the estimation, we introduced a new family of causal covariance
functions that encode three basic assumptions about the effective connectivity kernel: (1) temporal
localization, (2) causality, and (3) temporal smoothness. The resulting estimated kernels reflect the
time-modulated causal influence that one region exerts on another. Using simulations, we showed
that GP CaKe produces effective connectivity estimates that are orders of magnitude more accurate
than those obtained using (regularized) multivariate autoregression. Furthermore, using MEG data,
we showed that GP CaKe is able to uncover interesting patterns of effective connectivity between
different brain regions, modulated by cognitive state.
The strategy for selecting the hyperparameters of the GP CaKe model depends on the specific study.
If they are hand-chosen they should be set in a conservative manner. For example, the temporal
localization should be longer than the highest biologically meaningful conduction delay. Analogously,
the smoothing parameter should be smaller than the time scale of the system of interest. In ideal
cases, such as for the analysis of the subthreshold postsynaptic response of the cellular membrane,
these values can be reasonably obtained from biophysical models. When prior knowledge is not
available, several off-the-shelf Bayesian hyperparameter selection or marginalization techniques can
be applied to GP CaKe directly since both the marginal likelihood and its gradient are available in
closed-form. In this paper, instead of proposing a particular hyper-parameter selection technique, we
decided to focus our exposition on the interpretability of the hyperparameters. In fact, biophysical
interpretability can help neuroscientists construct informed hyperprior distributions.
Despite its high performance, the current version of the GP CaKe method has some limitations.
First, the method can only be used on signals whose intrinsic dynamics are well approximated
by linear stochastic differential equations. Real-world neural recordings are often a mixture of
several independent dynamic components. In this case the signal needs to be preprocessed using
a dynamic decomposition technique [17]. The second limitation is that the intrinsic dynamics are
currently estimated from the univariate signals. This procedure can lead to biases when the neuronal
populations are strongly coupled. Therefore, future developments should focus on the integration of
dynamic decomposition with connectivity estimation within an overarching Bayesian model.
The model can be extended in several directions. First, the causal structure of the neural dynamical
system can be constrained using structural information in a hierarchical Bayesian model. Here,
structural connectivity may be provided as an a priori constraint, for example derived from diffusionweighted MRI [18], or learned from the functional data simultaneously [19]. This allows the model
to automatically remove connections that do not reflect a causal interaction, thereby regularizing
8
the estimation. Alternatively, the anatomical constraints on causal interactions may be integrated
into a spatiotemporal model of the brain cortex by using partial integro-differential neural field
equations [20] and spatiotemporal causal kernels. In addition, the nonparametric modeling of the
causal kernel can be integrated into a more complex and biophysically realistic model where the
differential equations are not assumed to be linear [12] or where the observed time series data are
filtered through a haemodynamic [21] or calcium impulse response function [22].
Finally, while our model explicitly refers to neuronal populations, we note that the applicability of
the GP CaKe framework is in no way limited to neuroscience and may also be relevant for fields such
as econometrics and computational biology.
References
[1] A Fornito and E T Bullmore. Connectomics: A new paradigm for understanding brain disease.
European Neuropsychopharmacology, 25:733?748, 2015.
[2] K Friston. Functional and effective connectivity: A review. Brain Connectivity, 1(1):13?35,
2011.
[3] S L Bressler and V Menon. Large-scale brain networks in cognition: Emerging methods and
principles. Trends in Cognitive Sciences, 14(6):277?290, 2010.
[4] K E Stephan and A Roebroeck. A short history of causal modeling of fMRI data. NeuroImage,
62(2):856?863, 2012.
[5] K Friston, R Moran, and A K Seth. Analysing connectivity with Granger causality and dynamic
causal modelling. Current Opinion in Neurobiology, 23(2):172?178, 2013.
[6] K Sameshima and L A Baccal?. Using partial directed coherence to describe neuronal ensemble
interactions. Journal of Neuroscience Methods, 94(1):93?103, 1999.
[7] M Kami?nski, M Ding, W A Truccolo, and S. L. Bressler. Evaluating causal relations in neural
systems: Granger causality, directed transfer function and statistical assessment of significance.
Biological Cybernetics, 85(2):145?157, 2001.
[8] M Dhamala, G Rangarajan, and M Ding. Analyzing information flow in brain networks with
nonparametric Granger causality. NeuroImage, 41(2):354?362, 2008.
[9] S L Bressler and A K Seth. Wiener?Granger causality: A well established methodology.
NeuroImage, 58(2):323?329, 2011.
[10] B Schelter, J Timmer, and M Eichler. Assessing the strength of directed influences among
neural signals using renormalized partial directed coherence. Journal of Neuroscience Methods,
179(1):121?130, 2009.
[11] K Friston, B Li, J Daunizeau, and K E Stephan. Network discovery with DCM. NeuroImage,
56(3):1202?1221, 2011.
[12] O David, S J Kiebel, L M Harrison, J Mattout, J M Kilner, and K J Friston. Dynamic causal
modeling of evoked responses in EEG and MEG. NeuroImage, 30(4):1255?1272, 2006.
[13] N L Hjort, C Holmes, P M?ller, and S G Walker. Bayesian Nonparametrics. Cambridge
University Press, 2010.
[14] C E Rasmussen. Gaussian Processes for Machine Learning. The MIT Press, 2006.
[15] L Ambrogioni and E Maris. Complex?valued Gaussian process regression for time series
analysis. arXiv preprint arXiv:1611.10073, 2016.
[16] U C T?uber. Critical dynamics: a field theory approach to equilibrium and non-equilibrium
scaling behavior. Cambridge University Press, 2014.
[17] L Ambrogioni, M A J van Gerven, and E Maris. Dynamic decomposition of spatiotemporal
neural signals. arXiv preprint arXiv:1605.02609, 2016.
9
[18] M Hinne, L Ambrogioni, R J Janssen, T Heskes, and M A J van Gerven. Structurally-informed
Bayesian functional connectivity analysis. NeuroImage, 86:294?305, 2014.
[19] M Hinne, R J Janssen, T Heskes, and M A J van Gerven. Bayesian estimation of conditional
independence graphs improves functional connectivity estimates. PLoS Computational Biology,
11(11):e1004534, 2015.
[20] S Coombes, P beim Graben, R Potthast, and J Wright. Neural Fields. Springer, 2014.
[21] K J Friston, A Mechelli, R Turner, and C J Price. Nonlinear responses in fMRI: the Balloon
model, Volterra kernels, and other hemodynamics. NeuroImage, 12(4):466?477, 2000.
[22] C Koch. Biophysics of computation: Information processing in single neurons. Computational
Neuroscience Series. Oxford University Press, 2004.
10
| 6696 |@word trial:2 mri:1 version:1 polynomial:2 coombes:1 d2:2 simulation:5 propagate:1 covariance:24 accounting:1 decomposition:5 thereby:1 moment:1 series:12 efficacy:1 selecting:1 outperforms:1 imaginary:1 recovered:2 mari:4 discretization:1 current:2 scaffolding:1 must:1 written:1 connectomics:1 kiebel:1 realistic:2 predetermined:1 shape:2 analytic:1 remove:1 interpretable:1 stationary:1 selected:1 parametrization:1 short:1 underestimating:1 filtered:1 characterization:1 unmixed:1 node:14 differential:19 become:2 combine:2 pathway:1 hermitian:2 manner:1 introduce:2 ica:2 expected:2 behavior:2 surge:1 brain:12 inspired:2 automatically:2 actual:1 considering:1 becomes:2 provided:1 estimating:1 linearity:1 bounded:1 notation:1 panel:1 kind:1 interpreted:1 emerging:1 proposing:1 informed:2 temporal:20 quantitative:1 every:1 demonstrates:1 unit:1 organize:1 timmer:1 t1:14 positive:2 understood:1 local:3 consequence:1 despite:1 analyzing:1 oxford:1 solely:1 abuse:1 approximately:1 black:1 exert:1 evoked:1 specifying:1 shaded:2 conversely:1 limited:1 range:2 kfi:1 directed:8 decided:1 block:2 integro:7 procedure:3 area:3 convenient:1 haemodynamic:2 induce:1 refers:1 cannot:3 selection:2 operator:9 influence:5 applying:1 equivalent:3 deterministic:1 imposed:1 yt:2 maximizing:1 center:2 attention:4 occipital:5 duration:1 independently:1 overarching:1 simplicity:2 unstructured:1 recovery:2 holmes:1 array:1 importantly:2 regularize:1 population:22 variation:1 construction:1 target:1 exact:2 trend:1 intuitiveness:1 particularly:1 approximated:2 econometrics:1 observed:1 preprint:2 ding:2 electrical:1 commonplace:1 region:3 wj:8 connected:1 plo:1 balloon:1 highest:2 substantial:2 disease:1 vanishes:3 asked:1 dt2:1 dynamic:26 renormalized:1 signature:1 depend:1 rewrite:1 localization:10 eric:1 easily:1 seth:2 indirect:1 neurotransmitter:1 effective:31 describe:3 radboud:4 corresponded:1 hyper:1 whose:4 lag:18 larger:1 valued:5 plausible:1 reconstruct:1 otherwise:1 bullmore:1 statistic:1 gp:41 transform:4 noisy:1 final:1 advantage:1 differentiable:1 biophysical:9 propose:1 interaction:10 coming:1 inserting:1 neighboring:1 tu:2 relevant:1 mixing:1 flexibility:2 inducing:1 validate:1 double:1 rangarajan:1 assessing:1 produce:1 help:1 coupling:1 kilner:1 axial:1 ij:5 eq:8 implemented:1 recovering:1 marcel:1 implies:1 uu:1 indicate:2 aud:2 direction:1 closely:1 stochastic:12 subsequently:1 opinion:1 truccolo:1 assign:1 decompose:1 biological:2 dendritic:1 adjusted:1 extension:3 correction:1 koch:1 ground:5 normal:1 visually:1 wright:1 equilibrium:2 cognition:1 driving:1 smallest:1 omitted:1 estimation:9 applicable:1 currently:1 bridge:1 successfully:1 tool:1 mit:1 clearly:1 sensor:1 gaussian:9 always:1 aim:2 shelf:1 encode:1 release:1 focus:4 derived:1 modelling:1 indicates:3 likelihood:2 greatly:1 contrast:1 baseline:2 inference:4 dependent:1 integrated:2 kc:1 relation:1 dual:1 flexible:1 among:1 exponent:1 priori:1 development:1 smoothing:6 integration:1 constrained:1 marginal:3 equal:3 construct:3 field:4 beach:1 biology:2 future:2 fmri:2 t2:13 overestimating:3 preserve:1 simultaneously:1 crosscovariance:1 individual:2 consisting:1 freedom:1 neuroscientist:1 interest:1 message:1 highly:1 investigate:1 introduces:1 mixture:3 nl:4 yielding:1 damped:2 accurate:1 integral:1 edge:2 partial:5 decoupled:1 damping:1 hyperprior:1 desired:2 causal:43 minimal:1 fitted:1 increased:1 column:2 modeling:8 ar:5 tractability:1 applicability:1 entry:5 usefulness:1 delay:5 too:1 conduction:3 spatiotemporal:3 nski:1 st:1 density:2 fundamental:1 peak:2 off:1 analogously:1 kami:1 connectivity:57 squared:6 reflect:3 again:1 recorded:1 cognitive:2 american:1 li:1 account:3 potential:1 coefficient:7 explicitly:1 depends:5 ornstein:1 onset:2 performed:2 break:1 stream:3 closed:4 characterizes:1 bayes:1 hf:1 recover:1 participant:4 ass:1 square:3 wiener:1 variance:7 ensemble:1 correspond:1 subthreshold:1 bayesian:16 biophysically:2 raw:1 drive:1 cybernetics:1 history:1 oscillatory:1 explain:1 whenever:1 synaptic:2 pp:2 frequency:12 fixate:2 associated:2 recovers:1 static:1 auditory:8 popular:1 knowledge:2 subsection:2 improves:1 hilbert:1 uncover:1 actually:1 back:1 feed:3 dt:6 follow:1 methodology:1 specify:2 response:4 nonparametrics:2 strongly:1 bressler:3 furthermore:4 angular:1 correlation:14 hand:2 nonlinear:1 assessment:1 propagation:1 lack:1 nonparametrically:1 indicated:2 impulse:1 menon:1 usa:1 effect:5 true:1 analytically:2 discounting:3 regularization:2 symmetric:2 white:2 deal:1 during:3 rhythm:1 generalized:1 complete:1 demonstrate:3 ridge:3 tn:1 dedicated:1 meaning:2 ranging:1 variational:1 novel:1 harmonic:2 fi:3 recently:1 functional:6 eichler:1 slight:1 approximates:1 resting:3 marginals:1 significant:1 measurement:1 cambridge:2 ai:6 smoothness:6 uv:2 resorting:1 heskes:2 dj:4 cortex:17 longer:1 dominant:1 posterior:9 multivariate:2 recent:1 showed:2 discretizing:1 binary:1 seen:3 determine:1 paradigm:1 ller:1 dashed:3 ii:2 signal:7 multiple:1 infer:1 cgp:2 smooth:3 faster:1 characterized:3 cross:3 long:2 luca:1 bigger:2 biophysics:1 regression:14 basic:3 dipolar:1 expectation:1 exerts:1 arxiv:4 kernel:47 uhlenbeck:1 addition:1 want:2 separately:1 interval:2 harrison:1 walker:1 source:11 macroscopic:1 biased:1 rest:1 daunizeau:1 strict:1 hz:1 recording:3 isolate:1 flow:2 call:2 axonal:1 structural:2 gerven:4 ideal:1 hjort:1 iii:2 identically:1 easy:1 stephan:2 xj:6 marginalization:1 independence:1 cn:2 tm:1 translates:1 shift:4 absent:2 expression:2 effort:1 penalty:1 mattout:1 algebraic:1 detailed:1 aimed:1 clear:2 nonparametric:10 band:1 induces:1 clip:1 generate:1 specifies:1 outperform:1 inhibitory:1 neuroscience:5 estimated:12 correctly:1 anatomical:2 hyperparameter:1 affected:1 nevertheless:1 demonstrating:1 clarity:1 preprocessed:1 pj:4 graph:3 relaxation:2 year:1 sum:1 powerful:1 family:2 throughout:1 oscillation:3 coherence:4 scaling:1 fl:1 pay:1 activity:6 strength:8 constraint:3 fourier:8 argument:1 structured:2 tv:2 disconnected:5 membrane:1 smaller:2 beneficial:1 remain:1 postsynaptic:2 suppressed:1 biologically:1 unregularized:1 equation:13 turn:1 granger:6 discus:1 tractable:1 autoregression:2 available:2 apply:1 hierarchical:2 enforce:3 appropriate:1 spectral:1 cake:25 top:1 denotes:2 leakage:4 mechelli:1 volterra:1 parametric:3 strategy:1 diagonal:2 said:1 gradient:1 dp:1 simulated:1 presynaptic:1 cellular:1 assuming:1 ru:4 meg:9 modeled:4 unfortunately:1 negative:5 neuroscientific:1 lagged:1 fdr:1 calcium:1 unknown:2 perform:1 neuron:2 convolution:2 observation:2 finite:1 t:5 kse:4 anti:3 parietal:4 neuropsychopharmacology:1 extended:1 communication:2 neurobiology:1 frame:1 gc:1 interacting:2 perturbation:1 varied:1 arbitrary:2 introduced:4 david:1 pair:1 specified:1 connection:17 learned:4 established:1 nip:1 able:1 usually:2 dynamical:3 pattern:2 summarize:1 reverts:1 max:1 interpretability:4 video:3 hemodynamics:1 critical:1 natural:1 friston:5 regularized:4 turner:1 mediated:1 coupled:1 deviate:1 prior:6 review:2 understanding:1 discovery:1 kf:7 multiplication:1 relative:1 fully:3 expect:1 par:2 mixed:1 interesting:1 limitation:2 var:10 localized:2 validation:1 undamped:1 degree:1 principle:1 elsewhere:1 excitatory:1 rasmussen:1 bias:2 allow:1 understand:1 wide:2 taking:1 van:4 distributed:1 default:2 xn:1 cortical:2 donders:4 rich:1 autoregressive:3 world:1 reside:1 commonly:1 instructed:3 evaluating:1 alpha:3 active:1 assumed:2 xi:4 alternatively:1 continuous:1 latent:2 channel:1 nature:2 transfer:3 learn:1 ca:1 reasonably:1 eeg:1 mse:6 complex:6 european:1 domain:10 roebroeck:1 significance:1 whole:1 noise:11 hyperparameters:3 quadrature:1 x1:1 neuronal:20 causality:11 referred:2 fig:9 tl:1 screen:2 neuroimage:7 structurally:1 exponential:1 vanish:1 theorem:1 formula:1 minute:1 specific:2 moran:1 intrinsic:6 janssen:2 circularly:2 adding:1 ci:13 magnitude:3 illustrates:1 television:1 gap:1 mf:3 electrophysiology:1 distinguishable:1 univariate:2 infinitely:3 explore:1 visual:2 expressed:1 scalar:1 watch:1 applies:1 springer:1 dt1:1 truth:5 extracted:2 dcm:6 conditional:1 goal:1 formulated:1 magnetoencephalography:1 consequently:1 exposition:1 oscillator:1 occ:2 price:1 analysing:1 djh:2 included:1 specifically:3 determined:2 corrected:1 wt:2 conservative:1 total:1 called:1 invariance:1 duality:1 uber:1 underestimation:1 meaningful:1 internal:1 mark:1 latter:1 modulated:2 audio:2 regularizing:1 |
6,297 | 6,697 | Decoupling ?when to update? from ?how to update?
Eran Malach
School of Computer Science
The Hebrew University, Israel
[email protected]
Shai Shalev-Shwartz
School of Computer Science
The Hebrew University, Israel
[email protected]
Abstract
Deep learning requires data. A useful approach to obtain data is to be creative and
mine data from various sources, that were created for different purposes. Unfortunately, this approach often leads to noisy labels. In this paper, we propose a meta
algorithm for tackling the noisy labels problem. The key idea is to decouple ?when
to update? from ?how to update?. We demonstrate the effectiveness of our algorithm by mining data for gender classification by combining the Labeled Faces
in the Wild (LFW) face recognition dataset with a textual genderizing service,
which leads to a noisy dataset. While our approach is very simple to implement,
it leads to state-of-the-art results. We analyze some convergence properties of the
proposed algorithm.
1
Introduction
In recent years, deep learning achieves state-of-the-art results in various different tasks, however,
neural networks are mostly trained using supervised learning, where a massive amount of labeled
data is required. While collecting unlabeled data is relatively easy given the amount of data available
on the web, providing accurate labeling is usually an expensive task. In order to overcome this
problem, data science becomes an art of extracting labels out of thin air. Some popular approaches
to labeling are crowdsourcing, where the labeling is not done by experts, and mining available
meta-data, such as text that is linked to an image in a webpage. Unfortunately, this gives rise to a
problem of abundant noisy labels - labels may often be corrupted [19], which might deteriorate the
performance of neural-networks [12].
Let us start with an intuitive explanation as to why noisy labels are problematic. Common neural
network optimization algorithms start with a random guess of what the classifier should be, and
then iteratively update the classifier based on stochastically sampled examples from a given dataset,
optimizing a given loss function such as the hinge loss or the logistic loss. In this process, wrong
predictions lead to an update of the classifier that would hopefully result in better classification
performance. While at the beginning of the training process the predictions are likely to be wrong,
as the classifier improves it will fail on less and less examples, thus making fewer and fewer updates.
On the other hand, in the presence of noisy labels, as the classifier improves the effect of the noise
increases - the classifier may give correct predictions, but will still have to update due to wrong
labeling. Thus, in an advanced stage of the training process the majority of the updates may actually
be due to wrongly labeled examples, and therefore will not allow the classifier to further improve.
To tackle this problem, we propose to decouple the decision of ?when to update? from the decision
of ?how to update?. As mentioned before, in the presence of noisy labels, if we update only when
the classifier?s prediction differs from the available label, then at the end of the optimization process,
these few updates will probably be mainly due to noisy labels. We would therefore like a different
update criterion, that would let us decide whether it is worthy to update the classifier based on a
given example. We would like to preserve the behavior of performing many updates at the beginning of the training process but only a few updates when we approach convergence. To do so, we
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
suggest to train two predictors, and perform update steps only in case of disagreement between them.
This way, when the predictors get better, the ?area? of their disagreement gets smaller, and updates
are performed only on examples that lie in the disagreement area, therefore preserving the desired
behavior of the standard optimization process. On the other hand, since we do not perform an update based on disagreement with the label (which may be due to a problem in the label rather than
a problem in the predictor), this method keeps the effective amount of noisy labels seen throughout
the training process at a constant rate.
The idea of deciding ?when to update? based on a disagreement between classifiers is closely related
to approaches for active learning and selective sampling - a setup in which the learner does not have
unlimited access to labeled examples, but rather has to query for each instance?s label, provided at
a given cost (see for example [34]). Specifically, the well known query-by-committee algorithm
maintains a version space of hypotheses and at each iteration, decides whether to query the label of
a given instance by sampling two hypotheses uniformly at random from the version space [35, 14].
Naturally, maintaining the version space of deep networks seems to be intractable. Our algorithm
maintains only two deep networks. The difference between them stems from the random initialization. Therefore, unlike the original query-by-committee algorithm, that samples from the version
space at every iteration, we sample from the original hypotheses class only once (at the initialization), and from there on, we update these two hypotheses using the backpropagation rule, when
they disagree on the label. To the best of our knowledge, this algorithm was not proposed/analyzed
previously, not in the active learning literature and especially not as a method for dealing with noisy
labels.
To show that this method indeed improves the robustness of deep learning to noisy labels, we conduct an experiment that aims to study a real-world scenario of acquiring noisy labels for a given
dataset. We consider the task of gender classification based on images. We did not have a dedicated
dataset for this task. Instead, we relied on the Labeled Faces in the Wild (LFW) dataset, which contains images of different people along with their names, but with no information about their gender.
To find the gender for each image, we use an online service to match a gender to a given name (as is
suggested by [25]), a method which is naturally prone to noisy labels (due to unisex names). Applying our algorithm to an existing neural network architecture reduces the effect of the noisy labels,
achieving better results than similar available approaches, when tested on a clean subset of the data.
We also performed a controlled experiment, in which the base algorithm is the perceptron, and show
that using our approach leads to a noise resilient algorithm, which can handle an extremely high
label noise rates of up to 40%. The controlled experiments are detailed in Appendix B.
In order to provide theoretical guarantees for our meta algorithm, we need to tackle two questions:
1. does this algorithm converge? and if so, how quickly? and 2. does it converge to an optimum? We
give a positive answer to the first question, when the base algorithm is the perceptron and the noise
is label flip with a constant probability. Specifically, we prove that the expected number of iterations
required by the resulting algorithm equals (up to a constant factor) to that of the perceptron in the
noise-free setting. As for the second question, clearly, the convergence depends on the initialization
of the two predictors. For example, if we initialize the two predictors to be the same predictor, the
algorithm will not perform any updates. Furthermore, we derive lower bounds on the quality of the
solution even if we initialize the two predictors at random. In particular, we show that for some
distributions, the algorithm?s error will be bounded away from zero, even in the case of linearly
separable data. This raises the question of whether a better initialization procedure may be helpful.
Indeed, we show that for the same distribution mentioned above, even if we add random label noise,
if we initialize the predictors by performing few vanilla perceptron iterations, then the algorithm
performs much better. Despite this worst case pessimism, we show that empirically, when working
with natural data, the algorithm converges to a good solution. We leave a formal investigation of
distribution dependent upper bounds to future work.
2
Related Work
The effects of noisy labels was vastly studied in many different learning algorithms (see for example
the survey in [13]), and various solutions to this problem have been proposed, some of them with
theoretically provable bounds, including methods like statistical queries, boosting, bagging and more
[21, 26, 7, 8, 29, 31, 23, 27, 3]. Our focus in this paper is on the problem of noisy labels in the context
of deep learning. Recently, there have been several works aiming at improving the resilience of deep
2
learning to noisy labels. To the best of our knowledge, there are four main approaches. The first
changes the loss function. The second adds a layer that tries to mimic the noise behavior. The third
groups examples into buckets. The fourth tries to clean the data as a preprocessing step. Beyond
these approaches, there are methods that assume a small clean data set and another large, noisy, or
even unlabeled, data set [30, 6, 38, 1]. We now list some specific algorithms from these families.
[33] proposed to change the cross entropy loss function by adding a regularization term that takes
into account the current prediction of the network. This method is inspired by a technique called
minimum entropy regularization, detailed in [17, 16]. It was also found to be effective by [12],
which suggested a further improvement of this method by effectively increasing the weight of the
regularization term during the training procedure.
[28] suggested to use a probabilistic model that models the conditional probability of seeing a wrong
label, where the correct label is a latent variable of the model. While [28] assume that the probability
of label-flips between classes is known in advance, a follow-up work by [36] extends this method
to a case were these probabilities are unknown. An improved method, that takes into account the
fact that some instances might be more likely to have a wrong label, has been proposed recently
in [15]. In particular, they add another softmax layer to the network, that can use the output of
the last hidden layer of the network in order to predict the probability of the label being flipped.
Unfortunately, their method involves optimizing the biases of the additional softmax layer by first
training it on a simpler setup (without using the last hidden layer), which implies two-phase training
that further complicates the optimization process. It is worth noting that there are some other works
that suggest methods that are very similar to [36, 15], with a slightly different objective or training
method [5, 20], or otherwise suggest a complicated process which involves estimation of the classdependent noise probabilities [32]. Another method from the same family is the one described in
[37], who suggests to differentiate between ?confusing? noise, where some features of the example
make it hard to label, or otherwise a completely random label noise, where the mislabeling has no
clear reason.
[39] suggested to train the network to predict labels on a randomly selected group of images from
the same class, instead of classifying each image individually. In their method, a group of images
is fed as an input to the network, which merges their inner representation in a deeper level of the
network, along with an attention model added to each image, and producing a single prediction.
Therefore, noisy labels may appear in groups with correctly labeled examples, thus diminishing
their impact. The final setup is rather complicated, involving many hyper-parameters, rather than
providing a simple plug-and-play solution to make an existing architecture robust to noisy labels.
From the family of preprocessing methods, we mention [4, 10], that try to eliminate instances that are
suspected to be mislabeled. Our method shares a similar motivation of disregarding contaminated
instances, but without the cost of complicating the training process by a preprocessing phase.
In our experiment we test the performance of our method against methods that are as simple as
training a vanilla version of neural network. In particular, from the family of modified loss function
we chose the two variants of the regularized cross entropy loss suggested by [33] (soft and hard
bootsrapping). From the family of adding a layer that models the noise, we chose to compare to one
of the models suggested in [15] (which is very similar to the model proposed by [36]), because this
model does not require any assumptions or complication of the training process. We find that our
method outperformed all of these competing methods, while being extremely simple to implement.
Finally, as mentioned before, our ?when to update? rule is closely related to approaches for active
learning and selective sampling, and in particular to the query-by-committee algorithm. In [14] a
thorough analysis is provided for various base algorithms implementing the query-by-committee
update rule, and particularly they analyze the perceptron base algorithm under some strong distributional assumptions. In other works, an ensemble of neural networks is trained in an active learning
setup to improve the generalization of neural networks [11, 2, 22]. Our method could be seen as
a simplified member of ensemble methods. As mentioned before, our motivation is very different
than the active learning scenario, since our main goal is dealing with noisy labels, rather than trying
to reduce the number of label queries. To the best of our knowledge, the algorithm we propose was
not used or analyzed in the past for the purpose of dealing with noisy labels in deep learning.
3
3
METHOD
As mentioned before, to tackle the problem of noisy labels, we suggest to change the update rule
commonly used in deep learning optimization algorithms in order to decouple the decision of ?when
to update? from ?how to update?. In our approach, the decision of ?when to update? does not depend
on the label. Instead, it depends on a disagreement between two different networks. This method
could be generally thought of as a meta-algorithm that uses two base classifiers, performing updates
according to a base learning algorithm, but only on examples for which there is a disagreement
between the two classifiers.
To put this formally, let X be an instance space and Y be the label space, and assume we sample
? over X ? Y, with possibly noisy labels. We wish to train a classifier
examples from a distribution D
h, coming from a hypothesis class H. We rely on an update rule, U , that updates h based on its
current value as well as a mini-batch of b examples. The meta algorithm receives as input a pair of
two classifiers, h1 , h2 2 H, the update rule, U , and a mini batch size, b. A pseudo-code is given in
Algorithm 1.
Note that we do not specify how to initialize the two base classifiers, h1 , h2 . When using deep
learning as the base algorithm, the easiest approach is maybe to perform a random initialization.
Another approach is to first train the two classifiers while following the regular ?when to update?
rule (which is based on the label y), possibly training each classifier on a different subset of the data,
and switching to the suggested update rule only in an advanced stage of the training process. We
later show that the second approach is preferable.
At the end of the optimization process, we can simply return one of the trained classifiers. If a
small accurately labeled test data is available, we can choose to return the classifier with the better
accuracy on the clean test data.
Algorithm 1 Update by Disagreement
input:
an update rule U
batch size b
two initial predictors h1 , h2 2 H
for t = 1, 2, . . . , N do
?b
draw mini-batch (x1 , y1 ), . . . , (xb , yb ) ? D
let S = {(xi , yi ) : h1 (xi ) 6= h2 (xi )}
h1
U (h1 , S)
h2
U (h2 , S)
end for
4
Theoretical analysis
Since a convergence analysis for deep learning is beyond our reach even in the noise-free setting,
we focus on analyzing properties of our algorithm for linearly separable data, which is corrupted by
random label noise, and while using the perceptron as a base algorithm.
Let X = {x 2 Rd : kxk ? 1}, Y = {?1}, and let D be a probability distribution over X ? Y,
such that there exists w? for which D({(x, y) : yhw? , xi < 1}) = 0. The distribution we observe,
? is a noisy version of D. Specifically, to sample (x, y?) ? D
? one should sample (x, y) ? D
denoted D,
and output (x, y) with probability 1 ? and (x, y) with probability ?. Here, ? is in [0, 1/2).
Finally, let H be the class of linear classifiers, namely, H = {x 7! sign(hw, xi) : w 2 Rd }. We
use the perceptron?s update rule with mini-batch size of 1. That is, given the classifier wt 2 Rd , the
update on example (xt , yt ) 2 X ? Y is: wt+1 = U (wt , (xt , yt )) := wt + yt xt .
As mentioned in the introduction, to provide a full theoretical analysis of this algorithm, we need to
account for two questions:
1. does this algorithm converge? and if so, how quickly?
2. does it converge to an optimum?
4
Theorem 1 below provides a positive answer for the first question. It shows that the number of
updates of our algorithm is only larger by a constant factor (that depends on the initial vectors and
the amount of noise) relatively to the bound for the vanilla perceptron in the noise-less case.
Theorem 1 Suppose that the ?Update by Disagreement? algorithm is run on a sequence of random
? and with initial vectors w(1) , w(2) . Denote K = maxi kw(i) k. Let T be the
N examples from D,
0
0
0
number of updates performed by the ?Update by Disagreement? algorithm.
? 2
?
Then, E[T ] ? 3(1(4 K+1)
2?)2 kw k where the expectation is w.r.t. the randomness of sampling from D.
Proof It will be more convenient to rewrite the algorithm as follows. We perform N iterations,
(i)
(i)
where at iteration t we receive (xt , y?t ), and update wt+1 = wt + ?t y?t xt , where
(
(1)
(2)
1 if sign(hwt , xt i) 6= sign(hwt , xt i)
?t =
0 otherwise
Observe that we can write y?t = ?t yt , where (xt , yt ) ? D, and ?t is a random variables with
?
P[?t = 1] = 1 ? and P[?t = 1] = ?. We
Palso use the notation vt = yt hw , xt i and v?t = ?t vt .
Our goal is to upper bound T? := E[T ] = E[ t ?t ].
We start with showing that
E
"
N
X
?t v?t
t=1
#
(1
(1)
2?)T
Indeed, since ?t is independent of ?t and vt , we get that:
E[?t v?t ] = E[?t ?t vt ] = E[?t ] ? E[?t vt ] = (1
where in the last inequality we used the fact that vt
Summing over t we obtain that Equation 1 holds.
Next, we show that for i 2 {1, 2},
(i)
kwt k2
?
(i)
kw0 k2
(1)
+
N
X
2?) E[?t vt ]
1 with probability 1 and ?t is non-negative.
(2)
?t (2kw0
t=1
(2)
2?) E[?t ]
(1
(1)
(2)
w0 k + 1)
(1)
(2)
(1)
(2)
Indeed, since the update of wt+1 and wt+1 is identical, we have that kwt+1 wt+1 k = kw0
w0 k
(1)
(2)
for every t. Now, whenever ?t = 1 we have that either yt hwt 1 , xt i ? 0 or yt hwt 1 , xt i ? 0.
(1)
Assume w.l.o.g. that yt hwt 1 , xt i ? 0. Then,
(1)
Second,
(1)
1
kwt k2 = kwt
(2)
(1) 2
1k
+ yt xt k2 = kwt
(1)
1 , xt i
+ 2yt hwt
(1) 2
1k
+ kxt k2 ? kwt
(2)
(2) 2
(2)
2
2
1 + yt xt k = kwt 1 k + 2yt hwt 1 , xt i + kxt k
(2)
(2)
(1)
kwt 1 k2 + 2yt hwt 1 wt 1 , xt i + kxt k2
(2)
(2)
(1)
(2)
(2)
kwt 1 k2 + 2 kwt 1 wt 1 k + 1 = kwt 1 k2 + 2 kw0
+1
kwt k2 = kwt
?
?
(i)
(i) 2
1k
Therefore, the above two equations imply 8i 2 {1, 2}, kwt k2 ? kwt
Summing over t we obtain that Equation 2 holds.
(1)
w0 k + 1
(2)
+ 2 kw0
(1)
w0 k + 1.
Equipped with Equation 1 and Equation 2 we are ready to prove the theorem.
(i)
(2)
(1)
Denote K = maxi kw0 k and note that kw0
w0 k ? 2K. We prove the theorem by providing
(i)
upper and lower bounds on E[hwt , w? i]. Combining the update rule with Equation 1 we get:
"N
#
X
(i)
(i)
(i)
?
?
E[hwt , w i] = hw , w i + E
?t v?t
hw , w? i + (1 2?)T?
K kw? k + (1 2?)T?
0
0
t=1
To construct an upper bound, first note that Equation 2 implies that
(i)
(i)
(2)
E[kwt k2 ] ? kw0 k2 + (2kw0
(1)
w0 k + 1)T? ? K 2 + (4 K + 1) T?
5
Using the above and Jensen?s inequality, we get that
q
q
(i)
(i)
(i)
E[hwt , w? i] ? E[kwt k kw? k] ? kw? k E[kwt k2 ] ? kw? k K 2 + (4 K + 1)T?
Comparing the upper and lower bounds, we obtain that
q
K kw? k + (1 2?)T? ? kw? k K 2 + (4 K + 1)T?
p
p
p
Using a + b ? a + b, the above implies that
p
p
(1 2?)T? kw? k (4 K + 1) T? 2 K kw? k ? 0
p
p
Denote ? = kw? k (4 K + 1), then the above also implies that (1 2?)T? ? T? ? ? 0.
Denote = ?/(1 2?), using standard algebraic manipulations, the above implies that
T? ? + 2 + 1.5 ? 3 2 ,
where we used the fact that kw? k must be at least 1 for the separability assumption to hold, hence
1. This concludes our proof.
The above theorem tells us that our algorithm converges quickly. We next address the second question, regarding the quality of the point to which the algorithm converges. As mentioned in the
(1)
(2)
introduction, the convergence must depend on the initial predictors. Indeed, if w0 = w0 , then
(1)
the algorithm will not make any updates. The next question is what happens if we initialize w0
(2)
and w0 at random. The lemma below shows that this does not suffice to ensure convergence to the
optimum, even if the data is linearly separable without noise. The proof for this lemma is given in
Appendix A.
Lemma 1 Fix some 2 (0, 1) and let d be an integer greater than 40 log(1/ ). There exists a
distribution over Rd ? {?1}, which is separable by a weight vector w? for which kw? k2 = d, such
that running the ?Update by Disagreement? algorithm, with the perceptron as the underlying update
(1)
(2)
rule, and with every coordinate of w0 , w0 initialized according to any symmetric distribution over
R, will yield a solution whose error is at least 1/8, with probability of at least 1
.
Trying to circumvent the lower bound given in the above lemma, one may wonder what would
(1)
(2)
happen if we will initialize w0 , w0 differently. Intuitively, maybe noisy labels are not such a big
(1)
(2)
problem at the beginning of the learning process. Therefore, we can initialize w0 , w0 by running
the vanilla perceptron for several iterations, and only then switch to our algorithm. Trivially, for
the distribution we constructed in the proof of Lemma 1, this approach will work just because in
(1)
(2)
the noise-free setting, both w0 and w0 will converge to vectors that give the same predictions
?
as w . But, what would happen in the noisy setting, when we flip the label of every example with
probability of ?? The lemma below shows that the error of the resulting solution is likely to be order
of ?3 . Here again, the proof is given in Appendix A.
? over Rd ? {?1} such that to
Lemma 2 Consider a vector w? 2 {?1}d and the distribution D
sample a pair (x, y?) we first choose x uniformly at random from {e1 , . . . , ed }, set y = hw? , ei i, and
(1)
(2)
set y? = y with probability 1 ? and y? = y with probability ?. Let w0 , w0 be the result of
? for any number of iterations.
running the vanilla perceptron algorithm on random examples from D
Suppose that we run the ?Update by Disagreement? algorithm for an additional arbitrary number
of iterations. Then, the error of the solution is likely to be ?(?3 ).
To summarize, we see that without making additional assumptions on the data distribution, it is
impossible to prove convergence of our algorithm to a good solution. In the next section we show
that for natural data distributions, our algorithm converges to a very good solution.
5
EXPERIMENTS
We now demonstrate the merit of our suggested meta-algorithm using empirical evaluation. Our
main experiment is using our algorithm with deep networks in a real-world scenario of noisy labels.
6
In particular, we use a hypothesis class of deep networks and a Stochastic Gradient Descent with momentum as the basis update rule. The task is classifying face images according to gender. As training
data, we use the Labeled Faces in the Wild (LFW) dataset for which we had a labeling of the name
of the face, but we did not have gender labeling. To construct gender labels, we used an external
service that provides gender labels based on names. This process resulted in noisy labels. We show
that our method leads to state-of-the-art results on this task, compared to competing noise robustness
methods. We also performed controlled experiments to demonstrate our algorithm?s performance on
linear classification with varying levels of noise. These results are detailed in Appendix B.
5.1
Deep Learning
We have applied our algorithm with a Stochastic Gradient Descent (SGD) with momentum as the
base update rule on the task of labeling images of faces according to gender. The images were taken
from the Labeled Faces in the Wild (LFW) benchmark [18]. This benchmark consists of 13,233
images of 5,749 different people collected from the web, labeled with the name of the person in the
picture. Since the gender of each subject is not provided, we follow the method of [25] and use a
service that determines a person?s gender by their name (if it is recognized), along with a confidence
level. This method gives rise to ?natural? noisy labels due to ?unisex? names, and therefore allows
us to experiment with a real-world setup of dataset with noisy labels.
Name
Confidence
Kim
88%
Morgan
64%
Joan
82%
Leslie
88%
Correct
Mislabeled
Figure 1: Images from the dataset tagged as female
We have constructed train and test sets as follows. We first took all the individuals on which the
gender service gave 100% confidence. We divided this set at random into three subsets of equal
size, denoted N1 , N2 , N3 . We denote by N4 the individuals on which the confidence level is in
[90%, 100%), and by N5 the individuals on which the confidence level is in [0%, 90%). Needless to
say that all the sets N1 , . . . , N5 have zero intersection with each other.
We repeated each experiment three times, where in every time we used a different Ni as the test set,
for i 2 {1, 2, 3}. Suppose N1 is the test set, then for the training set we used two configurations:
1. A dataset consisting of all the images that belong to names in N2 , N3 , N4 , N5 , where unrecognized names were labeled as male (since the majority of subjects in LFW are males).
2. A dataset consisting of all the images that belong to names in N2 , N3 , N4 .
We use a network architecture suggested by [24], using an available tensorflow implementation1 .
It should be noted that we did not change any parameters of the network architecture or the optimization process, and use the default parameters in the implementation. Since the amount of male
and female subjects in the dataset is not balanced, we use an objective of maximizing the balanced
accuracy [9] - the average accuracy obtained on either class.
Training is done for 30,000 iterations on 128 examples mini-batch. In order to make the networks
disagreement meaningful, we initialize the two networks by training both of them normally (updating on all the examples) until iteration #5000, where we switch to training with the ?Update by
Disagreement? rule. Due to the fact that we are not updating on all examples, we decrease the weight
of batches with less than 10% of the original examples in the original batch to stabilize gradients. 2 .
1
2
https://github.com/dpressel/rude-carnie.
Code is available online on https://github.com/emalach/UpdateByDisagreement.
7
We inspect the balanced accuracy on our test data during the training process, comparing our method
to a vanilla neural network training, as well as to soft and hard bootstrapping described in [33] and
to the s-model described in [15], all of which are using the same network architecture. We use the
initialization parameters for [33, 15] that were suggested in the original papers. We show that while
in other methods, the accuracy effectively decreases during the training process due to overfitting the
noisy labels, in our method this effect is less substantial, allowing the network to keep improving.
We study two different scenarios, one in which a small clean test data is available for model selection,
and therefore we can choose the iteration with best test accuracy, and a more realistic scenario where
there is no clean test data at hand. For the first scenario, we observe the balanced accuracy of the best
available iteration. For the second scenario, we observe the balanced accuracy of the last iteration.
As can be seen in Figure 2 and the supplementary results listed in Table 1 in Appendix B, our
method outperforms the other methods in both situations. This is true for both datasets, although, as
expected, the improvement in performance is less substantial on the cleaner dataset.
The second best algorithm is the s-model described in [15]. Since our method can be applied to
any base algorithm, we also applied our method on top of the s-model. This yields even better
performance, especially when the data is less noisy, where we obtain a significant improvement.
Dataset #1 - more noise
Dataset #2 - less noise
Figure 2: Balanced accuracy of all methods on clean test data, trained on the two different datasets.
6
Discussion
We have described an extremely simple approach for supervised learning in the presence of noisy
labels. The basic idea is to decouple the ?when to update? rule from the ?how to update? rule. We
achieve this by maintaining two predictors, and update based on their disagreement. We have shown
that this simple approach leads to state-of-the-art results.
Our theoretical analysis shows that the approach leads to fast convergence rate when the underlying
update rule is the perceptron. We have also shown that proving that the method converges to an optimal solution must rely on distributional assumptions. There are several immediate open questions
that we leave to future work. First, suggesting distributional assumptions that are likely to hold in
practice and proving that the algorithm converges to an optimal solution under these assumptions.
Second, extending the convergence proof beyond linear predictors. While obtaining absolute convergence guarantees seems beyond reach at the moment, coming up with oracle based convergence
guarantees may be feasible.
Acknowledgements: This research is supported by the European Research Council (TheoryDL
project).
8
References
[1] Rie Kubota Ando and Tong Zhang. Two-view feature generation model for semi-supervised
learning. In Proceedings of the 24th international conference on Machine learning, pages
25?32. ACM, 2007.
[2] Les E Atlas, David A Cohn, Richard E Ladner, Mohamed A El-Sharkawi, Robert J Marks,
ME Aggoune, and DC Park. Training connectionist networks with queries and selective sampling. In NIPS, pages 566?573, 1989.
[3] Pranjal Awasthi, Maria Florina Balcan, and Philip M Long. The power of localization for
efficiently learning linear separators with noise. In Proceedings of the 46th Annual ACM Symposium on Theory of Computing, pages 449?458. ACM, 2014.
[4] Ricardo Barandela and Eduardo Gasca. Decontamination of training samples for supervised
pattern recognition methods. In Joint IAPR International Workshops on Statistical Techniques
in Pattern Recognition (SPR) and Structural and Syntactic Pattern Recognition (SSPR), pages
621?630. Springer, 2000.
[5] Alan Joseph Bekker and Jacob Goldberger. Training deep neural-networks based on unreliable labels. In Acoustics, Speech and Signal Processing (ICASSP), 2016 IEEE International
Conference on, pages 2682?2686. IEEE, 2016.
[6] Avrim Blum and Tom Mitchell. Combining labeled and unlabeled data with co-training. In
Proceedings of the eleventh annual conference on Computational learning theory, pages 92?
100. ACM, 1998.
[7] Jakramate Bootkrajang and Ata Kab?an. Label-noise robust logistic regression and its applications. In Joint European Conference on Machine Learning and Knowledge Discovery in
Databases, pages 143?158. Springer, 2012.
[8] Jakramate Bootkrajang and Ata Kab?an. Boosting in the presence of label noise. arXiv preprint
arXiv:1309.6818, 2013.
[9] Kay Henning Brodersen, Cheng Soon Ong, Klaas Enno Stephan, and Joachim M Buhmann.
The balanced accuracy and its posterior distribution. In Pattern recognition (ICPR), 2010 20th
international conference on, pages 3121?3124. IEEE, 2010.
[10] Carla E. Brodley and Mark A. Friedl. Identifying mislabeled training data. Journal of Artificial
Intelligence Research, 11:131?167, 1999.
[11] David Cohn, Les Atlas, and Richard Ladner. Improving generalization with active learning.
Machine learning, 15(2):201?221, 1994.
[12] David Flatow and Daniel Penner. On the robustness of convnets to training on noisy
labels. http://cs231n.stanford.edu/reports/flatow_penner_report.
pdf, 2017.
[13] Beno??t Fr?enay and Michel Verleysen. Classification in the presence of label noise: a survey.
IEEE transactions on neural networks and learning systems, 25(5):845?869, 2014.
[14] Yoav Freund, H Sebastian Seung, Eli Shamir, and Naftali Tishby. Selective sampling using the
query by committee algorithm. Machine learning, 28(2-3):133?168, 1997.
[15] Jacob Goldberger and Ehud Ben-Reuven. Training deep neural networks using a noise adaptation layer. Under review for ICLR, 2017.
[16] Yves Grandvalet and Yoshua Bengio. Entropy regularization. Semi-supervised learning, pages
151?168, 2006.
[17] Yves Grandvalet, Yoshua Bengio, et al. Semi-supervised learning by entropy minimization. In
NIPS, volume 17, pages 529?536, 2004.
9
[18] Gary B Huang, Manu Ramesh, Tamara Berg, and Erik Learned-Miller. Labeled faces in the
wild: A database for studying face recognition in unconstrained environments. Technical
report, Technical Report 07-49, University of Massachusetts, Amherst, 2007.
[19] Panagiotis G Ipeirotis, Foster Provost, and Jing Wang. Quality management on amazon mechanical turk. In Proceedings of the ACM SIGKDD workshop on human computation, pages
64?67. ACM, 2010.
[20] Pravin Kakar and Alex Yong-Sang Chia. Probabilistic learning from mislabelled data for multimedia content recognition. In Multimedia and Expo (ICME), 2015 IEEE International Conference on, pages 1?6. IEEE, 2015.
[21] Michael Kearns. Efficient noise-tolerant learning from statistical queries. Journal of the ACM
(JACM), 45(6):983?1006, 1998.
[22] Anders Krogh, Jesper Vedelsby, et al. Neural network ensembles, cross validation, and active
learning. Advances in neural information processing systems, 7:231?238, 1995.
[23] Jan Larsen, L Nonboe, Mads Hintz-Madsen, and Lars Kai Hansen. Design of robust neural
network classifiers. In Acoustics, Speech and Signal Processing, 1998. Proceedings of the
1998 IEEE International Conference on, volume 2, pages 1205?1208. IEEE, 1998.
[24] Gil Levi and Tal Hassner. Age and gender classification using convolutional neural networks.
In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pages 34?42, 2015.
[25] Philip Masek and Magnus Thulin. Evaluation of face recognition apis and libraries. Master?s
thesis, University of Gothenburg, 2015.
[26] Ross A McDonald, David J Hand, and Idris A Eckley. An empirical comparison of three
boosting algorithms on real data sets with artificial class noise. In International Workshop on
Multiple Classifier Systems, pages 35?44. Springer, 2003.
[27] Aditya Krishna Menon, Brendan van Rooyen, and Nagarajan Natarajan. Learning from binary
labels with instance-dependent corruption. arXiv preprint arXiv:1605.00751, 2016.
[28] Volodymyr Mnih and Geoffrey E Hinton. Learning to label aerial images from noisy data.
In Proceedings of the 29th International Conference on Machine Learning (ICML-12), pages
567?574, 2012.
[29] Nagarajan Natarajan, Inderjit S Dhillon, Pradeep K Ravikumar, and Ambuj Tewari. Learning
with noisy labels. In Advances in neural information processing systems, pages 1196?1204,
2013.
[30] Kamal Nigam and Rayid Ghani. Analyzing the effectiveness and applicability of co-training. In
Proceedings of the ninth international conference on Information and knowledge management,
pages 86?93. ACM, 2000.
[31] Giorgio Patrini, Frank Nielsen, Richard Nock, and Marcello Carioni. Loss factorization,
weakly supervised learning and label noise robustness. arXiv preprint arXiv:1602.02450,
2016.
[32] Giorgio Patrini, Alessandro Rozza, Aditya Menon, Richard Nock, and Lizhen Qu. Making neural networks robust to label noise: a loss correction approach. arXiv preprint
arXiv:1609.03683, 2016.
[33] Scott Reed, Honglak Lee, Dragomir Anguelov, Christian Szegedy, Dumitru Erhan, and Andrew Rabinovich. Training deep neural networks on noisy labels with bootstrapping. arXiv
preprint arXiv:1412.6596, 2014.
[34] Burr Settles. Active learning literature survey. University of Wisconsin, Madison, 52(5566):11, 2010.
10
[35] H Sebastian Seung, Manfred Opper, and Haim Sompolinsky. Query by committee. In Proceedings of the fifth annual workshop on Computational learning theory, pages 287?294. ACM,
1992.
[36] Sainbayar Sukhbaatar, Joan Bruna, Manohar Paluri, Lubomir Bourdev, and Rob Fergus. Training convolutional networks with noisy labels. arXiv preprint arXiv:1406.2080, 2014.
[37] Tong Xiao, Tian Xia, Yi Yang, Chang Huang, and Xiaogang Wang. Learning from massive
noisy labeled data for image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 2691?2699, 2015.
[38] Xiaojin Zhu. Semi-supervised learning literature survey. Computer Sciences TR 1530, 2005.
[39] Bohan Zhuang, Lingqiao Liu, Yao Li, Chunhua Shen, and Ian Reid. Attend in groups:
a weakly-supervised deep learning framework for learning from web data. arXiv preprint
arXiv:1611.09960, 2016.
11
| 6697 |@word version:6 seems:2 open:1 jacob:2 palso:1 sgd:1 mention:1 tr:1 moment:1 initial:4 configuration:1 contains:1 liu:1 daniel:1 past:1 existing:2 outperforms:1 current:2 comparing:2 com:2 bootkrajang:2 goldberger:2 tackling:1 must:3 realistic:1 happen:2 klaas:1 christian:1 atlas:2 mislabelled:1 update:57 sukhbaatar:1 intelligence:1 fewer:2 guess:1 selected:1 beginning:3 manfred:1 provides:2 boosting:3 complication:1 simpler:1 zhang:1 along:3 constructed:2 symposium:1 prove:4 consists:1 wild:5 eleventh:1 burr:1 theoretically:1 deteriorate:1 expected:2 indeed:5 paluri:1 behavior:3 inspired:1 equipped:1 increasing:1 becomes:1 provided:3 project:1 bounded:1 notation:1 suffice:1 underlying:2 pravin:1 israel:2 what:4 easiest:1 bootstrapping:2 eduardo:1 guarantee:3 pseudo:1 thorough:1 every:5 collecting:1 tackle:3 preferable:1 classifier:23 wrong:5 k2:15 normally:1 appear:1 producing:1 reid:1 before:4 service:5 positive:2 giorgio:2 resilience:1 attend:1 aiming:1 switching:1 despite:1 analyzing:2 might:2 chose:2 initialization:6 studied:1 suggests:1 co:2 factorization:1 tian:1 mads:1 practice:1 implement:2 differs:1 backpropagation:1 procedure:2 jan:1 area:2 empirical:2 thought:1 convenient:1 confidence:5 regular:1 seeing:1 suggest:4 get:5 eckley:1 unlabeled:3 needle:1 selection:1 wrongly:1 put:1 context:1 applying:1 impossible:1 yt:14 maximizing:1 attention:1 survey:4 shen:1 amazon:1 identifying:1 rule:18 kay:1 proving:2 handle:1 beno:1 coordinate:1 shamir:1 play:1 suppose:3 massive:2 us:1 hypothesis:6 recognition:10 expensive:1 particularly:1 updating:2 natarajan:2 malach:2 distributional:3 labeled:14 database:2 preprint:7 wang:2 worst:1 sompolinsky:1 decrease:2 aggoune:1 mentioned:7 balanced:7 substantial:2 environment:1 alessandro:1 seung:2 mine:1 ong:1 cs231n:1 trained:4 raise:1 depend:2 rewrite:1 weakly:2 iapr:1 localization:1 learner:1 completely:1 basis:1 mislabeled:3 icassp:1 joint:2 differently:1 various:4 train:5 fast:1 effective:2 jesper:1 query:12 artificial:2 labeling:7 tell:1 hyper:1 shalev:1 whose:1 larger:1 supplementary:1 stanford:1 say:1 kai:1 otherwise:3 syntactic:1 mislabeling:1 noisy:40 final:1 online:2 differentiate:1 sequence:1 kxt:3 took:1 propose:3 coming:2 fr:1 adaptation:1 combining:3 achieve:1 intuitive:1 webpage:1 convergence:11 optimum:3 extending:1 jing:1 converges:6 leave:2 ben:1 derive:1 andrew:1 ac:2 bourdev:1 school:2 krogh:1 strong:1 c:1 involves:2 implies:5 closely:2 correct:3 nock:2 stochastic:2 lars:1 human:1 settle:1 implementing:1 resilient:1 require:1 hassner:1 nagarajan:2 fix:1 generalization:2 investigation:1 sainbayar:1 manohar:1 correction:1 hold:4 magnus:1 deciding:1 predict:2 achieves:1 enno:1 purpose:2 estimation:1 hwt:11 outperformed:1 panagiotis:1 label:68 hansen:1 ross:1 council:1 individually:1 minimization:1 awasthi:1 clearly:1 aim:1 modified:1 rather:5 brodersen:1 varying:1 focus:2 joachim:1 improvement:3 maria:1 mainly:1 sigkdd:1 brendan:1 kim:1 helpful:1 dependent:2 el:1 anders:1 eliminate:1 diminishing:1 hidden:2 selective:4 classification:7 denoted:2 verleysen:1 art:5 softmax:2 initialize:8 equal:2 once:1 construct:2 beach:1 sampling:6 identical:1 flipped:1 kw:13 marcello:1 icml:1 park:1 thin:1 kamal:1 future:2 mimic:1 report:3 contaminated:1 connectionist:1 richard:4 few:3 yoshua:2 randomly:1 preserve:1 resulted:1 kwt:18 individual:3 phase:2 consisting:2 n1:3 ando:1 mining:2 mnih:1 evaluation:2 male:3 analyzed:2 pradeep:1 xb:1 accurate:1 conduct:1 initialized:1 abundant:1 desired:1 theoretical:4 complicates:1 instance:7 soft:2 penner:1 yoav:1 rabinovich:1 leslie:1 cost:2 applicability:1 subset:3 predictor:12 wonder:1 tishby:1 reuven:1 answer:2 corrupted:2 st:1 person:2 international:9 amherst:1 huji:2 probabilistic:2 lee:1 pessimism:1 michael:1 quickly:3 yao:1 vastly:1 again:1 thesis:1 management:2 choose:3 possibly:2 huang:2 stochastically:1 external:1 expert:1 return:2 ricardo:1 michel:1 sang:1 account:3 suggesting:1 volodymyr:1 szegedy:1 li:1 stabilize:1 depends:3 performed:4 try:3 h1:6 later:1 view:1 analyze:2 linked:1 start:3 relied:1 maintains:2 complicated:2 shai:1 implementation1:1 il:2 air:1 accuracy:10 ni:1 yves:2 who:1 efficiently:1 ensemble:3 yield:2 miller:1 convolutional:2 accurately:1 worth:1 corruption:1 randomness:1 reach:2 whenever:1 ed:1 sebastian:2 yhw:1 against:1 mohamed:1 bekker:1 tamara:1 turk:1 naturally:2 proof:6 vedelsby:1 larsen:1 sampled:1 dataset:15 popular:1 massachusetts:1 mitchell:1 knowledge:5 improves:3 nielsen:1 actually:1 supervised:9 follow:2 tom:1 specify:1 improved:1 yb:1 rie:1 done:2 furthermore:1 just:1 stage:2 until:1 convnets:1 hand:4 working:1 receives:1 web:3 ei:1 cohn:2 hopefully:1 icme:1 logistic:2 quality:3 menon:2 usa:1 effect:4 name:12 kab:2 true:1 regularization:4 hence:1 tagged:1 symmetric:1 iteratively:1 dhillon:1 during:3 naftali:1 noted:1 criterion:1 trying:2 decontamination:1 pdf:1 demonstrate:3 mcdonald:1 performs:1 dedicated:1 patrini:2 dragomir:1 balcan:1 image:17 recently:2 common:1 empirically:1 volume:2 belong:2 lizhen:1 significant:1 anguelov:1 honglak:1 rd:5 vanilla:6 trivially:1 unconstrained:1 had:1 bruna:1 access:1 base:11 add:3 posterior:1 recent:1 female:2 madsen:1 optimizing:2 chunhua:1 scenario:7 manipulation:1 meta:6 inequality:2 binary:1 vt:7 yi:2 preserving:1 seen:3 minimum:1 additional:3 greater:1 morgan:1 krishna:1 recognized:1 converge:5 signal:2 semi:4 full:1 multiple:1 reduces:1 stem:1 alan:1 technical:2 match:1 plug:1 cross:3 long:2 chia:1 divided:1 e1:1 ravikumar:1 controlled:3 impact:1 prediction:7 involving:1 regression:1 variant:1 n5:3 basic:1 lfw:5 expectation:1 florina:1 arxiv:14 iteration:14 vision:2 receive:1 source:1 unlike:1 probably:1 subject:3 henning:1 member:1 effectiveness:2 integer:1 extracting:1 structural:1 presence:5 noting:1 manu:1 bengio:2 easy:1 stephan:1 yang:1 switch:2 gave:1 architecture:5 competing:2 inner:1 idea:3 reduce:1 regarding:1 lubomir:1 whether:3 lingqiao:1 algebraic:1 speech:2 deep:18 useful:1 generally:1 detailed:3 clear:1 listed:1 cleaner:1 maybe:2 amount:5 tewari:1 http:3 problematic:1 gil:1 sign:3 correctly:1 write:1 group:5 key:1 four:1 levi:1 blum:1 achieving:1 clean:7 year:1 run:2 eli:1 fourth:1 master:1 extends:1 throughout:1 family:5 decide:1 draw:1 decision:4 appendix:5 confusing:1 bound:9 layer:7 haim:1 cheng:1 oracle:1 annual:3 xiaogang:1 alex:1 expo:1 n3:3 unlimited:1 yong:1 tal:1 extremely:3 performing:3 separable:4 relatively:2 kubota:1 rude:1 according:4 creative:1 icpr:1 aerial:1 smaller:1 slightly:1 separability:1 joseph:1 n4:3 making:3 happens:1 qu:1 rob:1 intuitively:1 bucket:1 taken:1 equation:7 previously:1 kw0:9 fail:1 committee:6 flip:3 merit:1 fed:1 end:3 studying:1 available:9 observe:4 away:1 disagreement:15 batch:8 robustness:4 original:5 bagging:1 top:1 running:3 ensure:1 hinge:1 maintaining:2 madison:1 especially:2 objective:2 question:9 added:1 sspr:1 eran:2 shais:1 gradient:3 iclr:1 majority:2 philip:2 w0:20 me:1 mail:1 collected:1 reason:1 provable:1 erik:1 code:2 reed:1 mini:5 providing:3 hebrew:2 setup:5 unfortunately:3 mostly:1 robert:1 frank:1 negative:1 rise:2 unrecognized:1 rooyen:1 implementation:1 design:1 carioni:1 unknown:1 perform:5 allowing:1 disagree:1 upper:5 inspect:1 ladner:2 datasets:2 benchmark:2 ramesh:1 descent:2 immediate:1 situation:1 hinton:1 y1:1 worthy:1 dc:1 provost:1 ninth:1 arbitrary:1 david:4 pair:2 required:2 namely:1 mechanical:1 acoustic:2 merges:1 textual:1 tensorflow:1 learned:1 nip:3 address:1 beyond:4 suggested:10 usually:1 below:3 pattern:6 scott:1 summarize:1 ambuj:1 including:1 explanation:1 power:1 natural:3 rely:2 regularized:1 circumvent:1 buhmann:1 ipeirotis:1 advanced:2 zhu:1 improve:2 classdependent:1 github:2 brodley:1 imply:1 library:1 picture:1 zhuang:1 created:1 ready:1 concludes:1 xiaojin:1 text:1 joan:2 literature:3 acknowledgement:1 discovery:1 review:1 wisconsin:1 freund:1 loss:9 generation:1 geoffrey:1 age:1 validation:1 h2:6 xiao:1 suspected:1 foster:1 grandvalet:2 classifying:2 share:1 pranjal:1 prone:1 ata:2 supported:1 last:4 free:3 soon:1 formal:1 allow:1 bias:1 perceptron:12 deeper:1 face:11 absolute:1 fifth:1 van:1 overcome:1 default:1 complicating:1 world:3 opper:1 xia:1 commonly:1 preprocessing:3 simplified:1 erhan:1 transaction:1 keep:2 dealing:3 unreliable:1 apis:1 active:8 decides:1 overfitting:1 tolerant:1 summing:2 xi:5 shwartz:1 fergus:1 latent:1 why:1 table:1 robust:4 ca:1 decoupling:1 obtaining:1 nigam:1 improving:3 spr:1 european:2 separator:1 ehud:1 did:3 main:3 linearly:3 motivation:2 noise:30 big:1 n2:3 repeated:1 ghani:1 x1:1 tong:2 momentum:2 wish:1 lie:1 third:1 hw:5 ian:1 theorem:5 dumitru:1 specific:1 xt:17 showing:1 kakar:1 jensen:1 maxi:2 list:1 disregarding:1 intractable:1 exists:2 workshop:5 avrim:1 adding:2 effectively:2 sharkawi:1 entropy:5 intersection:1 carla:1 simply:1 likely:5 jacm:1 kxk:1 aditya:2 inderjit:1 chang:1 acquiring:1 gender:14 springer:3 gary:1 determines:1 acm:9 conditional:1 goal:2 feasible:1 change:4 hard:3 content:1 specifically:3 uniformly:2 wt:11 decouple:4 lemma:7 kearns:1 called:1 multimedia:2 meaningful:1 formally:1 berg:1 people:2 mark:2 tested:1 crowdsourcing:1 |
6,298 | 6,698 | Self-Normalizing Neural Networks
G?nter Klambauer
Thomas Unterthiner
Andreas Mayr
Sepp Hochreiter
LIT AI Lab & Institute of Bioinformatics,
Johannes Kepler University Linz
A-4040 Linz, Austria
{klambauer,unterthiner,mayr,hochreit}@bioinf.jku.at
Abstract
Deep Learning has revolutionized vision via convolutional neural networks (CNNs)
and natural language processing via recurrent neural networks (RNNs). However,
success stories of Deep Learning with standard feed-forward neural networks
(FNNs) are rare. FNNs that perform well are typically shallow and, therefore cannot
exploit many levels of abstract representations. We introduce self-normalizing
neural networks (SNNs) to enable high-level abstract representations. While
batch normalization requires explicit normalization, neuron activations of SNNs
automatically converge towards zero mean and unit variance. The activation
function of SNNs are ?scaled exponential linear units? (SELUs), which induce
self-normalizing properties. Using the Banach fixed-point theorem, we prove that
activations close to zero mean and unit variance that are propagated through many
network layers will converge towards zero mean and unit variance ? even under
the presence of noise and perturbations. This convergence property of SNNs allows
to (1) train deep networks with many layers, (2) employ strong regularization
schemes, and (3) to make learning highly robust. Furthermore, for activations
not close to unit variance, we prove an upper and lower bound on the variance,
thus, vanishing and exploding gradients are impossible. We compared SNNs on
(a) 121 tasks from the UCI machine learning repository, on (b) drug discovery
benchmarks, and on (c) astronomy tasks with standard FNNs, and other machine
learning methods such as random forests and support vector machines. For FNNs
we considered (i) ReLU networks without normalization, (ii) batch normalization,
(iii) layer normalization, (iv) weight normalization, (v) highway networks, and (vi)
residual networks. SNNs significantly outperformed all competing FNN methods
at 121 UCI tasks, outperformed all competing methods at the Tox21 dataset, and
set a new record at an astronomy data set. The winning SNN architectures are often
very deep.
1
Introduction
Deep Learning has set new records at different benchmarks and led to various commercial applications
[21, 26]. Recurrent neural networks (RNNs) [15] achieved new levels at speech and natural language
processing, for example at the TIMIT benchmark [10] or at language translation [29], and are already
employed in mobile devices [24]. RNNs have won handwriting recognition challenges (Chinese and
Arabic handwriting) [26, 11, 4] and Kaggle challenges, such as the ?Grasp-and Lift EEG? competition.
Their counterparts, convolutional neural networks (CNNs) [20] excel at vision and video tasks. CNNs
are on par with human dermatologists at the visual detection of skin cancer [8]. The visual processing
for self-driving cars is based on CNNs [16], as is the visual input to AlphaGo which has beaten one
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
BatchNorm Depth 8
BatchNorm Depth 16
BatchNorm Depth 32
SNN Depth 8
SNN Depth 16
SNN Depth 32
1
10
2
10
3
10
4
10
5
Training loss
10
0
250
500
750
1000
Iterations
1250
1500
1750
BatchNorm Depth 8
BatchNorm Depth 16
BatchNorm Depth 32
SNN Depth 8
SNN Depth 16
SNN Depth 32
100
10
1
10
2
10
3
10
4
10
5
Training loss
100
2000
0
250
500
750
1000
Iterations
1250
1500
1750
2000
Figure 1: The left panel and the right panel show the training error (y-axis) for feed-forward neural
networks (FNNs) with batch normalization (BatchNorm) and self-normalizing networks (SNN) across
update steps (x-axis) on the MNIST dataset the CIFAR10 dataset, respectively. We tested networks
with 8, 16, and 32 layers and learning rate 1e-5. FNNs with batch normalization exhibit high variance
due to perturbations. In contrast, SNNs do not suffer from high variance as they are more robust to
perturbations and learn faster.
of the best human GO players [27]. At vision challenges, CNNs are constantly winning, for example
at the large ImageNet competition [19, 13], but also almost all Kaggle vision challenges, such as the
?Diabetic Retinopathy? and the ?Right Whale? challenges [7, 12].
However, looking at Kaggle challenges that are not related to vision or sequential tasks, gradient
boosting, random forests, or support vector machines (SVMs) are winning most of the competitions.
Deep Learning is notably absent, and for the few cases where FNNs won, they are shallow. For
example, the HIGGS challenge, the Merck Molecular Activity challenge, and the Tox21 Data
challenge were all won by FNNs with at most four hidden layers. Surprisingly, it is hard to find
success stories with FNNs that have many hidden layers, though they would allow for different levels
of abstract representations of the input [2].
To robustly train very deep CNNs, batch normalization evolved into a standard to normalize neuron
activations to zero mean and unit variance [17]. Layer normalization [1] also ensures zero mean
and unit variance, while weight normalization [25] ensures zero mean and unit variance if in the
previous layer the activations have zero mean and unit variance. Natural neural networks [6] also aim
at normalizing the variance of activations by reparametrization of the weights. However, training with
normalization techniques is perturbed by stochastic gradient descent (SGD), stochastic regularization
(like dropout), and the estimation of the normalization parameters. Both RNNs and CNNs can
stabilize learning via weight sharing, therefore they are less prone to these perturbations. In contrast,
FNNs trained with normalization techniques suffer from these perturbations and have high variance in
the training error (see Figure 1). This high variance hinders learning and slows it down. Furthermore,
strong regularization, such as dropout, is not possible as it would further increase the variance
which in turn would lead to divergence of the learning process. We believe that this sensitivity to
perturbations is the reason that FNNs are less successful than RNNs and CNNs.
Self-normalizing neural networks (SNNs) are robust to perturbations and do not have high variance
in their training errors (see Figure 1). SNNs push neuron activations to zero mean and unit variance
thereby leading to the same effect as batch normalization, which enables to robustly learn many
layers. SNNs are based on scaled exponential linear units ?SELUs? which induce self-normalizing
properties like variance stabilization which in turn avoids exploding and vanishing gradients.
2
Self-normalizing Neural Networks (SNNs)
Normalization and SNNs. For a neural network with activation function f , we consider two
consecutive layers that are connected by a weight matrix W . Since the input to a neural network
is a random variable, the activations x in the lower layer, the network inputs z = W x, and the
activations y = f (z) in the higher layer are random variables as well. We assume that all activations
2
xi of the lower layer have mean ? := E(xi ) and variance ? := Var(xi ). An activation y in the higher
layer has mean ?
? := E(y) and variance ?? := Var(y). Here E(.) denotes the expectation and Var(.)
the variance of a random variable. A single activation y = f (z) has net input z = wT x. For n units
with activation xiP
, 1 6 i 6 n in the lower layer, we define n timesP
the mean of the weight vector
n
n
w ? Rn as ? := i=1 wi and n times the second moment as ? := i=1 wi2 .
We consider the mapping g that maps mean and variance of the activations from one layer to mean
and variance of the activations in the next layer g : (?, ?) 7? (?
?, ??). Normalization techniques
like batch, layer, or weight normalization ensure a mapping g that keeps (?, ?) and (?
?, ??) close to
predefined values, typically (0, 1).
Definition 1 (Self-normalizing neural net). A neural network is self-normalizing if it possesses a
mapping g : ? 7? ? for each activation y that maps mean and variance from one layer to the next
and has a stable and attracting fixed point depending on (?, ? ) in ?. Furthermore, the mean and
the variance remain in the domain ?, that is g(?) ? ?, where ? = {(?, ?) | ? ? [?min , ?max ], ? ?
[?min , ?max ]}. When iteratively applying the mapping g, each point within ? converges to this fixed
point.
Therefore, we consider activations of a neural network to be normalized, if both their mean and their
variance across samples are within predefined intervals. If mean and variance of x are already within
these intervals, then also mean and variance of y remain in these intervals, i.e., the normalization is
transitive across layers. Within these intervals, the mean and variance both converge to a fixed point
if the mapping g is applied iteratively.
Therefore, SNNs keep normalization of activations when propagating them through layers of the
network. The normalization effect is observed across layers of a network: in each layer the activations
are getting closer to the fixed point. The normalization effect can also observed be for two fixed
layers across learning steps: perturbations of lower layer activations or weights are damped in the
higher layer by drawing the activations towards the fixed point. If for all y in the higher layer, ? and
? of the corresponding weight vector are the same, then the fixed points are also the same. In this
case we have a unique fixed point for all activations y. Otherwise, in the more general case, ? and
? differ for different y but the mean activations are drawn into [?min , ?max ] and the variances are
drawn into [?min , ?max ].
Constructing Self-Normalizing Neural Networks. We aim at constructing self-normalizing neural networks by adjusting the properties of the function g. Only two design choices are available for
the function g: (1) the activation function and (2) the initialization of the weights.
For the activation function, we propose ?scaled exponential linear units? (SELUs) to render a FNN as
self-normalizing. The SELU activation function is given by
x
if x > 0
selu(x) = ?
.
(1)
?ex ? ? if x 6 0
SELUs allow to construct a mapping g with properties that lead to SNNs. SNNs cannot be derived
with (scaled) rectified linear units (ReLUs), sigmoid units, tanh units, and leaky ReLUs. The
activation function is required to have (1) negative and positive values for controlling the mean, (2)
saturation regions (derivatives approaching zero) to dampen the variance if it is too large in the lower
layer, (3) a slope larger than one to increase the variance if it is too small in the lower layer, (4) a
continuous curve. The latter ensures a fixed point, where variance damping is equalized by variance
increasing. We met these properties of the activation function by multiplying the exponential linear
unit (ELU) [5] with ? > 1 to ensure a slope larger than one for positive net inputs.
For the weight initialization, we propose ? = 0 and ? = 1 for all units in the higher layer. The
next paragraphs will show the advantages of this initialization. Of course, during learning these
assumptions on the weight vector will be violated. However, we can prove the self-normalizing
property even for weight vectors that are not normalized, therefore, the self-normalizing property can
be kept during learning and weight changes.
Deriving the Mean and Variance Mapping Function g. We assume that the xi are independent
from each other but share the same mean ? and variance ?. Of course, the independence assumptions
is not fulfilled in general. We will elaborate on the independence assumption below. The network
3
T
input
which we can infer the following moments E(z) =
Pn z in the higher layer is z = w x for P
n
i=1 wi xi ) = ? ? , where we used the independence
i=1 wi E(xi ) = ? ? and Var(z) = Var(
of the xi . The net input z is a weighted sum of independent, but not necessarily identically distributed
variables xi ,?
for which the central limit theorem
? (CLT) states that z approaches a normal distribution:
z ? N (??, ?? ) with density pN (z; ??, ?? ). According to the CLT, the larger n, the closer is z
to a normal distribution. For Deep Learning, broad layers with hundreds of neurons xi are common.
Therefore the assumption that z is normally distributed is met well for most currently used neural
networks (see Supplementary Figure S7). The function g maps the mean and variance of activations
in the lower layer to the mean ?
? = E(y) and variance ?? = Var(y) of the activations y in the next
layer:
Z ?
?
?
?
?
g:
7?
: ?
?(?, ?, ?, ? ) =
selu(z) pN (z; ??, ?? ) dz
(2)
?
??
??
Z ?
?
?2 .
??(?, ?, ?, ? ) =
selu(z)2 pN (z; ??, ?? ) dz ? ?
??
These integrals can be analytically computed and lead to following mappings of the moments:
1
??
?
? = ? (??) erf ? ?
+
(3)
2
2 ??
!
r
(??)2
?? + ??
??
2?
?
??+ ??
2 erfc
? ?
?? e 2(?? ) + ??
?e
? ? erfc ? ?
+
?
2 ??
2 ??
??
??
1
?? + ??
+ ?2 ?2e??+ 2 erfc ? ?
(4)
?? = ?2 (??)2 + ?? 2 ? erfc ? ?
2
2 ??
2 ??
!
r
(??)2
?
?? + 2??
??
2
2
? 2(?? )
2(??+?? )
+e
erfc ? ?
(??) ?? e
? (?
?)
+ erfc ? ?
+
?
2 ??
2 ??
Stable and Attracting Fixed Point (0, 1) for Normalized Weights. We assume a normalized
weight vector w with ? = 0 and ? = 1. Given a fixed point (?, ?), we can solve equations Eq. (3)
and Eq. (4) for ? and ?. We chose the fixed point (?, ?) = (0, 1), which is typical for activation
normalization. We obtain the fixed point equations ?
? = ? = 0 and ?? = ? = 1 that we solve for ?
and ? and obtain the solutions ?01 ? 1.6733 and ?01 ? 1.0507, where the subscript 01 indicates
that these are the parameters for fixed point (0, 1). The analytical expressions for ?01 and ?01 are
given in Supplementary Eq. (8). We are interested whether the fixed point (?, ?) = (0, 1) is stable
and attracting. If the Jacobian of g has a norm smaller than 1 at the fixed point, then g is a contraction
mapping and the fixed point is stable. The (2x2)-Jacobian J (?, ?) of g : (?, ?) 7? (?
?, ??) evaluated at
the fixed point (0, 1) with ?01 and ?01 is J (0, 1) = ((0.0, 0.088834), (0.0, 0.782648)). The spectral
norm of J (0, 1) (its largest singular value) is 0.7877 < 1. That means g is a contraction mapping
around the fixed point (0, 1) (the mapping is depicted in Figure 2). Therefore, (0, 1) is a stable
fixed point of the mapping g. The norm of the Jacobian also determines the convergence rate as a
consequence of the Banach fixed point theorem. The convergence rate around the fixed point (0,1) is
about 0.78. In general, the convergence rate depends on ?, ?, ?, ? and is between 0.78 and 1.
Stable and Attracting Fixed Points for Unnormalized Weights. A normalized weight vector w
cannot be ensured during learning. For SELU parameters ? = ?01 and ? = ?01 , we show in the next
theorem that if (?, ? ) is close to (0, 1), then g still has an attracting and stable fixed point that is close
to (0, 1). Thus, in the general case there still exists a stable fixed point which, however, depends
on (?, ? ). If we restrict (?, ?, ?, ? ) to certain intervals, then we can show that (?, ?) is mapped to
the respective intervals. Next we present the central theorem of this paper, from which follows that
SELU networks are self-normalizing under mild conditions on the weights.
Theorem 1 (Stable and Attracting Fixed Points). We assume ? = ?01 and ? = ?01 . We restrict the
range of the variables to the following intervals ? ? [?0.1, 0.1], ? ? [?0.1, 0.1], ? ? [0.8, 1.5], and
? ? [0.95, 1.1], that define the functions? domain ?. For ? = 0 and ? = 1, the mapping Eq. (2) has
the stable fixed point (?, ?) = (0, 1), whereas for other ? and ? the mapping Eq. (2) has a stable
and attracting fixed point depending on (?, ? ) in the (?, ?)-domain: ? ? [?0.03106, 0.06773] and
4
Figure 2: For ? = 0 and ? = 1, the mapping g of mean ? (x-axis) and variance ? (y-axis) to the
next layer?s mean ?
? and variance ?? is depicted. Arrows show in which direction (?, ?) is mapped by
g : (?, ?) 7? (?
?, ??). The fixed point of the mapping g is (0, 1).
? ? [0.80009, 1.48617]. All points within the (?, ?)-domain converge when iteratively applying the
mapping Eq. (2) to this fixed point.
Proof. We provide a proof sketch (see detailed proof in Supplementary Material). With the Banach
fixed point theorem we show that there exists a unique attracting and stable fixed point. To this end,
we have to prove that a) g is a contraction mapping and b) that the mapping stays in the domain, that
is, g(?) ? ?. The spectral norm of the Jacobian of g can be obtained via an explicit formula for the
largest singular value for a 2 ? 2 matrix. g is a contraction mapping if its spectral norm is smaller
than 1. We perform a computer-assisted proof to evaluate the largest singular value on a fine grid and
ensure the precision of the computer evaluation by an error propagation analysis of the implemented
algorithms on the according hardware. Singular values between grid points are upper bounded by the
mean value theorem. To this end, we bound the derivatives of the formula for the largest singular
value with respect to ?, ?, ?, ?. Then we apply the mean value theorem to pairs of points, where one
is on the grid and the other is off the grid. This shows that for all values of ?, ?, ?, ? in the domain ?,
the spectral norm of g is smaller than one. Therefore, g is a contraction mapping on the domain ?.
Finally, we show that the mapping g stays in the domain ? by deriving bounds on ?
? and ??. Hence,
the Banach fixed-point theorem holds and there exists a unique fixed point in ? that is attained.
Consequently, feed-forward neural networks with many units in each layer and with the SELU
activation function are self-normalizing (see definition 1), which readily follows from Theorem 1. To
give an intuition, the main property of SELUs is that they damp the variance for negative net inputs
and increase the variance for positive net inputs. The variance damping is stronger if net inputs are
further away from zero while the variance increase is stronger if net inputs are close to zero. Thus, for
large variance of the activations in the lower layer the damping effect is dominant and the variance
decreases in the higher layer. Vice versa, for small variance the variance increase is dominant and the
variance increases in the higher layer.
However, we cannot guarantee that mean and variance remain in the domain ?. Therefore, we next
treat the case where (?, ?) are outside ?. It is especially crucial to consider ? because this variable
has much stronger influence than ?. Mapping ? across layers to a high value corresponds to an
exploding gradient, since the Jacobian of the activation of high layers with respect to activations
in lower layers has large singular values. Analogously, mapping ? across layers to a low value
corresponds to an vanishing gradient. Bounding the mapping of ? from above and below would avoid
both exploding and vanishing gradients. Theorem 2 states that the variance of neuron activations of
SNNs is bounded from above, and therefore ensures that SNNs learn robustly and do not suffer from
exploding gradients.
Theorem 2 (Decreasing ?). For ? = ?01 , ? = ?01 and the domain ?+ : ?1 6 ? 6 1, ?0.1 6 ? 6
0.1, 3 6 ? 6 16, and 0.8 6 ? 6 1.25, we have for the mapping of the variance ??(?, ?, ?, ?, ?, ?)
given in Eq. (4): ??(?, ?, ?, ?, ?01 , ?01 ) < ?.
The proof can be found in Supplementary Material. Thus, when mapped across many layers, the
variance in the interval [3, 16] is mapped to a value below 3. Consequently, all fixed points (?, ?)
5
of the mapping g (Eq. (2)) have ? < 3. Analogously, Theorem 3 states that the variance of neuron
activations of SNNs is bounded from below, and therefore ensures that SNNs do not suffer from
vanishing gradients.
Theorem 3 (Increasing ?). We consider ? = ?01 , ? = ?01 and the domain ?? : ?0.1 6 ? 6 0.1,
and ?0.1 6 ? 6 0.1. For the domain 0.02 6 ? 6 0.16 and 0.8 6 ? 6 1.25 as well as for the
domain 0.02 6 ? 6 0.24 and 0.9 6 ? 6 1.25, the mapping of the variance ??(?, ?, ?, ?, ?, ?) given
in Eq. (4) increases: ??(?, ?, ?, ?, ?01 , ?01 ) > ?.
The proof can be found in the Supplementary Material. All fixed points (?, ?) of the mapping g
(Eq. (2)) ensure for 0.8 6 ? that ?? > 0.16 and for 0.9 6 ? that ?? > 0.24. Consequently, the variance
mapping Eq. (4) ensures a lower bound on the variance ?. Therefore SELU networks control the
variance of the activations and push it into an interval, whereafter the mean and variance move
toward the fixed point. Thus, SELU networks are steadily normalizing the variance and subsequently
normalizing the mean, too. In all experiments, we observed that self-normalizing neural networks
push the mean and variance of activations into the domain ? .
Since SNNsP
have a fixed point at zero mean and unit variance for normalized weights
Initialization.
Pn
n
? = i=1 wi = 0 and ? = i=1 wi2 = 1 (see above), we initialize SNNs such that these constraints
are fulfilled in expectation. We draw the weights from a Gaussian distribution with E(wi ) = 0 and
variance Var(wi ) = 1/n. Uniform and truncated Gaussian distributions with these moments led to
networks with similar behavior. The ?MSRA initialization? is similar since it uses zero mean and
variance 2/n to initialize the weights [14]. The additional factor 2 counters the effect of rectified
linear units.
New Dropout Technique. Standard dropout randomly sets an activation x to zero with probability
1 ? q for 0 < q 6 1. In order to preserve the mean, the activations are scaled by 1/q during
training. If x has mean E(x) = ? and variance Var(x) = ?, and the dropout variable d follows a
binomial distribution B(1, q), then the mean E(1/qdx) = ? is kept. Dropout fits well to rectified
linear units, since zero is in the low variance region and corresponds to the default value. For scaled
exponential linear units, the default and low variance value is limx??? selu(x) = ??? = ?0 .
Therefore, we propose ?alpha dropout?, that randomly sets inputs to ?0 . The new mean and new
variance is E(xd + ?0 (1 ? d)) = q? + (1 ? q)?0 , and Var(xd + ?0 (1 ? d)) = q((1 ? q)(?0 ?
?)2 + ?). We aim at keeping mean and variance to their original values after ?alpha dropout?, in
order to ensure the self-normalizing property even for ?alpha dropout?. The affine transformation
a(xd + ?0 (1 ? d)) + b allows to determine parameters a and b such that mean and variance are kept
to their values: E(a(x ? d + ?0 (1 ? d)) + b) = ? and Var(a(x ? d + ?0 (1 ? d)) + b) = ? . In
contrast to dropout, a and b will depend on ? and ?, however our SNNs converge to activations with
?1/2
zero mean and unit variance. With ? = 0 and ? = 1, we obtain a = q + ?02 q(1 ? q)
and
?1/2
02
0
b = ? q + ? q(1 ? q)
((1 ? q)? ). The parameters a and b only depend on the dropout rate
1 ? q and the most negative activation ?0 . Empirically, we found that dropout rates 1 ? q = 0.05 or
0.10 lead to models with good performance. ?Alpha-dropout? fits well to scaled exponential linear
units by randomly setting activations to the negative saturation value.
Applicability of the central limit theorem and independence assumption. In the derivative of
the mapping
(Eq. (2)), we used the central limit theorem (CLT) to approximate the network inputs
Pn
z = i=1 wi xi with a normal distribution. We justified normality because network inputs represent
a weighted sum of the inputs xi , where for Deep Learning n is typically large. The Berry-Esseen
theorem states that the convergence rate to normality is n?1/2 [18]. In the classical version of the CLT,
the random variables have to be independent and identically distributed, which typically does not
hold for neural networks. However, the Lyapunov CLT does not require the variable to be identically
distributed anymore. Furthermore, even under weak dependence, sums of random variables converge
in distribution to a Gaussian distribution [3].
Optimizers. Empirically, we found that SGD, momentum, Adadelta and Adamax worked well for
training SNNs, whereas for Adam we had to adjust the parameters (?2 = 0.99, = 0.01) to obtain
proficient networks.
6
3
Experiments
We compare SNNs to other deep networks at different benchmarks. Hyperparameters such as number
of layers (blocks), neurons per layer, learning rate, and dropout rate, are adjusted by grid-search
for each dataset on a separate validation set (see Supplementary Section S4). We compare the
following FNN methods: (1) ?MSRAinit?: FNNs without normalization and with ReLU activations
and ?Microsoft weight initialization? [14]. (2) ?BatchNorm?: FNNs with batch normalization [17].
(3) ?LayerNorm?: FNNs with layer normalization [1]. (4) ?WeightNorm?: FNNs with weight
normalization [25]. (5) ?Highway?: Highway networks [28]. (6) ?ResNet?: Residual networks
[13] adapted to FNNs using residual blocks with 2 or 3 layers with rectangular or diavolo shape.
(7) ?SNNs?: Self normalizing networks with SELUs with ? = ?01 and ? = ?01 and the proposed
dropout technique and initialization strategy.
121 UCI Machine Learning Repository datasets. The benchmark comprises 121 classification
datasets from the UCI Machine Learning repository [9] from diverse application areas, such as
physics, geology, or biology. The size of the datasets ranges between 10 and 130, 000 data points
and the number of features from 4 to 250. In abovementioned work [9], there were methodological
mistakes [30] which we avoided here. Each compared FNN method was optimized with respect to
its architecture and hyperparameters on a validation set that was then removed from the subsequent
analysis. The selected hyperparameters served to evaluate the methods in terms of accuracy on
the pre-defined test sets. The accuracies are reported in the Supplementary Table S8. We ranked
the methods by their accuracy for each prediction task and compared their average ranks. SNNs
significantly outperform all competing networks in pairwise comparisons (paired Wilcoxon test
across datasets) as reported in Table 1 (left panel).
Table 1: Left: Comparison of seven FNNs on 121 UCI tasks. We consider the average rank difference
to rank 4, which is the average rank of seven methods with random predictions. The first column gives
the method, the second the average rank difference, and the last the p-value of a paired Wilcoxon test
whether the difference to the best performing method is significant. SNNs significantly outperform
all other methods. Right: Comparison of 24 machine learning methods (ML) on the UCI datasets
with more than 1000 data points. The first column gives the method, the second the average rank
difference to rank 12.5, and the last the p-value of a paired Wilcoxon test whether the difference
to the best performing method is significant. Methods that were significantly worse than the best
method are marked with ?*?. SNNs outperform all competing methods.
Method
FNN method comparison
avg. rank diff. p-value
SNN
MSRAinit
LayerNorm
Highway
ResNet
WeightNorm
BatchNorm
-0.756
-0.240*
-0.198*
0.021*
0.273*
0.397*
0.504*
Method
ML method comparison
avg. rank diff. p-value
SNN
SVM
RandomForest
MSRAinit
LayerNorm
Highway
...
2.7e-02
1.5e-02
1.9e-03
5.4e-04
7.8e-07
3.5e-06
-6.7
-6.4
-5.9
-5.4*
-5.3
-4.6*
...
5.8e-01
2.1e-01
4.5e-03
7.1e-02
1.7e-03
...
We further included 17 machine learning methods representing diverse method groups [9] in the
comparison and the grouped the data sets into ?small? and ?large? data sets (for details see Supplementary Section S4.2). On 75 small datasets with less than 1000 data points, random forests and
SVMs outperform SNNs and other FNNs. On 46 larger datasets with at least 1000 data points, SNNs
show the highest performance followed by SVMs and random forests (see right panel of Table 1, for
complete results see Supplementary Tables S9 and S10). Overall, SNNs have outperformed state of
the art machine learning methods on UCI datasets with more than 1,000 data points.
Typically, hyperparameter selection chose SNN architectures that were much deeper than the selected
architectures of other FNNs, with an average depth of 10.8 layers, compared to average depths of 6.0
for BatchNorm, 3.8 WeightNorm, 7.0 LayerNorm, 5.9 Highway, and 7.1 for MSRAinit networks. For
ResNet, the average number of blocks was 6.35. SNNs with many more than 4 layers often provide
the best predictive accuracies across all neural networks.
7
Drug discovery: The Tox21 challenge dataset. The Tox21 challenge dataset comprises about
12,000 chemical compounds whose twelve toxic effects have to be predicted based on their chemical
structure. We used the validation sets of the challenge winners for hyperparameter selection (see
Supplementary Section S4) and the challenge test set for performance comparison. We repeated
the whole evaluation procedure 5 times to obtain error bars. The results in terms of average AUC
are given in Table 2. In 2015, the challenge organized by the US NIH was won by an ensemble
of shallow ReLU FNNs which achieved an AUC of 0.846 [23]. Besides FNNs, this ensemble
also contained random forests and SVMs. Single SNNs came close with an AUC of 0.845?0.003.
The best performing SNNs have 8 layers, compared to the runner-ups ReLU networks with layer
normalization with 2 and 3 layers. Also batchnorm and weightnorm networks, typically perform best
with shallow networks of 2 to 4 layers (Table 2). The deeper the networks, the larger the difference in
performance between SNNs and other methods (see columns 5?8 of Table 2). The best performing
method is an SNN with 8 layers.
Table 2: Comparison of FNNs at the Tox21 challenge dataset in terms of AUC. The rows represent
different methods and the columns different network depth and for ResNets the number of residual
blocks (6 and 32 blocks were omitted due to computational constraints). The deeper the networks,
the more prominent is the advantage of SNNs. The best networks are SNNs with 8 layers.
#layers / #blocks
method
2
3
4
6
8
16
32
SNN
Batchnorm
WeightNorm
LayerNorm
Highway
MSRAinit
ResNet
83.7 ? 0.3
80.0 ? 0.5
83.7 ? 0.8
84.3 ? 0.3
83.3 ? 0.9
82.7 ? 0.4
82.2 ? 1.1
84.4 ? 0.5
79.8 ? 1.6
82.9 ? 0.8
84.3 ? 0.5
83.0 ? 0.5
81.6 ? 0.9
80.0 ? 2.0
84.2 ? 0.4
77.2 ? 1.1
82.2 ? 0.9
84.0 ? 0.2
82.6 ? 0.9
81.1 ? 1.7
80.5 ? 1.2
83.9 ? 0.5
77.0 ? 1.7
82.5 ? 0.6
82.5 ? 0.8
82.4 ? 0.8
80.6 ? 0.6
81.2 ? 0.7
84.5 ? 0.2
75.0 ? 0.9
81.9 ? 1.2
80.9 ? 1.8
80.3 ? 1.4
80.9 ? 1.1
81.8 ? 0.6
83.5 ? 0.5
73.7 ? 2.0
78.1 ? 1.3
78.7 ? 2.3
80.3 ? 2.4
80.2 ? 1.1
81.2 ? 0.6
82.5 ? 0.7
76.0 ? 1.1
56.6 ? 2.6
78.8 ? 0.8
79.6 ? 0.8
80.4 ? 1.9
na
Astronomy: Prediction of pulsars in the HTRU2 dataset. Since a decade, machine learning
methods have been used to identify pulsars in radio wave signals [22]. Recently, the High Time
Resolution Universe Survey (HTRU2) dataset has been released with 1,639 real pulsars and 16,259
spurious signals. Currently, the highest AUC value of a 10-fold cross-validation is 0.976 which has
been achieved by Naive Bayes classifiers followed by decision tree C4.5 with 0.949 and SVMs with
0.929. We used eight features constructed by the PulsarFeatureLab as used previously [22]. We
assessed the performance of FNNs using 10-fold nested cross-validation, where the hyperparameters
were selected in the inner loop on a validation set (for details see Supplementary Section S4). Table 3
reports the results in terms of AUC. SNNs outperform all other methods and have pushed the
state-of-the-art to an AUC of 0.98.
Table 3: Comparison of FNNs and reference methods at HTRU2 in terms of AUC. The first, fourth
and seventh column give the method, the second, fifth and eight column the AUC averaged over 10
cross-validation folds, and the third and sixth column the p-value of a paired Wilcoxon test of the
AUCs against the best performing method across the 10 folds. FNNs achieve better results than Naive
Bayes (NB), C4.5, and SVM. SNNs exhibit the best performance and set a new record.
method
SNN
MSRAinit
WeightNorm
Highway
FNN methods
AUC
0.9803 ? 0.010
0.9791 ? 0.010
0.9786* ? 0.010
0.9766* ? 0.009
p-value
3.5e-01
2.4e-02
9.8e-03
method
LayerNorm
BatchNorm
ResNet
FNN methods
AUC
0.9762* ? 0.011
0.9760 ? 0.013
0.9753* ? 0.010
p-value
1.4e-02
6.5e-02
6.8e-03
ref. methods
method AUC
NB
C4.5
SVM
0.976
0.946
0.929
SNNs and convolutional neural networks. In initial experiments with CNNs, we found that SELU
activations work well at image classification tasks: On MNIST, SNN-CNNs (2x Conv, MaxPool, 2x
fully-connected, 30 Epochs) reach 99.2%?0.1 accuracy (ReLU: 99.2%?0.1) and on CIFAR10 (2x
Conv, MaxPool, 2x Conv, MaxPool, 2x fully-connected, 200 Epochs) SNN-CNNs reach 82.5?0.8%
8
accuracy (ReLU: 76.1?1.0%). This finding unsurprising since even standard ELUs without the
self-normalizing property have been shown to improve CNN training and accuracy[5].
4
Conclusion
To summarize, self-normalizing networks work well with the following configuration:
? SELU activation with parameters ? ? 1.0507 and ? ? 1.6733,
? inputs normalized to zero mean and unit variance,
? network weights initialized with variance 1/n, and
? regularization with ?alpha-dropout?.
We have introduced self-normalizing neural networks for which we have proved that neuron activations are pushed towards zero mean and unit variance when propagated through the network.
Additionally, for activations not close to unit variance, we have proved an upper and lower bound
on the variance mapping. Consequently, SNNs do not face vanishing and exploding gradient problems. Therefore, SNNs work well for architectures with many layers, allowed us to introduce a
novel regularization scheme and learn very robustly. On 121 UCI benchmark datasets, SNNs have
outperformed other FNNs with and without normalization techniques, such as batch, layer, and
weight normalization, or specialized architectures, such as Highway or Residual networks. SNNs
also yielded the best results on drug discovery and astronomy tasks. The best performing SNN
architectures are typically very deep in contrast to other FNNs.
References
[1] Ba, J. L., Kiros, J. R., and Hinton, G. (2016). Layer normalization. arXiv preprint arXiv:1607.06450.
[2] Bengio, Y. (2013). Deep learning of representations: Looking forward. In Proceedings of the First
International Conference on Statistical Language and Speech Processing, pages 1?37, Berlin, Heidelberg.
[3] Bradley, R. C. (1981). Central limit theorems under weak dependence. Journal of Multivariate Analysis,
11(1):1?16.
[4] Cire?san, D. and Meier, U. (2015). Multi-column deep neural networks for offline handwritten chinese
character classification. In 2015 International Joint Conference on Neural Networks (IJCNN), pages 1?6.
IEEE.
[5] Clevert, D.-A., Unterthiner, T., and Hochreiter, S. (2015). Fast and accurate deep network learning by exponential linear units (ELUs). 5th International Conference on Learning Representations, arXiv:1511.07289.
[6] Desjardins, G., Simonyan, K., Pascanu, R., et al. (2015). Natural neural networks. In Advances in Neural
Information Processing Systems, pages 2071?2079.
[7] Dugan, P., Clark, C., LeCun, Y., and Van Parijs, S. (2016). Phase 4: Dcl system using deep learning
approaches for land-based or ship-based real-time recognition and localization of marine mammals-distributed
processing and big data applications. arXiv preprint arXiv:1605.00982.
[8] Esteva, A., Kuprel, B., Novoa, R., Ko, J., Swetter, S., Blau, H., and Thrun, S. (2017). Dermatologist-level
classification of skin cancer with deep neural networks. Nature, 542(7639):115?118.
[9] Fern?ndez-Delgado, M., Cernadas, E., Barro, S., and Amorim, D. (2014). Do we need hundreds of classifiers
to solve real world classification problems. Journal of Machine Learning Research, 15(1):3133?3181.
[10] Graves, A., Mohamed, A., and Hinton, G. (2013). Speech recognition with deep recurrent neural networks.
In IEEE International conference on acoustics, speech and signal processing (ICASSP), pages 6645?6649.
[11] Graves, A. and Schmidhuber, J. (2009). Offline handwriting recognition with multidimensional recurrent
neural networks. In Advances in neural information processing systems, pages 545?552.
[12] Gulshan, V., Peng, L., Coram, M., Stumpe, M. C., Wu, D., Narayanaswamy, A., Venugopalan, S., Widner,
K., Madams, T., Cuadros, J., et al. (2016). Development and validation of a deep learning algorithm for
detection of diabetic retinopathy in retinal fundus photographs. JAMA, 316(22):2402?2410.
9
[13] He, K., Zhang, X., Ren, S., and Sun, J. (2015a). Deep residual learning for image recognition. In IEEE
Conference on Computer Vision and Pattern Recognition (CVPR).
[14] He, K., Zhang, X., Ren, S., and Sun, J. (2015b). Delving deep into rectifiers: Surpassing human-level
performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer
Vision (ICCV), pages 1026?1034.
[15] Hochreiter, S. and Schmidhuber, J. (1997). Long short-term memory. Neural Computation, 9(8):1735?
1780.
[16] Huval, B., Wang, T., Tandon, S., et al. (2015). An empirical evaluation of deep learning on highway driving.
arXiv preprint arXiv:1504.01716.
[17] Ioffe, S. and Szegedy, C. (2015). Batch normalization: Accelerating deep network training by reducing
internal covariate shift. In Proceedings of The 32nd International Conference on Machine Learning, pages
448?456.
[18] Korolev, V. and Shevtsova, I. (2012). An improvement of the Berry?Esseen inequality with applications to
Poisson and mixed Poisson random sums. Scandinavian Actuarial Journal, 2012(2):81?105.
[19] Krizhevsky, A., Sutskever, I., and Hinton, G. (2012). Imagenet classification with deep convolutional
neural networks. In Advances in Neural Information Processing Systems, pages 1097?1105.
[20] LeCun, Y. and Bengio, Y. (1995). Convolutional networks for images, speech, and time series. The
handbook of brain theory and neural networks, 3361(10):1995.
[21] LeCun, Y., Bengio, Y., and Hinton, G. (2015). Deep learning. Nature, 521(7553):436?444.
[22] Lyon, R., Stappers, B., Cooper, S., Brooke, J., and Knowles, J. (2016). Fifty years of pulsar candidate
selection: From simple filters to a new principled real-time classification approach. Monthly Notices of the
Royal Astronomical Society, 459(1):1104?1123.
[23] Mayr, A., Klambauer, G., Unterthiner, T., and Hochreiter, S. (2016). DeepTox: Toxicity prediction using
deep learning. Frontiers in Environmental Science, 3:80.
[24] Sak, H., Senior, A., Rao, K., and Beaufays, F. (2015). Fast and accurate recurrent neural network acoustic
models for speech recognition. arXiv preprint arXiv:1507.06947.
[25] Salimans, T. and Kingma, D. P. (2016). Weight normalization: A simple reparameterization to accelerate
training of deep neural networks. In Advances in Neural Information Processing Systems, pages 901?909.
[26] Schmidhuber, J. (2015). Deep learning in neural networks: An overview. Neural Networks, 61:85?117.
[27] Silver, D., Huang, A., Maddison, C., et al. (2016). Mastering the game of Go with deep neural networks
and tree search. Nature, 529(7587):484?489.
[28] Srivastava, R. K., Greff, K., and Schmidhuber, J. (2015). Training very deep networks. In Advances in
Neural Information Processing Systems, pages 2377?2385.
[29] Sutskever, I., Vinyals, O., and Le, Q. V. (2014). Sequence to sequence learning with neural networks. In
Advances in Neural Information Processing Systems, pages 3104?3112.
[30] Wainberg, M., Alipanahi, B., and Frey, B. J. (2016). Are random forests truly the best classifiers? Journal
of Machine Learning Research, 17(110):1?5.
10
| 6698 |@word mild:1 arabic:1 repository:3 version:1 cnn:1 norm:6 stronger:3 nd:1 contraction:5 mammal:1 sgd:2 thereby:1 delgado:1 moment:4 initial:1 configuration:1 ndez:1 series:1 jku:1 bradley:1 activation:51 readily:1 subsequent:1 shape:1 enables:1 hochreit:1 update:1 selected:3 device:1 proficient:1 marine:1 vanishing:6 short:1 record:3 boosting:1 pascanu:1 kepler:1 zhang:2 constructed:1 prove:4 paragraph:1 introduce:2 cernadas:1 barro:1 pairwise:1 peng:1 notably:1 behavior:1 kiros:1 multi:1 brain:1 decreasing:1 automatically:1 snn:17 lyon:1 increasing:2 conv:3 bounded:3 panel:4 evolved:1 astronomy:4 transformation:1 finding:1 guarantee:1 multidimensional:1 xd:3 blau:1 ensured:1 scaled:7 classifier:3 control:1 unit:29 normally:1 mayr:3 positive:3 frey:1 treat:1 limit:4 consequence:1 mistake:1 subscript:1 rnns:5 chose:2 initialization:7 range:2 averaged:1 unique:3 lecun:3 block:6 optimizers:1 procedure:1 area:1 linz:2 drug:3 empirical:1 significantly:4 ups:1 pre:1 induce:2 cannot:4 close:8 selection:3 s9:1 nb:2 impossible:1 applying:2 influence:1 map:3 dz:2 sepp:1 go:2 rectangular:1 resolution:1 survey:1 deriving:2 toxicity:1 reparameterization:1 controlling:1 commercial:1 tandon:1 us:1 adadelta:1 recognition:7 observed:3 preprint:4 wang:1 region:2 ensures:6 connected:3 hinders:1 sun:2 decrease:1 counter:1 removed:1 highest:2 principled:1 intuition:1 trained:1 depend:2 predictive:1 localization:1 icassp:1 joint:1 accelerate:1 various:1 train:2 actuarial:1 fast:2 equalized:1 lift:1 outside:1 whose:1 larger:5 supplementary:11 solve:3 cvpr:1 drawing:1 otherwise:1 erf:1 simonyan:1 advantage:2 sequence:2 net:8 analytical:1 propose:3 clevert:1 uci:8 loop:1 achieve:1 competition:3 normalize:1 getting:1 sutskever:2 convergence:5 adam:1 converges:1 silver:1 resnet:5 depending:2 recurrent:5 batchnorm:13 geology:1 propagating:1 eq:12 strong:2 implemented:1 predicted:1 elus:2 met:2 differ:1 direction:1 lyapunov:1 cnns:11 stochastic:2 subsequently:1 stabilization:1 human:3 coram:1 enable:1 filter:1 material:3 alphago:1 require:1 adjusted:1 frontier:1 assisted:1 hold:2 around:2 considered:1 normal:3 mapping:32 driving:2 desjardins:1 consecutive:1 omitted:1 released:1 estimation:1 outperformed:4 radio:1 tanh:1 currently:2 highway:10 largest:4 grouped:1 vice:1 dampen:1 weighted:2 gaussian:3 aim:3 pn:6 avoid:1 mobile:1 derived:1 improvement:1 methodological:1 rank:9 indicates:1 contrast:4 xip:1 typically:7 hidden:2 spurious:1 interested:1 overall:1 classification:8 development:1 art:2 initialize:2 construct:1 beach:1 whale:1 biology:1 lit:1 broad:1 report:1 employ:1 few:1 randomly:3 preserve:1 divergence:1 phase:1 microsoft:1 detection:2 limx:1 highly:1 evaluation:3 grasp:1 adjust:1 runner:1 brooke:1 truly:1 damped:1 predefined:2 accurate:2 integral:1 closer:2 cifar10:2 respective:1 damping:3 tree:2 iv:1 unterthiner:4 initialized:1 column:8 rao:1 applicability:1 rare:1 hundred:2 uniform:1 krizhevsky:1 successful:1 seventh:1 too:3 unsurprising:1 reported:2 perturbed:1 damp:1 st:1 density:1 tox21:5 sensitivity:1 twelve:1 international:6 stay:2 off:1 physic:1 maxpool:3 analogously:2 na:1 central:5 huang:1 worse:1 derivative:3 leading:1 szegedy:1 huval:1 retinal:1 stabilize:1 vi:1 depends:2 higgs:1 lab:1 alipanahi:1 wave:1 relus:2 bayes:2 reparametrization:1 dcl:1 slope:2 timit:1 gulshan:1 accuracy:7 convolutional:5 variance:76 ensemble:2 identify:1 weak:2 handwritten:1 nter:1 fern:1 ren:2 venugopalan:1 multiplying:1 served:1 rectified:3 reach:2 sharing:1 definition:2 sixth:1 against:1 steadily:1 mohamed:1 proof:6 handwriting:3 propagated:2 dataset:9 adjusting:1 proved:2 austria:1 astronomical:1 car:1 organized:1 feed:3 higher:8 attained:1 evaluated:1 though:1 furthermore:4 layernorm:6 sketch:1 propagation:1 believe:1 usa:1 effect:6 normalized:7 counterpart:1 regularization:5 analytically:1 hence:1 chemical:2 iteratively:3 during:4 self:23 game:1 auc:13 won:4 unnormalized:1 prominent:1 complete:1 greff:1 image:3 novel:1 recently:1 nih:1 sigmoid:1 common:1 specialized:1 empirically:2 overview:1 winner:1 banach:4 s8:1 he:2 surpassing:1 significant:2 monthly:1 versa:1 ai:1 kaggle:3 grid:5 language:4 had:1 retinopathy:2 stable:12 scandinavian:1 attracting:8 dominant:2 wilcoxon:4 multivariate:1 dugan:1 revolutionized:1 compound:1 certain:1 ship:1 schmidhuber:4 inequality:1 success:2 came:1 additional:1 employed:1 converge:6 determine:1 exploding:6 ii:1 clt:5 signal:3 infer:1 faster:1 cross:3 long:2 molecular:1 paired:4 prediction:4 ko:1 vision:7 expectation:2 poisson:2 arxiv:9 iteration:2 normalization:33 represent:2 esseen:2 resnets:1 hochreiter:4 achieved:3 justified:1 whereas:2 fine:1 interval:9 singular:6 crucial:1 fifty:1 posse:1 presence:1 iii:1 identically:3 bengio:3 independence:4 relu:6 fit:2 architecture:7 competing:4 approaching:1 restrict:2 andreas:1 inner:1 randomforest:1 absent:1 shift:1 msra:1 whether:3 expression:1 accelerating:1 narayanaswamy:1 s7:1 suffer:4 render:1 speech:6 deep:29 detailed:1 johannes:1 s4:4 hardware:1 svms:5 outperform:5 notice:1 fulfilled:2 weightnorm:6 per:1 diverse:2 hyperparameter:2 group:1 four:1 drawn:2 kept:3 sum:4 year:1 fourth:1 almost:1 wu:1 knowles:1 draw:1 decision:1 pushed:2 dropout:16 layer:60 bound:5 followed:2 fold:4 yielded:1 activity:1 adapted:1 ijcnn:1 constraint:2 worked:1 s10:1 x2:1 min:4 performing:6 according:2 across:11 remain:3 smaller:3 character:1 mastering:1 wi:7 shallow:4 toxic:1 iccv:1 equation:2 previously:1 turn:2 end:2 available:1 apply:1 eight:2 away:1 spectral:4 salimans:1 robustly:4 anymore:1 sak:1 batch:10 thomas:1 original:1 denotes:1 binomial:1 ensure:5 exploit:1 chinese:2 erfc:6 especially:1 classical:1 society:1 skin:2 move:1 already:2 cire:1 strategy:1 dependence:2 abovementioned:1 exhibit:2 gradient:10 separate:1 mapped:4 berlin:1 thrun:1 maddison:1 seven:2 reason:1 toward:1 besides:1 slows:1 negative:4 ba:1 design:1 perform:3 upper:3 neuron:8 datasets:9 benchmark:6 descent:1 truncated:1 hinton:4 looking:2 rn:1 perturbation:8 amorim:1 introduced:1 pair:1 required:1 meier:1 optimized:1 imagenet:3 c4:3 acoustic:2 kingma:1 nip:1 bar:1 below:4 pattern:1 wi2:2 challenge:15 summarize:1 fundus:1 saturation:2 max:4 memory:1 video:1 royal:1 natural:4 ranked:1 residual:6 normality:2 scheme:2 representing:1 improve:1 axis:4 excel:1 transitive:1 naive:2 epoch:2 discovery:3 bioinf:1 berry:2 graf:2 loss:2 par:1 fully:2 mixed:1 var:10 dermatologist:2 clark:1 validation:8 affine:1 story:2 share:1 land:1 translation:1 row:1 cancer:2 prone:1 course:2 surprisingly:1 last:2 keeping:1 offline:2 allow:2 deeper:3 senior:1 institute:1 face:1 fifth:1 leaky:1 distributed:5 van:1 curve:1 depth:15 default:2 world:1 avoids:1 klambauer:3 forward:4 avg:2 san:1 avoided:1 alpha:5 approximate:1 beaufays:1 keep:2 elu:1 ml:2 ioffe:1 handbook:1 xi:11 continuous:1 search:2 diabetic:2 decade:1 table:11 additionally:1 learn:4 nature:3 delving:1 robust:3 ca:1 eeg:1 forest:6 heidelberg:1 necessarily:1 constructing:2 domain:14 main:1 universe:1 arrow:1 bounding:1 noise:1 hyperparameters:4 whole:1 big:1 repeated:1 ref:1 allowed:1 elaborate:1 cooper:1 precision:1 momentum:1 comprises:2 explicit:2 exponential:7 winning:3 candidate:1 jacobian:5 third:1 theorem:19 down:1 formula:2 rectifier:1 covariate:1 beaten:1 svm:3 normalizing:25 exists:3 mnist:2 sequential:1 push:3 depicted:2 led:2 photograph:1 visual:3 snns:43 jama:1 vinyals:1 contained:1 srivastava:1 corresponds:3 nested:1 determines:1 constantly:1 environmental:1 fnn:7 marked:1 consequently:4 towards:4 hard:1 change:1 included:1 typical:1 diff:2 reducing:1 fnns:27 wt:1 adamax:1 player:1 internal:1 support:2 latter:1 assessed:1 bioinformatics:1 violated:1 evaluate:2 tested:1 ex:1 |
6,299 | 6,699 | Learning to Pivot with Adversarial Networks
Gilles Louppe
New York University
[email protected]
Michael Kagan
SLAC National Accelerator Laboratory
[email protected]
Kyle Cranmer
New York University
[email protected]
Abstract
Several techniques for domain adaptation have been proposed to account for
differences in the distribution of the data used for training and testing. The majority
of this work focuses on a binary domain label. Similar problems occur in a scientific
context where there may be a continuous family of plausible data generation
processes associated to the presence of systematic uncertainties. Robust inference
is possible if it is based on a pivot ? a quantity whose distribution does not depend
on the unknown values of the nuisance parameters that parametrize this family
of data generation processes. In this work, we introduce and derive theoretical
results for a training procedure based on adversarial networks for enforcing the
pivotal property (or, equivalently, fairness with respect to continuous attributes) on
a predictive model. The method includes a hyperparameter to control the tradeoff between accuracy and robustness. We demonstrate the effectiveness of this
approach with a toy example and examples from particle physics.
1
Introduction
Machine learning techniques have been used to enhance a number of scientific disciplines, and they
have the potential to transform even more of the scientific process. One of the challenges of applying
machine learning to scientific problems is the need to incorporate systematic uncertainties, which
affect both the robustness of inference and the metrics used to evaluate a particular analysis strategy.
In this work, we focus on supervised learning techniques where systematic uncertainties can be
associated to a data generation process that is not uniquely specified. In other words, the lack of
systematic uncertainties corresponds to the (rare) case that the process that generates training data is
unique, fully specified, and an accurate representative of the real world data. By contrast, a common
situation when systematic uncertainty is present is when the training data are not representative
of the real data. Several techniques for domain adaptation have been developed to create models
that are more robust to this binary type of uncertainty. A more generic situation is that there are
several plausible data generation processes, specified as a family parametrized by continuous nuisance
parameters, as is typically found in scientific domains. In this broader context, statisticians have for
long been working on robust inference techniques based on the concept of a pivot ? a quantity whose
distribution is invariant with the nuisance parameters (see e.g., (Degroot and Schervish, 1975)).
Assuming a probability model p(X, Y, Z), where X are the data, Y are the target labels, and Z are the
nuisance parameters, we consider the problem of learning a predictive model f (X) for Y conditional
on the observed values of X that is robust to uncertainty in the unknown value of Z. We introduce a
flexible learning procedure based on adversarial networks (Goodfellow et al., 2014) for enforcing that
f (X) is a pivot with respect to Z. We derive theoretical results proving that the procedure converges
towards a model that is both optimal and statistically independent of the nuisance parameters (if
that model exists) or for which one can tune a trade-off between accuracy and robustness (e.g., as
driven by a higher level objective). In particular, and to the best of our knowledge, our contribution is
the first solution for imposing pivotal constraints on a predictive model, working regardless of the
31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
Adversary r
Classifier f
Z
?1 (f (X; ?f ); ?r )
f (X; ?f )
...
X
...
?2 (f (X; ?f ); ?r )
P(?1 , ?2 , . . . )
...
p?r (Z|f (X; ?f ))
Lf (?f )
?f
?r
Lr (?f , ?r )
Figure 1: Architecture for the adversarial training of a binary classifier f against a nuisance parameters Z. The
adversary r models the distribution p(z|f (X; ?f ) = s) of the nuisance parameters as observed only through
the output f (X; ?f ) of the classifier. By maximizing the antagonistic objective Lr (?f , ?r ), the classifier f
forces p(z|f (X; ?f ) = s) towards the prior p(z), which happens when f (X; ?f ) is independent of the nuisance
parameter Z and therefore pivotal.
type of the nuisance parameter (discrete or continuous) or of its prior. Finally, we demonstrate the
effectiveness of the approach with a toy example and examples from particle physics.
2
Problem statement
We begin with a family of data generation processes p(X, Y, Z), where X ? X are the data, Y ? Y
are the target labels, and Z ? Z are the nuisance parameters that can be continuous or categorical. Let
us assume that prior to incorporating the effect of uncertainty in Z, our goal is to learn a regression
function f : X ? S with parameters ?f (e.g., a neural network-based probabilistic classifier) that
minimizes a loss Lf (?f ) (e.g., the cross-entropy). In classification, values s ? S = R|Y| correspond
to the classifier scores used for mapping hard predictions y ? Y, while S = Y for regression.
We augment our initial objective so that inference based on f (X; ?f ) will be robust to the value
z ? Z of the nuisance parameter Z ? which remains unknown at test time. A formal way of enforcing
robustness is to require that the distribution of f (X; ?f ) conditional on Z (and possibly Y ) be
invariant with the nuisance parameter Z. Thus, we wish to find a function f such that
p(f (X; ?f ) = s|z) = p(f (X; ?f ) = s|z 0 )
(1)
0
for all z, z ? Z and all values s ? S of f (X; ?f ). In words, we are looking for a predictive function
f which is a pivotal quantity with respect to the nuisance parameters. This implies that f (X; ?f ) and
Z are independent random variables.
As stated in Eqn. 1, the pivotal quantity criterion is imposed with respect to p(X|Z) where Y is
marginalized out. In some situations however (see e.g., Sec. 5.2), class conditional independence of
f (X; ?f ) on the nuisance Z is preferred, which can then be stated as requiring
p(f (X; ?f ) = s|z, y) = p(f (X; ?f ) = s|z 0 , y)
(2)
for one or several specified values y ? Y.
3
Method
Joint training of adversarial networks was first proposed by (Goodfellow et al., 2014) as a way to
build a generative model capable of producing samples from random noise z. More specifically, the
authors pit a generative model g : Rn ? Rp against an adversarial classifier d : Rp ? [0, 1] whose
antagonistic objective is to recognize real data X from generated data g(Z). Both models g and d are
trained simultaneously, in such a way that g learns to produce samples that are difficult to identify by
d, while d incrementally adapts to changes in g. At the equilibrium, g models a distribution whose
samples can be identified by d only by chance. That is, assuming enough capacity in d and g, the
distribution of g(Z) eventually converges towards the real distribution of X.
2
Algorithm 1 Adversarial training of a classifier f against an adversary r.
? ?
Inputs: training data {xi , yi , zi }N
i=1 ; Outputs: ?f , ?r .
1: for t = 1 to T do
2:
for k = 1 to K do
3:
Sample minibatch {xm , zm , sm = f (xm ; ?f )}M
m=1 of size M ;
4:
With ?f fixed, update r by ascending its stochastic gradient ??r E(?f , ?r ) :=
??r
M
X
log p?r (zm |sm );
m=1
5:
6:
7:
end for
Sample minibatch {xm , ym , zm , sm = f (xm ; ?f )}M
m=1 of size M ;
With ?r fixed, update f by descending its stochastic gradient ??f E(?f , ?r ) :=
??f
M
X
? log p?f (ym |xm ) + log p?r (zm |sm ) ,
m=1
where p?f (ym |xm ) denotes 1(ym = 0)(1 ? sm ) + 1(ym = 1)sm ;
8: end for
In this work, we repurpose adversarial networks as a means to constrain the predictive model f
in order to satisfy Eqn. 1. As illustrated in Fig. 1, we pit f against an adversarial model r :=
p?r (z|f (X; ?f ) = s) with parameters ?r and associated loss Lr (?f , ?r ). This model takes as input
realizations s of f (X; ?f ) and produces as output a function modeling the posterior probability
density p?r (z|f (X; ?f ) = s). Intuitively, if p(f (X; ?f ) = s|z) varies with z, then the corresponding
correlation can be captured by r. By contrast, if p(f (X; ?f ) = s|z) is invariant with z, as we require,
then r should perform poorly and be close to random guessing. Training f such that it additionally
minimizes the performance of r therefore acts as a regularization towards Eqn. 1.
If Z takes discrete values, then p?r can be represented as a probabilistic classifier R ? R|Z| whose
j th output (for j = 1, . . . , |Z|) is the estimated probability mass p?r (zj |f (X; ?f ) = s). Similarly, if
Z takes continuous values, then we can model the posterior probability density p(z|f (X; ?f ) = s)
with a sufficiently flexible parametric family of distributions P(?1 , ?2 , . . . ), where the parameters ?j
depend on f (X, ?f ) and ?r . The adversary r may take any form, i.e. it does not need to be a neural
network, as long as it exposes a differentiable function p?r (z|f (X; ?f ) = s) of sufficient capacity
to represent the true distribution. Fig. 1 illustrates a concrete example where p?r (z|f (X; ?f ) = s)
is a mixture of gaussians, as modeled with a mixture density network (Bishop, 1994)). The j th
output corresponds to the estimated value of the corresponding parameter ?j of that distribution (e.g.,
the mean, variance and mixing coefficients of its components). The estimated probability density
p?r (z|f (X; ?f ) = s) can then be evaluated for any z ? Z and any score s ? S.
As with generative adversarial networks, we propose to train f and r simultaneously, which we carry
out by considering the value function
E(?f , ?r ) = Lf (?f ) ? Lr (?f , ?r )
(3)
that we optimize by finding the minimax solution
??f , ??r = arg min max E(?f , ?r ).
(4)
?f
?r
Without loss of generality, the adversarial training procedure to obtain (??f , ??r ) is formally presented
in Algorithm 1 in the case of a binary classifier f : Rp ? [0, 1] modeling p(Y = 1|X). For reasons
further explained in Sec. 4, Lf and Lr are respectively set to the expected value of the negative
log-likelihood of Y |X under f and of Z|f (X; ?f ) under r:
Lf (?f ) = Ex?X Ey?Y |x [? log p?f (y|x)],
(5)
Lr (?f , ?r ) = Es?f (X;?f ) Ez?Z|s [? log p?r (z|s)].
(6)
The optimization algorithm consists in using stochastic gradient descent alternatively for solving
Eqn. 4. Finally, in the case of a class conditional pivot, the settings are the same, except that the
adversarial term Lr (?f , ?r ) is restricted to Y = y.
3
4
Theoretical results
In this section, we show that in the setting of Algorithm 1 where Lf and Lr are respectively set
to expected value of the negative log-likelihood of Y |X under f and of Z|f (X; ?f ) under r, the
minimax solution of Eqn. 4 corresponds to a classifier f which is a pivotal quantity.
In this setting, the nuisance parameter Z is considered as a random variable of prior p(Z), and
our goal is to find a function f (?; ?f ) such that f (X; ?f ) and Z are independent random variables.
Importantly, classification of Y with respect to X is considered in the context where Z is marginalized
out, which means that the classifier minimizing Lf is optimal with respect to Y |X, but not necessarily
with Y |X, Z. Results hold for a nuisance parameters Z taking either categorical or continuous values.
By abuse of notation, H(Z) denotes the differential entropy in this latter case. Finally, the proposition
below is derived in a non-parametric setting, by assuming that both f and r have enough capacity.
Proposition 1. If there exists a minimax solution (??f , ??r ) for Eqn. 4 such that E(??f , ??r ) =
H(Y |X) ? H(Z), then f (?; ??f ) is both an optimal classifier and a pivotal quantity.
Proof. For fixed ?f , the adversary r is optimal at
?
??r = arg max E(?f , ?r ) = arg min Lr (?f , ?r ),
?r
?r
(7)
in which case p ?? (z|f (X; ?f ) = s) = p(z|f (X; ?f ) = s) for all z and all s, and Lr reduces to the
?r
expected entropy Es?f (X;?f ) [H(Z|f (X; ?f ) = s)] of the conditional distribution of the nuisance
parameters. This expectation corresponds to the conditional entropy of the random variables Z and
f (X; ?f ) and can be written as H(Z|f (X; ?f )). Accordingly, the value function E can be restated
as a function depending on ?f only:
E 0 (?f ) = Lf (?f ) ? H(Z|f (X; ?f )).
(8)
In particular, we have the lower bound
H(Y |X) ? H(Z) ? Lf (?f ) ? H(Z|f (X; ?f ))
where the equality holds at ??f = arg min? E 0 (?f ) when:
(9)
f
? ??f minimizes the negative log-likelihood of Y |X under f , which happens when ??f are the
parameters of an optimal classifier. In this case, Lf reduces to its minimum value H(Y |X).
? ??f maximizes the conditional entropy H(Z|f (X; ?f )), since H(Z|f (X; ?)) ? H(Z) from
the properties of entropy. Note that this latter inequality holds for both the discrete and the
differential definitions of entropy.
By assumption, the lower bound is active, thus we have H(Z|f (X; ?f )) = H(Z) because of the
second condition, which happens exactly when Z and f (X; ?f ) are independent variables. In other
words, the optimal classifier f (?; ??f ) is also a pivotal quantity.
Proposition 1 suggests that if at each step of Algorithm 1 the adversary r is allowed to reach its
optimum given f (e.g., by setting K sufficiently high) and if f is updated to improve Lf (?f ) ?
H(Z|f (X; ?f )) with sufficiently small steps, then f should converge to a classifier that is both
optimal and pivotal, provided such a classifier exists. Therefore, the adversarial term Lr can be
regarded as a way to select among the class of all optimal classifiers a function f that is also pivotal.
Despite the former theoretical characterization of the minimax solution of Eqn. 4, let us note that
formal guarantees of convergence towards that solution by Algorithm 1 in the case where a finite
number K of steps is taken for r remains to be proven.
In practice, the assumption of existence of an optimal and pivotal classifier may not hold because the
nuisance parameter directly shapes the decision boundary. In this case, the lower bound
H(Y |X) ? H(Z) < Lf (?f ) ? H(Z|f (X; ?f ))
(10)
is strict: f can either be an optimal classifier or a pivotal quantity, but not both simultaneously. In
this situation, it is natural to rewrite the value function E as
E? (?f , ?r ) = Lf (?f ) ? ?Lr (?f , ?r ),
4
(11)
3.0
p(f(X)|Z = ? ?)
p(f(X)|Z = 0)
p(f(X)|Z = + ?)
3.5
3.0
p(f(X))
2.5
2.0
1.5
1.0
0.5
0.00.0
0.2
0.4
f(X)
0.6
1.0
0.9
0.8
2.0
Z= +?
0.7
1.5
0.6
1.0
Z=0
0.5
0.5
0.4
0.0
Z= ??
0.3
0.5
0.2
1.01.0 0.5 0.0 0.5 1.0 1.5 2.0 0.1
2.5
0.8
1.0
4.0
?0
?1 |Z = z
3.5
3.0
2.5
p(f(X))
4.0
3.0
p(f(X)|Z = ? ?)
p(f(X)|Z = 0)
p(f(X)|Z = + ?)
2.5
2.0
?0
?1 |Z = z
0.84
Z=?
1.5
0.72
0.60
2.0
1.0
1.5
0.5
1.0
0.0
0.5
0.5
0.00.0
1.01.0 0.5 0.0 0.5 1.0 1.5 2.0 0.12
0.2
0.4
f(X)
0.6
0.8
1.0
Z=0
Z= ??
0.48
0.36
0.24
Figure 2: Toy example. (Left) Conditional probability densities of the decision scores at Z = ??, 0, ? without
adversarial training. The resulting densities are dependent on the continuous parameter Z, indicating that f is
not pivotal. (Middle left) The associated decision surface, highlighting the fact that samples are easier to classify
for values of Z above ?, hence explaining the dependency. (Middle right) Conditional probability densities of
the decision scores at Z = ??, 0, ? when f is built with adversarial training. The resulting densities are now
almost identical to each other, indicating only a small dependency on Z. (Right) The associated decision surface,
illustrating how adversarial training bends the decision function vertically to erase the dependency on Z.
where ? ? 0 is a hyper-parameter controlling the trade-off between the performance of f and its
independence with respect to the nuisance parameter. Setting ? to a large value will preferably
enforces f to be pivotal while setting ? close to 0 will rather constraint f to be optimal. When the
lower bound is strict, let us note however that there may exist distinct but equally good solutions ?f , ?r
minimizing Eqn. 11. In this zero-sum game, an increase in accuracy would exactly be compensated
by a decrease in pivotality and vice-versa. How to best navigate this Pareto frontier to maximize a
higher-level objective remains a question open for future works.
Interestingly, let us finally emphasize that our results hold using only the (1D) output s of f (?; ?f ) as
input to the adversary. We could similarly enforce an intermediate representation of the data to be
pivotal, e.g. as in (Ganin and Lempitsky, 2014), but this is not necessary.
5
Experiments
In this section, we empirically demonstrate the effectiveness of the approach with a toy example
and examples from particle physics. Notably, there are no other other approaches to compare to in
the case of continuous nuisance parameters, as further explained in Sec. 6. In the case of binary
parameters, we do not expect results to be much different from previous works. The source code to
reproduce the experiments is available online 1 .
5.1
A toy example with a continous nuisance parameter
As a guiding toy example, let us consider the binary classification of 2D data drawn from multivariate
gaussians with equal priors, such that
1
?0.5
x ? N (0, 0),
when Y = 0,
(12)
?0.5
1
1 0
x|Z = z ? N (1, 1 + z),
when Y = 1.
(13)
0 1
The continuous nuisance parameter Z here represents our uncertainty about the location of the mean
of the second gaussian. Our goal is to build a classifier f (?; ?f ) for predicting Y given X, but such
that the probability distribution of f (X; ?f ) is invariant with respect to the nuisance parameter Z.
Assuming a gaussian prior z ? N (0, 1), we generate data {xi , yi , zi }N
i=1 , from which we train a
neural network f minimizing Lf (?f ) without considering its adversary r. The network architecture
comprises 2 dense hidden layers of 20 nodes respectively with tanh and ReLU activations, followed
by a dense output layer with a single node with a sigmoid activation. As shown in Fig. 2, the resulting
classifier is not pivotal, as the conditional probability densities of its decision scores f (X; ?f ) show
1
https://github.com/glouppe/paper-learning-to-pivot
5
7
6
67.5
68.0
68.5
69.0
69.5
70.0
70.5
? = 0|Z = 0
?=0
?=1
? = 10
? = 500
5
1.42
1.41
1.40
1.39
1.38
1.37
1.36
4
AMS
Lf
Lr
Lf ? ?Lr
8
0.70
0.65
0.60
0.55
0.50
0.45
3
2
1
0
0
50
100
150
1
0.0
200
T
Figure 3: Toy example. Training curves for Lf (?f ),
Lr (?f , ?r ) and Lf (?f ) ? ?Lr (?f , ?r ). Initialized
with a pre-trained classifier f , adversarial training
was performed for 200 iterations, mini-batches of
size M = 128, K = 500 and ? = 50.
0.2
0.4
0.6
threshold on f(X)
0.8
1.0
Figure 4: Physics example. Approximate median
significance as a function of the decision threshold
on the output of f . At ? = 10, trading accuracy
for independence to pileup results in a net benefit in
terms of statistical significance.
large discrepancies between values z of the nuisance parameters. While not shown here, a classifier
trained only from data generated at the nominal value Z = 0 would also not be pivotal.
Let us now consider the joint training of f against an adversary r implemented as a mixture density
network modeling Z|f (X; ?f ) as a mixture of five gaussians. The network architecture of r comprises
2 dense hidden layers of 20 nodes with ReLU activations, followed by an output layer of 15 nodes
corresponding to the means, standard deviations and mixture coefficients of the gaussians. Output
nodes for the mean values come with linear activations, output nodes for the standard deviations
with exponential activations to ensure positivity, while output nodes for the mixture coefficients
implement the softmax function to ensure positivity and normalization. When running Algorithm 1
as initialized with the classifier f obtained previously, adversarial training effectively reshapes the
decision function so it that becomes almost independent on the nuisance parameter, as shown in
Fig. 2. The conditional probability densities of the decision scores f (X; ?f ) are now very similar to
each other, indicating only a residual dependency on the nuisance, as theoretically expected. The
dynamics of adversarial training is illustrated in Fig. 3, where the losses Lf , Lr and Lf ? ?Lr are
evaluated after each iteration. In the first iterations, we observe that the global objective Lf ? ?Lr
is minimized by making the classifier less accurate, hence the corresponding increase of Lf , but
which results in a classifier that is more pivotal, hence the associated increase of Lr and the total
net benefit. As learning goes, minimizing E requires making predictions that are more accurate,
hence decreasing Lf , or that are even less dependent on Z, hence shaping p?r towards the prior p(Z).
Indeed, Lf eventually starts decreasing, while remaining bounded from below by min?f Lf (?f ) as
approximated by the dashed line in the first
? plot. Similarly, Lr tends towards the differential entropy
H(Z) of the prior (where H(Z) = log(? 2?e) = 1.419 in the case of a standard normal), as shown
by the dashed line in the second plot. Finally, let us note that the ideal situation of a classifier that
is both optimal and pivotal is unreachable for this problem, as shown in the third plot by the offset
between Lf ? ?Lr and the dashed line approximating H(Y |X) ? ?H(Z).
5.2
High energy physics examples
Binary Case Experiments at high energy colliders like the LHC (Evans and Bryant, 2008) are
searching for evidence of new particles beyond those described by the Standard Model (SM) of
particle physics. A wide array of theories predict the existence of new massive particles that would
decay to known particles in the SM such as the W boson. The W boson is unstable and can decay to
two quarks, each of which produce collimated sprays of particles known as jets. If the exotic particle is
heavy, then the W boson will be moving very fast, and relativistic effects will cause the two jets from
its decay to merge into a single ?W -jet?. These W -jets have a rich internal substructure. However,
jets are also produced ubiquitously at high energy colliders through more mundane processes in the
6
SM, which leads to a challenging classification problem that is beset with a number of sources of
systematic uncertainty. The classification challenge used here is common in jet substructure studies
(see e.g. (CMS Collaboration, 2014; ATLAS Collaboration, 2015, 2014)): we aim to distinguish
normal jets produced copiously at the LHC (Y = 0) and from W -jets (Y = 1) potentially coming
from an exotic process. We reuse the datasets used in (Baldi et al., 2016a).
Challenging in its own right, this classification problem is made all the more difficult by the presence
of pileup, or multiple proton-proton interactions occurring simultaneously with the primary interaction.
These pileup interactions produce additional particles that can contribute significant energies to jets
unrelated to the underlying discriminating information. The number of pileup interactions can vary
with the running conditions of the collider, and we want the classifier to be robust to these conditions.
Taking some liberty, we consider an extreme case with a categorical nuisance parameter, where
Z = 0 corresponds to events without pileup and Z = 1 corresponds to events with pileup, for which
there are an average of 50 independent pileup interactions overlaid.
We do not expect that we will be able to find a function f that simultaneously minimizes the
classification loss Lf and is pivotal. Thus, we need to optimize the hyper-parameter ? of Eqn. 11 with
respect to a higher-level objective. In this case, the natural higher-level context is a hypothesis test of
a null hypothesis with no Y = 1 events against an alternate hypothesis that is a mixture of Y = 0 and
Y = 1 events. In the absence of systematic uncertainties, optimizing Lf simultaneously optimizes
the power of a classical hypothesis test in the Neyman-Pearson sense. When we include systematic
uncertainties we need to balance the classification performance against the robustness to uncertainty
in Z. Since we are still performing a hypothesis test against the null, we only wish to impose the
pivotal property on Y = 0 events. To this end, we use as a higher level objective the Approximate
Median Significance (AMS), which is a natural generalization of the power of a hypothesis test when
systematic uncertainties are taken into account (see Eqn. 20 of Adam-Bourdarios et al. (2014)).
For several values of ?, we train a classifier using Algorithm 1 but consider the adversarial term
Lr conditioned on Y = 0 only, as outlined in Sec. 2. The architecture of f comprises 3 hidden
layers of 64 nodes respectively with tanh, ReLU and ReLU activations, and is terminated by a single
final output node with a sigmoid activation. The architecture of r is the same, but uses only ReLU
activations in its hidden nodes. As in the previous example, adversarial training is initialized with
f pre-trained. Experiments are performed on a subset of 150000 samples for training while AMS
is evaluated on an independent test set of 5000000 samples. Both training and testing samples are
weighted such that the null hypothesis corresponded to 1000 of Y = 0 events and the alternate
hypothesis included an additional 100 Y = 1 events prior to any thresholding on f . This allows us
to probe the efficacy of the method proposed here in a representative background-dominated high
energy physics environment. Results reported below are averages over 5 runs.
As Fig. 4 illustrates, without adversarial training (at ? = 0|Z = 0 when building a classifier at the
nominal value Z = 0 only, or at ? = 0 when building a classifier on data sampled from p(X, Y, Z)),
the AMS peaks at 7. By contrast, as the pivotal constraint is made stronger (for ? > 0) the AMS
peak moves higher, with a maximum value around 7.8 for ? = 10. Trading classification accuracy
for robustness to pileup thereby results in a net benefit in terms of the power of the hypothesis
test. Setting ? too high however (e.g. ? = 500) results in a decrease of the maximum AMS, by
focusing the capacity of f too strongly on independence with Z, at the expense of accuracy. In effect,
optimizing ? yields a principled and effective approach to control the trade-off between accuracy and
robustness that ultimately maximizes the power of the enveloping hypothesis test.
Continous Case Recently, an independent group has used our approach to learn jet classifiers that
are independent of the jet mass (Shimmin et al., 2017), which is a continuous attribute. The results of
their studies show that the adversarial training strategy works very well for real-world problems with
continuous attributes, thus enhancing the sensitivity of searches for new physics at the LHC.
6
Related work
Learning to pivot can be related to the problem of domain adaptation (Blitzer et al., 2006; Pan et al.,
2011; Gopalan et al., 2011; Gong et al., 2013; Baktashmotlagh et al., 2013; Ajakan et al., 2014;
Ganin and Lempitsky, 2014), where the goal is often stated as trying to learn a domain-invariant
representation of the data. Likewise, our method also relates to the problem of enforcing fairness
7
in classification (Kamishima et al., 2012; Zemel et al., 2013; Feldman et al., 2015; Edwards and
Storkey, 2015; Zafar et al., 2015; Louizos et al., 2015), which is stated as learning a classifier that is
independent of some chosen attribute such as gender, color or age. For both families of methods, the
problem can equivalently be stated as learning a classifier which is a pivotal quantity with respect
to either the domain or the selected feature. As an example, unsupervised domain adaptation with
labeled data from a source domain and unlabeled data from a target domain can be recast as learning
a predictive model f (i.e., trained to minimize Lf evaluated on labeled source data only) that is also a
pivot with respect to the domain Z (i.e., trained to maximize Lr evaluated on both source and target
data). In this context, (Ganin and Lempitsky, 2014; Edwards and Storkey, 2015) are certainly among
the closest to our work, in which domain invariance and fairness are enforced through an adversarial
minimax setup composed of a classifier and an adversarial discriminator. Following this line of work,
our method can be regarded as a unified generalization that also supports a continuously parametrized
family of domains or as enforcing fairness over continuous attributes.
Most related work is based on the strong and limiting assumption that Z is a binary random variable
(e.g., Z = 0 for the source domain, and Z = 1 for the target domain). In particular, (Pan et al., 2011;
Gong et al., 2013; Baktashmotlagh et al., 2013; Zemel et al., 2013; Ganin and Lempitsky, 2014;
Ajakan et al., 2014; Edwards and Storkey, 2015; Louizos et al., 2015) are all based on the minimization
of some form of divergence between the two distributions of f (X)|Z = 0 and f (X)|Z = 1. For this
reason, these works cannot directly be generalized to non-binary or continuous nuisance parameters,
both from a practical and theoretical point of view. Notably, Kamishima et al. (2012) enforces
fairness through a prejudice regularization term based on empirical estimates of p(f (X)|Z). While
this approach is in principle sufficient for handling non-binary nuisance parameters Z, it requires
accurate empirical estimates of p(f (X)|Z = z) for all values z, which quickly becomes impractical
as the cardinality of Z increases. By contrast, our approach models the conditional dependence
through an adversarial network, which allows for generalization without necessarily requiring an
exponentially growing number of training examples.
A common approach to account for systematic uncertainties in a scientific context (e.g. in high energy
physics) is to take as fixed a classifier f built from training data for a nominal value z0 of the nuisance
parameter, and then propagate uncertainty by estimating p(f (x)|z) with a parametrized calibration
procedure. Clearly, this classifier is however not optimal for z 6= z0 . To overcome this issue, the
classifier f is sometimes built instead on a mixture of training data generated from several plausible
values z0 , z1 , . . . of the nuisance parameter. While this certainly improves classification performance
with respect to the marginal model p(X, Y ), there is no reason to expect the resulting classifier to
be pivotal, as shown previously in Sec. 5.1. As an alternative, parametrized classifiers (Cranmer
et al., 2015; Baldi et al., 2016b) directly take (nuisance) parameters as additional input variables,
hence ultimately providing the most statistically powerful approach for incorporating the effect
of systematics on the underlying classification task. In practice, parametrized classifiers are also
computationally expensive to build and evaluate. In particular, calibrating their decision function,
i.e. approximating p(f (x, z)|y, z) as a continuous function of z, remains an open challenge. By
contrast, constraining f to be pivotal yields a classifier that can be directly used in a wider range of
applications, since the dependence on the nuisance parameter Z has already been eliminated.
7
Conclusions
In this work, we proposed a flexible learning procedure for building a predictive model that is
independent of continuous or categorical nuisance parameters by jointly training two neural networks
in an adversarial fashion. From a theoretical perspective, we motivated the proposed algorithm by
showing that the minimax value of its value function corresponds to a predictive model that is both
optimal and pivotal (if that models exists) or for which one can tune the trade-off between power and
robustness. From an empirical point of view, we confirmed the effectiveness of our method on a toy
example and a particle physics example.
In terms of applications, our solution can be used in any situation where the training data may not be
representative of the real data the predictive model will be applied to in practice. In the scientific
context, the presence of systematic uncertainty can be incorporated by considering a family of data
generation processes, and it would be worth revisiting those scientific problems that utilize machine
learning in light of this technique. The approach also extends to cases where independence of the
predictive model with respect to observed random variables is desired, as in fairness for classification.
8
Acknowledgements
We would like to thank the authors of (Baldi et al., 2016a) for sharing the data used in their studies.
KC and GL are both supported through NSF ACI-1450310, additionally KC is supported through
PHY-1505463 and PHY-1205376. MK is supported by the US Department of Energy (DOE) under
grant DE-AC02-76SF00515 and by the SLAC Panofsky Fellowship.
References
Adam-Bourdarios, C., Cowan, G., Germain, C., Guyon, I., K?gl, B., and Rousseau, D. (2014). The
higgs boson machine learning challenge. In NIPS 2014 Workshop on High-energy Physics and
Machine Learning, volume 42, page 37.
Ajakan, H., Germain, P., Larochelle, H., Laviolette, F., and Marchand, M. (2014). Domain-adversarial
neural networks. arXiv preprint arXiv:1412.4446.
ATLAS Collaboration (2014). Performance of Boosted W Boson Identification with the ATLAS
Detector. Technical Report ATL-PHYS-PUB-2014-004, CERN, Geneva.
ATLAS
Collaboration (2015). Identification of boosted, hadronically-decaying W and Z bosons in
?
s = 13 TeV Monte Carlo Simulations for ATLAS. Technical Report ATL-PHYS-PUB-2015-033,
CERN, Geneva.
Baktashmotlagh, M., Harandi, M., Lovell, B., and Salzmann, M. (2013). Unsupervised domain
adaptation by domain invariant projection. In Proceedings of the IEEE International Conference
on Computer Vision, pages 769?776.
Baldi, P., Bauer, K., Eng, C., Sadowski, P., and Whiteson, D. (2016a). Jet substructure classification
in high-energy physics with deep neural networks. Physical Review D, 93(9):094034.
Baldi, P., Cranmer, K., Faucett, T., Sadowski, P., and Whiteson, D. (2016b). Parameterized neural
networks for high-energy physics. Eur. Phys. J., C76(5):235.
Bishop, C. M. (1994). Mixture density networks.
Blitzer, J., McDonald, R., and Pereira, F. (2006). Domain adaptation with structural correspondence
learning. In Proceedings of the 2006 conference on empirical methods in natural language
processing, pages 120?128. Association for Computational Linguistics.
CMS Collaboration (2014). Identification techniques for highly boosted W bosons that decay into
hadrons. JHEP, 12:017.
Cranmer, K., Pavez, J., and Louppe, G. (2015). Approximating likelihood ratios with calibrated
discriminative classifiers. arXiv preprint arXiv:1506.02169.
Degroot, M. H. and Schervish, M. J. (1975). Probability and statistics. 1st edition.
Edwards, H. and Storkey, A. J. (2015). Censoring representations with an adversary. arXiv preprint
arXiv:1511.05897.
Evans, L. and Bryant, P. (2008). LHC Machine. JINST, 3:S08001.
Feldman, M., Friedler, S. A., Moeller, J., Scheidegger, C., and Venkatasubramanian, S. (2015).
Certifying and removing disparate impact. In Proceedings of the 21th ACM SIGKDD International
Conference on Knowledge Discovery and Data Mining, pages 259?268. ACM.
Ganin, Y. and Lempitsky, V. (2014). Unsupervised Domain Adaptation by Backpropagation. arXiv
preprint arXiv:1409.7495.
Gong, B., Grauman, K., and Sha, F. (2013). Connecting the dots with landmarks: Discriminatively
learning domain-invariant features for unsupervised domain adaptation. In Proceedings of The
30th International Conference on Machine Learning, pages 222?230.
9
Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., and
Bengio, Y. (2014). Generative adversarial nets. In Advances in Neural Information Processing
Systems, pages 2672?2680.
Gopalan, R., Li, R., and Chellappa, R. (2011). Domain adaptation for object recognition: An
unsupervised approach. In Computer Vision (ICCV), 2011 IEEE International Conference on,
pages 999?1006. IEEE.
Kamishima, T., Akaho, S., Asoh, H., and Sakuma, J. (2012). Fairness-aware classifier with prejudice
remover regularizer. Machine Learning and Knowledge Discovery in Databases, pages 35?50.
Louizos, C., Swersky, K., Li, Y., Welling, M., and Zemel, R. (2015). The variational fair autoencoder.
arXiv preprint arXiv:1511.00830.
Pan, S. J., Tsang, I. W., Kwok, J. T., and Yang, Q. (2011). Domain adaptation via transfer component
analysis. Neural Networks, IEEE Transactions on, 22(2):199?210.
Shimmin, C., Sadowski, P., Baldi, P., Weik, E., Whiteson, D., Goul, E., and S?gaard, A. (2017).
Decorrelated Jet Substructure Tagging using Adversarial Neural Networks.
Zafar, M. B., Valera, I., Rodriguez, M. G., and Gummadi, K. P. (2015). Fairness constraints: A
mechanism for fair classification. arXiv preprint arXiv:1507.05259.
Zemel, R. S., Wu, Y., Swersky, K., Pitassi, T., and Dwork, C. (2013). Learning fair representations.
ICML (3), 28:325?333.
10
| 6699 |@word illustrating:1 middle:2 stronger:1 open:2 simulation:1 propagate:1 eng:1 thereby:1 carry:1 venkatasubramanian:1 phy:2 initial:1 score:6 efficacy:1 pub:2 salzmann:1 interestingly:1 com:1 activation:8 written:1 evans:2 shape:1 plot:3 atlas:5 update:2 generative:4 selected:1 accordingly:1 lr:24 characterization:1 node:10 location:1 contribute:1 five:1 differential:3 spray:1 consists:1 baldi:6 introduce:2 theoretically:1 tagging:1 indeed:1 notably:2 expected:4 growing:1 decreasing:2 considering:3 cardinality:1 erase:1 begin:1 provided:1 notation:1 becomes:2 maximizes:2 mass:2 bounded:1 null:3 exotic:2 underlying:2 cm:2 unrelated:1 minimizes:4 developed:1 unified:1 finding:1 impractical:1 guarantee:1 preferably:1 act:1 bryant:2 exactly:2 grauman:1 classifier:45 control:2 grant:1 producing:1 vertically:1 tends:1 despite:1 abuse:1 merge:1 ajakan:3 suggests:1 challenging:2 pit:2 repurpose:1 range:1 statistically:2 unique:1 practical:1 enforces:2 testing:2 practice:3 implement:1 lf:29 backpropagation:1 procedure:6 empirical:4 projection:1 word:3 pre:2 cannot:1 close:2 unlabeled:1 bend:1 context:7 applying:1 descending:1 optimize:2 imposed:1 compensated:1 maximizing:1 go:1 regardless:1 restated:1 pouget:1 array:1 importantly:1 regarded:2 proving:1 searching:1 atl:2 antagonistic:2 updated:1 limiting:1 target:5 controlling:1 nominal:3 massive:1 us:1 goodfellow:3 hypothesis:10 storkey:4 approximated:1 expensive:1 recognition:1 quark:1 relativistic:1 labeled:2 database:1 observed:3 louppe:3 preprint:6 tsang:1 revisiting:1 trade:4 decrease:2 principled:1 environment:1 warde:1 dynamic:1 ultimately:2 trained:6 depend:2 solving:1 rewrite:1 predictive:10 joint:2 represented:1 regularizer:1 train:3 distinct:1 fast:1 effective:1 chellappa:1 monte:1 zemel:4 corresponded:1 hyper:2 pearson:1 whose:5 stanford:1 plausible:3 remover:1 statistic:1 transform:1 jointly:1 final:1 online:1 differentiable:1 net:4 propose:1 interaction:5 coming:1 adaptation:10 zm:4 realization:1 moeller:1 mixing:1 poorly:1 adapts:1 convergence:1 optimum:1 produce:4 adam:2 converges:2 object:1 wider:1 derive:2 depending:1 ganin:5 gong:3 boson:7 blitzer:2 strong:1 edward:4 implemented:1 implies:1 trading:2 come:1 larochelle:1 collider:3 liberty:1 attribute:5 stochastic:3 tev:1 require:2 generalization:3 proposition:3 rousseau:1 frontier:1 hold:5 sufficiently:3 considered:2 around:1 normal:2 equilibrium:1 mapping:1 predict:1 overlaid:1 vary:1 friedler:1 label:3 tanh:2 expose:1 collimated:1 vice:1 create:1 weighted:1 minimization:1 clearly:1 gaussian:2 aim:1 rather:1 asoh:1 boosted:3 broader:1 derived:1 focus:2 likelihood:4 contrast:5 adversarial:30 sigkdd:1 am:6 sense:1 inference:4 dependent:2 typically:1 hidden:4 kc:2 reproduce:1 issue:1 arg:4 flexible:3 classification:15 augment:1 unreachable:1 among:2 softmax:1 marginal:1 equal:1 aware:1 beach:1 eliminated:1 identical:1 represents:1 unsupervised:5 fairness:8 icml:1 future:1 discrepancy:1 minimized:1 report:2 mirza:1 jhep:1 composed:1 simultaneously:6 national:1 recognize:1 lhc:4 divergence:1 statistician:1 highly:1 mining:1 dwork:1 certainly:2 mixture:9 extreme:1 farley:1 light:1 accurate:4 capable:1 necessary:1 initialized:3 desired:1 theoretical:6 mk:1 classify:1 modeling:3 deviation:2 subset:1 rare:1 too:2 reported:1 dependency:4 varies:1 calibrated:1 eur:1 st:2 density:12 peak:2 sensitivity:1 discriminating:1 international:4 systematic:11 off:4 physic:13 probabilistic:2 discipline:1 michael:1 ym:5 continuously:1 concrete:1 enhance:1 quickly:1 connecting:1 possibly:1 positivity:2 toy:8 li:2 account:3 potential:1 de:1 sec:5 includes:1 coefficient:3 satisfy:1 performed:2 view:2 higgs:1 slac:3 start:1 decaying:1 substructure:4 contribution:1 minimize:1 accuracy:7 variance:1 likewise:1 correspond:1 identify:1 yield:2 cern:2 identification:3 produced:2 carlo:1 confirmed:1 worth:1 detector:1 reach:1 phys:3 sharing:1 decorrelated:1 definition:1 against:8 energy:10 bourdarios:2 associated:6 proof:1 sampled:1 knowledge:3 color:1 improves:1 shaping:1 focusing:1 higher:6 supervised:1 evaluated:5 strongly:1 generality:1 correlation:1 working:2 eqn:10 lack:1 incrementally:1 minibatch:2 rodriguez:1 scientific:8 building:3 usa:1 calibrating:1 concept:1 effect:4 requiring:2 true:1 former:1 regularization:2 equality:1 hence:6 laboratory:1 illustrated:2 game:1 nuisance:34 uniquely:1 criterion:1 generalized:1 trying:1 lovell:1 demonstrate:3 mcdonald:1 variational:1 kyle:2 recently:1 common:3 sigmoid:2 empirically:1 physical:1 exponentially:1 volume:1 association:1 louizos:3 significant:1 versa:1 imposing:1 feldman:2 outlined:1 similarly:3 particle:11 akaho:1 language:1 dot:1 moving:1 calibration:1 surface:2 pitassi:1 posterior:2 multivariate:1 own:1 closest:1 perspective:1 optimizing:2 optimizes:1 driven:1 inequality:1 binary:10 ubiquitously:1 yi:2 captured:1 minimum:1 additional:3 impose:1 hadron:1 ey:1 converge:1 maximize:2 dashed:3 relates:1 multiple:1 reduces:2 technical:2 jet:13 cross:1 long:3 equally:1 estimating:1 gummadi:1 impact:1 prediction:2 regression:2 enhancing:1 metric:1 expectation:1 vision:2 arxiv:12 iteration:3 represent:1 normalization:1 sometimes:1 background:1 want:1 fellowship:1 scheidegger:1 median:2 source:6 strict:2 degroot:2 enveloping:1 cowan:1 effectiveness:4 structural:1 yang:1 presence:3 ideal:1 intermediate:1 enough:2 constraining:1 bengio:1 beset:1 affect:1 independence:5 zi:2 relu:5 architecture:5 identified:1 ac02:1 tradeoff:1 pivot:8 motivated:1 reuse:1 york:2 cause:1 deep:1 gopalan:2 tune:2 generate:1 http:1 exist:1 zj:1 nsf:1 estimated:3 discrete:3 hyperparameter:1 group:1 threshold:2 drawn:1 utilize:1 schervish:2 sum:1 enforced:1 run:1 parameterized:1 uncertainty:17 powerful:1 sakuma:1 extends:1 family:8 almost:2 guyon:1 swersky:2 wu:1 decision:11 mundane:1 bound:4 layer:5 followed:2 distinguish:1 courville:1 correspondence:1 marchand:1 occur:1 constraint:4 constrain:1 certifying:1 dominated:1 generates:1 min:4 performing:1 department:1 alternate:2 pan:3 making:2 happens:3 intuitively:1 iccv:1 invariant:7 explained:2 restricted:1 handling:1 taken:2 computationally:1 neyman:1 remains:4 previously:2 eventually:2 mechanism:1 ascending:1 end:3 parametrize:1 gaussians:4 available:1 systematics:1 probe:1 observe:1 kwok:1 generic:1 enforce:1 batch:1 robustness:8 alternative:1 rp:3 existence:2 denotes:2 running:2 ensure:2 remaining:1 include:1 linguistics:1 marginalized:2 laviolette:1 build:3 approximating:3 classical:1 objective:8 move:1 question:1 quantity:9 kagan:1 already:1 strategy:2 parametric:2 primary:1 dependence:2 sha:1 guessing:1 gradient:3 thank:1 capacity:4 majority:1 parametrized:5 landmark:1 unstable:1 reason:3 enforcing:5 ozair:1 assuming:4 code:1 modeled:1 mini:1 providing:1 minimizing:4 balance:1 ratio:1 equivalently:2 difficult:2 setup:1 statement:1 potentially:1 expense:1 stated:5 negative:3 disparate:1 unknown:3 perform:1 gilles:1 datasets:1 sm:9 finite:1 descent:1 situation:6 looking:1 incorporated:1 rn:1 germain:2 specified:4 continous:2 discriminator:1 z1:1 proton:2 nip:2 beyond:1 adversary:10 able:1 below:3 xm:6 challenge:4 built:3 max:2 recast:1 power:5 event:7 natural:4 force:1 predicting:1 valera:1 residual:1 minimax:6 improve:1 github:1 categorical:4 autoencoder:1 prior:9 review:1 acknowledgement:1 discovery:2 fully:1 loss:5 expect:3 discriminatively:1 accelerator:1 generation:6 proven:1 age:1 sufficient:2 thresholding:1 principle:1 pareto:1 heavy:1 collaboration:5 censoring:1 gl:2 supported:3 formal:2 explaining:1 wide:1 taking:2 cranmer:5 benefit:3 bauer:1 boundary:1 curve:1 overcome:1 world:2 rich:1 author:2 made:2 welling:1 transaction:1 approximate:2 emphasize:1 geneva:2 preferred:1 baktashmotlagh:3 global:1 active:1 xi:2 discriminative:1 alternatively:1 aci:1 continuous:16 search:1 additionally:2 learn:3 transfer:1 robust:6 ca:1 whiteson:3 necessarily:2 zafar:2 domain:24 significance:3 dense:3 terminated:1 noise:1 edition:1 allowed:1 fair:3 pivotal:26 xu:1 fig:6 representative:4 fashion:1 guiding:1 wish:2 comprises:3 exponential:1 pereira:1 third:1 learns:1 z0:3 sadowski:3 removing:1 bishop:2 navigate:1 harandi:1 showing:1 nyu:2 offset:1 decay:4 abadie:1 evidence:1 reshapes:1 exists:4 incorporating:2 workshop:1 effectively:1 illustrates:2 occurring:1 conditioned:1 easier:1 entropy:8 ez:1 highlighting:1 gender:1 corresponds:7 chance:1 kamishima:3 acm:2 conditional:12 lempitsky:5 goal:4 towards:7 absence:1 hard:1 change:1 included:1 specifically:1 except:1 prejudice:2 total:1 invariance:1 e:2 indicating:3 formally:1 select:1 internal:1 support:1 latter:2 incorporate:1 evaluate:2 ex:1 |
Subsets and Splits