ieee transactions on pattern analysis and machine intelligence 1 exploiting negative...

14
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1 Exploiting Negative Evidence for Deep Latent Structured Models Thibaut Durand, Nicolas Thome and Matthieu Cord Abstract—The abundance of image-level labels and the lack of large scale detailed annotations (e.g. bounding boxes, segmentation masks) promotes the development of weakly supervised learning (WSL) models. In this work, we propose a novel framework for WSL of deep convolutional neural networks dedicated to learn localized features from global image-level annotations. The core of the approach is a new latent structured output model equipped with a pooling function which explicitly models negative evidence, e.g. a cow detector should strongly penalize the prediction of the bedroom class. We show that our model can be trained end-to-end for different visual recognition tasks: multi-class and multi-label classification, and also structured average precision (AP) ranking. Extensive experiments highlight the relevance of the proposed method: our model outperforms state-of-the art results on six datasets. We also show that our framework can be used to improve the performance of state-of-the-art deep models for large scale image classification on ImageNet. Finally, we evaluate our model for weakly supervised tasks: in particular, a direct adaptation for weakly supervised segmentation provides a very competitive model. Index Terms—Weakly Supervised Learning, Convolutional Networks, Structured Outputs, Image Classification, Ranking, Localization 1 I NTRODUCTION O VER the last few years, deep learning and Convolu- tional Neural Networks (ConvNets) [1] have become a key ingredient of visual recognition systems. They have been successfully applied to various visual recognition tasks, e.g. image classification [1], [2], [3], object detec- tion [4], [5], [6], semantic segmentation [7], [8], [9]. Learning standard architectures (AlexNet [1], VGG16 [2], ResNet [3]) requires a huge number of training examples, which limits the number of potential training datasets. But it has been shown that these networks can be efficiently transfered on small datasets [10], [11]: the common practice is to use models pre-trained on large scale datasets, e.g. ImageNet [12], and to fine-tune them on the target domain. As objects could be small and appear at different lo- cations in the image, several frameworks [13], [14] rely on bounding boxes to train object-centric classifiers, and apply the classifiers by searching over different locations within the images. However, these rich annotations rapidly become costly to obtain and difficult to scale up [15], [16]. Another approach is to use Weakly Supervised Learning (WSL) models. The idea is to simultaneously learned a model to classify and localize objects, with only image- level labels. This is a challenging task, because the only available information for training is the presence / absence of the objects in the image [17], [18]. There is no information about the location or the size of the objects in the image. The aim of this paper is to learn localized representations for image classification with only image-level labels during training, indicating the presence or absence of a category. T. Durand and M. Cord are with the Sorbonne Universit´ es, UPMC Univ Paris 06, CNRS, LIP6 UMR 7606, 4 place Jussieu, 75005 Paris, France. E-mail: {thibaut.durand, matthieu.cord}@lip6.fr N. Thome is with the CEDRIC - Conservatoire National des Arts et etiers, 292 rue St Martin, 75003 Paris, France. E-mail: [email protected] We call this problem weakly supervised because our model first localizes the discriminative regions, and then uses the predicted regions to perform image classification. We show that these localized representations can be used for object localization and semantic segmentation. We use the term WSL whatever the evaluated prediction. Recently, several methods have been proposed for WSL of deep ConvNets with image-level labels [17], [18], [19], [20], [21], [22]. The key issue is to determine how to pool the regions to have a score per class. The output of the ConvNet is a detection map for each category, so to train it with stan- dard classification loss, it is necessary to aggregate the maps into a global prediction for each class. This pooling issue is also present in WSL structured models [23], [24], [25], [26]. The most popular pooling is the max pooling [17], [19], [23], which selects the best region to perform prediction. In the case of binary classification, this pooling is an instantiation of the Multiple Instance Learning (MIL) paradigm [27]. In another way, some pooling strategies propose to use all regions to perform prediction, by marginalizing over the regions [18], [24], [25]. In this work, we propose a new approach to auto- matically learn localized features, with a pooling strategy explicitly encoding negative evidence. Our pooling function use both maximum and minimum scoring regions. The maximum regions seek discriminative regions for the class whereas the minimum regions seek regions indicating the absence of the class i.e. that provide a counter-evidence for the presence of a class. For instance, a cow detector should strongly penalize the prediction of the bedroom class. For a multi-label classification task, our model learn positive and negative correlation between object classes. We propose several instantiations of our model for clas- sification and structured Average Precision (AP) ranking. In particular, we show that our pooling function allows to solve exactly and efficiently both inference and loss-

Upload: others

Post on 28-Mar-2021

4 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1 Exploiting Negative ...cedric.cnam.fr/~thomen/papers/Durand_PAMI_2018.pdf · 2017. 12. 26. · IEEE TRANSACTIONS ON

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1

Exploiting Negative Evidence for Deep LatentStructured Models

Thibaut Durand, Nicolas Thome and Matthieu Cord

Abstract—The abundance of image-level labels and the lack of large scale detailed annotations (e.g. bounding boxes, segmentationmasks) promotes the development of weakly supervised learning (WSL) models. In this work, we propose a novel framework for WSL ofdeep convolutional neural networks dedicated to learn localized features from global image-level annotations. The core of the approachis a new latent structured output model equipped with a pooling function which explicitly models negative evidence, e.g. a cow detectorshould strongly penalize the prediction of the bedroom class. We show that our model can be trained end-to-end for different visualrecognition tasks: multi-class and multi-label classification, and also structured average precision (AP) ranking. Extensive experimentshighlight the relevance of the proposed method: our model outperforms state-of-the art results on six datasets. We also show that ourframework can be used to improve the performance of state-of-the-art deep models for large scale image classification on ImageNet.Finally, we evaluate our model for weakly supervised tasks: in particular, a direct adaptation for weakly supervised segmentation providesa very competitive model.

Index Terms—Weakly Supervised Learning, Convolutional Networks, Structured Outputs, Image Classification, Ranking, Localization

F

1 INTRODUCTION

O VER the last few years, deep learning and Convolu-tional Neural Networks (ConvNets) [1] have become

a key ingredient of visual recognition systems. They havebeen successfully applied to various visual recognitiontasks, e.g. image classification [1], [2], [3], object detec-tion [4], [5], [6], semantic segmentation [7], [8], [9]. Learningstandard architectures (AlexNet [1], VGG16 [2], ResNet [3])requires a huge number of training examples, which limitsthe number of potential training datasets. But it has beenshown that these networks can be efficiently transfered onsmall datasets [10], [11]: the common practice is to usemodels pre-trained on large scale datasets, e.g. ImageNet[12], and to fine-tune them on the target domain.

As objects could be small and appear at different lo-cations in the image, several frameworks [13], [14] relyon bounding boxes to train object-centric classifiers, andapply the classifiers by searching over different locationswithin the images. However, these rich annotations rapidlybecome costly to obtain and difficult to scale up [15], [16].Another approach is to use Weakly Supervised Learning(WSL) models. The idea is to simultaneously learned amodel to classify and localize objects, with only image-level labels. This is a challenging task, because the onlyavailable information for training is the presence / absenceof the objects in the image [17], [18]. There is no informationabout the location or the size of the objects in the image.The aim of this paper is to learn localized representationsfor image classification with only image-level labels duringtraining, indicating the presence or absence of a category.

• T. Durand and M. Cord are with the Sorbonne Universites, UPMC UnivParis 06, CNRS, LIP6 UMR 7606, 4 place Jussieu, 75005 Paris, France.E-mail: {thibaut.durand, matthieu.cord}@lip6.fr

• N. Thome is with the CEDRIC - Conservatoire National des Arts etMetiers, 292 rue St Martin, 75003 Paris, France.E-mail: [email protected]

We call this problem weakly supervised because our modelfirst localizes the discriminative regions, and then uses thepredicted regions to perform image classification. We showthat these localized representations can be used for objectlocalization and semantic segmentation. We use the termWSL whatever the evaluated prediction.

Recently, several methods have been proposed for WSLof deep ConvNets with image-level labels [17], [18], [19],[20], [21], [22]. The key issue is to determine how to pool theregions to have a score per class. The output of the ConvNetis a detection map for each category, so to train it with stan-dard classification loss, it is necessary to aggregate the mapsinto a global prediction for each class. This pooling issue isalso present in WSL structured models [23], [24], [25], [26].The most popular pooling is the max pooling [17], [19], [23],which selects the best region to perform prediction. In thecase of binary classification, this pooling is an instantiationof the Multiple Instance Learning (MIL) paradigm [27]. Inanother way, some pooling strategies propose to use allregions to perform prediction, by marginalizing over theregions [18], [24], [25].

In this work, we propose a new approach to auto-matically learn localized features, with a pooling strategyexplicitly encoding negative evidence. Our pooling functionuse both maximum and minimum scoring regions. Themaximum regions seek discriminative regions for the classwhereas the minimum regions seek regions indicating theabsence of the class i.e. that provide a counter-evidence forthe presence of a class. For instance, a cow detector shouldstrongly penalize the prediction of the bedroom class. For amulti-label classification task, our model learn positive andnegative correlation between object classes.

We propose several instantiations of our model for clas-sification and structured Average Precision (AP) ranking.In particular, we show that our pooling function allowsto solve exactly and efficiently both inference and loss-

Page 2: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1 Exploiting Negative ...cedric.cnam.fr/~thomen/papers/Durand_PAMI_2018.pdf · 2017. 12. 26. · IEEE TRANSACTIONS ON

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2

augmented inference problems in the AP ranking case,which is not the case with max pooling [28]. We apply ourweakly supervised structured model to end-to-end trainingof deep ConvNets with weak supervision. We design a fullyconvolutional network architecture which enables fast re-gion feature computation by convolutional sharing. Finally,we evaluate our model on different applications: classifica-tion, ranking, pointwise localization and segmentation. Ourmodel outperforms previous state-of-the-art results on sixclassification datasets. We also perform an analysis of ourmodel, including an ablation study.Contributions. This paper extends two conference papers[29] and [30]. Our contributions are threefold. First, we in-troduce a unified framework for pooling, which generalizesstandard WSL models as special cases, including our model[29] (Section 3.5). This study enables a better understandingof the difference between structured WSL models and theirrespective pooling functions, and is supplemented by adetailed experimental comparison in Section 6.3.4. Secondly,we propose a fully convolutional network based on ResNet-101 (Section 4.1) instead of VGG16, to learn feature mapswith more accurate resolution than in [30], which is crucialfor localization. Finally, we extend the experimental valida-tion of [29], [30] at several levels. We perform experimentson the large scale ILSVRC 2012 classification dataset [12],showing a large improvement with respect to similar net-works with the same number of parameters (Section 6.2).We also show that our model is able to learn localized fea-tures and can be successfully applied to weakly supervisedobject localization (Section 6.4) and semantic segmentation(Section 6.5).

The paper is organized as follows. In Section 2, we givean overview of the most relevant related work. Section 3describes our weakly supervised structured model, which isbased on negative evidence. We also provide a comparisonbetween our model and existing weakly supervised struc-tured models. In Section 4, we propose a fully convolutionalnetwork architecture, with a new prediction layer based onthe weakly supervised structured model of Section 3. Wedetail the learning and instantiations for classification andranking in Section 5. Section 6 presents the experimentalstudies.

2 STATE-OF-THE-ART

Deep ConvNets. The computer vision community is cur-rently witnessing a revolutionary change, essentially causedby ConvNets and deep learning. Beyond the outstandingsuccess reached in the context of large scale classification(ImageNet [12] or Places [31]), deep features also prove to bevery effecticient for transfer learning: state-of-the-art resultson standard benchmarks are nowadays obtained with deepfeatures [10]. Several studies reveal that performances canfurther be improved by collecting large datasets that aresemantically closer to the target domain [31], or by fine-tuning the network [11].

Despite their excellent performances, current ConvNetarchitectures only carry limited invariance properties. [32]has shown that, although a small amount of shift invarianceis built into the models through subsampling (pooling) lay-ers, strong invariance is generally not dealt with. Recently,

attempts have been made to overcome this limitation. Somemethods revisit the BoW model with deep features as localregion activations [33], [34] or design BoW layers [35]. Thedrawback of these models is that background regions areencoded into the final representation, decreasing its dis-criminative power. Another option to gain strong invarianceis to explicitly align image regions, e.g. by using WeaklySupervised Learning (WSL) models.

2.1 Pooling scheme for WSL

Learning object detectors with image-level annotations isa common WSL problem. Most WSL methods are basedon the Multiple Instance Learning (MIL) [27] paradigm: animage is regarded as a bag of instances (regions), and thereis an asymmetric relationship between the bag and instancelabels. A bag is positive if it contains at least one positiveinstance, and negative if all its instances are negative. MILmodels thus perform image prediction through its maxscoring region. The Latent SVM (LSVM) [36] is the mostpopular instantiation of MIL for computer vision, and itsuse in Deformable Part Model (DPM) [36] showed excellentperformances for object detection. [17], [19], [21] use aspatial max pooling for WSL of deep ConvNets. Regardingnon-convex optimization issues, a multi-fold MIL procedureis introduced in [37].

A limitation of the max pooling is related to its sensitivityto noise in the region scores, because it only uses the mostdiscriminative region [20]. To increase robustness, severalapproaches propose to use several regions. The authors of[18] use the global average pooling (GAP), and show that thispooling can find all the discriminative regions of a category.[20], [22], [38] observe that this pooling have problems toidentifying the extent of the object: the models trained withmax pooling tend to underestimate object sizes, while themodels trained with GAP overestimate them. [22] proposesa trade-off between max and average pooling by using a log-sum-exp pooling (LSE). Similarly, [38] introduces the globalweighted rank-pooling (GWRP), where max pooling and GAPare special cases.

Recently, interesting MIL extensions have been intro-duced in [39], [40], [41], [42]. All these methods use a bagprediction strategy which departs from the standard maxscoring function in MIL, especially due to the relaxation ofthe common Negative instances in Negative bags (NiN) MILassumption. In the Learning with Label Proportion (LLP)framework [39], only label ratios between ⊕/ instancesin the bags are provided during training. In [40], the LLPmethod of [39] is explicitly applied to MIL problems, in thecontext of video event detection. LLP outperforms baselinemethods (mi/MI-SVM [43]), especially by its capacity torelax the NiN assumption. In [41], the authors questionthe NiN assumption by claiming that it is often violatedin practice during image annotation: human rather labelimages based on their dominant concept than on the actualpresence of the concept in each sub-region. To support thedominant concept annotation, the authors in [41] introducea prediction function selecting the top scoring instances ineach bag.

Other approaches depart from the NiN assumption bytracking the negative evidence of a class with regions [42],

Page 3: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1 Exploiting Negative ...cedric.cnam.fr/~thomen/papers/Durand_PAMI_2018.pdf · 2017. 12. 26. · IEEE TRANSACTIONS ON

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 3

[44]. The main idea is to learn mutual exclusion constraints,model scene subcategories where the positive object classis unlikely to be found, or to capture specific parts whichpotentially indicate the presence of an object of a similarbut distinct class. [44] proposes a generalization of LSVM byincluding negative latent variables. In [42], the authors intro-duce a WSL formulation specific to multi-class classification,where negative evidence is explicitly encoded by augment-ing the model parameters to represent the positive/negativecontribution of a part to a class.

In this paper, we incorporate the idea of negative evi-dence in the weakly supervised structured model. Our pre-diction function uses max+min region scores. The min scor-ing region accounts for the concept of negative evidence,and is capitalized on for learning a more robust model. Inthe experiments, we report that our region selection strategyoutperforms approaches pooling over all regions.

2.2 Deep structured models & WSL

Deep structured models. Many problems in real-world ap-plications (e.g. semantic segmentation, ranking) involve pre-dicting a collection of random variables that are related toeach other. Several approaches propose to combine Markovrandom fields (MRFs) with deep learning algorithms toestimate complex representations while taking into accountthe dependencies between the output random variables.In [45], the authors use a two step approach to combineConvNets and fully-connected CRFs [46]. To jointly learnthe ConvNet and CRF parameters, [47] reformulates themean-field approximate inference for the fully-connectedCRF as a Recurrent Neural Networks (RNN), and introducesCRF-RNN, a network that can be trained end-to-end. Theauthors show that joint learning of the ConvNet and theCRF parameters results in significant performance gains. Atthe same time, [48] proposes a general training algorithmto learn structured models, where MRF potentials are deepnetworks. [49] shows that for some kind of energy functions,proximal methods can be efficiently solved as a RNN feed-foward pass. Recently, [50] introduces the Structured Predic-tion Energy Networks (SPENs), where a deep architecture isused to define an energy function of candidate labels. [51]proposes direct loss minimization approach, and show itseffectiveness for AP ranking problems.WSL of deep structured models. The main idea of WSL isto model the missing information with latent/hidden vari-ables. The most popular approach is the Latent StructuralSVM (LSSVM) [23], which extends the Latent SVM [36]. TheLSSVM model performs prediction by maximizing the jointposterior probability over the output and latent variables.The parameters of the model are learned by minimizing anupper bound on the empirical risk. One drawback of theLSSVM prediction is that the maximization over the latentvariables is not robust to the inherent uncertainty of thelatent variables. To address this issue, the Hidden Condi-tional Random Fields (HCRFs) [24] and the Marginal Struc-tured SVM (MSSVM) [25] marginalize the latent variablesto estimate the probability of the output variables. The ε-framework introduced in [26] proposes a trade-off betweenmaximization and marginalization by using a log-sum-exppooling. Other works explicitly model the uncertainty over

the latent space [52], [53] and propose to predict the outputvariables by minimizing an entropy-based uncertainty mea-sure. To put into perspective the differences between modelsand pooling functions, we introduce a general frameworkwhich includes LSSVM, HCRF, MSSVM, ε-framework andour model as special cases.WSL ranking. In this paper, we also tackle the importantproblem of learning to rank, since many computer visiontasks are evaluated with ranking metrics, e.g. Average Pre-cision (AP) in PASCAL VOC. [54] shows that the learningobjective should be tailored to the evaluation loss in orderto obtain the best performance with respect to this loss.Optimizing ranking models with AP is challenging becausethe AP loss is non-differentiable and non-decomposable(it cannot be expressed as simple sums over the examplelabels). In the fully supervised case, an elegant instantiationof structural SVM is introduced in [55], making it possibleto optimize a convex upper bound over the AP. On thecontrary, few works tackle the problem of weakly super-vised ranking from the latent structured output perspective,with the exception of [28]. In [28], the authors introduceLAPSVM, and point out that directly using LSSVM [23] forthis purpose is not practical, mainly because no algorithmfor solving the loss-augmented inference problem exists.LAPSVM introduces a tractable optimization by defining anad-hoc prediction rule dedicated to ranking: first the latentvariables are fixed, and then an optimal ranking with fixedlatent variables is found. We show that our WSL modeloffers the ability to solve the loss-augmented inference withan elegant symmetrization due to the max+min predictionfunction.WSL segmentation. We focus on segmentation modelslearned with only labels indicating the presence or absenceof a class in the image. Many methods are based on theMIL framework: MIL-FCN [21] extends MIL to multi-classsegmentation, while MIL-Base [20] generalizes MIL-FCNwith a LSE pooling to speed up the convergence. EM-Adapt[56] includes an adaptive bias into the MIL framework,that boosts classes known to be present and suppressesall the others. Constrained CNN (CCNN) [57] uses a lossfunction optimized for any set of linear constraints on theoutput space of a ConvNet, to control the proportion of thelabels of each class. Recently, [38] introduces a more complexnetwork architecture, that is optimized with a combinationof three losses.

3 NEGATIVE EVIDENCE MODEL

We present here our latent structured model based onnegative evidence. We begin by introducing the notations,then our prediction function, the learning formulation andthe intuitions. Finally, we compare our model with exitinglatent models.

3.1 Notations

We first give some basic notations used in the (latent) struc-tured output learning framework. We consider an inputspace X , that can be arbitrary, and a structured outputspace Y . For (x,y) ∈ X × Y , we are interested in theproblem of learning a discriminant function of the form:

Page 4: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1 Exploiting Negative ...cedric.cnam.fr/~thomen/papers/Durand_PAMI_2018.pdf · 2017. 12. 26. · IEEE TRANSACTIONS ON

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 4

f : X → Y . In order to incorporate hidden parametersthat are not available at training time, we augment thedescription between an input/output pair with a latentvariable h ∈ H. We define a scoring function Fw(x,y,h),with depends on the input data x ∈ X , the output y ∈ Y ,the latent variable h ∈ H and some parameters w ∈ Rd. Ourgoal is to learn a prediction function fw, parametrized by w,so that the predicted output y depends on Fw(x,y,h) ∈ R.During training, we assume that we are given a set of Ntraining pairs D = {(xi,y?i ) ∈ X × Y : i ∈ {1, . . . , N}},where y?i is the ground-truth label of example i. Our goalis to optimize w in order to minimize a user-supplied lossfunction ∆(y?i ,y) over the training set.

3.2 Negative Evidence Model

As mentioned in the introduction, the main intuition of ournegative evidence model is to equip each possible outputy ∈ Y with a pair of latent variables (h+

i,y,h−i,y). h+

i,y

(resp. h−i,y) corresponding to the maximum (resp. minimum)scoring latent value, for input xi and output y:

h+i,y =arg max

h∈HFw(xi,y,h) (1)

h−i,y =arg minh∈H

Fw(xi,y,h) (2)

For an input/output pair (xi,y), the scoring of the model,sw(xi,y), sums h+

i,y and h−i,y scores, as follows:

sw(xi,y) =1

2

(Fw(xi,y,h

+i,y) + Fw(xi,y,h

−i,y))

(3)

Finally, our prediction is:

y = fw(xi) = arg maxy∈Y

sw(xi,y) (4)

This maximization in Eq. (4) is known as the inferenceproblem. Regarding the soring function in Eq. (3), we arehere considering deep ConvNets models for Fw. This gener-alizes the MANTRA model in [29], using a log-linear scoringfunction: Fw(x,y,h) = 〈w,Ψ(x,y,h)〉 where Ψ(x,y,h) isa joint feature map that describes the relation between inputx, output y, and latent variable h.

3.3 Learning Formulation

During training, we enforce the following constraints:

∀y 6= y?i , sw(xi,y?i ) ≥ ∆(y?i ,y) + sw(xi,y) (5)

Each constraint in Eq. (5) requires the scoring valuesw(xi,y

?i ) for the correct output y?i to be larger than the

scoring value sw(xi,y) for each incorrect output y 6= y?i ,plus a margin of ∆(y?i ,y). ∆(y?i ,y), a user-specified loss,makes it possible to incorporate domain knowledge into thepenalization.

To give some insights of how the model parameters canbe adjusted to fulfill constraints in Eq. (5), let us notice that:

• sw(xi,y?i ), i.e. the score for the correct output y?i ,

can be increased if we find statistically high scoringvariables h+

i,y?i

, which represent strong evidence forthe presence of y?i , while enforcing h−i,y?

ivariables

not having large negative scores.

• sw(xi,y), i.e. the score for an incorrect output y, canbe decreased if we find low scoring variables h+

i,y,limiting evidence of the presence of y, while seekingh−i,y variables with large negatives scores, supportingthe absence of output y.

To allow some constraints in Eq. (5) to be violated, weintroduce the following loss function:

`w(xi,yi) = maxy∈Y

[∆(y?i ,y) + sw(xi,y)− sw(xi,y?i )] (6)

We show that `w(xi,y?i ) in Eq. (6) is an upper bound of

∆(y,y?i ) in Appendix A.Using the standard max margin regularization term

‖w‖2, our primal objective function is defined as follows:

P(w) =λ

2‖w‖2 +

1

N

N∑i=1

`w(xi,y?i ) (7)

where λ is the regularization parameter.

3.4 Negative Evidence Intuition

To illustrate the rationale of the approach, let us considera multi-class classification instantiation of our negative ev-idence model, where x is the image, y is the label andthe latent variables h correspond to region locations1. h+

is the max scoring latent value for each class y, i.e. theregion which best represents class y. h− is the min scoringlatent value, and can thus be regarded as an indicator of theabsence of class y in the image.

To highlight the importance of the pair (h+,h−), weshow in Figure 1, for an image of the class bedroom of MIT67dataset [58], the heatmap representing the classificationscores for each latent location using the bedroom classifier,the airport inside classifier and the dining room classifier.The h+ (resp. h−) regions are boxed in green (resp. red).As we can see, the prediction score for the correct classclassifier (bedroom) is large, since the model finds stronglocal evidence h+ of it presence, and no clear evidence of itsabsence (medium score Fw(x,y=bedroom,h−y ) = 0.1). Fora wrong class very different like airport inside, the predictionscore is very low, because there is not region similar to anairport. For a wrong class with similar objects like diningroom, the maximum score for the dining room classifier iscomparable with bedroom classifier: the model heavily fireson discriminative objects (bed for bedroom and chair fordining room). The prediction score sw for the dining roomclassifier is significantly lower than for the bedroom classifier,because it also finds clear evidence of the absence of diningroom, here bed (Fw(x,y = dining room,h−y ) = −1.7).As a consequence, our negative evidence model correctlypredicts the class bedroom. Another example on the fine-grained bird classification task is shown in Figure 3. Inthe Section 6.3.2, we analyze experimentally the selectedregions for multi-label classification task and we show thatour model can learn the positive and negative correlationsbetween the classes.

1. This analysis can be extended to any problems where h is a part ofx, e.g. h is a paragraph of a text x.

Page 5: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1 Exploiting Negative ...cedric.cnam.fr/~thomen/papers/Durand_PAMI_2018.pdf · 2017. 12. 26. · IEEE TRANSACTIONS ON

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 5

y? = bedroom y =bedroom y =airport inside y =dining roomoriginal image sw(x,y) = 2.1 sw(x,y) = −1.7 sw(x,y) = 0.4

Fw(x,y,h+y )=2: high Fw(x,y,h+

y )=−0.1: low Fw(x,y,h+y )=2.1: high

Fw(x,y,h−y )=0.1: medium Fw(x,y,h−y )=−1.6: low Fw(x,y,h−y )=−1.7: low

Fig. 1: Negative evidence intuition. The heatmaps and the predicted regions (h+y in green, h−y in red) for different learned class

models (bedroom, airport inside and dining room) are shown on a bedroom image x. Fw(x,y,h+y ) (resp. Fw(x,y,h−y )) is the score of

the h+y (resp. h−y ) region for class y, and sw(x,y) is the predicted score for class y. The bedroom and dining room models have high

score for max regions because each model focus on objects discriminative for the class (bed for bedroom and chair for dining room).The min region brings complementary information to max region: the min regions score of dining room have a low score becausethe dining room model has found a negative evidence (bed) for the absence of dining room class.

3.5 DiscussionTo put into perspective connections between negative evi-dence and existing latent structured models, we introducethe following generalized scoring function, with ”inversetemperature” β+

h and β−h parameters smoothing betweenmax, softmax and average:

s(β+

h ,β−h )

w (x,y) =1

2β+h

log1

|H|∑h∈H

exp[β+h Fw(x,y,h)] (8)

+1

2β−hlog

1

|H|∑h∈H

exp[β−h Fw(x,y,h)]

As shown in Table 1, the scoring function in Eq. (8)includes several existing models as special cases. Whenβ+h = β−h → +∞, it maximizes over latent variables and is

equivalent to LSSVM [23] or max pooling for deep ConvNets[17]. When β+

h = β−h = 1, is is equivalent HCRF [24] orMSSVM [25], which marginalize over latent variables. GAP[18] (β+

h = β−h → 0) also sum over latent variables, butunlike HCRF or MSSVM, all the latent variables have thesame importance. The ε-framework [26] proposes a trade-off between max and average. This pooling strategy is alsoused to learn ConvNets [20], [22]. The prediction function isequivalent to our model when β+

h → +∞ and β−h → −∞,and pools over both maximum and minimum scores.

From the prediction function in Eq. (8), the condi-tional probability of output y given an input x can be

defined as follows: P (y|x) ∝ exp

[βy · s

(β+h ,β−h )

w (x,y)

]. In

Section 6.3.4, we provide a systematic comparison of thedifferent pooling functions given in Table 1 to highlight theirstrengths and weaknesses in different contexts.

4 RESNET-WELDON NETWORK ARCHITECTURE

Based on the model presented in previous section, we pro-pose ResNet-WELDON, a new weakly supervised learningdedicated to learn localized visual features by using onlyimage-level labels during training. The proposed networkarchitecture is decomposed into two sub-networks: a deepfeature extraction network based on Fully Convolutional

Model β+h β−h

HCRF [24] / MSSVM [25] 1 1GAP [18] → 0 → 0LSSVM [23] / max [17] +∞ +∞Our model +∞ -∞ε-framework [26] / LSE [20], [22] β+

h = β−h ∈ (1,+∞[

TABLE 1: Model comparison with corresponding parameters.

Network (FCN) and a prediction network, as illustrated inFigure 2. The feature extraction net purpose is to extracta fixed-size deep descriptor for each region in the image,while the prediction net outputs a structured output.Notation. We note F lw(x,y,h) the output of the layer l atthe location h of the feature map (or category) y for theinput image x. w are the parameters of the ConvNet.

4.1 Feature extraction network

The feature extraction network is dedicated to compute afixed-size representation for any region of the input image.When using ConvNets as feature extractors, the most naiveoption is to process input regions independently, i.e. toresize each region to match the size of a full image forConvNet architectures trained on large scale databases suchas ImageNet (e.g. 224×224 for ResNet-101 [3]). This is theapproach followed in R-CNN [4], or in MANTRA [29]. Thisis, however, highly inefficient since feature computation in(close) neighbor regions is not shared. Recent improvementsin SPP nets [33], fully convolutional network (FCN) [7]or fast R-CNN [5] process images of any size by usingonly convolutional/pooling layers of ConvNets trained onImageNet, subsequently applying max pooling to map eachregion into a fixed-size vector. They convolutionalize stan-dard classification networks (AlexNet, VGG16) by replacingfully connected by convolution layers.

To have higher spatial resolution on the top of thenetwork, we use the recently introduced the ResNet-101 [3],which is by design fully convolutional. ResNet-101 has 100convolutional layers followed by global average pooling

Page 6: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1 Exploiting Negative ...cedric.cnam.fr/~thomen/papers/Durand_PAMI_2018.pdf · 2017. 12. 26. · IEEE TRANSACTIONS ON

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 6

Fig. 2: ResNet-WELDON deep architecture is decomposed into two sub-networks: a feature extraction network (left) and aprediction network (right) The feature extraction network is based on ResNet-101 to extract local features from whole imageswith good spatial resolution. Then a transfert layer is used to learn class-specific heatmaps (car, motorbike and person), and finallya prediction layer aggregates the heatmaps to produce a single score for each class. Finally, we show for each class the 3 regionswith the highest score on the right.

and fully-connected layer. To have feature map with spatialinformation, we remove the fully-connected layer (as usu-ally done in the litterature), and the global average pooling,which has not learnable parameter. The architecture withonly the convolutional layers (and spatial pooling) allowsto process images of arbitrary sizes, and the sharing ofintermediate features over overlapping image regions. Withthis architecture, the spatial information is naturally pre-served throughout the network: for an input image sizeof 224 × 224, the output size is 7 × 7. Spatial resolutionimpacts the localization and discriminability of the learnedrepresentations. We thus expect the resolution of the featuremaps to be a key component for our model: finer maps keepmore spatial resolution and lead to more specific regions.Moreover, ResNet is more effective at image classificationwhile being parameter- and time-efficient than VGG16.

The input of the feature extraction network is an RGBimage hi×wi, where hi (resp. wi) is the height (resp. width).The output is a h × w × 2048 feature map, where h = hi

32and w = wi

32 are number of sliding window positions in thehorizontal and vertical direction in the image, respectively(see Figure 2). The weights of the feature extraction networkare initialized on ImageNet.

4.2 Prediction network designThis part aims at selecting relevant regions to properlypredict the global (structured) image label.

4.2.1 Transfer layerThe first layer of the prediction network is a transfert layer.Its goal is to transfer weights of the feature extractionnetwork from large scale datasets to new target datasets.It transforms the output of the feature extraction networkF fe into a feature map F t of size h × w × C , where Cis the number of categories (see Figure 2). This layer is

convolutional layer, composed of C filters, each of size1 × 1 × 2048. Due to the kernel size of the convolution,this layer preserves the spatial resolution of the featuremaps. The output of this layer can be seen as localizationheatmaps. In Figure 2, we show the heatmaps for differentcategories: car, motorbike and person.

4.2.2 Weakly-Supervised Prediction (WSP) layerThe second layer is a spatial pooling layer s aggregatesthe score maps into classification scores: for each outputy ∈ {1, . . . , C}, the score over the h × w regions areaggregated into a single scalar value. We note F tw(xi,y,h)is the score of region h from image xi for category y, andH = {1, . . . , ri} the region index set, and ri is the numberof regions for image xi. The output s of the predictionlayer is a vector 1 × 1 × C . As mentioned in Section 2, thestandard approach for WSL inherited from MIL is to selectthe max scoring region. The output score is the score of theregion with the maximum score. We propose to improvethis strategy in two complementary directions: use negativeevidence and several instances.WELDON Pooling. This pooling improves max poolingby incorporate negative evidence. The prediction consistsin summing the max and min scoring regions. Based onrecent MIL insights on learning with top instances [41], wealso propose to extend the selection of a single region tomultiple regions. Formally, let hz ∈ {0, 1} be the binaryvariable denoting the selection of the zth region from layerF t. We propose the scoring function stop, which selects thek+ highest scoring regions as follows:

stopw,k+(F t(xi,y)) =1

k+

ri∑z=1

h+z Ftw(xi,y, z) (9)

where h+ = arg maxh∈{0,1}ri

ri∑z=1

hzFtw(xi,y, z) s.t.

ri∑z=1

hz=k+

Page 7: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1 Exploiting Negative ...cedric.cnam.fr/~thomen/papers/Durand_PAMI_2018.pdf · 2017. 12. 26. · IEEE TRANSACTIONS ON

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 7

where F tw(xi,y, z) is the value of the zth region score forclass y.

Beyond the relaxation of the NiN assumption, whichis sometimes inappropriate (see Section 2), the intuitionbehind F top is to provide a more robust region selectionstrategy. Indeed, using a single area for training the modelnecessarily increases the risk of selecting outliers.

To incorporate negative evidence in our prediction func-tion, we propose the scoring function slow, which selects thek− lowest scoring regions as follows:

sloww,k−(F t(xi,y)) =1

k−

ri∑z=1

h−z Ftw(xi,y, z) (10)

where h−= arg minh∈{0,1}ri

ri∑z=1

hzFtw(xi,y, z) s.t.

ri∑z=1

hz=k−

The final prediction simply consists in summing F top

and F low:

sw(xi,y)=1

2

(stopw,k+(F t(xi,y)) + sloww,k−(F t(xi,y))

)(11)

This prediction function is equivalent to MANTRA predic-tion function whenever k+ = k− = 1.

5 LEARNING & INSTANTIATIONS

As shown in Figure 2, the WELDON model outputss ∈ RC . This vector represents a structured output, whichcan be used in a multi-class or multi-label classificationframework, but also in a ranking problem formulation.

5.1 Training formulationIn this paper, we consider three different structured pre-diction for WELDON, and their associated loss functionsduring training.

5.1.1 ClassificationMulti-class classification. In this simple case, C is thenumber of classes and the output space is Y = {1, . . . , C}.We use the usual softmax activation function on top of thespatial aggregation s. The probability of class y for image xis: P (y|x) = exp(sw(x,y))/

∑y′∈Y exp(sw(x,y′)) with its

corresponding log loss during training.

Multi-label classification. In the case of multiple labels,we use a one-against-all strategy, as [17]. For C differentclasses, we train theC binary classifiers jointly, using logisticregression for prediction P (y|x) = (1+exp(−sw(x,y)))−1,with its associated log loss.

5.1.2 Ranking: Average PrecisionWe also tackle the problem of optimizing ranking metrics,and especially Average Precision (AP). We use a latentstructured output ranking formulation, following [55]: ourinput is a set of N training images x = {xi : i ∈ 1, . . . , N},with their binary labels yi, and our goal is to predict aranking matrix y ∈ Y of sizeN×N providing an ordering ofthe training examples. Our ranking feature map for categoryc is expressed as:

Fwc(x,y,h)=

∑p∈P

∑n∈N

ypn(F few (xp,c,hpn)−F few (xn,c,h

np))

(12)

where P (resp. N ) is the set of positive (resp. negative)examples. hpn (resp. hnp) is a vector which represent theselected region for image xp (resp xn) when we consider thecouple of image (p, n), and h is the set of selected regionsfor all pair of examples (p, n) ∈ P ×N

h = {(hpn, hnp) ∈ {0, 1}rp × {0, 1}rn , (13)rp∑z=1

hpnz = k,rn∑z=1

hnpz = k, (p, n) ∈ P ×N}

where rp is the number of regions for image xp (resp. xn).F few (xp,c,h

pn) is the score for category c of region hpn ofimage xp

During training, we aim at minimizing the followingloss: ∆ap(y

?,y) = 1 − AP (y?,y), where y? is the ground-truth ranking. Since AP is non-smooth, we use the follow-ing surrogate (upper-bound) loss:

`w(x,y?) = maxy∈Y

[∆ap(y?,y) + sw(x,y)]− sw(x,y?) (14)

The maximization in Eq (14) is generally referred to asLoss-Augmented Inference (LAI). Exhaustive maximizationis intractable due to the huge size of the structured outputspace. The problem is even exacerbated in the WSL setting,see [28]. We exhibit here the following result for WELDON:

Proposition 1. For each training example, let us denote s(i) =stopw,k(F tw(xi, c))+s

loww,k(F tw(xi, c)) in Eq (11). Inference and LAI

for the WELDON ranking model can be solved exactly by sortingexamples in descending order of score s(i).

The proof is given in Appendix B and comes from an el-egant symmetrization of the problem due to the max + minoperation. Proposition 1 shows that the optimization overregions, i.e. score s(i), decouples from the maximizationover output variables y. This reduces inference and LAI op-timization to fully supervised problems. Inference solutiondirectly corresponds to s(i) sorting. It also allows to use ourmodel with different loss functions, as soon as there is analgorithm to solve the loss-augmented inference in the fullysupervised setting. To solve it with ∆ap, we use the greedyalgorithm proposed by [55], which finds a globally optimalsolution. Note that it is possible to use faster methods [59]to address large-scale problem if required.

5.2 OptimizationOur model is based on the architecture ResNet-101 [3]. Weinitialize it from a model pre-trained on ImageNet [12]and train it with Stochastic Gradient Descent (SGD) withmomentum with image-level labels only. All the layers ofthe network are fine tuned. For multi-class and multi-labelpredictions, error gradients are well-known. For the rankinginstantiation, we detail the gradient:

∂`

∂w=∂sw(x, y)

∂w− ∂sw(x,y?)

∂w

where y is the LAI solution. When learning ResNet-WELDON, the gradients are backpropagated through thespatial pooling layer only within the selected regions, allother gradients being discarded. The selection of relevantregions for backpropagation is a key to learn preciselylocalized features without any spatial supervision [22].

Page 8: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1 Exploiting Negative ...cedric.cnam.fr/~thomen/papers/Durand_PAMI_2018.pdf · 2017. 12. 26. · IEEE TRANSACTIONS ON

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 8

6 EXPERIMENTS

Our deep ConvNet architecture is based on ResNet-101. Weevaluate our ResNet-WELDON strategy on several Com-puter Vision benchmarks corresponding to various visualrecognition tasks. Absolute comparison with state-of-the-artmethods is provided in Section 6.1, while Section 6.2 reportsresults on the large scale dataset ILSVRC. Section 6.3 ana-lyzes the impact of the different improvements for trainingdeep WSL ConvNets. Finally, we evaluate our model onthe challenging weakly supervised segmentation applica-tion. The code is publicly available at https://github.com/durandtibo/weldon.resnet.pytorch.Experimental Setup. In order to get results in very differ-ent recognition contexts, several datasets are used: objectrecognition (PASCAL VOC 2007 [60], PASCAL VOC 2012[61], MS COCO [62]), scene categorization (MIT67 [58]),action recognition (VOC 2012 Action [61]) and fine-grainedrecognition (CUB-200 [63]). For MIT67, CUB-200, VOC 2007and 2012, the performances are evaluated following thestandard protocol. On CUB-200, we follow the standardprotocol without the bounding boxes and part annotations.On VOC 2012, we used VOC evaluation server to evaluate.On MS COCO, we follow the protocol in [17] to performclassification experiments. On VOC 2012 Action, we use thesame weakly supervised protocol as in [29], with evaluationon the val set. We also evaluate our model on the the val setof ILSVRC [12]. Table 2 summarizes dataset information.

Dataset Train Test Classes Eval.

VOC07 5,011 4,952 20 mAPVOC12 11,540 10,991 20 mAPVOCAction 2,296 2,292 10 mAPCOCO 82,783 40,504 80 mAPMIT67 5,360 1,340 67 accuracyCUB-200 5,994 5,794 200 accuracyILSVRC 2012 1,281,167 50,000 1000 accuracy

TABLE 2: Dataset information: number of train and test images,number of classes and evaluation measures (mAP: mean Aver-age Precision).

6.1 State-of-the-art comparisonFirstly, we compare the proposed ResNet-WELDON modelto state-of-the-art methods. We use different image size asinput of our model, and scale combination is performedusing an Object-Bank [64] strategy –as done in [30]. Theimage size and the values of k+/k− are given in Table 3.An analysis of the number of selected regions is donein Section 6.3.4, showing further improvements by carefultuning. Results are gathered in Table 4 and Table 5. Inthis section, we note ResNet-max (resp. ResNet-GAP) thespecial case of ResNet-WELDON where the spatial poolingis equivalent to the max pooling (resp. GAP).Object datasets. We report in Table 4 the performancesfor object datasets, and we can see that ResNet-WELDONoutperforms all recent methods based on deep features bya large margin. More specifically, the improvement com-pared to deep features computed on the whole image ([2], [3], [11]) is significant: there is an improvement overthe best method [3] of 5.2 pt (resp. 4.2 and 9.8) on VOC2007 (resp. VOC 2012 and MS COCO). Note that since our

Image size Size before pooling k+, k−

224× 224 7× 7 5280× 280 9× 9 10320× 320 10× 10 20374× 374 12× 12 30448× 448 14× 14 50560× 560 18× 18 75747× 747 24× 24 100

TABLE 3: Multi-scale setup. We detail the input image sizes,along with the sizes of the feature maps before spatial poolingand the parameter values used in the spatial pooling.

Method VOC07 VOC12 MSCOCO

Return Devil [11] 82.4 - -VGG16 [2] 89.3 89.0 -SPP net [33] 82.4 - -NUS-HCP [65] 85.2 84.2 -Nonlinear Embeddings [66] 86.1 - -ResNet-101 [3] ? 89.8 89.2 72.5

DeepMIL [17] - 86.3 62.8MANTRA [29] 85.8 - -WELDON [30] 90.2 - 68.8ProNet [22] - 89.3 70.9RRSVM [67] 92.9 - -SPLeaP [68] 88.0 - -

ResNet-max 92.0 90.9 78.9ResNet-WELDON 95.0 93.4 80.7

TABLE 4: mAP results on object recognition datasets. ResNet-WELDON and state-of-the-art methods results are reported.Half at the top shows the performances using global image rep-resentation, whereas the half at the bottom shows performancesfor models based on regions selection. ? means that the resultsare obtained by fine-tuning the network on the dataset with theonline code https://github.com/facebook/fb.resnet.torch.

model used a ResNet-101 architecture, the performance gaindirectly measures the relevance of using a WSL method,which selects localized evidence for performing prediction,rather than relying on the whole image information. BothResNet-101 and ResNet-WELDON have the same numberof parameters. Compared to SPP net [33], the improvementof 12 pt on VOC 2007 highlights the superiority of region se-lection based on supervised information, rather than usinghandcrafted aggregation with spatial-pooling BoW models.

The most important comparison is the improvementover other recent WSL methods on deep features [17], [22],[29], [30], [67]. We outperform the deep WSL ConvNetin [17], the approach which is the most closely connected toours, by 7.1 pt (resp. 17.9) on VOC 2012 (resp. MS COCO).This big improvement illustrates the positive impact ofincorporating MIL relaxations for WSL training of deepConvNets, i.e. negative evidence scoring and top-instanceselection. We also note a significant gain of 4.1 pt (resp9.8) on VOC 2012 (resp. MS COCO) with ProNet [22], thatrelaxes the max pooling with a log-sum-exp pooling. Unlikeour model, these models use a VGG16, but a fair comparisonof the pooling of these methods with the same featureextraction network is done in Section 6.3.4. We also reportthe results of ResNet-max, where the only difference withrespect to ResNet-WELDON is that the spatial pooling isa max pooling. We note that ResNet-WELDON is 2 or 3pt better than ResNet-max on the three datasets. ResNet-

Page 9: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1 Exploiting Negative ...cedric.cnam.fr/~thomen/papers/Durand_PAMI_2018.pdf · 2017. 12. 26. · IEEE TRANSACTIONS ON

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 9

WELDON also outperforms by 2.1 pt the RRSVM [67] onVOC 2007, which learn a constrained aggregation operatoron all the regions. Compared to [30], the improvement of4.8 pt (resp. 11.9) on VOC 2007 (resp. MS COCO) essentiallyshows the importance of the FCN ResNet-101 that preservesspatial information throughout the network, and allowsfiner maps to learn more specific regions.

Method CUB-200 MIT67 VOCAct

CaffeNet Places [31] - 68.2 -MOP CNN [34] - 68.9 -Nonlinear Embeddings [66] 63.1 - -Compact Bilinear Pooling [69] 84.0 76.2 -ResNet-101 [3] ? 72.5 78.0 77.9

Two-level attention [70] 69.7 - -STN [71] 84.1 - -MANTRA [29] - 76.6 -Negative parts [42] - 77.1 -MetaObject-CNN [72] - 78.9 -NAC [73] 81.0 - -GAP GoogLeNet [18] 63.0 66.6 -WELDON [30] - 78.0 75.0Part-Stacked CNN [74] † 76.6 - -SPLeaP [68] - 73.5 -

ResNet-max 76.1 68.3 79.9ResNet-GAP 82.7 85.3 85.5ResNet-WELDON 85.6 84.0 86.4

TABLE 5: Results on scene, action and fine-grained datasets.The performances on MIT67 and CUB-200 (resp. VOC 2012Action) are evaluated with multi-class accuracy (resp. mAP).ResNet-WELDON and state-of-the-art methods results are re-ported. Half at the top shows the performances using globalimage representation, whereas the half at the bottom showsperformances for models based on regions selection. † usespart-annotations during training.

Scene, action and fine-grained datasets. We also validateour model on scene, action and fine-grained classification.The results are reported in Table 5 and illustrate the bigimprovement of ResNet-WELDON compared to deep fea-tures computed on the whole image [2], [3], [31] and globalimage representation with deep features computed on im-age regions: MOP CNN [34] and Compact Bilinear Pooling[69] – these models use a VGG16. It is worth noticingthat ResNet-WELDON also outperforms recent part-basedmethods [68], [72] including negative evidence during train-ing [29], [42], but most of these models use a VGG16 archi-tecture. This validates that our region selection approach isbetter than using all regions. ResNet-WELDON also signifi-cantly outperforms, on CUB-200 and MIT67, the recent GAPGoogLeNet [18], which used a global average pooling. OnCUB-200, we can also note that our model is 9 pt betterthan Part-Stacked CNN [74], which uses bounding boxesand part annotations during training. This validates that ourmodel can automatically find discriminative regions, evenin the case of fine-grained classification. Finally, we reportthe results of ResNet-max and ResNet-GAP, which have thesame architeture as ResNet-WELDON, except the spatialpooling. ResNet-WELDON and ResNet-GAP significantlyoutperform ResNet-max on the three datasets. We can notethat the ResNet-GAP is better than ResNet-WELDON onMIT67. In Section 6.3.4, we show that using a lot of regionsis important on MIT67 dataset.

6.2 Large-scale Image ClassificationWe also evaluate ResNet-WELDON on ILSVRC classifica-tion challenge [12] to show the scalability and the effi-ciency of our model for large-scale image classification. Ta-ble 6 summarizes the classification performances of ResNet-WELDON and existing models. To have a fair comparisonbetween models, we only report results for single model.For our model, we use a mono-scale model with an inputimage size 448× 448, and k+ = k− = 50.

We can see that ResNet-WELDON outperforms mostof existing models trained using whole image (VGG16[2], GoogleNet [75], ResNet-152 [3]) and regions (RRSVM[67], GoogleNet-GAP [18]). Our model have similar per-formances that ResNeXt-101 [76], which proposes a newresidual block with a multi-branch architecture. ResNet-WELDON is slightly worse than Inception-ResNet-v2 (12crops) [77] that combines both ResNet and Inception archi-tectures. Better results can be obtained by learning ensembleof models.

We also reported results for different ResNets. TheResNet-101 is directly comparable to our model, becauseit corresponds to our initial model. It is important to notethat with the same number of parameters and a verysimilar architecture, our model have a significant perfor-mance gain with respect to ResNet-101 (1 crop): +3.2 pt(resp. +2.0) on top-1 (resp. top-5) error. We also report theresults with multi-crops post-processing, which a widelyused post-processing to boost performances. Compared toour approach, multi-crops strategy extracts regions with afixed grid, whereas our model automatically selects relevantregions. The important gain validates the relevance of ourregion selection approach. We also compare our model todeeper ResNet models: ResNet-WELDON is +0.9 pt (resp.+0.7) better than the deeper model ResNet-200, which haveabout the double of parameters. We can also note thatResNet-WELDON prediction is simple because it needs only1 forward on the image to predict image label, whereas themulti-crops prediction needs several forwards on differentimage regions.

Model Top-1 error Top-5 error

VGG16 (144 crops) [2] 24.4 7.2GoogleNet (144 crops) [75] - 7.89ResNet-152 (10 crops) [3] 21.43 5.71RRSVM [67] 22.9 6.7GoogleNet-GAP [18] 35.0 13.2Inception-ResNet-v2 (12 crops) [77] 18.7 4.1ResNeXt-101 (1 crop) [76] 19.1 4.4

ResNet-101 † (1 crop) 22.44 6.21ResNet-101 † (10 crops) 21.08 5.35ResNet-152 † (10 crops) 20.69 5.21ResNet-200 † (10 crops) 20.15 4.93

ResNet-WELDON 19.21 4.23

TABLE 6: Classification error on the ILSVRC validation set withsingle model. ResNet-101 is our initial model. † is the resultsof pretrained model given at https://github.com/facebook/fb.resnet.torch.

6.3 ResNet-WELDON AnalysisIn this section, we analyze our model. In Section 6.3.1,we analyze the impact of the different contributions with

Page 10: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1 Exploiting Negative ...cedric.cnam.fr/~thomen/papers/Durand_PAMI_2018.pdf · 2017. 12. 26. · IEEE TRANSACTIONS ON

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 10

an ablation study. Then, we show some visual results inSection 6.3.2. In Section 6.3.3, we compare our rankingmodel with existing model. Finally, we analyze our poolingfunction and we compare it with standard pooling functions(Section 6.3.4).

6.3.1 Ablation study

We analyze the impact on prediction performances of thedifferent contributions of ResNet-WELDON. Our baselinemodel a) is the WSL ConvNet model using an aggregationfunction s=max at the prediction layer (Figure 2). We presentresults for an input image of size 448 × 448, but similarbehaviors are observed for other scales. It gives a networksimilar to [17], trained at a single scale. To measure the im-portance of the difference between ResNet-WELDON anda), we perform a systematic evaluation on the performancewhen the following variations are incorporated:b) Fine-tuning (FT) the network on the target dataset.c) Use of k top instances instead of the max. We use k = 30.d) Incorporation of negative evidence through max+min

aggregation function. When b)+c) are combined, we usek lowest-instances instead of the min, with k = 30.

a) max b) FT c) k=30 d) min VOC07 VOCAct MIT67 CUB-200

X 86.8 71.8 62.4 66.0X X 91.3 77.9 65.4 72.8X X X 91.7 82.1 72,2 78.9X X X 92.2 82.4 69.6 78.2X X X X 93.7 85.4 77.2 82.4

TABLE 7: Ablation study of our WSL deep ConvNet contribu-tions on object (VOC 2007), action (VOC 2012 Action), scene(MIT67) and fine-grained (CUB-200) datasets.

The results are reported in Table 7 for object (VOC 2007),action (VOC 2012 action), scene (MIT67) and fine-grained(CUB-200) datasets. From this systematic evaluation, we candraw the following conclusions:

1) The fine-tuning (b)) significantly impacts performances,with +4.5 pt (resp. +6.1, +3.0 and +6.8) gain on VOC2007 (resp. VOC 2012 Action, MIT67 and CUB-200). Itvalidates that jointly updating all network parametersis crucial, in particular for fine-grained datasets.

2) Both c) and d) improvements result in a very largeperformance gain on all datasets w.r.t the baseline maxpooling with fine-tuning. But the trends change accord-ing to the datasets. d) leads to a better improvementthan c) on VOC 2007 and VOC 2012 Action datasets:d) has a gain of +0.9 pt (resp. +4.5) whereas c) has again of +0.3 pt (resp. +4.2) on VOC 2007 (resp. VOC2012 Action). On the contrary, on MIT67 and CUB-200,c) leads to a better improvement than d): c) has a gainof +6.8 pt (resp. +6.1) whereas c) has a gain of +4.2 pt(resp. +5.4) on MIT67 (resp. CUB-200).

3) Combining c) and d) improvements further boost per-formances with respect to the best of c) or d): +5 pt onMIT67, +3.5 pt on CUB-200, +1.5 pt on VOC 2007 and3 pt on VOC 2012 Action. This shows the complemen-tarity of these two extensions at the aggregation level.This confirms that our pooling with several instancesand negative evidence is relevant.

6.3.2 Visual resultsTo illustrate the region selection policy performed byResNet-WELDON, we show in Figure 3 the top 3 positive(resp. top 3 negative) regions selected by the model in green(resp. red), on the CUB-200 dataset. We show the results fortwo similar bird species, where the main difference is thatthe indigo bunting is completely blue, whereas the paintedbunting is multicolor. The painted bunting model has highscores for all regions, because all regions are correlated withthe class. On the contrary, the indigo bunting model has highscore on the head – because the head is blue – but very lowscore on the tail because it is not blue. The red-black tail is aclear evidence of the absence of indigo bunting class.

painted bunting (1.9) indigo bunting (0.2)top-3=1.8 low-3=0.1 top-3=1.9 low-3=-1.7

Fig. 3: Visual results of ResNet-WELDON with the final predic-tion score. We visualize the predicted regions for the groundtruth model (left column) and a model of a similar incorrectclass (right column). The top 3 positive (resp. top 3 negative)regions selected by the model are in green (resp. red).

In Figure 4, we show the heatmaps of 2 classes: motorbikewhich is present in the image and bottle which is absent.The motorbike model outputs high scores for the motorbikeregions and medium scores for the other regions includingthe person regions. The motorbike model learns that theperson class is positively correlated with the motorbike class.This positive correlation is confirmed by the co-occurrencematrix on VOC 2007 shown in Figure 5: 70% of the motorbikeimages also contain a person. On the contrary, we observethat motorbike regions have very low scores for bottle modeland act as negative evidence for the bottle class, becausethe model learns that motorbike never occurs with a bottle.This negative correlation is confirmed by the co-occurrencematrix. An interesting behavior of the negative evidencemodel is that it can model positive and negative correlationbetween object classes.

6.3.3 Ranking analysisIn this section, we compare models optimized with clas-sification loss and ranking AP loss. We report results formodel using max pooling and max+min pooling (k = 1).The results are shown in Table 8 and are obtained with aninput image size 448 × 448. To have a fair comparison, weuse the same network architecture for all experiments.

We compare our model with the LAPSVM [28], whichis, to the best of our knowledge, the only method that op-timizes an AP-based loss function over weakly superviseddata. We note that both methods optimizing AP rankingduring training have better results than classification base-line on all datasets. For our model, we note a significant

Page 11: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1 Exploiting Negative ...cedric.cnam.fr/~thomen/papers/Durand_PAMI_2018.pdf · 2017. 12. 26. · IEEE TRANSACTIONS ON

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 11

motorbike bottle

Fig. 4: Visual results of ResNet-WELDON. We show theheatmaps of 2 categories: motorbike which is present in theimage and bottle which is absent. For each class, we show themaximum (resp. minimum) region in green (resp. red) with itscorresponding score. The model learns that the person class ispositively correlated with motorbike class. On the contrary, thebottle model learns that the motorbike is a negative evidence ofthe class: the motorbike has a very low score, which shows theabsence of bottle class.

Fig. 5: Normalized co-occurrence matrix between the classesof PASCAL VOC 2007 (trainval). For each category (row), weshow the percentage of other objects (column) that appear inthe same image. For instance, we observe that in horse images,there is usually a person (77%), but there is never an aeroplane(0%).

improvement of +2.2 pt on VOC 2007, +2.2 pt on VOC 2012Action and +1 pt on MS COCO. This validates that optimiz-ing AP ranking during training is better than optimizingclassification when the performances are evaluated withAP. We can also see that our ranking model outperformsLAPSVM model: +3.3 pt on VOC 2007, +7.4 pt on VOC 2012Action and +0.8 pt on MS COCO.

Dataset VOC07 VOCAct MS COCO

max + classif. loss 86.8 71.8 77.4max + AP loss (LAPSVM [28]) 87.9 73.3 77.9max+min + classif. loss 89.9 78.5 77.7max+min + AP loss 91.2 80.7 78.7

TABLE 8: Comparison of optimization with classification lossand ranking AP loss.

6.3.4 Pooling analysis

In this section, we analysis our pooling function, and wecompare it with standard pooling functions presented inSection 3.5. We report results for an input image of size448 × 448, but similar behaviors are observed for otherscales. To have fair comparison, all the experiments usesthe same network (ResNet-101). We analyse the impact ofthe number of selected instances. We show in Figure 6the performance with respect to the proportion of selectedregions. We can note that the Global Average Pooling (GAP)[18] (resp. MANTRA pooling) is a special case of WELDONpooling, when the proportion of selected regions is 1 (resp.∼0).

Firstly, we can see that negative evidence is important,because on all dataset, MANTRA is better than max pooling.In particular, we observe a large improvement of +4.5 pton VOC 2012 Action dataset, where the context plays animportant role [78]. We can also see that region selectionis important: on all dataset except MIT67, the WELDONpooling with a proportion of selected instances in [0.2, 0.8]is equal or better than GAP. The WELDON pooling hassimilar or better results than GAP by using only 20% ofregions. Using more regions (50%) gives better results thanGAP: +0.4 pt on VOC 2007, +0.3 pt on VOC 2012, +0.6 pton VOC 2012 Action, +1.9 pt and +2.1 pt on CUB-200. OnMIT67, we can see that using a large number of regions isbetter (≥ 80%). This shows that a large number of regionsare discriminants.

We also compare our pooling function to max and LSEpooling functions. The LSE pooling is a soft extension of maxpooling. On all datasets, LSE is better than max pooling:+6 pt on CUB-200 and +5.5 on MIT67. This shows thatusing several regions is more robust than using only thebest region. We also note that LSE pooling performancesare closed to MANTRA pooling performances. MANTRApooling, which used only 2 regions, is as efficient as theLSE pooling which used all regions. Except on MS COCO,we observe that the GAP is better than LSE which is betterthan max: using more regions increase the robustness. OnMS COCO, we see that the region selection is importantbecause a lot of objects have small sizes: the gain is +3 ptbetween WELDON with 20% of regions and GAP.

6.3.5 ResNet-WELDON vs. WELDON

We show the improvements of ResNet-WELDON overWELDON [30]. In Tables 4 and 5, we observe that ResNet-WELDON based on ResNet-101 delivers better classificationresults than WELDON based on VGG16. We also comparethe training time (forward+backward) of both architecturesfor different image sizes. The results in Table 9 are evaluatedfor one epoch on VOC 2007. We see that the ResNet-WELDON is faster than WELDON for all image sizes. Wenote that the larger the image size, the greater the improve-ment. The limitation of the FCN VGG16 is because of thefirst fully-connected layer that has 4,096 filters of large 7× 7spatial size and becomes the computational bottleneck afterconverting the VGG16 to a FCN [79]. These results showthat ResNet-WELDON is better and faster than WELDON.

Page 12: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1 Exploiting Negative ...cedric.cnam.fr/~thomen/papers/Durand_PAMI_2018.pdf · 2017. 12. 26. · IEEE TRANSACTIONS ON

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 12

Fig. 6: Pooling analysis. We compare different spatial pooling strategies on 6 datasets for an input image of size 448 × 448. Thex-axis shows the proportion of selected regions and the y-axis the performance. We can see that MANTRA always outperformsthe max pooling, which validates the relevance of negative evidence. We can also see that our spatial performes equally or betterthan GAP with only 20% of regions.

Image size 224 320 448 560 747

WELDON [30] 91 193 391 682 1302ResNet-WELDON 75 150 286 480 866Speed-up +21% +29% +36% +42% +50%

TABLE 9: Training time (s) for one epoch with different imagesizes on VOC 2007.

6.4 Pointwise object localizationIn this section, we evaluate the localization performances ofour model on VOC 2012 validation set [61] and MS COCOvalidation set [62]. We use the pointwise localization metric,which is a standard evaluation metric for weakly supervisedmodels introduced by [17]. We report the results of ourmodel and three other methods in Table 10. Our model istrained with an input image 448 × 448 and k+ = k− = 30.The DeepMIL [17] is a MIL-based architecture – max pool-ing. In spite of its simple architecture, our model outper-forms the complex cascaded architecture of ProNet [22] andWSLocalization [80], which used a complex strategy basedon search- trees to predict locations. Our model outperformsthese models on the two datasets, but these models use olderarchitecture. To have a fair comparison, we also report theresult of our model for max pooling, and call it ResNet-max. This model is equivalent to DeepMIL and has lowerperformances than ResNet-WELDON on the two datasets.

Method VOC 2012 MS COCO

DeepMIL [17] 74.5 41.2ProNet [22] 77.7 46.4WSLocalization [80] 79.7 49.2

ResNet-max 80.2 50.6ResNet-WELDON 81.5 51.7

TABLE 10: Pointwise object localization performances (MAP)on VOC 2012 and MS COCO.

6.5 Weakly supervised image segmentation

In this section, we show that our model can be applied toweakly supervised image segmentation, while being trainedfrom global image labels only. We evaluate our model onthe VOC 2012 image segmentation dataset [61], consistingof 20 foreground object classes and one background class.We train our model with the train set (1,464 images) andthe extra annotations provided by [81] (resulting in an aug-mented set of 10,582 images), and test it on the validation set(1,449 images). The network is trained with an input image448×448 and k+ = k− = 50. The performance is measuredin terms of pixel Intersection-over-Union (IoU) averagedacross the 21 categories. As in existing methods, we usea fully connected CRF (FC-CRF) [46] to post-process thefinal segmentation mask. We use the class-specific heatmapsgenerated by the network as unary potentials of the FC-CRF.

Method Mean IoU

MIL-FCN [21] 24.9MIL-Base+ILP+SP-sppxl [20] 36.6EM-Adapt +FC-CRF [56] 33.8CCNN + FC-CRF [57] 35.3SEC [38] 50.7

ResNet-WELDON + FC-CRF 42.1

TABLE 11: Comparison of weakly supervised semantic segmen-tation methods on VOC 2012.

Results. The result of our method is presented in Table 11,and some predictions are shown in Figure 7. We compareit to weakly supervised methods that use only image labelsduring training. Firstly, we note a large improvement withrespect to MIL models based on max pooling [21] (+17.2pt) or its soft extension [20] (+5.5 pt). This validates theefficient of our negative evidence pooling for segmentation.ResNet-WELDON also significantly outperforms the recentCCNN [57] that uses a loss function to optimize for any set

Page 13: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1 Exploiting Negative ...cedric.cnam.fr/~thomen/papers/Durand_PAMI_2018.pdf · 2017. 12. 26. · IEEE TRANSACTIONS ON

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 13

Image Ground truth Our prediction

Fig. 7: Segmentation examples on VOC 2012. We show theprediction of our model after FC-CRF and the ground truth.

of linear constraints on the output space of the ConvNet.[38] achieves the best results by using a quite more complexstrategy. The training scheme incorporates different terms,which are specifically tailored to segmentation: one enforcesthe segmentation mask to match low-level image bound-aries, another one incorporates prior knowledge to supportpredicted classes to occupy a certain image proportion. Incontrast, our model is generic, and is trained in the samemanner for the classification and segmentation. It would bepossible to use the specific segmentation priors of [38] in ourmodel to boost performances.

7 CONCLUSION

We propose a new structured output latent variable model,based on negative evidence. Based on this model, we pro-pose ResNet-WELDON, a new weakly supervised learn-ing dedicated to learn discriminative localized visual fea-tures by using only image-level labels during training.ResNet-WELDON model uses a fully convolutional archi-tecture, where the spatial information is naturally preservedthroughout the network. Extensive experiments have shownthe effectiveness of our model for image classification. Weshow the scalability and the efficiency of our model on largescale ILSVRC dataset. We also present a detailed experi-mental comparison of different pooling functions on severaldatasets. Finally, we show that ResNet-WELDON can besuccessfully apply for weakly supervised segmentation.

REFERENCES

[1] A. Krizhevsky, I. Sutskever, and G. Hinton, “Imagenet classifica-tion with deep convolutional neural networks,” in NIPS, 2012.

[2] K. Simonyan and A. Zisserman, “Very Deep Convolutional Net-works for Large-Scale Image Recognition,” in ICLR, 2015.

[3] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning forimage recognition,” in CVPR, 2016.

[4] R. Girshick, J. Donahue, T. Darrell, and J. Malik, “Region-basedconvolutional networks for accurate object detection and segmen-tation,” TPAMI, 2016.

[5] R. Girshick, “Fast R-CNN,” in ICCV, 2015.[6] J. Dai, Y. Li, K. He, and J. Sun, “R-FCN: Object detection via region-

based fully convolutional networks,” in NIPS, 2016.[7] J. Long, E. Shelhamer, and T. Darrell, “Fully Convolutional Net-

works for Semantic Segmentation,” in CVPR, 2015.[8] L. Chen, G. Papandreou, I. Kokkinos, K. Murphy, and A. L. Yuille,

“Semantic Image Segmentation with Deep Convolutional Netsand Fully Connected CRFs,” in ICLR, 2015.

[9] J. Dai, K. He, Y. Li, S. Ren, and J. Sun, “Instance-sensitive fullyconvolutional networks,” in ECCV, 2016.

[10] H. Azizpour, A. S. Razavian, J. Sullivan, A. Maki, and S. Carlsson,“Factors of transferability for a generic convnet representation,”TPAMI, 2016.

[11] K. Chatfield, K. Simonyan, A. Vedaldi, and A. Zisserman, “Returnof the Devil in the Details: Delving Deep into Convolutional Nets,”in BMVC, 2014.

[12] O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma,Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, A. C. Berg, andL. Fei-Fei, “ImageNet large scale visual recognition challenge,”International Journal of Computer Vision (IJCV), 2015.

[13] M. Oquab, L. Bottou, I. Laptev, and J. Sivic, “Learning and Trans-ferring Mid-Level Image Representations using ConvolutionalNeural Networks,” in CVPR, 2014.

[14] R. Girshick, J. Donahue, T. Darrell, and J. Malik, “Rich feature hier-archies for accurate object detection and semantic segmentation,”in CVPR, 2014.

[15] M. Blaschko, P. Kumar, and B. Taskar, “Tutorial: Visual learningwith weak supervision,” CVPR 2013.

[16] A. Bearman, O. Russakovsky, V. Ferrari, and L. Fei-Fei, “What’s thePoint: Semantic Segmentation with Point Supervision,” in ECCV,2016.

[17] M. Oquab, L. Bottou, I. Laptev, and J. Sivic, “Is object localizationfor free? Weakly-supervised learning with convolutional neuralnetworks,” in CVPR, 2015.

[18] B. Zhou, A. Khosla, A. Lapedriza, A. Oliva, and A. Torralba,“Learning Deep Features for Discriminative Localization,” inCVPR, 2016.

[19] G. Papandreou, I. Kokkinos, and P.-A. Savalle, “Modeling Localand Global Deformations in Deep Learning: Epitomic Convolu-tion, Multiple Instance Learning, and Sliding Window Detection,”in CVPR, 2015.

[20] P. O. Pinheiro and R. Collobert, “From image-level to pixel-levellabeling with convolutional networks,” in CVPR, 2015.

[21] D. Pathak, E. Shelhamer, J. Long, and T. Darrell, “Fully Convo-lutional Multi-Class Multiple Instance Learning,” in ICLR (Work-shop), 2015.

[22] C. Sun, M. Paluri, R. Collobert, R. Nevatia, and L. Bourdev,“ProNet: Learning to Propose Object-Specific Boxes for CascadedNeural Networks,” in CVPR, 2016.

[23] C.-N. Yu and T. Joachims, “Learning structural svms with latentvariables,” in ICML, 2009.

[24] A. Quattoni, S. Wang, L.-P. Morency, M. Collins, and T. Darrell,“Hidden conditional random fields,” TPAMI, 2007.

[25] W. Ping, Q. Liu, and A. Ihler, “Marginal structured svm withhidden variables,” in ICML, 2014.

[26] A. G. Schwing, T. Hazan, M. Pollefeys, and R. Urtasun, “EfficientStructured Prediction with Latent Variables for General GraphicalModels,” in ICML, 2012.

[27] T. G. Dietterich, R. H. Lathrop, and T. Lozano-Perez, “Solvingthe multiple instance problem with axis-parallel rectangles,” Artif.Intell., 1997.

[28] A. Behl, P. Mohapatra, C. V. Jawahar, and M. P. Kumar, “Opti-mizing average precision using weakly supervised data,” TPAMI,2015.

[29] T. Durand, N. Thome, and M. Cord, “MANTRA: MinimumMaximum Latent Structural SVM for Image Classification andRanking,” in ICCV, 2015.

[30] ——, “WELDON: Weakly Supervised Learning of Deep Convolu-tional Neural Networks,” in CVPR, 2016.

[31] B. Zhou, A. Lapedriza, J. Xiao, A. Torralba, and A. Oliva, “Learn-ing Deep Features for Scene Recognition using Places Database.”in NIPS, 2014.

[32] N. Zhang, M. Paluri, M. Ranzato, T. Darrell, and L. Bourdev,“PANDA: Pose Aligned Networks for Deep Attribute Modeling,”in ECCV, 2014.

[33] K. He, X. Zhang, S. Ren, and J. Sun, “Spatial pyramid poolingin deep convolutional networks for visual recognition,” in ECCV,2014.

[34] Y. Gong, L. Wang, R. Guo, and S. Lazebnik, “Multi-scale orderlesspooling of deep convolutional activation features,” in ECCV, 2014.

[35] R. Arandjelovic, P. Gronat, A. Torii, T. Pajdla, and J. Sivic,“NetVLAD: CNN architecture for weakly supervised place recog-nition,” in CVPR, 2016.

[36] P. F. Felzenszwalb, R. B. Girshick, D. McAllester, and D. Ramanan,“Object Detection with Discriminatively Trained Part Based Mod-els,” in TPAMI, 2010.

Page 14: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1 Exploiting Negative ...cedric.cnam.fr/~thomen/papers/Durand_PAMI_2018.pdf · 2017. 12. 26. · IEEE TRANSACTIONS ON

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 14

[37] R. G. Cinbis, J. Verbeek, and C. Schmid, “Weakly SupervisedObject Localization with Multi-fold Multiple Instance Learning,”in TPAMI, 2016.

[38] A. Kolesnikov and C. H. Lampert, “Seed, expand and constrain:Three principles for weakly-supervised image segmentation,” inECCV, 2016.

[39] F. X. Yu, D. Liu, S. Kumar, T. Jebara, and S.-F. Chang, “∝svm forlearning with label proportions.” in ICML, 2013.

[40] K.-T. Lai, F. X. Yu, M.-S. Chen, and S.-F. Chang, “Video eventdetection by inferring temporal instance labels,” in CVPR, 2014.

[41] W. Li and N. Vasconcelos, “Multiple Instance Learning for SoftBags via Top Instances,” in CVPR, 2015.

[42] S. N. Parizi, A. Vedaldi, A. Zisserman, and P. F. Felzenszwalb,“Automatic discovery and optimization of parts for image classi-fication,” in ICLR, 2015.

[43] S. Andrews, I. Tsochantaridis, and T. Hofmann, “Support vectormachines for multiple-instance learning,” in NIPS, 2003.

[44] H. Azizpour, M. Arefiyan, S. N. Parizi, and S. Carlsson, “Spotlightthe negatives: A generalized discriminative latent model,” inBMVC, 2015.

[45] L.-C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, and A. L.Yuille, “Semantic Image Segmentation with Deep ConvolutionalNets and Fully Connected CRFs,” in ICLR, 2015.

[46] P. Krahenbuhl and V. Koltun, “Efficient inference in fully con-nected crfs with gaussian edge potentials,” in NIPS, 2011.

[47] S. Zheng, S. Jayasumana, B. Romera-Paredes, V. Vineet, Z. Su,D. Du, C. Huang, and P. Torr, “Conditional random fields asrecurrent neural networks,” in ICCV, 2015.

[48] L. Chen, A. G. Schwing, A. L. Yuille, and R. Urtasun, “Learningdeep structured models,” in ICML, 2015.

[49] S. Wang, S. Fidler, and R. Urtasun, “Proximal deep structuredmodels,” in NIPS, 2016.

[50] D. Belanger and A. McCallum, “Structured prediction energynetworks,” in ICML, 2016.

[51] Y. Song, A. G. Schwing, R. S. Zemel, and R. Urtasun, “Trainingdeep neural networks via direct loss minimization,” in ICML, 2016.

[52] K. Miller, M. P. Kumar, B. Packer, D. Goodman, and D. Koller,“Max-margin min-entropy models,” in AISTATS, 2012.

[53] D. Bouchacourt, S. Nowozin, and M. Pawan Kumar, “Entropy-based latent structured output prediction,” in ICCV, 2015.

[54] D. Bouchacourt, P. K. Mudigonda, and S. Nowozin, “Disco nets :Dissimilarity coefficients networks,” in NIPS, 2016.

[55] Y. Yue, T. Finley, F. Radlinski, and T. Joachims, “A support vectormethod for optimizing average precision,” in SIGIR, 2007.

[56] G. Papandreou, L.-C. Chen, K. Murphy, and A. L. Yuille, “Weakly-and semi-supervised learning of a DCNN for semantic imagesegmentation,” in ICCV, 2015.

[57] D. Pathak, P. Krahenbuhl, and T. Darrell, “Constrained Convolu-tional Neural Networks for Weakly Supervised Segmentation,” inICCV, 2015.

[58] A. Quattoni and A. Torralba, “Recognizing indoor scenes,” inCVPR, 2009.

[59] P. Mohapatra, C. Jawahar, and M. P. Kumar, “Efficient optimiza-tion for average precision svm,” in NIPS, 2014.

[60] M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, andA. Zisserman, “The PASCAL Visual Object Classes Challenge 2007(VOC2007) Results.”

[61] ——, “The PASCAL Visual Object Classes Challenge 2012(VOC2012) Results.”

[62] T.-Y. Lin, M. Maire, S. Belongie, J. Hays, P. Perona, D. Ramanan,P. Dollar, and C. L. Zitnick, “Microsoft COCO: Common Objects inContext,” in ECCV, 2014. [Online]. Available: http://mscoco.org

[63] C. Wah, S. Branson, P. Welinder, P. Perona, and S. Belongie, “TheCaltech-UCSD Birds-200-2011 Dataset,” Tech. Rep., 2011.

[64] L.-J. Li, H. Su, E. P. Xing, and L. Fei-Fei, “Object Bank: A High-Level Image Representation for Scene Classification & SemanticFeature Sparsification,” in NIPS, 2010.

[65] Y. Wei, W. Xia, J. Huang, B. Ni, J. Dong, Y. Zhao, and S. Yan,“CNN: single-label to multi-label,” CoRR, vol. abs/1406.5726,2014. [Online]. Available: http://arxiv.org/abs/1406.5726

[66] G. Sharma and B. Schiele, “Scalable nonlinear embeddings forsemantic category-based image retrieval,” in ICCV, 2015.

[67] Z. Wei and M. Hoai, “Region Ranking SVM for Image Classifica-tion,” in CVPR, June 2016.

[68] P. Kulkarni, F. Jurie, J. Zepeda, P. Perez, and L. Chevallier, “Spleap:Soft pooling of learned parts for image classification,” in ECCV,2016.

[69] Y. Gao, O. Beijbom, N. Zhang, and T. Darrell, “Compact BilinearPooling,” in CVPR, 2016.

[70] T. Xiao, Y. Xu, K. Yang, J. Zhang, Y. Peng, and Z. Zhang, “Theapplication of two-level attention models in deep convolutionalneural network for fine-grained image classification,” in CVPR,June 2015.

[71] M. Jaderberg, K. Simonyan, A. Zisserman et al., “Spatial Trans-former Networks,” in NIPS, 2015.

[72] R. Wu, B. Wang, W. Wang, and Y. Yu, “Harvesting discriminativemeta objects with deep cnn features for scene classification,” inICCV, 2015.

[73] M. Simon and E. Rodner, “Neural activation constellations: Unsu-pervised part model discovery with convolutional networks,” inICCV, December 2015.

[74] S. Huang, Z. Xu, D. Tao, and Y. Zhang, “Part-stacked cnn for fine-grained visual categorization,” in CVPR, June 2016.

[75] C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov,D. Erhan, V. Vanhoucke, and A. Rabinovich, “Going deeper withconvolutions,” in CVPR, 2015.

[76] S. Xie, R. Girshick, P. Dollr, Z. Tu, and K. He, “Aggregatedresidual transformations for deep neural networks,” Nov 2016.[Online]. Available: http://arxiv.org/pdf/1611.05431v1

[77] C. Szegedy, S. Ioffe, and V. Vanhoucke, “Inception-v4, inception-resnet and the impact of residual connections on learning,” 2016.

[78] G. Gkioxari, R. Girshick, and J. Malik, “Contextual action recogni-tion with r* cnn,” in CVPR, 2015.

[79] L.-C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, and A. L.Yuille, “Semantic image segmentation with deep convolutionalnets and fully connected crfs,” in ICLR, 2015.

[80] A. J. Bency, H. Kwon, H. Lee, S. Karthikeyan, and B. S. Manjunath,“Weakly supervised localization using deep feature maps,” inECCV, 2016.

[81] B. Hariharan, P. Arbelaez, L. D. Bourdev, S. Maji, and J. Malik,“Semantic contours from inverse detectors.” in ICCV, 2011.

Thibaut Durand received the Ph.D. in ComputerVision and Machine Learning from the Univer-sity of Pierre et Marie Curie, France, in 2017.He received an M.Sc. in Electrical Engineeringby ENSEA, France, and an M.Sc. degrees incomputer science from the University of Cergy-Pontoise, France, in 2013.

Nicolas Thome is a full professor at Conserva-toire Nationnal des Arts et Metiers (Cnam Paris).He received the Ph.D. degree in computer sci-ence from the University of Lyon, France in 2007,and has been associate professor at UPMC-Paris 6 from 2008 to 2016. His research interestsinclude machine learning for computer vision,including applications for semantic understand-ing of multimedia data. He is involved in sev-eral French (ANR), European and international(Canada, Singapore, Brazil) research projects.

He is being coordinator of an ANR project on weakly supervised learningfor image retrieval in 2013-2018.

Matthieu Cord is a full professor at SorbonneUniversity. He received the PhD degree com-puter science from the UCP, France, beforeworking as postdoc at KU Leuven, Belgium.His research interests include computer vision,deep learning and artificial intelligence. He de-veloped several interactive learning-based ap-proaches for CBIR and many models for patternrecognition using deep architectures. Recently,he focused on multimodal (vison and language)understanding. M. Cord is (co-)author of more

than 100 international, peer-reviewed publications among including twoedited books. He is involved in several French, European and interna-tional research projects. In 2009, he was nominated to the prestigiousIUF (French Research Institute).