dense captioning with joint inference and visual...

10
Dense Captioning with Joint Inference and Visual Context Linjie Yang Kevin Tang Jianchao Yang Li-Jia Li * Snap Inc. Google Inc. {linjie.yang, kevin.tang, jianchao.yang}@snap.com [email protected] Abstract Dense captioning is a newly emerging computer vision topic for understanding images with dense language de- scriptions. The goal is to densely detect visual concepts (e.g., objects, object parts, and interactions between them) from images, labeling each with a short descriptive phrase. We identify two key challenges of dense captioning that need to be properly addressed when tackling the problem. First, dense visual concept annotations in each image are associ- ated with highly overlapping target regions, making accu- rate localization of each visual concept challenging. Sec- ond, the large amount of visual concepts makes it hard to recognize each of them by appearance alone. We propose a new model pipeline based on two novel ideas, joint infer- ence and context fusion, to alleviate these two challenges. We design our model architecture in a methodical manner and thoroughly evaluate the variations in architecture. Our final model, compact and efficient, achieves state-of-the-art accuracy on Visual Genome [23] for dense captioning with a relative gain of 73% compared to the previous best al- gorithm. Qualitative experiments also reveal the semantic capabilities of our model in dense captioning. 1. Introduction The computer vision community has recently witnessed the success of deep neural networks for image caption- ing, in which a sentence is generated to describe a given image. Challenging as it seems, a list of pioneering ap- proaches [8] [10] [41] [43] have achieved remarkable suc- cess on datasets such as Flicker30k [44] and MS COCO [5]. For evaluation, metrics in natural language processing are employed to measure the similarity between ground truth captions and predictions, such as BLEU [35], Meteor [2], and CIDEr [40]. However, the holistic image descriptions from these datasets are either limited to the salient objects of the images, or tend to broadly depict the entire visual scene. A picture is worth a thousand words, and these holistic im- age descriptions are far from a complete visual understand- * This work was done when Li-Jia Li was with Snap Inc.. a baseball player umpire wearing blue shirt catcher wearing a blue shirt a baseball game black and white shoe a catcher’s mask baseball in the air a tree man wearing a blue shirt a black helmet on a baseball player blue shirt on umpire spectators in the stands a crowd of people watching the game a man in a red shirt the umpire is wearing a helmet the man is wearing black shoes white lines on the field the catcher’s shin guards (b) (a) Without context: a desktop computer With context: a modern building <start> à woman à playing à frisbee (c) Figure 1: Illustration of our approach for dense captioning. (a) For a region proposal, the bounding box can adapt and improve with the caption word by word. In this example, the bounding box is guided by the caption to include the frisbee, even though the initial position was ambiguous. (b) The object in the red box is hard to recognize as a building without the context of the whole image. (c) An example image overlaid with the most confident region descriptions by our model. ing of the image. Furthermore, giving one description for an entire image can sometimes be quite subjective, making the evaluation of captioning often ambiguous. Recently, Johnson et al. [20] propose to use a dense de- scription of image regions as a better interpretation of the visual content, known as dense captioning. Human anno- 2193

Upload: others

Post on 01-Jun-2020

5 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Dense Captioning With Joint Inference and Visual Contextopenaccess.thecvf.com/content_cvpr_2017/papers/... · Dense Captioning with Joint Inference and Visual Context Linjie Yang

Dense Captioning with Joint Inference and Visual Context

Linjie Yang Kevin Tang Jianchao Yang Li-Jia Li∗

Snap Inc. Google Inc.

{linjie.yang, kevin.tang, jianchao.yang}@snap.com [email protected]

Abstract

Dense captioning is a newly emerging computer vision

topic for understanding images with dense language de-

scriptions. The goal is to densely detect visual concepts

(e.g., objects, object parts, and interactions between them)

from images, labeling each with a short descriptive phrase.

We identify two key challenges of dense captioning that need

to be properly addressed when tackling the problem. First,

dense visual concept annotations in each image are associ-

ated with highly overlapping target regions, making accu-

rate localization of each visual concept challenging. Sec-

ond, the large amount of visual concepts makes it hard to

recognize each of them by appearance alone. We propose

a new model pipeline based on two novel ideas, joint infer-

ence and context fusion, to alleviate these two challenges.

We design our model architecture in a methodical manner

and thoroughly evaluate the variations in architecture. Our

final model, compact and efficient, achieves state-of-the-art

accuracy on Visual Genome [23] for dense captioning with

a relative gain of 73% compared to the previous best al-

gorithm. Qualitative experiments also reveal the semantic

capabilities of our model in dense captioning.

1. Introduction

The computer vision community has recently witnessed

the success of deep neural networks for image caption-

ing, in which a sentence is generated to describe a given

image. Challenging as it seems, a list of pioneering ap-

proaches [8] [10] [41] [43] have achieved remarkable suc-

cess on datasets such as Flicker30k [44] and MS COCO [5].

For evaluation, metrics in natural language processing are

employed to measure the similarity between ground truth

captions and predictions, such as BLEU [35], Meteor [2],

and CIDEr [40]. However, the holistic image descriptions

from these datasets are either limited to the salient objects of

the images, or tend to broadly depict the entire visual scene.

A picture is worth a thousand words, and these holistic im-

age descriptions are far from a complete visual understand-

∗This work was done when Li-Jia Li was with Snap Inc..

a baseball player

umpire wearing blue shirt

catcher wearing a blue shirt a baseball game

black and white shoe

a catcher’s mask

baseball in the air

a tree

man wearing a blue shirt

a black helmet on a baseball player

blue shirt on umpire

spectators in the stands

a crowd of people watching the game

a man in a red shirt

the umpire is wearing a helmet

the man is wearing black shoes

white lines on the field

the catcher’s shin guards

(b) (a)

Without context: a desktop computer With context: a modern building

<start> à woman à playing à frisbee

(c)

Figure 1: Illustration of our approach for dense captioning.

(a) For a region proposal, the bounding box can adapt and

improve with the caption word by word. In this example,

the bounding box is guided by the caption to include the

frisbee, even though the initial position was ambiguous. (b)

The object in the red box is hard to recognize as a building

without the context of the whole image. (c) An example

image overlaid with the most confident region descriptions

by our model.

ing of the image. Furthermore, giving one description for

an entire image can sometimes be quite subjective, making

the evaluation of captioning often ambiguous.

Recently, Johnson et al. [20] propose to use a dense de-

scription of image regions as a better interpretation of the

visual content, known as dense captioning. Human anno-

12193

Page 2: Dense Captioning With Joint Inference and Visual Contextopenaccess.thecvf.com/content_cvpr_2017/papers/... · Dense Captioning with Joint Inference and Visual Context Linjie Yang

tators are required to exhaustively label bounding boxes

over different levels of visual concepts (e.g., objects, object

parts, and interactions between them). Compared to global

image descriptions, dense local descriptions are more ob-

jective and less affected by annotator preference. The lo-

cal descriptions provide a rich and dense semantic labeling

of the visual elements, which can benefit other tasks such

as semantic segmentation [27] and visual question answer-

ing [1] [29]. For convenience, we refer to image regions

associated with annotated visual concepts as regions of in-

terest in the remaining text.

The exploration of dense captioning is only just begin-

ning. An end-to-end neural network is used in [20] to pre-

dict descriptions based on region proposals generated by a

region proposal network [37]. For each region proposal, the

network produces three elements separately: a region-of-

interest probability (similar to the detection score in object

detection), a phrase to describe the content, and a bound-

ing box offset. The major difference dense captioning has

from traditional object detection is that it has an open set

of targets (not limited to valid objects), and includes parts

of objects and multi-object interactions. Because of this,

two types of challenges emerge when predicting region cap-

tions.

First, the target bounding boxes become much denser

than object detection with limited categories (e.g. 20 cat-

egories for PASCAL VOC [9]). Take the Visual Genome

dataset as an example. The statistics of the maximum

Intersection-over-Union (IoU) between ground truth bound-

ing boxes can be seen in Fig. 2(a), from which we see more

than half of the bounding boxes have maximum IoU larger

than 0.31. Fig. 2(b) shows an image overlaid with all ground

truth bounding boxes. Here, we can visually see that any re-

gion proposal can easily have multiple overlapping regions

of interest. Therefore, it is necessary to localize a target

region with the guidance of the description.

Second, since there are a huge number of visual con-

cepts being described, some of the target regions are visu-

ally ambiguous without information about their context. In

Visual Genome, the number of different object categories is

18, 136 [23], which includes a long list of visually similar

object pairs such as “street light” and “pole”, “horse” and

“donkey”, and “chair” and “bench”.

Thus, we believe that tackling these two challenges can

greatly benefit the task of dense captioning. We carefully

design our dense captioning model to address the above two

problems by introducing two key components. The first

component is joint inference, where pooled features from

regions of interest are fed into a recurrent neural network

to predict region descriptions, and the localization bound-

1Note that because a large portion of overlapping bounding boxes refer

to the same object and have high IoU ratios, we have merged the bounding

boxes with IoU larger than 0.7 together into one.

(a) (b)

Figure 2: (a) Distribution of maximum IoUs between

bounding boxes in ground truth annotations. (b) Sample

image overlaid with all ground truth bounding boxes.

ing boxes are jointly predicted from the the pooled fea-

tures with recurrent inputs from the predicted descriptions.

Fig. 1(a) shows an example of a step-by-step localization

process with joint inference, where the localization bound-

ing box gradually adapts to the correct position using the

predicted descriptions. The second component is context

fusion, where pooled features from regions of interest are

combined with context features to predict better region de-

scriptions. An example is shown in Fig. 1(b), where the

object in the red bounding box is described as a desktop

without visual cues from the surrounding context. We de-

sign several different network structures to implement the

two key components respectively, and conduct extensive

experiments to explore the benefits and characteristics of

each. Our unified model achieves a mean average precision

(mAP) accuracy of 9.31% on Visual Genome V1.0, a rela-

tive gain of 73% over the previous state-of-the-art approach

by [20]. An example image with the most confident region

descriptions from our model is shown in Fig. 1(c).

To reiterate, the contributions of this work are two-fold:

• We design network structures that incorporate two

novel ideas, joint inference and context fusion, to ad-

dress the challenges we identified in dense captioning.

• We conduct an extensive set of experiments to explore

the capabilities of the different model structures, and

analyze the underlying mechanisms for each. With

this, we are able to obtain a compact and effective

model with state-of-the-art performance.

2. Related Work

Recent image captioning models often utilize a convo-

lutional neural network (CNN) [24] as an image encoder

and a recurrent neural network (RNN) [42] as a decoder

for predicting a sentence [8] [21] [41]. RNNs have been

widely used in language modeling [4] [13] [32] [39]. Some

image captioning approaches, though targeted at a global

description, also build relationships with local visual ele-

ments. Karpathy et al. [21] [22] learn an embedding with a

2194

Page 3: Dense Captioning With Joint Inference and Visual Contextopenaccess.thecvf.com/content_cvpr_2017/papers/... · Dense Captioning with Joint Inference and Visual Context Linjie Yang

proposals

RegionProposalNetwork

convlayers

regionfeatures

contextfeature

detec9on

scores

bounding

boxes

cap9ons

regiondetec+onnetwork localiza+onandcap+oningnetwork

Figure 3: Our framework consists of two stages: a region

detection network and a localization and captioning net-

work.

latent alignment between image regions and word phrases.

Fang et al. [10] first detect words from images using multi-

ple instance learning, then incorporate the words in a maxi-

mum entropy language model. A soft-attention mechanism

is also proposed to cast attention over different image re-

gions when predicting each word [19] [43].

Recent object detection algorithms based on deep learn-

ing often show a two-stage paradigm: region proposal and

detection [11] [12] [37]. Faster R-CNN [37] is the most re-

lated to our work, as it utilizes a Region Proposal Network

(RPN) to generate region proposals and a detection network

to predict object categories and bounding box offsets. The

two networks can share convolutional features and can be

trained with an approximate fast joint training algorithm. A

recent improvement to faster R-CNN is the incorporation of

context information using a four-direction RNN on the con-

volutional feature map [3]. Visual context can greatly help

tasks such as object detection [3] [7] [33] and semantic seg-

mentation [33]. Another direction is to remove the RPN and

directly produce detection results [26] [36] to further speed

up the algorithm.

The task of dense captioning was first proposed in [20],

in which a spatial transformer network [15] is used to fa-

cilitate joint training of the whole network. A related ap-

plication is also proposed to detect an arbitrary phrase in

images using the dense captioning model. The experiments

are conducted on the Visual Genome dataset [23], which

provides not only region descriptions but also objects, at-

tributes, question answering pairs, etc. Also closely related

are other recent topics such as localizing a phrase in a spe-

cific image [14] [30] [34], generating an unambiguous de-

scription for a specific region [30] [45], or detecting visual

relationships in images [25] [28].

3. Our Model

Dense captioning is similar to object detection in that

it also needs to localize the regions of interest in an im-

region

feature

man

<SOS>

playing tennis <EOS>

Figure 4: An illustration of the unrolled LSTM for re-

gion captioning. <SOS> and <EOS> denote the start-of-

sentence and end-of-sentence tokens, respectively.

age, but differs in that it replaces the fixed number of ob-

ject categories with a much larger set of visual concepts de-

scribed by phrases. Therefore, we can borrow successful

recipes from the object detection literature in designing our

dense captioning algorithm. In this work, our dense cap-

tioning model pipeline is inspired by the efficient faster R-

CNN framework [37]. Fig. 3 illustrates our dense caption-

ing framework, which includes a region detection network

adopted from faster R-CNN and a localization and caption-

ing network. In this section, we will design different local-

ization and captioning network architectures step by step in

searching for the right formula. Our baseline model directly

combines the faster R-CNN framework for region detection

and long short-term memory (LSTM) [13] for captioning.

3.1. Baseline model

Faster R-CNN [37] uses a two-stage neural network to

detect objects based on the image feature maps, which are

generated by a fully convolutional neural network. In the

first stage, the network uses a RPN to generate region pro-

posals that are highly likely to be the regions of interest,

then it generates fixed-length feature vectors for each region

proposal using Region-Of-Interest (ROI) pooling layers. In

the second stage, the feature vectors are fed into another

network to predict object categories as well as the bound-

ing box offsets. Since the gradients cannot be propagated

through the proposal coordinates, exact joint training is not

viable for faster R-CNN. Instead, it can be trained by alter-

natively updating parameters with gradients from the RPN

and the final prediction network, or by approximate joint

training which updates the parameters with gradients from

the two parts jointly.

Our baseline model for dense captioning directly uses

the proposal detection network from faster R-CNN in the

first stage. For the second stage of localization and cap-

tioning, we use the model structure in Fig. 5(a), where the

region features are used to produce detection scores and

bounding box offsets, as well as fed into an LSTM to gen-

erate region descriptions. The LSTM predicts a word at

each time step and uses this prediction to predict the next

word. Fig. 4 shows an example of using such a recurrent

process to generate descriptions. We use the structure of

VGG-16 [38] for the convolutional layers, which generates

2195

Page 4: Dense Captioning With Joint Inference and Visual Contextopenaccess.thecvf.com/content_cvpr_2017/papers/... · Dense Captioning with Joint Inference and Visual Context Linjie Yang

region

features

LSTM

nextwordbboxoffset

t=T t=0..T

region

features

LSTM

nextwordbboxoffset

t=T t=0..T

region

features

LSTM

nextwordbboxoffset

t=T t=0..T

LSTM

(b) (c) (d)

region

features

nextwordbboxoffset

t=0..T

(a)

LSTM

region

features

Figure 5: Baseline model and several model designs for joint inference of bounding box offset and region description. The

four structures are (a) Baseline model (b) S-LSTM (c) SC-LSTM (d) T-LSTM. See detailed description in text.

feature maps 16× smaller than the input image. Follow-

ing faster R-CNN [37], pretrained weights from the Ima-

geNet Classification challenge [6] are used. Also follow-

ing previous work [20] [21] [31] [41], the region feature is

only fed into the LSTM at the first time step, followed by

a special start-of-sentence token, and then by the embed-

ded feature vectors of the predicted words one by one. This

model is similar to the model in [20] except that their model

replaces the ROI pooling layer with a bilinear interpola-

tion module so that gradients can be propagated through

bounding box coordinates. In contrast, our baseline model

uses approximate joint training that is proven to be effective

for object detection and instance-level semantic segmenta-

tion [17] [18] [37]. In our experiments, we observe that the

baseline model with approximate joint training is very ef-

fective and already outperforms the previous state-of-the-art

method [20]. A potential reason is that although bilinear in-

terpolation allows for exact end-to-end training, the model

may be harder to train due to the transformation properties

of the gradients.

3.2. Joint inference for accurate localization

In this section, we explore the model design for joint in-

ference of bounding box localization and region description

for a given region proposal. Due to the large number of open

phrases and dense bounding boxes, we find it is necessary

to combine the two in order to improve both localization

and captioning. We fix the first stage of the proposal detec-

tion network in Fig. 3 to be the same as our baseline model,

and focus on designing a joint localization and captioning

network for the second stage.

To make the predictor of the bounding box offset aware

of the semantic information in the associated region, we

make the bounding box offset an output of an LSTM en-

coded with region descriptions. Several designs are shown

in Fig. 5. Shared-LSTM (S-LSTM) (Fig. 5(b)) directly uses

the existing LSTM to predict the bounding box offset at

the last time step of the caption. This model embeds the

captioning model and the location information in the same

hidden space. Shared-Concatenation-LSTM (SC-LSTM)

(Fig. 5(c)) concatenates the output of the LSTM and region

features to predict the offset, so the prediction of the offset

is directly guided by the region features. Twin-LSTM (T-

LSTM) (Fig. 5(d)) uses two LSTMs to predict the bound-

ing box offset and description separately. This model sepa-

rates the embedded hidden spaces of the captioning model

and the location predictor. The two LSTMs are denoted

as location-LSTM and caption-LSTM, and both receive the

embedded representation of the last predicted word as input.

In all three models, the bounding box offset is predicted at

the last time step of the description, when the “next word”

is an end-of-sentence token and the description is finished.

Thus the network obtains information about the whole de-

scription at the time of predicting the bounding box offset.

3.3. Context fusion for accurate description

Visual context is important for understanding a local re-

gion in an image, where it has already shown to bene-

fit tasks such as object detection and semantic segmenta-

tion [3] [7] [33]. Despite the exploration of context fea-

tures in these tasks, there is limited work on the integration

of context features into sequential prediction tasks such as

image captioning. We concentrate on finding the optimal

way to combine context features and local features in the

sequential prediction task of dense captioning, rather than

investigating better representations of context information.

Thus, we resort to a simple but effective implementation of

context features, which utilizes a global ROI pooling feature

vector as the visual context. Since the bounding box offset

is not directly connected to the context feature, we only use

the context feature to assist in caption prediction, which in

turn will influence localization through joint inference as

discussed in the previous section.

In this work, we experiment with two variants of com-

bining local features and context features, which are shown

in Fig. 6 and termed as early-fusion and late-fusion. Early-

fusion (Fig. 6(a)) directly combines the region feature and

context feature together before feeding into the LSTM,

while late-fusion (Fig.6(b)) uses an extra LSTM to generate

a recurrent representation of the context feature, and then

combines it with the local feature. The context feature rep-

resentation is combined with the region feature represen-

2196

Page 5: Dense Captioning With Joint Inference and Visual Contextopenaccess.thecvf.com/content_cvpr_2017/papers/... · Dense Captioning with Joint Inference and Visual Context Linjie Yang

region

features

LSTM

nextword

context

feature

LSTM

fusion operator

region

features

LSTM

nextword

context

feature

fusion operator

(b) (a)

Figure 6: Model structures for region description assisted

by context features. (a) Early-fusion. (b) Late-fusion. The

fusion operator denoted by the red dot can be concatenation,

summation, multiplication, etc.

region

features

LSTM

nextwordbboxoffset

LSTM

context

feature

LSTM

(a)

region

feature

playing tennis <EOS>

context

feature

man

bbox offset

(b)

Figure 7: Illustrations of an integrated model. (a) The inte-

grated model of T-LSTM and the late-fusion context model.

(b) An unrolled example of the model in (a). Start-of-

sentence token is omitted for clarity.

tation via a fusion operator for both variants. We experi-

mented with concatenation, summation, and multiplication.

After each word is selected, its embedded representation is

fed back into the caption-LSTM to guide the generation of

the next word. Such fusion designs can be easily integrated

with any of the models in Fig. 5.

3.4. Integrated model

The aforementioned model structures of joint inference

and context fusion can be easily plugged together to pro-

duce an integrated model. For example, the integration of

T-LSTM and the late-fusion context model can be viewed

in Fig. 7. Note that a single word is predicted at each time

step and the bounding box offset is predicted at the last time

step of the caption, after all words have been encoded into

the location-LSTM. Different integrated models are differ-

ent instantiations of the model pipeline we show in Fig. 3.

Table 1: Comparison of our final model with previous best

result on Visual Genome V1.0 and V1.2.

Visual Genome V1.0 V1.2

Model Johnson et al. [20] Ours Gain Ours

mAP 5.39 9.31 73% 9.96

Finally, training our dense captioning model boils down

to minimizing the following loss function L,

L = Lcap + αLdet + βLbbox, (1)

whereLcap, Ldet, and Lbbox denote caption prediction loss,

detection loss and bounding box regression loss, respec-

tively, with α and β the weighting coefficients. Lcap is a

cross-entropy term for word prediction at each time step

of the sequential model, Ldet is a two-class cross-entropy

loss for foreground / background regions, while Lbbox is a

smoothed-L1 loss [37]. Ldet and Lbbox are computed both

in the region proposal network and the final prediction. For

those models using an LSTM for predicting bounding box

offset, the second Lbbox is calculated at the last time-step of

the LSTM output.

4. Experiments

4.1. Evaluation Dataset

We use the Visual Genome dataset [23] as the evaluation

benchmark. Visual Genome has two versions: V1.0 and

V1.2. V1.2 is a cleaner version of V1.0, while V1.0 is used

by [20]. For comparison purposes, we conduct experiments

mainly on V1.0, and report additional results on V1.2. We

use the same train/val/test splits as in [20] for both V1.0 and

V1.2, i.e., 77398 images for training and 5000 images each

for validation and test. We use the same evaluation metric

of mean Average Precision (mAP) as [20], which measures

localization and description accuracy jointly. Average pre-

cision is computed for different IoU thresholds for local-

ization accuracy, and different Meteor [2] score thresholds

for language similarity, then averaged to produce the mAP

score. For localization, IoU thresholds .3, .4, .5, .6, .7 are

used. For language similarity, Meteor score thresholds 0,

.05, .1, .15, .2, .25 are used. A comparison of our final

model using the structure in Fig. 7 with the previous best

result can be seen in Tab. 1, which shows that we achieve

a 73% relative gain compared to the previous best method.

In the following sections, we first introduce the training and

evaluation details, then evaluate and compare the joint infer-

ence models and integrated models under different structure

designs. The influence of hyper-parameters in evaluation is

also explored.

4.2. Model training and evaluation

In training, we use approximate joint training for all

models. We use stochastic gradient descent with a mini-

2197

Page 6: Dense Captioning With Joint Inference and Visual Contextopenaccess.thecvf.com/content_cvpr_2017/papers/... · Dense Captioning with Joint Inference and Visual Context Linjie Yang

Table 2: The mAP performance of baseline and joint inference models on Visual Genome V1.0. First row is the performance

with CNN and RPN fixed, second row is the performance of corresponding models with end-to-end training.

model Johnson et al. [20] baseline S-LSTM SC-LSTM T-LSTM

fixed-CNN&RPN - 5.26 5.15 5.57 5.64

end-to-end 5.39 6.85 6.47 6.83 8.03

Table 3: The mAP performance of integrated models with

combinations of joint inference models and context fusion

structures on Visual Genome V1.0.

model S-LSTM SC-LSTM T-LSTM

early-fusion[·, ·] 6.74 7.18 8.24

⊕ 6.54 7.29 8.16

⊗ 6.69 7.04 8.19

late-fusion

[·, ·] 7.50 7.72 8.49

⊕ 7.19 7.47 8.53

⊗ 7.57 7.64 8.60

Table 4: The mAP performance of different dense caption-

ing models on Visual Genome V1.2.

model baseline S-LSTM T-LSTM

no context

6.98

6.44 8.16

late-fusion

[·, ·] 7.76 9.03

⊕ 7.06 8.71

⊗ 7.63 8.52

batch size of 1 to train the whole network. The input im-

age is re-sized to have a longer side of 720 pixels. Initial

learning rate is set to 0.001 and halved every 100K itera-

tions, and momentum is set to 0.98. Weight decay is not

used in training. We begin fine-tuning the CNN layers after

200K iterations (∼3 epochs) and finish training after 600K

iterations (∼9 epochs). The first seven convolutional layers

are fixed for efficiency, with the other convolutional lay-

ers fine-tuned. We found that training models with context

fusion from scratch tends not to converge well, so we fine-

tune these models from their non-context counterparts, with

a total of 600K training iterations. We only use descriptions

with no more than 10 words for efficiency. We use the most

frequent 10000 words as the vocabulary and replace other

words with an <UNK> tag. For sequential modeling, We

use an LSTM with 512 hidden nodes. For the RPN, we use

12 anchor boxes for generating the anchor positions in each

cell of the feature map, and 256 boxes are sampled in each

forward pass of training. For the loss function, we fix values

of α, β in Eq. (1) to 0.1 and 0.01, respectively.

In evaluation, we follow the settings of [20] for fair com-

parison. First, 300 boxes with the highest predicted con-

fidence after non-maximum suppression (NMS) with IoU

ratio 0.7 are generated. Then, the corresponding region fea-

tures are fed into the second stage of the network, which

produces detection scores, bounding boxes, and region de-

scriptions. We use efficient beam-1 search to produce re-

gion descriptions, where the word with the highest proba-

bility is selected at each time step. With another round of

NMS with IoU ratio 0.3, the remaining regions and their

descriptions are used as the final results.

4.3. Joint inference models

We evaluate the baseline and three joint inference mod-

els in this section. All models are trained end-to-end with

the convolutional layers and the RPN. To further clarify the

effect of different model designs, we also conduct experi-

ments to evaluate the performance of the models based on

the same region proposals and image features. Towards this

end, we fix the weights of the CNN to those of VGG16 and

use a hold-out region proposal network also trained on Vi-

sual Genome based on the fixed CNN weights. The results

of the end-to-end trained models and the fixed-CNN&RPN

models are shown in Tab. 2.

T-LSTM performs best for joint inference. Among

the three different structures for joint inference, T-LSTM

has the best performance for both end-to-end training (mAP

8.03), and fixed-CNN&RPN training (mAP 5.64). The end-

to-end model of T-LSTM outperforms the baseline model

by more than 1% in mAP, while the others are even worse

than the baseline model. By using a shared LSTM to predict

both the caption and bounding box offset, S-LSTM unifies

the language representation and the target location informa-

tion into a single hidden space, which is quite challenging

since they are from completely different domains. Even as-

sisted by the original region feature, the shared LSTM so-

lution does not show much improvement, only on par with

the baseline (mAP 6.83). By separating the hidden space,

i.e. using two LSTMs targeted at the two tasks respectively,

the T-LSTM model yields much better performance (mAP

8.03 vs 6.47). Compared with the baseline model, T-LSTM

is better at both localization and captioning. Fig. 8 shows

several example predictions of bounding box and captions

from one region proposal for the baseline model and the T-

LSTM model. Fig. 8(a) (b) shows that T-LSTM improves

on localization thanks to the guidance of the encoded cap-

tion information, while Fig. 8(c) (d) shows that T-LSTM

is also better at predicting the descriptions, which reveals

that location information helps to improve captioning. Al-

though bounding box prediction does not feed information

to the captioning process in the forward pass, it does in-

fluence captioning through back-propagation in the training

stage. Considering all these factors, we believe that sepa-

rating the hidden space using T-LSTM is most suitable for

2198

Page 7: Dense Captioning With Joint Inference and Visual Contextopenaccess.thecvf.com/content_cvpr_2017/papers/... · Dense Captioning with Joint Inference and Visual Context Linjie Yang

(a) (b) (c) (d)

Figure 8: Qualitative comparisons between baseline and T-LSTM. In each image, the yellow box, the red box, and the blue

box are the region proposal, the prediction of the baseline model, and the prediction of the T-LSTM model, respectively.

the joint inference of caption and location.

4.4. Integrated models

We evaluate the integrated models with different de-

signs for both joint inference and context fusion in this sec-

tion. For joint inference models, we evaluate three vari-

ants: S-LSTM, SC-LSTM, and T-LSTM. For context fu-

sion, we compare the different settings proposed in Sec-

tion 3.3, where we evaluate early-fusion and late-fusion

with different fusion operators: concatenation, summation,

and multiplication. For early-fusion with concatenation, we

plug in a fully-connected layer after the concatenated fea-

ture to reduce it to the same input dimension as the LSTM.

The mAP results of different variants of models are shown

in Tab. 3.

Effectiveness of context fusion. In all models, context

information helps to improve mAP ranging from 0.07 (S-

LSTM, early-fusion, summation) to 1.10 (S-LSTM, late-

fusion, multiplication). The three types of fusion meth-

ods all yield improvements in mAP for different models.

Generally, concatenation and multiplication are more ef-

fective than summation, but the margin is subtle. With T-

LSTM and late-fusion with multiplication, we obtain the

best mAP performance 8.60 in this set of experiments. We

refer to this model as T-LSTM-mult for brevity in the re-

maining text. Fig. 9 shows example predictions for compar-

ison of T-LSTM without context fusion and T-LSTM-mult.

In Fig. 9(a) (b) (c), T-LSTM-mult gives a better caption than

the model without context. Without context, these objects

are very hard to recognize even by humans. We can also

observe from these examples that the context information

employed by the model is not limited to the surrounding

part of the region proposal, but from the whole image. In

Fig. 9(d), the context model interestingly gives an incorrect

but reasonable prediction since it is misled by the context

which is full of sheep.

Late-fusion is better than early-fusion. Comparing

early-fusion and late-fusion of context information, we find

that late-fusion is better than early-fusion for all pairs of

corresponding models. Also, early fusion only outperforms

Table 5: The chosen hyper-parameters and the performance

on Visual Genome V1.0 and V1.2 respectively.

#proposal NMS r1 NMS r2 mAP

V1.0100 0.5 0.4 8.67

300 0.6 0.5 9.31

V1.2100 0.5 0.5 9.47

300 0.6 0.5 9.96

its no-context counterparts by a small margin. One disad-

vantage of early-fusion is that it directly combines the lo-

cal and context features that have quite differing visual ele-

ments, making it unlikely able to decorrelate the visual ele-

ment into the local region or the context region in the later

stages of the model.

Intermediate location predictions. Since we only add

the regression target to the last time step of the location-

LSTM, it is not clear what the bounding box predictions

from the previous time steps will be like. We test the bound-

ing box predictions from these time steps, and find them to

be fairly good. Fig. 10 shows examples of the predicted

bounding box location at different time steps for the T-

LSTM-mult model. Generally, the bounding box prediction

at the first time step is already close to the region of inter-

est. As words are fed into the location-LSTM, it gradually

adjusts the bounding box to a tight localization of the object

being described.

4.5. Results on Visual Genome V1.2

We also conduct experiments on Visual Genome V1.2

using the same train/val/test split as V1.0. The mAP perfor-

mances are shown in Tab. 4. Here, we see similar results as

on V1.0, which further verifies the advantage of T-LSTM

over S-LSTM (mAP 8.16 vs 6.44 for no-context), and that

context fusion greatly improves performance for both mod-

els. For context fusion, we can see that the T-LSTM model

with late concatenation achieves the best result with mAP

9.03. We refer to this model as T-LSTM-concat. Over-

all, the accuracies are higher than those on Visual Genome

V1.0, likely due to the cleaner ground truth labels.

2199

Page 8: Dense Captioning With Joint Inference and Visual Contextopenaccess.thecvf.com/content_cvpr_2017/papers/... · Dense Captioning with Joint Inference and Visual Context Linjie Yang

(a) (b) (c) (d)

Figure 9: Qualitative comparisons of T-LSTM and T-LSTM-mult. In each image, the red box and the blue box are the

prediction of the no-context and context model, respectively, with their predicted captions. Region proposals are omitted.

Figure 10: Bounding box predictions at different time steps of the caption using T-LSTM-mult. In each image, different

colors of boxes denote the outputs of different time steps, with the brighter the color the later in time. The corresponding

words fed into the location-LSTM are shown in the legends. <SOS> is the start-of-sentence token.

4.6. Best practice: hyper­parameters

The evaluation pipeline for dense captioning, a two-stage

process of target prediction (region proposal and region de-

scription along with location refinement), involves several

hyper-parameters that can influence the accuracy. These pa-

rameters include the number of proposals given by the RPN

and the IoU ratio used by NMS both in the RPN and the

final prediction. For these parameters, we use the same set-

tings as [20] for all evaluations above. However, we are also

interested in the impact of these parameters on our results.

Parameters such as number of proposals is worth investigat-

ing because it can be used to find a trade-off between speed

and performance. Also, the NMS thresholds used by [20]

seem to overly suppress the predicted bounding box, espe-

cially since the ground truth regions are very dense (Fig. 2).

We use T-LSTM-mult for Visual Genome V1.0 and T-

LSTM-concat for V1.2 as prototypes and find the best

hyper-parameters for each by validating on the validation

set. For the number of proposals given by the RPN, we

validate between 100 and 300 proposals. We also validate

to find the optimal IoU ratios used in the NMS thresholds

for RPN and for final prediction, denoted as NMS r1 and

NMS r2, respectively. NMS r1 is chosen from the range

0.4 ∼ 0.9, and NMS r2 is chosen from the range 0.3 ∼ 0.8.

The results and corresponding optimal hyper-parameter set-

tings are shown in Tab. 5.

With the validated hyper-parameters, we achieve even

better mAP performance with 9.31 on Visual Genome V1.0

and 9.96 on Visual Genome V1.2 using 300 proposals,

which sets the new state-of-the-art. With only 100 propos-

als, our model achieves mAP 8.67 on Visual Genome V1.0

and 9.47 on Visual Genome V1.2. Comparing the running

times, we find that a 600 × 720 image takes 350ms and

200ms for 300 and 100 proposals on a GTX TITAN GPU,

respectively. The LSTM computations take around 30% of

the total time consumption. Our implementation is devel-

oped using Caffe [16].

5. Conclusions

In this work, we have proposed a novel model structure

which incorporates two ideas, joint inference and context

fusion, to address specific challenges in dense captioning.

To find an exact model realization incorporating these two

approaches, we design our model step by step and propose

different variants for each component. We evaluate the dif-

ferent models extensively, and gain intuitions on the effec-

tiveness of each component. Finally, we find a model which

utilizes the two approaches effectively and achieves state-

of-the-art performance on the Visual Genome dataset. The

feature representation learned by these models can poten-

tially benefit other computer vision tasks requiring dense

visual understanding such as object detection, semantic seg-

mentation, and caption localization. The extensive compar-

ison of different model structures can hopefully help guide

model design in other tasks involving sequential modeling.

2200

Page 9: Dense Captioning With Joint Inference and Visual Contextopenaccess.thecvf.com/content_cvpr_2017/papers/... · Dense Captioning with Joint Inference and Visual Context Linjie Yang

References

[1] S. Antol, A. Agrawal, J. Lu, M. Mitchell, D. Batra,

C. Lawrence Zitnick, and D. Parikh. Vqa: Visual question

answering. In ICCV, 2015.

[2] S. Banerjee and A. Lavie. Meteor: An automatic metric for

mt evaluation with improved correlation with human judg-

ments. In ACL Workshop, 2005.

[3] S. Bell, C. L. Zitnick, K. Bala, and R. Girshick. Inside-

outside net: Detecting objects in context with skip pooling

and recurrent neural networks. CVPR, 2016.

[4] Y. Bengio, R. Ducharme, P. Vincent, and C. Jauvin. A neu-

ral probabilistic language model. JMLR, 3(Feb):1137–1155,

2003.

[5] X. Chen, H. Fang, T.-Y. Lin, R. Vedantam, S. Gupta,

P. Dollar, and C. L. Zitnick. Microsoft coco captions:

Data collection and evaluation server. arXiv preprint

arXiv:1504.00325, 2015.

[6] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-

Fei. Imagenet: A large-scale hierarchical image database. In

CVPR, 2009.

[7] S. K. Divvala, D. Hoiem, J. H. Hays, A. A. Efros, and

M. Hebert. An empirical study of context in object detec-

tion. In CVPR, 2009.

[8] J. Donahue, L. Anne Hendricks, S. Guadarrama,

M. Rohrbach, S. Venugopalan, K. Saenko, and T. Dar-

rell. Long-term recurrent convolutional networks for visual

recognition and description. In CVPR, 2015.

[9] M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn,

and A. Zisserman. The PASCAL Visual Object Classes

Challenge 2012 (VOC2012) Results. http://www.pascal-

network.org/challenges/VOC/voc2012/workshop/index.html.

[10] H. Fang, S. Gupta, F. Iandola, R. K. Srivastava, L. Deng,

P. Dollar, J. Gao, X. He, M. Mitchell, J. C. Platt, et al. From

captions to visual concepts and back. In CVPR, 2015.

[11] R. Girshick. Fast r-cnn. In ICCV, 2015.

[12] R. Girshick, J. Donahue, T. Darrell, and J. Malik. Rich fea-

ture hierarchies for accurate object detection and semantic

segmentation. In CVPR, 2014.

[13] S. Hochreiter and J. Schmidhuber. Long short-term memory.

Neural computation, 9(8):1735–1780, 1997.

[14] R. Hu, H. Xu, M. Rohrbach, J. Feng, K. Saenko, and T. Dar-

rell. Natural language object retrieval. CVPR, 2016.

[15] M. Jaderberg, K. Simonyan, A. Zisserman, et al. Spatial

transformer networks. In NIPS, 2015.

[16] Y. Jia, E. Shelhamer, J. Donahue, S. Karayev, J. Long, R. Gir-

shick, S. Guadarrama, and T. Darrell. Caffe: Convolu-

tional architecture for fast feature embedding. arXiv preprint

arXiv:1408.5093, 2014.

[17] H. Jiang and E. G. Learned-Miller. Face detection with the

faster R-CNN. CoRR, abs/1606.03473, 2016.

[18] J. S. Jifeng Dai, Kaiming He. Instance-aware semantic seg-

mentation via multi-task network cascades. In CVPR, 2016.

[19] J. Jin, K. Fu, R. Cui, F. Sha, and C. Zhang. Aligning where to

see and what to tell: image caption with region-based atten-

tion and scene factorization. CoRR, abs/1506.06272, 2015.

[20] J. Johnson, A. Karpathy, and L. Fei-Fei. Densecap: Fully

convolutional localization networks for dense captioning.

CVPR, 2016.

[21] A. Karpathy and L. Fei-Fei. Deep visual-semantic align-

ments for generating image descriptions. In CVPR, 2015.

[22] A. Karpathy, A. Joulin, and L. Fei Fei. Deep fragment em-

beddings for bidirectional image sentence mapping. In NIPS,

2014.

[23] R. Krishna, Y. Zhu, O. Groth, J. Johnson, K. Hata, J. Kravitz,

S. Chen, Y. Kalantidis, L.-J. Li, D. A. Shamma, et al.

Visual genome: Connecting language and vision using

crowdsourced dense image annotations. arXiv preprint

arXiv:1602.07332, 2016.

[24] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-

based learning applied to document recognition. Proceed-

ings of the IEEE, 86(11):2278–2324, 1998.

[25] Y. Li, W. Ouyang, and X. Wang. Vip-cnn: A visual phrase

reasoning convolutional neural network for visual relation-

ship detection. CoRR, abs/1702.07191, 2017.

[26] W. Liu, D. Anguelov, D. Erhan, C. Szegedy, and S. Reed.

Ssd: Single shot multibox detector. arXiv preprint

arXiv:1512.02325, 2015.

[27] J. Long, E. Shelhamer, and T. Darrell. Fully convolutional

networks for semantic segmentation. In CVPR, 2015.

[28] C. Lu, R. Krishna, M. Bernstein, and L. Fei-Fei. Visual rela-

tionship detection with language priors. In ECCV, 2016.

[29] M. Malinowski, M. Rohrbach, and M. Fritz. Ask your neu-

rons: A neural-based approach to answering questions about

images. In ICCV, 2015.

[30] J. Mao, J. Huang, A. Toshev, O. Camburu, A. L. Yuille, and

K. Murphy. Generation and comprehension of unambiguous

object descriptions. In CVPR, 2016.

[31] J. Mao, W. Xu, Y. Yang, J. Wang, and A. L. Yuille. Explain

images with multimodal recurrent neural networks. arXiv

preprint arXiv:1410.1090, 2014.

[32] T. Mikolov, M. Karafiat, L. Burget, J. Cernocky, and S. Khu-

danpur. Recurrent neural network based language model. In

Interspeech, 2010.

[33] R. Mottaghi, X. Chen, X. Liu, N.-G. Cho, S.-W. Lee, S. Fi-

dler, R. Urtasun, and A. Yuille. The role of context for object

detection and semantic segmentation in the wild. In CVPR,

2014.

[34] V. K. Nagaraja, V. I. Morariu, and L. S. Davis. Modeling

context between objects for referring expression understand-

ing. In ECCV, 2016.

[35] K. Papineni, S. Roukos, T. Ward, and W.-J. Zhu. Bleu: a

method for automatic evaluation of machine translation. In

ACL, 2002.

[36] J. Redmon, S. Divvala, R. Girshick, and A. Farhadi. You

only look once: Unified, real-time object detection. CVPR,

2016.

[37] S. Ren, K. He, R. Girshick, and J. Sun. Faster r-cnn: Towards

real-time object detection with region proposal networks. In

NIPS, 2015.

[38] K. Simonyan and A. Zisserman. Very deep convolutional

networks for large-scale image recognition. arXiv preprint

arXiv:1409.1556, 2014.

2201

Page 10: Dense Captioning With Joint Inference and Visual Contextopenaccess.thecvf.com/content_cvpr_2017/papers/... · Dense Captioning with Joint Inference and Visual Context Linjie Yang

[39] I. Sutskever, J. Martens, and G. E. Hinton. Generating text

with recurrent neural networks. In ICML, 2011.

[40] R. Vedantam, C. Lawrence Zitnick, and D. Parikh. Cider:

Consensus-based image description evaluation. In CVPR,

2015.

[41] O. Vinyals, A. Toshev, S. Bengio, and D. Erhan. Show and

tell: A neural image caption generator. In ICCV, 2015.

[42] P. J. Werbos. Generalization of backpropagation with appli-

cation to a recurrent gas market model. Neural Networks,

1(4):339–356, 1988.

[43] K. Xu, J. Ba, R. Kiros, K. Cho, A. Courville, R. Salakhut-

dinov, R. S. Zemel, and Y. Bengio. Show, attend and tell:

Neural image caption generation with visual attention. In

NIPS, 2015.

[44] P. Young, A. Lai, M. Hodosh, and J. Hockenmaier. From im-

age descriptions to visual denotations: New similarity met-

rics for semantic inference over event descriptions. TACL,

2:67–78, 2014.

[45] L. Yu, P. Poirson, S. Yang, A. C. Berg, and T. L. Berg. Mod-

eling context in referring expressions. In ECCV, 2016.

2202