spatial information guided convolution for real-time rgbd ... · how to effectively use the extra...

11
1 Spatial Information Guided Convolution for Real-Time RGBD Semantic Segmentation Lin-Zhuo Chen, Zheng Lin, Ziqin Wang, Yong-Liang Yang, and Ming-Ming Cheng Abstract—3D spatial information is known to be beneficial to the semantic segmentation task. Most existing methods take 3D spatial data as an additional input, leading to a two-stream segmentation network that processes RGB and 3D spatial infor- mation separately. This solution greatly increases the inference time and severely limits its scope for real-time applications. To solve this problem, we propose Spatial information guided Convolution (S-Conv), which allows efficient RGB feature and 3D spatial information integration. S-Conv is competent to infer the sampling offset of the convolution kernel guided by the 3D spatial information, helping the convolutional layer adjust the receptive field and adapt to geometric transformations. S-Conv also incorporates geometric information into the feature learning process by generating spatially adaptive convolutional weights. The capability of perceiving geometry is largely enhanced without much affecting the amount of parameters and computational cost. We further embed S-Conv into a semantic segmentation network, called Spatial information Guided convolutional Net- work (SGNet), resulting in real-time inference and state-of-the- art performance on NYUDv2 and SUNRGBD datasets. Index Terms—Spatial information, geometric transformations, receptive field, RGBD semantic segmentation. I. I NTRODUCTION With the development of 3D sensing technologies, RGBD data with spatial information (depth, 3D coordinates) is easily accessible. As a result, RGBD semantic segmentation for high-level scene understanding becomes extremely important, benefiting a wide range of applications such as automatic driving [1], SLAM [2], and robotics. Due to the effective- ness of Convolutional Neural Network (CNN) and additional spatial information, recent advances demonstrate enhanced performance on indoor scene segmentation tasks [3], [4]. Nevertheless, there remains a significant challenge caused by the complexity of the environment and the extra efforts for considering spatial data, especially for applications that require real-time inference. A common approach treats 3D spatial information as an additional input of the network to extract features, followed by combining the features of RGB images [3], [5]–[9], as shown in Fig. 1 (a). This approach achieves promising results at the cost of significantly increasing the parameter number and computational time, thus being unsuitable for real-time tasks. Meanwhile, several works [3], [5], [8], [10], [11] encode raw spatial information into three channels (HHA) composed of horizontal disparity, height above ground, and norm angle. LZ Chen, Z Lin, MM Cheng (corresponding author, [email protected]) are with Colledge of Computer Science, Nankai University, Tianjin, China. Ziqin Wang are with University of Sydney. YL Yang are with University of Bath. + Guide Spatial information guided Convolution (S-Conv) Depth Image Loss Image Depth Loss (a) (b) Fig. 1: The network architecture of different approaches. (a) The conventional two-stream structure. (b) The proposed SGNet. It can be seen that the approach in (a) largely increases parameter number and inference time due to processing spatial information, thus less suitable for real-time applications. We replace the convolution with our S-Conv in (b) where the kernel distribution and weights of the convolution are adaptive to the spatial information. S-Conv greatly enhances the spatial awareness of the network with few additional parameters and computations, thus can efficiently utilize spatial information. Best viewed in color. However, the conversion from raw data to HHA is also time- consuming [8]. It is worth noting that indoor scenes have more complex spatial relations than outdoor scenes. This requires a stronger adaptive ability of the network to deal with geometric transfor- mations. However, due to the fixed structure of the convolution kernel, the 2D convolution in the aforementioned methods cannot well adapt to spatial transformation and adjust the receptive field inherently, limiting the accuracy of semantic segmentation. Although alleviation can be made by revised pooling operation and prior data augmentation [12], [13], a better spatially adaptive sampling mechanism for conducting convolution is still desirable. Moreover, the color and texture of objects in indoor scenes are not always representative. Instead, the geometry structure often plays a vital role in semantic segmentation. For example, to recognize the fridge and wall, the geometric structure is the primary cue due to the similar texture. However, such arXiv:2004.04534v1 [cs.CV] 9 Apr 2020

Upload: others

Post on 21-Jul-2020

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Spatial Information Guided Convolution for Real-Time RGBD ... · How to effectively use the extra geometry information (depth, 3D coordinates) is the key of RGBD semantic seg-mentation

1

Spatial Information Guided Convolution forReal-Time RGBD Semantic Segmentation

Lin-Zhuo Chen, Zheng Lin, Ziqin Wang, Yong-Liang Yang, and Ming-Ming Cheng

Abstract—3D spatial information is known to be beneficialto the semantic segmentation task. Most existing methods take3D spatial data as an additional input, leading to a two-streamsegmentation network that processes RGB and 3D spatial infor-mation separately. This solution greatly increases the inferencetime and severely limits its scope for real-time applications.To solve this problem, we propose Spatial information guidedConvolution (S-Conv), which allows efficient RGB feature and3D spatial information integration. S-Conv is competent to inferthe sampling offset of the convolution kernel guided by the 3Dspatial information, helping the convolutional layer adjust thereceptive field and adapt to geometric transformations. S-Convalso incorporates geometric information into the feature learningprocess by generating spatially adaptive convolutional weights.The capability of perceiving geometry is largely enhanced withoutmuch affecting the amount of parameters and computationalcost. We further embed S-Conv into a semantic segmentationnetwork, called Spatial information Guided convolutional Net-work (SGNet), resulting in real-time inference and state-of-the-art performance on NYUDv2 and SUNRGBD datasets.

Index Terms—Spatial information, geometric transformations,receptive field, RGBD semantic segmentation.

I. INTRODUCTION

With the development of 3D sensing technologies, RGBDdata with spatial information (depth, 3D coordinates) is easilyaccessible. As a result, RGBD semantic segmentation forhigh-level scene understanding becomes extremely important,benefiting a wide range of applications such as automaticdriving [1], SLAM [2], and robotics. Due to the effective-ness of Convolutional Neural Network (CNN) and additionalspatial information, recent advances demonstrate enhancedperformance on indoor scene segmentation tasks [3], [4].Nevertheless, there remains a significant challenge caused bythe complexity of the environment and the extra efforts forconsidering spatial data, especially for applications that requirereal-time inference.

A common approach treats 3D spatial information as anadditional input of the network to extract features, followedby combining the features of RGB images [3], [5]–[9], asshown in Fig. 1 (a). This approach achieves promising resultsat the cost of significantly increasing the parameter numberand computational time, thus being unsuitable for real-timetasks. Meanwhile, several works [3], [5], [8], [10], [11] encoderaw spatial information into three channels (HHA) composedof horizontal disparity, height above ground, and norm angle.

LZ Chen, Z Lin, MM Cheng (corresponding author, [email protected])are with Colledge of Computer Science, Nankai University, Tianjin, China.

Ziqin Wang are with University of Sydney.YL Yang are with University of Bath.

+

Guide

Spatial information guided Convolution (S-Conv)

Depth

Image

Loss

Image

Depth

Loss

(a)

(b)Fig. 1: The network architecture of different approaches.(a) The conventional two-stream structure. (b) The proposedSGNet. It can be seen that the approach in (a) largely increasesparameter number and inference time due to processing spatialinformation, thus less suitable for real-time applications. Wereplace the convolution with our S-Conv in (b) where thekernel distribution and weights of the convolution are adaptiveto the spatial information. S-Conv greatly enhances the spatialawareness of the network with few additional parameters andcomputations, thus can efficiently utilize spatial information.Best viewed in color.

However, the conversion from raw data to HHA is also time-consuming [8].

It is worth noting that indoor scenes have more complexspatial relations than outdoor scenes. This requires a strongeradaptive ability of the network to deal with geometric transfor-mations. However, due to the fixed structure of the convolutionkernel, the 2D convolution in the aforementioned methodscannot well adapt to spatial transformation and adjust thereceptive field inherently, limiting the accuracy of semanticsegmentation. Although alleviation can be made by revisedpooling operation and prior data augmentation [12], [13], abetter spatially adaptive sampling mechanism for conductingconvolution is still desirable.

Moreover, the color and texture of objects in indoor scenesare not always representative. Instead, the geometry structureoften plays a vital role in semantic segmentation. For example,to recognize the fridge and wall, the geometric structure isthe primary cue due to the similar texture. However, such

arX

iv:2

004.

0453

4v1

[cs

.CV

] 9

Apr

202

0

Page 2: Spatial Information Guided Convolution for Real-Time RGBD ... · How to effectively use the extra geometry information (depth, 3D coordinates) is the key of RGBD semantic seg-mentation

2

spatial information is ignored by 2D convolution on RGB data.The depth-aware convolution [14] is proposed to address thisproblem. It forces pixels with similar depths as the center ofthe kernel to have higher weight than others. Nevertheless, thisprior is handcrafted and may lead to sub-optimal results.

It can be seen that there is a contradiction between thefixed structure of 2D convolution and the varying spatial trans-formation, along with the efficiency bottleneck of separatelyprocessing RGB and spatial data. To overcome the limitationsmentioned above, we propose a novel operation, called Spatialinformation guided Convolution(S-Conv), which adaptivelychanges according to the spatial information (see Fig. 1 (b)).

Specifically, this operation can generate convolution kernelswith different sampling distributions adapting to spatial infor-mation, boosting the spatial adaptability and the receptive fieldregulation of the network. Furthermore, S-Conv establishesa link between the convolution weights and the underlyingspatial relationship with their corresponding pixel, incorporat-ing the geometric information into the convolution weights tobetter capture the spatial structure of the scene.

The proposed S-Conv is light yet flexible and achievessignificant performance improvements with only few addi-tional parameters and computation costs, making it suitablefor real-time applications. We conduct extensive experimentsto demonstrate the effectiveness and efficiency of S-Conv.We first design the ablation study and compare S-Conv withdeformable convolution [12], [13] and depth-aware convo-lution [14], exhibiting the advantages of S-Conv. We alsoverify the applicability of S-Conv to spatial transformationsby testing its influence on different types of spatial data withdepth, HHA and 3D coordinates. We demonstrate that spatialinformation is more suitable to generate offset than RGBfeature which is used by deformable convolution [12], [13].Finally, benefiting from the adaptability to spatial transfor-mation and the effectiveness of perceiving spatial structure,our network equipped with S-Conv, named Spatial informa-tion Guided convolutional Network (SGNet), achieves high-quality results with real-time inference on NYUDv2 [15] andSUNRGBD [16], [17] datasets.

We highlight our contributions as follows:• We propose a novel S-Conv operator that can adaptively

adjust receptive field while effectively adapting to spa-tial transformation, and can perceive intricate geometricpatterns with low cost.

• Based on S-Conv, we propose a new SGNet that achievescompetitive RGBD segmentation performance in real-time on NYUDv2 [15] and SUNRGBD [16], [17]datasets.

II. RELATED WORK

A. Semantic Segmentation

The recent advances of semantic segmentation benefit alot from the development of convolutional neural network(CNN) [18], [19]. FCN [3] is the pioneer of leveragingCNN for semantic segmentation. It leads to convincing resultsand serves as the basic framework for many tasks. Withthe research efforts in the field, the recent methods can

be classified into two categories according to the networkarchitecture, including atrous convolution-based methods [4],[20], and encoder-decoder based methods [21]–[24].Atrous convolution: The standard approach relies on strideconvolutions or poolings to reduce the output stride of theCNN backbone and enables a large receptive field. However,the resolution of the resulting feature map is reduced [25],and many details are lost. One approach exploits atrousconvolution to alleviate the conflict by enhancing the receptivefield while keeping the resolution of the feature map [4], [20],[22], [25], [26]. We use atrous convolution based backbone inthe proposed SGNet.Encoder-decoder architecture: The other approach utilizesthe encoder-decoder structure [21]–[24], [27], which learnsa decoder to recover the prediction details gradually. De-convNet [24] employs a series of deconvolutional layers toproduce a high-resolution prediction. SegNet [23] achievesbetter results by using pooling indices in the encoder to guidethe recovery process in the decoder. RefineNet [21] fuses low-level features in the encoder with the decoder to refine theprediction. While this method can achieve more precise results,it requires longer inference time.

B. RGBD Semantic Segmentation

How to effectively use the extra geometry information(depth, 3D coordinates) is the key of RGBD semantic seg-mentation. A number of works focus on how to extract moreinformation from geometry, which is treated as additionalinput in [6]–[9], [28]. Two-stream network is used in [5], [7]–[9], [11] to process RGB image and geometry informationseparately, and combines the two results in the last layer. Thesemethods achieve promising results at the expense of doublingthe parameters and computational cost. 3D CNNs or 3D KNNgraph networks are also used to take geometry information intoaccount [29]–[31]. Besides, various deep learning methods on3D point cloud [32]–[37] are also explored. However, thesemethods cost a lot of memory and are computationally expen-sive. Another stream incorporates geometric information intoexplicit operations. Cheng et al. [38] use geometry informationto build a feature affinity matrix acting in average poolingand up-pooling. Lin et al. [39] splits the image into differentbranches based on geometry information. Wang et al. [14]propose Depth-aware CNN, which adds depth prior to theconvolutional weights. Although it improves feature extractionby convolution, the prior is handcrafted but not learned fromdata. Other approaches, such as multi-task learning [6], [40]–[44] or spatial-temporal analysis [45], are further used toimprove segmentation accuracy. The proposed S-Conv aimsto efficiently utilize spatial information to improve the featureextraction ability. It can significantly enhance the performancewith high efficiency due to using only a small amount ofparameters.

C. Dynamic structure in CNN

Using dynamic structure to deal with varying input ofCNN has also been explored. Dilation Convolution is used

Page 3: Spatial Information Guided Convolution for Real-Time RGBD ... · How to effectively use the extra geometry information (depth, 3D coordinates) is the key of RGBD semantic seg-mentation

3

Spatial Projector

Spatial information

Offset Generator

S

X

Feature map

Feature map operation

Pixel operation

Sample

Convolution

Multiplication

Weight Generator

S

S

X f

.3Equ .4Equ.6Equ

Fig. 2: The illustration of the Spatial information guided Convolution (S-Conv). Firstly, the input 3D spatial informationis projected by the spatial projector to match the input feature map. Secondly, the adaptive convolution kernel distributionis generated by the offset generator. Finally, the projected spatial information is sampled according to the kernel distributionand fed into the weight generator to generate adaptive convolution weights.

in [4], [20] to increase the receptive field size without reduc-ing feature map resolution. Spatial transformer network [46]adapts spatial transformation by warping feature map. Dy-namic filter [47] adaptively changes its weights accordingto the input. Besides, self-attention based methods [48]–[50]generate attention maps from the intermediate feature map toadjust response at each location or capture long-range con-textual information adaptively. Some generalizations of con-volution from 2D image to 3D point cloud are also presented.PointCNN [37] is a seminal work that enables CNN on a setof unordered 3D points. There are other improvements [34]–[36] on utilizing neural networks to effectively extract deepfeatures from 3D point sets. Deformable convolution [12],[13] can generate different distribution with adaptive weights.Nevertheless, their input is an intermediate feature map ratherthan spatial information. Our work experimentally verifies thatbetter results can be obtained based on spatial information inSec. IV.

III. S-CONV AND SGNET

In this section, we first elaborate on the details of Spatialinformation guided Convolution (S-Conv), which is a general-ization of conventional RGB-based convolution by involvingspatial information in the RGBD scenario. Then, we discussthe relation between our S-Conv and other approaches. Finally,we describe the network architecture of Spatial informationGuided convolutional Network (SGNet), which is equippedwith S-Conv for RGBD semantic segmentation.

A. Spatial information guided Convolution

For completeness, we first review the conventional convolu-tion operation. We use Ai(j),A ∈ Rc×h×w to denote a tensor,where i is the index corresponding to the first dimension, andj ∈ R2 indicates the two indices for the second and thirddimensions. Non-scalar values are highlighted in bold forconvenience.

For an input feature map X ∈ Rc×h×w. We describe itin 2D for simplicity, thus X ∈ R1×h×w. Note that it is

straightforward to extend to the 3D case. The conventionalconvolution applied on X to get Y can be formulated as thefollowing:

Y(p) =

kh∗kw∑i=1

Wi ·X(p + di), (1)

where W ∈ Rkh∗kw is the convolution kernel with size kh ∗kw, p ∈ R2 is the 2D convolution center, d ∈ Rkh∗kw×2

denotes the kernel distribution around p. For 3×3 convolution:

d = {[−1,−1], [−1, 0], ..., [0, 1], [1, 1]}. (2)

From the above equation, we can see that the convolutionkernel is constant over X. In other words, W and d are fixed,meaning the convolution is content-agnostic.

In the RGBD context, we want to involve 3D spatialinformation efficiently by using adaptive convolution kernels.

We first generate the offset according to the spatial in-formation, then use the spatial information corresponding tothe given offset to generate new spatially adaptive weights.Our S-Conv requires two inputs. One is the feature map Xwhich is the same as conventional convolution. The other isthe spatial information S ∈ Rc′×h×w. In practice, S can beHHA (c′ = 3), 3D coordinates (c′ = 3), or depth (c′ = 1).The method of encoding depth into 3D coordinates and HHAis the same as [31]. Note that the input spatial information isnot included in the feature map.

As the first step of S-Conv, we project the input spatialinformation into a high-dimensional feature space, which canbe expressed as:

S′ = φ(S). (3)

where φ is a spatial transformation function, and S′ ∈R64×h×w, which has a higher dimension than S.

Then, we take the transformed spatial information S′ intoconsideration, perceive its geometric structure, and generatethe distribution (offset of pixel coordinate in x− and y−axis)

Page 4: Spatial Information Guided Convolution for Real-Time RGBD ... · How to effectively use the extra geometry information (depth, 3D coordinates) is the key of RGBD semantic seg-mentation

4

Loca

l patch

2D Conv view

S-Conv view

Fig. 3: The illustration of weights W in 2D convolutionand W d in S-Conv. The yellow dot indicates the point whosespatial position changes along the arrow. Illustration of 2Dconvolution is on the top, and S-Conv is on the bottom. Theconventional 2D convolution operation orderly places localpoints in a regular grid with fixed weights, while ignoringthe spatial information. We can see that the spatial positionvariation of the yellow point can not be reflected in the weight.Our S-Conv can be regarded as placing a local patch intoa weight space, which is generated by the spatial guidanceof that patch. Hence the weight of each point establishes alink with its spatial location, effectively capturing the spatialvariation of the local patch. The spatial relationship betweenthe yellow point and other points can be reflected in theadaptive weights.

of convolution kernels at different p. This processes can beexpressed as:

∆d = η(S′), (4)

where ∆d ∈ Rkh∗kw×h′×w′×2, h′, w′ represent the featuremap size after convolution, and kh, kw are the kernel size. ηis a non-linear function which can be implemented by a seriesof convolutions.

After generating the distribution of kernel for each possiblep using ∆d(p), we boost its feature extraction ability byestablishing the link between the geometric structure and theconvolution weight. More specifically, we sample the geomet-ric information of the pixels corresponding to the convolutionkernel after shifting:

Sd(p) = {S′(p + di + ∆di(p))|i=1,2,...,kh∗kw}, (5)

where ∆d(p) is the spatial distribution of convolution kernelsat p. Sd(p) ∈ R64∗kh∗kw is the spatial information corre-sponding to the feature map of the convolution kernel centeredon p after transformation.

Finally, we generate convolution weights according to thefinal spatial information as the following:

Wd(p) = σ(f(Sd(p))) ·W, (6)

where f is a non-linear function that can be implemented as aseries of convolution layers with non-linear activation function,W ∈ Rkh∗kw indicates the convolution weights, which can beupdated by the gradient descent algorithm. Wd(p) ∈ Rkh∗kw

denotes the spatially adaptive weights for convolution centeredat p.

Overall, our generalized S-Conv is formulated as:

Y(p) =

kh∗kw∑i=1

Wdi (p) ·X(p + di + ∆di(p)). (7)

We can see that Wdi establishes the correlation between spatial

information and convolution weights. Moreover, convolutionkernel distribution is also relevant to the spatial informationthrough ∆d. Note that Wd

i (p) and ∆di(p) are not constant,meaning the generalized convolution is adaptive to differentp. Also, as ∆d is typically fractional, we use bilinear interpo-lation to compute X(p + di + ∆di(p)) as in [12], [46]. Themain formulae discussed above are labeled in Fig. 2.

B. Relation to other approaches

2D convolution is the special case of the proposed S-Conv without geometry information and corresponding offset.Specifically, without geometry information, the center pointand its neighboring points have fixed positional relation inimage space. And we do not need Wd to capture the varyingspatial relation. This also shows that our S-Conv has very goodcompatibility to handle the 2D case. While for the RGBD case,our S-Conv can extract feature at the point level and is notlimited to the discrete grid by introducing spatially adaptiveweights as shown in Fig. 3. Deformable convolution [12], [13]also alleviates this problem by generating different distributionweights. Nevertheless, their distributions are inferred from2D feature maps instead of 3D spatial information as in ourcase. We will verify through experiments that our methodachieves better results than deformable convolution [12], [13].Compared with the 3D KNN graph-based method, our S-Convselects neighboring pixels adaptively instead of using the KNNgraph, which is not flexible and computationally expensive.

C. SGNet architecture

Our semantic segmentation network, called SGNet, isequipped with S-Conv and consists of a backbone and decoder.The structure of SGNet is illustrated in Fig. 4. We useResNet101 [51] as our backbone, and replace the first andthe last two conventional convolutions (3 × 3 filter) of eachlayer with our S-Conv. We add a series of convolutions toextract the feature further and then use bilinear up-samplingto produce the final segmentation probability map, whichcorresponds to the decoder part of the SGNet. The φ inEqu. (3) is implemented as three 3 × 3 convolution layers,i.e. Conv(3, 64) - Conv(64, 64) - Conv(64, 64) with non-linearactivation function. The η in Equ. (4) and the f in Equ. (6) areimplemented as single and two convolution layers separately.The S-Conv implementation is modified from deformableconvolution [12], [13] We add deep supervision between layer3 and layer 4 to improve the network optimization capability,which is the same as PSPNet [27].

Page 5: Spatial Information Guided Convolution for Real-Time RGBD ... · How to effectively use the extra geometry information (depth, 3D coordinates) is the key of RGBD semantic seg-mentation

5

U

|---------------------------Backbone-----------------------------------|----------Decoder--------|

Max-Pooling Convolution S-Convolution U Upsample

|------Conv------|--Layer1--|--Layer2--|--Layer3--|--Layer4--|

...

----Final Result-----||------RGB-Image------

Fig. 4: The network architecture of SGNet equipped with S-Conv for RGBD semantic segmentation. The SGNet consistsof a backbone network and a decoder. The deep supervision is added between layer 3 and layer 4 to improve networkoptimization.

IV. EXPERIMENTS

In this section, we first validate the performance of S-Convby analyzing its usage in different layers; conducting ablationstudy/comparison with its alternatives; and evaluating resultsof using different input information to generate offset. Then wecompare our SGNet equipped with S-Conv with other state-of-the-art semantic segmentation methods on NYUDV2 andSUNRGBD datasets. Finally, we visualize the depth adaptivereceptive field in each layer and the segmentation results,demonstrating that the proposed S-Conv can well exploitspatial information.Datasets and metrics: We evaluate the performance of S-Conv operator and SGNet segmentation method on two publicdatasets:• NYUDv2 [15] : This dataset has 1,449 RGB images

with corresponding depth maps and pixel-wise labels. 795images are used for training, while 654 images are usedfor testing as in [52]. The 40-class settings are used forexperiments.

• SUN-RGBD [16], [17]: This dataset contains 10,335RGBD images with semantic labels organized in 37categories. 5,285 images are used for training, and 5050images are used for testing.

We use three common metrics for evaluation, includingpixel accuracy (Acc), mean accuracy (mAcc), and mean in-tersection over union (mIoU). The three metrics are definedas the following:

Acc =∑i

piig,

mAcc =1

pC,

mIoU =1

pC

∑i

piigi +

∑j pji − pii

,

(8)

where pij is the amount of pixels which are predicted as classj with ground truth i, pC is the number of classes, and gi isthe number of pixels whose ground truth class is i. The depth

map is used as the default format of spatial information unlessspecified otherwise.Implementation details: We use dilated ResNet101 [51] pre-trained on ImageNet [53] as the backbone network for featureextraction following [4], and the output stride is 16 by default.The whole system is implemented based on PyTorch. The SGDoptimizer is adopted for training with the learning rate policyas used in [4], [22]: lr = initial lr × (1 − iter

total iter )power,where the initial learning rate is 5e-3 for NYUDv2 and 1e-3 forSUNRGBD, and the weight decay is 5e-4. This learning policyupdates the learning rate for every 40 epochs for NYUDv2 and10 epochs for SUNRGBD. We use ReLU activation function,and the batch size is 8. Following [5], we employ general dataaugmentation strategies, including random scaling, randomcropping, and random flipping. The crop size is 480 × 640.During testing, we down-sample the image to the training cropsize, and its prediction map is up-sampled to the original size.We use cross-entropy loss in both datasets, and reweight [54]training loss of each class in SUNRGBD due to its extremelyunbalanced label distribution. We train the network by 500epochs for the NYUDv2 dataset and 200 epochs for theSUNRGBD dataset on two NVIDIA 1080Ti GPUs.

A. Analysis of S-Conv

We design ablation studies on NYUDv2 [15] dataset. TheResNet101 with a simple decoder and deep supervision is usedas the baseline.Replace convolution with S-Conv: We evaluate the effective-ness of S-Conv by replacing the conventional convolution (of3 × 3 filter) in different layers. We first replace convolutionin layer 3, then extend the explored rules to other layers. TheFPS (Frames per Second) is tested on NVIDIA 1080Ti withinput image size 425×560×3 following [14]. The results areshown in Tab. I.

We can draw the following two conclusions from the resultsin the Tab. I. 1) The inference speed of the baseline networkis fast, but its performance is poor. Replacing convolutionwith S-Conv can improve the results of the baseline network

Page 6: Spatial Information Guided Convolution for Real-Time RGBD ... · How to effectively use the extra geometry information (depth, 3D coordinates) is the key of RGBD semantic seg-mentation

6

wall floor

cabine

tbe

dcha

irsof

atab

ledo

or

window

books

helf

pictur

e

counte

rblin

dsdesk

shelve

scur

tain

dresse

rpill

owmirro

r

floorm

at

clothe

scei

lingbooks

fridge tv

pape

rtow

el

showerbo

xbo

ardpe

rsonsta

ndtoi

letsinklam

p

batht

ubbagoth

er1oth

er2oth

er35

0

5

10

15

20

25

mIo

U im

prov

emen

t (%

)

Fig. 5: Per-category IoU improvement of S-Conv on NYUDv2 dataset.

S-Conv layer3 0 layer3 1 layer3 2 layer3 20 layer3 21 layer3 22 other layers mIoU(%) param(M) FPS43.0 56.8 37

X 47.0 56.9 37Baseline X X X 46.6 57.2 36

(ResNet101) X X X 46.5 57.2 36X X X 47.8 57.2 36X X X X 49.0 58.3 28

TABLE I: The results of replacing convolution (of 3 × 3 filter) of different layers with S-Conv on NYUDv2 dataset.”layerx y” means the 3× 3 convolution of y-th residual block in x-th layer.

Model Acc. mAcc. mIoUBaseline 72.1 54.6 43.0Baseline+OG 73.9 58.2 46.3Baseline+SP+OG 75.2 60.0 48.4Baseline+SP+WG 74.5 58.4 46.8Baseline+SP+OG+WG 75.5 60.9 49.0

TABLE II: Ablation study of SGNet on NYUDv2 [15]dataset. DCV2: Deformable Convolution V2 [13], OG: Offsetgenerator of S-Conv, WG: Weight generator of S-Conv, SP:Spatial projection of S-Conv.

Model Acc. mAcc. mIoU.Baseline 72.1 54.6 43.0Baseline+DCV2 73.0 56.1 44.5Baseline+HHANet 73.5 56.8 45.4Baseline+DAC 73.8 57.1 45.4Baseline+SP+WG 74.5 58.4 46.8Baseline+S-Conv(SGNet) 75.5 60.9 49.0

TABLE III: The comparison results on NYUDv2 testdataset. DAC: Depth-aware Convolution [14]. SP: Spatialprojector in S-Conv. WG: Weight generator in S-Conv.

with a little bit more parameters and computational time. 2)In addition to the first convolution in layer 3 whose strideis 2, the effect of replacing the later convolution is better.The main reason would be that spatial information can betterguide down-sampling operation in the first convolution. Thuswe choose to replace the first convolution and the last two

Information Acc. mAcc. mIoU.Depth 75.5 60.9 49.0

RGB Feature 73.9 58.5 46.4HHA 75.7 60.8 48.9

3D coordinates 75.3 61.2 48.5

TABLE IV: Comparison of using different types of spatialinformation on NYUDv2 Dataset.

convolutions of each layer with S-Conv. We generalize therules found in layer 3 to other layers and achieve better results.The above experiments show that our S-Conv can significantlyimprove network performance with only a few parameters. Itis worth noting that our network has no spatial informationstream. The spatial information only affects the distributionand weight of convolution kernel.

We also show the IoU improvement of S-Conv on mostcategories in Fig. 5. It’s obvious that our S-Conv improves IoUin most categories, especially for objects lacking representativetexture information such as mirror, board and bathtub. Thereare also clear improvements for objects with rich spatialtransformation, such as chairs and tables. This shows that ourS-Conv can make good use of spatial information during theinference process.Architecture ablation: To evaluate the effectiveness of eachcomponent in our proposed S-Conv, we design ablation stud-ies. The results are shown in Tab. II. By default, we replacethe first convolution and the last two convolutions of eachlayer according to Tab. I. We can see that the offset generator,spatial projection module, and weight generator of S-Conv allcontribute to the improvement of the results.

Page 7: Spatial Information Guided Convolution for Real-Time RGBD ... · How to effectively use the extra geometry information (depth, 3D coordinates) is the key of RGBD semantic seg-mentation

7

Network Backbone MS SI Acc. mAcc. mIoU. fps param (M)FCN [3]CV PR15 2×VGG16 HHA 65.4 46.1 34.0 8 272.2LSD-GF [38]CV PR17 2×VGG16 HHA 71.9 60.7 45.9 - -RefineNet [21]CV PR17 ResNet152 3 - 73.6 58.9 46.5 16 129.5RDFNet [5]ICCV 17 2×ResNet152 3 HHA 76.0 62.8 50.1 9 200.1RDFNet [5]ICCV 17 2×ResNet101 3 HHA 75.6 62.2 49.1 11 169.1CFNet [39]ICCV 17 2×ResNet152 3 HHA - - 47.7 - -3DGNN [31]ICCV 17 VGG16 HHA - 55.2 42.0 5 47.2D-CNN [14]ECCV 18 2×VGG16 HHA - 56.3 43.9 13 92.0D-CNN [14]ECCV 18 VGG16 Depth - 53.6 41.0 26 47.0D-CNN [14]ECCV 18 2×ResNet152 Depth - 61.1 48.4 - -ACNet [28]arXiv19 2×ResNet50 Depth - - 48.3 18 116.6SGNet ResNet101 depth 75.5 60.9 49.0 28 58.3SGNet-8s ResNet101 depth 76.4 62.7 50.3 12 58.3SGNet ResNet101 3 depth 76.4 62.1 50.3 28 58.3SGNet-8s ResNet101 3 depth 76.8 63.1 51.0 12 58.3

TABLE V: Comparison results on NYUDv2 test dataset. MS: Multi-scale test; SI: Spatial information. The input imagesize for forward speed testing is 425× 560 using NVIDIA 1080Ti.

5 10 15 20 25 30 35FPS

30.0

32.5

35.0

37.5

40.0

42.5

45.0

47.5

50.0

mIo

U

SCNetSCNet-8sRDFNet*

DCNet

DCNet

ACNet

FCN

3DGNN

real-time

Fig. 6: FPS, mIoU, and the number of parameters ofdifferent methods on NYUDv2 test dataset with inputimage size 425×560 using NVIDIA 1080Ti. The radius ofthe circle corresponds to the number of parameters of themodel. For the model marked with ∗, its mIoU is obtained bymulti-scale test, and its inference time is obtained by single-scale forward. The results of DCNet [14] and 3DGNN [31]are from [14]. Our SGNet can achieve real-time inference andcompetitive performance. The SGNet-8s with output stride is 8can get better results than RDFNet which use more parametersand multi-scale test.

Comparison with alternatives: Most methods [5], [8], [28],[54] use a two-stream network to extract features from twodifferent modalities and then combine them. Our S-Convfocuses on advancing the feature extraction process of thenetwork by utilizing spatial information. Here we compareour S-Conv with two-stream network, deformable convolu-tion [12], [13], and depth-aware convolution [14]. We use asimple baseline which consists of a ResNet101 network withdeep supervision and a simple decoder. We add an additional

ResNet101 network, called HHANet, to extract HHA featuresand fuse it with our baseline features at the final layer of a two-stream network. To compare with depth-aware convolution anddeformable convolution, similar to SGNet, we replace the firstconvolution and the last two convolutions of each layer inthe baseline. The results are shown in Tab. III. We find thatour S-Conv achieves better results than two-stream network,deformable convolution, and depth-aware convolution. Thisdemonstrates that our S-Conv can effectively utilizes spatialinformation. The baseline equipped with weight generatorcan also achieve better results than depth-aware convolution,indicating that learning weights from spatial information isnecessary.Spatial information comparison: We also evaluate the impactof different formats of spatial information on S-Conv. Theresults are shown in Tab. IV. We can see that depth informationleads to comparable results with HHA and 3D coordinates,and better results than intermediate RGB features which areused by deformable convolution [12], [13]. This shows theadvantage of using spatial information for offset and weightgeneration over RGB features. However, converting depth toHHA is time-consuming [8]. Hence 3D coordinates and depthmap are more suitable for real-time segmentation using SGNet.

B. Comparison with state-of-the-art

We compare our SGNet with other state-of-the-art methodson NYUDv2 [15] and SUNRGBD [16], [17] datasets. Thearchitecture of SGNet is shown in Fig. 4.NYUDv2 dataset: The comparison results can be foundin Tab. V and Fig. 6. The input image size for single-scalespeed testing is 425 × 560 following [14]. We tested thesingle-scale speed of other methods under the same conditionsusing NVIDIA 1080Ti. Note that some methods in Tab. Vdo not report parameter quantities or open source. So we justlisted the mIoU of these methods. Instead of using additionalnetworks to extract spatial features, our SGNet can achieve

Page 8: Spatial Information Guided Convolution for Real-Time RGBD ... · How to effectively use the extra geometry information (depth, 3D coordinates) is the key of RGBD semantic seg-mentation

8

RGB Depth Layer1 1 Layer1 2 Layer1 3 Layer2 1

(e)

(d)

(c)

(b)

(a)

Fig. 7: The visualization of receptive filed in S-Conv.

Network Backbone MS SI Acc. mAcc. mIoU. fps param (M)LSD-GF [38]CV PR17 2×VGG16 HHA - 58.0 - - -RefineNet [21]CV PR17 ResNet152 3 - 80.6 58.5 45.9 16 129.53DGNN [31]ICCV 17 VGG16 3 HHA - 57.0 45.9 5 47.2RDFNet [5]ICCV 17 2×ResNet152 3 HHA 81.5 60.1 47.7 9 200.1CFNet [39]ICCV 17 2×ResNet152 3 HHA - - 48.1 - -D-CNN [14]ECCV 18 2×VGG16 HHA - 53.5 42.0 12.5 92.0ACNet [28]arXiv19 2×ResNet50 HHA - - 48.1 8 272.2SGNet ResNet101 depth 81.0 59.6 47.1 28 58.3SGNet ResNet101 3 depth 81.8 60.9 48.5 28 58.3

TABLE VI: Comparison results on SUNRGBD test dataset. MS: Multi-scale test; SI: Spatial information.

competitive performance and real-time inference. This benefitsfrom S-Conv which can make use of spatial informationefficiently with only a small amount of extra parameters andcomputation cost. Moreover, our S-Conv can achieve goodresults without using HHA information, making it suitablefor real-time tasks. After using multi-scale test which is thesame as RDFNet [5] and CFNet [39], we exceed all methodsbased on complicated data fusion and achieve state-of-the-art performance using fewer parameters. This verifies theefficiency of our S-Conv in utilizing spatial information. Atthe expense of a little bit more reasoning time by changingthe output stride of SGNet to 8 noted as SGNet-8, the proposedSGNet can achieve better results than other methods andRDFNet which uses multi-scale test, HHA information and

two resnet152 backbones. After using multi-scale test whichis used by other methods, SGNet’s performance can be furtherimproved.SUNRGBD dataset: The comparison results on the SUN-RGBD dataset are shown in Tab. VI. It is worth notingthat some methods in Tab. V did not report results on theSUNRGBD dataset. The inference time and parameter numberof models in Tab. VI are the same as those in Tab. V. OurSGNet can achieve competitive results in real-time comparedwith models that do not have real-time performance.

C. Qualitative Performance

Visualization of receptive filed in S-Conv: Appropriatereceptive field is very important for scene recognition. We

Page 9: Spatial Information Guided Convolution for Real-Time RGBD ... · How to effectively use the extra geometry information (depth, 3D coordinates) is the key of RGBD semantic seg-mentation

9

RGB Depth GT Baseline SGNet SGNet-8s

(e)

(d)

(c)

(b)

(a)

Fig. 8: The qualitative semantic segmentation comparison results on NYUDv2 test dataset.

RGB GT Baseline

(d)

(c)

(b)

(a)

Fig. 9: The qualitative semantic segmentation comparisonresults on SUNRGBD test dataset.

visualize the input adaptive receptive filed of SGNet in dif-ferent layers generated by S-Conv. Specifically, we get thereceptive field of each pixel by summing up the norm of theiroffsets during the S-Conv operation, then we normalize eachvalue to [0, 255] and visualize the result using a gray-scaleimage. The results are shown in Fig. 7. The brighter the pixel,the larger the receptive field. We observe that the receptivefields of different convolutions vary adaptively with the depthof the input image. For example, in layer1 1, the receptivefield is inversely proportional to the depth, which is oppositeto layer1 2. The combination of the adaptive receptive fieldlearned at each layer can help the network better resolve indoorscenes with complex spatial relations.

Qualitative comparison results: We show qualitative compar-ison results on NYUDv2 test dataset in Fig. 8. For the visualresults in Fig. 8 (a), the bathtub and the wall have insufficienttexture, which cannot be easily distinguished by the baselinemethod. Some objects may have reflections such as the tablein Fig. 8 (b), which is also challenging for the baseline.SGNet, however, can recognize it well by incorporating spatialinformation with the help of S-Conv. The chairs in Fig. 8(c, d) are hard to be recognized by RGB data due to thelow contrast and confused texture, while they can be easilyrecovered by SGNet benefiting from the equipped S-Conv.In the meantime, SGNet can recover the object’s geometricshape nicely, as demonstrated by the chairs of Fig. 8 (e). Wealso show qualitative results on SUNRGBD test dataset inFig. 9. It can be seen that our SGNet can also achieve precise

Page 10: Spatial Information Guided Convolution for Real-Time RGBD ... · How to effectively use the extra geometry information (depth, 3D coordinates) is the key of RGBD semantic seg-mentation

10

segmentation on SUNRGBD.

V. CONCLUSION

In this paper, we propose a novel Spatial information guidedConvolution (S-Conv) operator. Compared with conventional2D convolution, it can adaptively adjust the convolutionweights and distributions according to the input spatial infor-mation, resulting in better awareness of the geometric structurewith only a few additional parameters and computation cost.We also propose Spatial information Guided convolutionalNetwork (SGNet) equipped with S-Conv that yields real-timeinference speed and achieves competitive results on NYUDv2and SUNRGBD datasets for RGBD semantic segmentation.We also compare the performance of using different inputs togenerate offset, demonstrating the advantage of using spatialinformation over RGB feature. Furthermore, we visualizethe depth-adaptive receptive filed in each layer to show ef-fectiveness. In the future, we will investigate the fusion ofdifferent modal information and the adaptive change of S-Conv structure simultaneously, making these two approachesbenefit each other. We will also explore the application of S-Conv in different fields, such as pose estimation and 3D objectdetection.

REFERENCES

[1] H. Zhao, X. Qi, X. Shen, J. Shi, and J. Jia, “Icnet for real-time semanticsegmentation on high-resolution images,” in European Conference onComputer Vision (ECCV), 2018, pp. 405–420.

[2] B. Bescos, J. M. Facil, J. Civera, and J. Neira, “Dynaslam: Tracking,mapping, and inpainting in dynamic scenes,” IEEE Robotics and Au-tomation Letters, vol. 3, no. 4, pp. 4076–4083, 2018.

[3] J. Long, E. Shelhamer, and T. Darrell, “Fully convolutional networks forsemantic segmentation,” in Computer Vision and Pattern Recognition(CVPR), 2015, pp. 3431–3440.

[4] L.-C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, and A. L. Yuille,“Deeplab: Semantic image segmentation with deep convolutional nets,atrous convolution, and fully connected crfs,” IEEE Trans. Pattern Anal.Mach. Intell., vol. 40, no. 4, pp. 834–848, 2017.

[5] S.-J. Park, K.-S. Hong, and S. Lee, “Rdfnet: Rgb-d multi-level residualfeature fusion for indoor semantic segmentation,” in The IEEE Interna-tional Conference on Computer Vision (ICCV), 2017, pp. 4980–4989.

[6] D. Eigen and R. Fergus, “Predicting depth, surface normals and semanticlabels with a common multi-scale convolutional architecture,” in TheIEEE International Conference on Computer Vision (ICCV), 2015, pp.2650–2658.

[7] L. Ma, J. Stuckler, C. Kerl, and D. Cremers, “Multi-view deep learningfor consistent semantic mapping with rgb-d cameras,” in IntelligentRobots and Systems (IROS). IEEE, 2017, pp. 598–605.

[8] C. Hazirbas, L. Ma, C. Domokos, and D. Cremers, “Fusenet: Incorporat-ing depth into semantic segmentation via fusion-based cnn architecture,”in Asian conference on computer vision (ACCV). Springer, 2016, pp.213–228.

[9] J. Wang, Z. Wang, D. Tao, S. See, and G. Wang, “Learning commonand specific features for rgb-d semantic segmentation with deconvolu-tional networks,” in European Conference on Computer Vision (ECCV).Springer, 2016, pp. 664–679.

[10] S. Gupta, R. Girshick, P. Arbelaez, and J. Malik, “Learning rich featuresfrom rgb-d images for object detection and segmentation,” in EuropeanConference on Computer Vision (ECCV). Springer, 2014, pp. 345–360.

[11] Z. Li, Y. Gan, X. Liang, Y. Yu, H. Cheng, and L. Lin, “Lstm-cf: Unifyingcontext modeling and fusion with lstms for rgb-d scene labeling,” inEuropean Conference on Computer Vision (ECCV). Springer, 2016,pp. 541–557.

[12] J. Dai, H. Qi, Y. Xiong, Y. Li, G. Zhang, H. Hu, and Y. Wei, “Deformableconvolutional networks,” in Proceedings of the IEEE internationalconference on computer vision, 2017, pp. 764–773.

[13] X. Zhu, H. Hu, S. Lin, and J. Dai, “Deformable convnets v2: Moredeformable, better results,” in Proceedings of the IEEE Conference onComputer Vision and Pattern Recognition, 2019, pp. 9308–9316.

[14] W. Wang and U. Neumann, “Depth-aware cnn for rgb-d segmentation,”in European Conference on Computer Vision (ECCV), 2018, pp. 135–150.

[15] N. Silberman, D. Hoiem, P. Kohli, and R. Fergus, “Indoor segmentationand support inference from rgbd images,” in European Conference onComputer Vision. Springer, 2012, pp. 746–760.

[16] S. Song, S. P. Lichtenberg, and J. Xiao, “Sun rgb-d: A rgb-d scene under-standing benchmark suite,” in Computer Vision and Pattern Recognition(CVPR), 2015, pp. 567–576.

[17] A. Janoch, S. Karayev, Y. Jia, J. T. Barron, M. Fritz, K. Saenko, andT. Darrell, “A category-level 3d object dataset: Putting the kinect towork,” in Consumer depth cameras for computer vision. Springer,2013, pp. 141–165.

[18] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classificationwith deep convolutional neural networks,” in NIPS, 2012, pp. 1097–1105.

[19] K. Simonyan and A. Zisserman, “Very deep convolutional networks forlarge-scale image recognition,” arXiv preprint arXiv:1409.1556, 2014.

[20] F. Yu and V. Koltun, “Multi-scale context aggregation by dilatedconvolutions,” arXiv preprint arXiv:1511.07122, 2015.

[21] G. Lin, A. Milan, C. Shen, and I. Reid, “Refinenet: Multi-path refine-ment networks for high-resolution semantic segmentation,” in ComputerVision and Pattern Recognition (CVPR), 2017, pp. 1925–1934.

[22] L.-C. Chen, Y. Zhu, G. Papandreou, F. Schroff, and H. Adam, “Encoder-decoder with atrous separable convolution for semantic image segmen-tation,” in European Conference on Computer Vision (ECCV), 2018, pp.801–818.

[23] V. Badrinarayanan, A. Kendall, and R. Cipolla, “Segnet: A deep con-volutional encoder-decoder architecture for image segmentation,” IEEEtransactions on pattern analysis and machine intelligence (TPAMI),vol. 39, no. 12, pp. 2481–2495, 2017.

[24] H. Noh, S. Hong, and B. Han, “Learning deconvolution network forsemantic segmentation,” in The IEEE International Conference onComputer Vision (ICCV), 2015, pp. 1520–1528.

[25] L.-C. Chen, G. Papandreou, F. Schroff, and H. Adam, “Rethinkingatrous convolution for semantic image segmentation,” arXiv preprintarXiv:1706.05587, 2017.

[26] M. Yang, K. Yu, C. Zhang, Z. Li, and K. Yang, “Denseaspp forsemantic segmentation in street scenes,” in Computer Vision and PatternRecognition (CVPR), 2018, pp. 3684–3692.

[27] H. Zhao, J. Shi, X. Qi, X. Wang, and J. Jia, “Pyramid scene parsingnetwork,” in Computer Vision and Pattern Recognition (CVPR), 2017,pp. 2881–2890.

[28] X. Hu, K. Yang, L. Fei, and K. Wang, “Acnet: Attention based networkto exploit complementary features for rgbd semantic segmentation,”arXiv preprint arXiv:1905.10089, 2019.

[29] S. Song, F. Yu, A. Zeng, A. X. Chang, M. Savva, and T. Funkhouser,“Semantic scene completion from a single depth image,” in ComputerVision and Pattern Recognition (CVPR), 2017, pp. 1746–1754.

[30] S. Song and J. Xiao, “Deep sliding shapes for amodal 3d object detectionin rgb-d images,” in Computer Vision and Pattern Recognition (CVPR),2016, pp. 808–816.

[31] X. Qi, R. Liao, J. Jia, S. Fidler, and R. Urtasun, “3d graph neuralnetworks for rgbd semantic segmentation,” in The IEEE InternationalConference on Computer Vision (ICCV), 2017, pp. 5199–5208.

[32] C. R. Qi, H. Su, K. Mo, and L. J. Guibas, “Pointnet: Deep learning onpoint sets for 3d classification and segmentation,” in Computer Visionand Pattern Recognition (CVPR), 2017, pp. 652–660.

[33] C. R. Qi, L. Yi, H. Su, and L. J. Guibas, “Pointnet++: Deep hierarchicalfeature learning on point sets in a metric space,” in Advances in neuralinformation processing systems (NIPS), 2017, pp. 5099–5108.

[34] L.-Z. Chen, X.-Y. Li, D.-P. Fan, M.-M. Cheng, K. Wang, and S.-P. Lu,“Lsanet: Feature learning on point sets by local spatial aware layer,”arXiv preprint arXiv:1905.05442, 2019.

[35] Y. Xu, T. Fan, M. Xu, L. Zeng, and Y. Qiao, “Spidercnn: Deep learningon point sets with parameterized convolutional filters,” in EuropeanConference on Computer Vision (ECCV), 2018, pp. 87–102.

[36] C. Wang, B. Samari, and K. Siddiqi, “Local spectral graph convolutionfor point set feature learning,” in European Conference on ComputerVision (ECCV), 2018, pp. 52–66.

[37] Y. Li, R. Bu, M. Sun, W. Wu, X. Di, and B. Chen, “Pointcnn: Con-volution on x-transformed points,” in Advances in neural informationprocessing systems (NIPS), 2018, pp. 828–838.

Page 11: Spatial Information Guided Convolution for Real-Time RGBD ... · How to effectively use the extra geometry information (depth, 3D coordinates) is the key of RGBD semantic seg-mentation

11

[38] Y. Cheng, R. Cai, Z. Li, X. Zhao, and K. Huang, “Locality-sensitivedeconvolution networks with gated fusion for rgb-d indoor semanticsegmentation,” in Computer Vision and Pattern Recognition (CVPR),2017, pp. 3029–3037.

[39] D. Lin, G. Chen, D. Cohen-Or, P.-A. Heng, and H. Huang, “Cascadedfeature network for semantic segmentation of rgb-d images,” in TheIEEE International Conference on Computer Vision (ICCV), Oct 2017.

[40] J. Jiao, Y. Wei, Z. Jie, H. Shi, R. W. Lau, and T. S. Huang, “Geometry-aware distillation for indoor semantic segmentation,” in Computer Visionand Pattern Recognition (CVPR), 2019, pp. 2869–2878.

[41] P. Wang, X. Shen, Z. Lin, S. Cohen, B. Price, and A. L. Yuille, “Towardsunified depth and semantic prediction from a single image,” in ComputerVision and Pattern Recognition (CVPR), 2015, pp. 2800–2809.

[42] J. Hoffman, S. Gupta, and T. Darrell, “Learning with side informationthrough modality hallucination,” in Computer Vision and Pattern Recog-nition (CVPR), 2016, pp. 826–834.

[43] I. Kokkinos, “Ubernet: Training a universal convolutional neural networkfor low-, mid-, and high-level vision using diverse datasets and limitedmemory,” in Proceedings of the IEEE Conference on Computer Visionand Pattern Recognition, 2017, pp. 6129–6138.

[44] Z. Zhang, Z. Cui, C. Xu, Y. Yan, N. Sebe, and J. Yang, “Pattern-affinitivepropagation across depth, surface normal and semantic segmentation,”in Computer Vision and Pattern Recognition (CVPR), June 2019.

[45] Y. He, W.-C. Chiu, M. Keuper, and M. Fritz, “Std2p: Rgbd semanticsegmentation using spatio-temporal data-driven pooling,” in ComputerVision and Pattern Recognition (CVPR), 2017, pp. 4837–4846.

[46] M. Jaderberg, K. Simonyan, A. Zisserman et al., “Spatial transformernetworks,” in Advances in neural information processing systems (NIPS),2015, pp. 2017–2025.

[47] X. Jia, B. De Brabandere, T. Tuytelaars, and L. V. Gool, “Dynamicfilter networks,” in Advances in neural information processing systems(NIPS), 2016, pp. 667–675.

[48] X. Li, W. Wang, X. Hu, and J. Yang, “Selective kernel networks,” inProceedings of the IEEE Conference on Computer Vision and PatternRecognition, 2019, pp. 510–519.

[49] X. Wang, R. Girshick, A. Gupta, and K. He, “Non-local neural net-works,” in Proceedings of the IEEE Conference on Computer Visionand Pattern Recognition, 2018, pp. 7794–7803.

[50] J. Hu, L. Shen, and G. Sun, “Squeeze-and-excitation networks,” inComputer Vision and Pattern Recognition (CVPR), June 2018.

[51] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for imagerecognition,” in Computer Vision and Pattern Recognition (CVPR), 2016,pp. 770–778.

[52] N. Silberman, D. Hoiem, P. Kohli, and R. Fergus, “Indoor segmentationand support inference from rgbd images,” in European Conference onComputer Vision (ECCV). Springer, 2012, pp. 746–760.

[53] O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma,Z. Huang, A. Karpathy, A. Khosla, M. Bernstein et al., “Imagenet largescale visual recognition challenge,” International journal of computervision (IJCV), vol. 115, no. 3, pp. 211–252, 2015.

[54] J. Jiang, L. Zheng, F. Luo, and Z. Zhang, “Rednet: Residual encoder-decoder network for indoor rgb-d semantic segmentation,” arXiv preprintarXiv:1806.01054, 2018.