an integrated deep learning framework for joint

43
An integrated deep learning framework for joint segmentation of blood pool and myocardium Xiuquan Du a,b , Yuhui Song b , Yueguo Liu b , Yanping Zhang a,b , Heng Liu c , Bo Chen d , Shuo Li e,* a Key Laboratory of Intelligent Computing and Signal Processing of Ministry of Education, Anhui University b School of Computer Science and Technology, Anhui University c Department of Gastroenterology, The First Affiliated Hospital of Anhui Medical University, Anhui, China d School of Health Science, Western University, London, ON N6A 3K7, Canada e Department of Medical Imaging, Western University, London, ON N6A 3K7, Canada Abstract Simultaneous and automatic segmentation of the blood pool and myocardium is an important precondition for early diagnosis and pre-operative planning in patients with complex congenital heart disease. However, due to the high diversity of cardiovascular structures and changes in mechanical properties caused by cardiac defects, the segmentation task still faces great challenges. To overcome these challenges, in this study we propose an integrated multi- task deep learning framework based on the dilated residual and hybrid pyra- mid pooling network (DRHPPN) for joint segmentation of the blood pool and myocardium. The framework consists of three closely connected pro- gressive sub-networks. An inception module is used to realize the initial multi-level feature representation of cardiovascular images. A dilated resid- ual network (DRN), as the main body of feature extraction and pixel classifi- cation, preliminary predicts segmentation regions. A hybrid pyramid pooling network (HPPN) is designed for facilitating the aggregation of local informa- tion to global information, which complements DRN. Extensive experiments on three-dimensional cardiovascular magnetic resonance (CMR) images (the available dataset of the MICCAI 2016 HVSMR challenge) demonstrate that our approach can accurately segment the blood pool and myocardium and * Corresponding author Email address: [email protected] (Shuo Li) Preprint submitted to Medical Image Analysis March 5, 2020

Upload: others

Post on 12-Apr-2022

6 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: An integrated deep learning framework for joint

An integrated deep learning framework for joint

segmentation of blood pool and myocardium

Xiuquan Dua,b, Yuhui Songb, Yueguo Liub, Yanping Zhanga,b, Heng Liuc,Bo Chend, Shuo Lie,∗

aKey Laboratory of Intelligent Computing and Signal Processing of Ministry ofEducation, Anhui University

bSchool of Computer Science and Technology, Anhui UniversitycDepartment of Gastroenterology, The First Affiliated Hospital of Anhui Medical

University, Anhui, ChinadSchool of Health Science, Western University, London, ON N6A 3K7, Canada

eDepartment of Medical Imaging, Western University, London, ON N6A 3K7, Canada

Abstract

Simultaneous and automatic segmentation of the blood pool and myocardiumis an important precondition for early diagnosis and pre-operative planningin patients with complex congenital heart disease. However, due to the highdiversity of cardiovascular structures and changes in mechanical propertiescaused by cardiac defects, the segmentation task still faces great challenges.To overcome these challenges, in this study we propose an integrated multi-task deep learning framework based on the dilated residual and hybrid pyra-mid pooling network (DRHPPN) for joint segmentation of the blood pooland myocardium. The framework consists of three closely connected pro-gressive sub-networks. An inception module is used to realize the initialmulti-level feature representation of cardiovascular images. A dilated resid-ual network (DRN), as the main body of feature extraction and pixel classifi-cation, preliminary predicts segmentation regions. A hybrid pyramid poolingnetwork (HPPN) is designed for facilitating the aggregation of local informa-tion to global information, which complements DRN. Extensive experimentson three-dimensional cardiovascular magnetic resonance (CMR) images (theavailable dataset of the MICCAI 2016 HVSMR challenge) demonstrate thatour approach can accurately segment the blood pool and myocardium and

∗Corresponding authorEmail address: [email protected] (Shuo Li)

Preprint submitted to Medical Image Analysis March 5, 2020

Page 2: An integrated deep learning framework for joint

achieve competitive performance compared with state-of-the-art segmenta-tion methods.

Keywords: Blood pool, Myocardium, Magnetic resonance image,Segmentation

1. Introduction

Congenital heart disease (CHD) is one of the causes of high mortality,especially among children, which seriously affects people’s normal life. Atthe same time, CHD surgery is complicated and poses certain risks. Hence,there is an urgent need for a safe and effective surgical auxiliary way to helpclinicians quickly to find lesions, thereby shortening the surgery time andreducing patients’ pain. Since cardiovascular magnetic resonance (CMR)images can intuitively show the specific morphology of the blood pool andsurrounding myocardium, clinicians are capable of judging patients’ cardiacdefects according to the distributions of the blood pool and myocardium, soas to make an accurate prediction of lesions easily. Thus, the simultaneoussegmentation of the blood pool and myocardium from 3D CMR images isan important premise of pre-operative planning and patient-specific heartmodels for patients with complex CHD (Pace et al., 2015). As shown inFigure 1, our task is to segment two portions from the axial slices of 3DCMR images: the blood pool and myocardium. As can be seen from Figure1(a), we artificially delineate the contours of the blood pool and myocardiumon original 2D slices. As the major component of cardiovascular structure,the blood pool is composed of some individuals with irregular shapes, whilethe myocardium comprises marginal areas surrounding the blood pool. Itis precisely the high complexity of shape and appearance that makes oursegmentation task extremely difficult. Figure 1(b) shows the segmentationresults by experienced experts.

In early clinical practice, a contrast agent is usually used to enhancethe image development, and then the boundaries of targets are manuallydescribed. Although this clinical method has been proved to be effectivein segmenting CMR images, sometimes it may not be able to obtain reli-able and accurate segmentation results due to the following limitations: 1)the anatomical variability of cardiac defects is large; 2) the signal intensitychanges greatly; and 3) the signal-to-noise ratio is low (Suinesiaputra et al.,2015; Zhao et al., 2017). In addition, manually segmenting the blood pool

4

Page 3: An integrated deep learning framework for joint

and myocardium is not only a waste of manpower and time but also proneto inter- and intra-observers (Afshin et al., 2012). Moreover, the employ-ment of a contrast agent produces many unpredictable side effects and placesgreat pressure on the patient’s body and mind (Ordovas and Higgins, 2011).Based on the above limitations, researchers began to explore automatic orsemi-automatic segmentation methods to replace the manual segmentationmethod. However, due to the high complexity and diversity of cardiovas-cular structures (Xue et al., 2017a), these methods still faces the followingchallenges.

Figure 1: (a) Original 2D slice: the blue line is the outline of the blood pool, and the greenline is the outline of the myocardium. (b) Segmentation result: the yellow represents thebackground, the blue represents the blood pool, and the green represents the myocardium.

• Complex changes of appearance intensity mainly including the follow-ing two forms.(1) Highly inconsistent appearance intensity of the same targets: theparts of the same category show a large difference in pixel intensity; itis easy to divide the two parts into different categories without corre-sponding labels.(2) The high similarity of appearance intensity of different targets: ina few cases, the appearance of the blood pool and myocardium mayshow very similar pixel intensity, which makes it difficult to determinethe correct category.

• Huge changes of boundaries mainly including the following two cases.(1) Huge changes in shape and appearance: the anatomical differencecaused by different heart movement states is great, which leads to dif-ferent shape and appearance presented by 3D image of each subject.(2) Local weak/no boundaries: the complexity of cardiovascular struc-ture determines that the contrast of boundaries is low or there is no

5

Page 4: An integrated deep learning framework for joint

Figure 2: Illustration of the challenges in segmenting 3D CMR images. (a) Highly inconsis-tent appearance intensity of same targets. (b) The high similarity of appearance intensityof different targets. (c) Huge changes in shape and appearance. (d) Local weak/no bound-aries.

contrast; these situations result in the absence of significant boundariesamong the blood pool, myocardium, and unconsidered background.

Several representative images are provided to illustrate the above chal-lenges. As shown in Figure 2(a), the white dotted line circled in the two slices

6

Page 5: An integrated deep learning framework for joint

should belong to the blood pool, but it can be seen that the appearance ofthe middle portion of the blood pool shows a very similar pixel intensityvalue as the myocardium. This not only makes it easy to classify both partsas myocardium but also introduces much interference to the classification ofthe blood pool. Examining the two images in Figure 2(b), in the left-handimage, the portion circled by the dotted line should be the blood pool, butit is easily classified as myocardium because it locates at the edge and looksvery similar to the myocardium. Regarding the right-hand image, the regionenclosed by the dotted line is obviously similar to the blood pool in appear-ance and size, but it is an unnecessary background area. Two other cases ofcomplexity are now described in detail. As shown in Figure 2(c), we haveselected the 100th slice of 5 subjects for display, and it is obvious that thereare significant differences in morphology and movement between images ofdifferent individuals. The segmentation of the two slices shown in Figure 2(d)is more difficult. The problems they present are the common problems thatmost medical images face in segmentation, namely weak boundary and noboundary. For the left-hand image, the area circled by the dotted line showsthe blurry boundary between the blood pool and myocardium, or betweenthe blood pool and background, or between the myocardium and background.Regarding the right-hand image, it is difficult to determine whether there isa boundary between the blood pool and background. In fact, the boundaryis indeed a difficult part in image segmentation. All of the problems posemany challenges to our multi-objective segmentation task.

To solve the aforementioned challenges and compensate for the short-comings of existing segmentation methods, we have proposed an integratedmulti-task deep learning framework based on the dilated residual and hybridpyramid pooling network (DRHPPN) for joint segmentation of the bloodpool and myocardium. The framework includes inception module, dilatedresidual network (DRN) and hybrid pyramid pooling network (HPPN). Theinception module is used to realize the multi-level feature representation ofCMR images. The multi-level feature maps by convolutional layers of threescales contain abundant feature information about the object’s shape, struc-ture, size, and boundary. The DRN focuses on the deep features of objectsand models the dynamic correlation between image sequences. The HPPNfurther uses the intermediate features of different scales obtained by the DRNto realize the aggregation of multiple local information maps into global infor-mation maps. While recovering the resolution of the original image throughup-sampling operation, a softmax classifier is adopted to realize the classi-

7

Page 6: An integrated deep learning framework for joint

fication at the pixel level, which would accurately segment the blood pooland myocardium, and provide a more effective reference basis for clinicaltreatment of CHD. The major contributions of this work are the following.

(1) We propose a new multi-task deep learning network based on the dilatedresidual and hybrid pyramid pooling network (DRHPPN), which couldsegment the blood pool and myocardium directly and simultaneously.

(2) To segment morphologically variable targets, blood pool and myocardium,effective components such as inception module, dilated convolution, andConvLSTM module are used to assist residual network in learning andextracting features and sequences’ correlations.

(3) We creatively feed the intermediate features of the DRN into a hybridpyramid pooling network (HPPN). Pooling layers of different scales en-able multi-level local enhancement features to converge into a mixedglobal enhancement feature map, which contains high-level semantic in-formation directly for accurate segmentation.

The rest of this paper is organized as follows. In Section 2, the relatedworks and methods of CMR image segmentation in recent years are intro-duced. In Section 3, we mainly describe our segmentation framework, includ-ing the principle and application of basic modules. In Section 4, we presentsegmentation results, including the comparative experiments, ablation stud-ies, and other applications. In the Section 5, we conclude the paper and putforward future research plans.

2. Related work

Although there is a large body of literature related to medical image seg-mentation, we mainly focus on the automatic segmentation methods of theblood pool and myocardium, which are closely bound up with our work. Byconsulting studies of segmentation of cardiovascular images, we have foundthat there are very few papers that are generated by the segmentation of theblood pool alone, which may be related to its special position and functionin the cardiovascular structures. Therefore, the relevant works we have sum-marized can be divided into two categories: (1) the segmentation methods ofthe myocardium, and (2) simultaneous segmentation of the blood pool andmyocardium.

Traditional image segmentation methods have been widely used in med-ical image analysis, especially in the segmentation of ventricles, and many

8

Page 7: An integrated deep learning framework for joint

Tab

le1:

Su

mm

ary

ofre

late

dw

orks.

Tar

get

Met

hods

Au

thor

sD

atas

etR

esu

lts

Myo

card

ium

Pro

bab

ilis

tic

atla

san

dth

eE

Mal

gori

thm

(Lor

enzo

-Val

ds

etal

.,20

04)

4DD

M:L

V=

0.95

Myo

=0.

83R

V=

0.91

LS

TM

-RN

N(X

uet

al.,

2017

)2D

+T

Acc

ura

cy=

94.3

Pre

cisi

on=

91.3

(LV

and

RV

)F

CN

(Qin

etal

.,20

18)

3DD

M:

LV

=0.

935

Myo

=0.

864

RV

=0.

886

HD

:LV

=3.

52M

yo=

13.4

3R

V=

11.3

8

Gra

ph

cut

(Son

get

al.,

2019

)2D

DM

=0.

899

HD

=1.

052

MA

D=

0.24

5

Cro

ss-c

onst

rain

edsh

ape

and

shap

ed

iscr

epan

cyco

mp

ensa

tion

(Liu

etal

.,20

19)

2DD

E:

DM

=78

.13

T2:

DM

=84

.97

and

3DM

RF

mod

el(T

ziri

tas,

2016

)3D

Bp:

DM

=0.

867

HD

=19

.723

Myo

:D

M=

0.61

2H

D=

13.1

99

Blo

odpo

olM

ult

i-at

las

and

leve

lse

ts(S

hah

zad

etal

.,20

16)

3DB

p:

DM

=0.

885

HD

=9.

408

Myo

:D

M=

0.74

7H

D=

5.09

1

Myo

card

ium

Ran

dom

fore

st(R

F)

(Mukhop

adhya

y,

2016

)3D

Bp:

DM

=0.

794

HD

=14

.634

Myo

:D

M=

0.49

5H

D=

12.7

96

3D-U

net

(Cic

eket

al.,

2016

)3D

Bp:

DM

=0.

926

HD

=8.

628

Myo

:D

M=

0.69

4H

D=

10.2

21

Dilat

edco

nvo

luti

on(W

olte

rink

etal

.,20

16)

3DB

p:

DM

=0.

926

HD

=7.

069

Myo

:D

M=

0.80

2H

D=

6.12

6D

ense

resi

du

alm

od

ule

and

dee

psu

per

vis

edle

arnin

gm

od

e(Y

uet

al.,

2017

)3D

Bp:

DM

=0.

931

HD

=9.

533

Myo

:D

M=

0.82

1H

D=

7.29

4

CN

N(T

aha

etal

.,20

18)

2DB

p:

DM

=0.

933

HD

=6.

791

Myo

:D

M=

0.78

5H

D=

6.06

1

∗ LV

:L

eft

Ven

tric

le/R

V:

Rig

ht

Ven

tric

le/B

p:

Blo

od

pool

/Myo

:M

yoca

rdiu

m/D

M:

Dic

eM

etri

c∗ H

D:

Hau

sdor

ffD

ista

nce

/M

AD

:M

ean

Abso

lute

Dis

tance

/DE

:D

elay

edE

nhan

cem

ent/

T2:

T2-

wei

ght

9

Page 8: An integrated deep learning framework for joint

effective segmentation methods have been derived. For example, (Mitchellet al., 2001) proposed a hybrid approach of an active shape model (ASM)and active appearance model (AAM) to automatically segment the left ven-tricle (LV) and right ventricle (RV), which is helpful to avoid the matchingfunction falling into a local minimum value. (Moolan-Feroze et al., 2014)proposed the mixed flow model and the segmentation framework of Markovrandom field (MRF) to segment the RV and generated the target shapepriority from manifold using image information and current segmentationestimation. (Grosgeorge et al., 2013) used the graph cut framework to guidethe segmentation of the RV in MR images. (Maier et al., 2012) adoptedthe two-stage segmentation method: first finding the excessive region imageof the RV in MR images, and then combining the segmentation region withthe intensity-based boundary term. Through the calculated region of interest(ROI), (Mahapatra and Buhmann, 2013) used a random forest (RF) classifierand graph cut method to classify the pixel, thus realizing the segmentation ofthe RV. (Woo et al., 2013) employed the level set and implicit shape priorityunder the variational framework to segment the LV. Although these tradi-tional methods have achieved good results in ventricular segmentation, theirperformance inevitably depends on the initially selected point and weightfactor, and some methods even require frequent user interaction (Petitjeanand Dacher, 2011; Xue et al., 2017b). These reasons cause the limitations ofthe use of traditional methods on clinical operation.

Deep convolutional neural networks (CNNs) have achieved great successin handwriting recognition, which promotes the extensive application of CNNin the medical imaging field. (Su et al., 2015) first proposed a fast CNN scan-ning method to segment cancer areas in breast MR images. However, theperformance of the traditional CNN-based segmentation method is limitedby the size of the selected pixel block, which can only extract local featuresand ignore global information. Therefore, (Long et al., 2014) proposed a fullyconvolutional neural network (FCN) to solve the input problem in instancesegmentation, which is also applicable to our medical image segmentation.The network can be fed the images of any size, and restore the original res-olution of the image through the deconvolutional layers, so as to directlyachieve pixel-level segmentation.

In recent years, many scholars have proposed some methods for segmenta-tion of the blood pool and myocardium in CMR images, including traditionalmethods and deep learning-based methods. To name just a few, (Lorenzo-Valds et al., 2004) proposed an approach based on the prior knowledge of

10

Page 9: An integrated deep learning framework for joint

atlas and the 4D MRF combined with the space-time background to segmentthe ventricle and myocardium of the cardiac MR images. (Song et al., 2019)applied the traditional graph cut method to segment myocardium. (Liu et al.,2019) used the cross-constraint shape method and shape discrepancy com-pensation principle to segment the myocardium over delayed enhancement(DE) and T2-weight images. Other traditional methods, such as the 3D ran-dom field model, was proposed by (Tziritas, 2016), and a method combiningmulti-atlas and level sets was proposed by (Shahzad et al., 2016). The ap-pearance model was proposed by (Mukhopadhyay, 2016), which uses randomforest (RF) to learn context awareness, and then makes a prediction based onappearance priori. These methods are all based on machine learning theoriesand applied to the segmentation of the blood pool and myocardium in CMRimages.

Traditional methods have not been widely used due to their limitations,but the emerging deep learning-based methods are able to compensate forthese shortcomings. For instance, (Xu et al., 2017) proposed an end-to-end deep learning framework (LSTM-RNN), that combines CNN and timerelation-based RNN module to accurately detect and segment the myocardialinfarction areas. (Qin et al., 2018) proposed a framework based on fully con-volutional neural network (FCN) to segment the ventricles and myocardiumas well as to evaluate the motion state of the heart. (Cicek et al., 2016)adopted a simple 3D-Unet network to perform multi-task segmentation di-rectly. (Wolterink et al., 2016) proposed a single dilated CNN for the problemthat CMR images contain a large number of small targets, but it often failsto highlight the local key features. In addition, (Yu et al., 2017) employedthe 3D form of the dense residual module and deep supervised learning mod-ule, which can segment the targets more accurately, but ignored the guidingrole of multi-level features contained by the high complexity of CMR images.(Taha et al., 2018) proposed a CNN consisting of an up-sampling moduleand a down-sampling module to simultaneously segment the blood vesselsand the myocardium around them. It is undeniable that these deep learning-based methods have become the mainstream methods in the field of medicalimage segmentation. Their emergence not only replaces the dominant posi-tion of traditional segmentation methods but also solves their limitations anddefects. Thus, this paper also proposes a new deep learning-based method tosolve the segmentation problems of the blood pool and myocardium in CMRimages.

11

Page 10: An integrated deep learning framework for joint

Figure 3: Illustration of DRHPPN network, which consists of inception module, dilatedresidual network (DRN) and hybrid pyramid pooling network (HPPN).

3. Method

The essence of a deep learning network lies in learning to create an ab-stract representation of the image data, which is expressed by a highly non-linear function f : x→ y, where f represents a transformation function frominput space x to output space y. The main components of a deep learningnetwork are convolutional layers, the formula of a convolutional layer is

al+1n = f

(M∑

m=1

W lnm ∗ alm + bl+1

n

)(1)

where al+1n and alm represent the nth feature map in the (l+1 )th layer and the

mth feature map in the lth layer, respectively, and M denotes the number offeature maps in layer l. W l

nm represents the 2D convolution kernel from themth feature map in the lth layer to the nth feature map in the (l+1 )th layer.∗ denotes the 2D convolution operation and bl+l

n represents the correspondingbias vector (Qi et al., 2019).

The framework of the dilated residual and hybrid pyramid pooling net-work (DRHPPN) is shown in Figure 3. First, an inception module is used toextract multi-level information of the input images, and obtain some shallow

12

Page 11: An integrated deep learning framework for joint

feature maps containing targets’ shape, location, and boundary information,which are fused into a rough visual feature map. Then, the fusion featuremap is put into a dilated residual network (DRN) with a ConvLSTM modulein the middle of the encoder and decoder. Under the action of the ConvL-STM module, the spatio-temporal information inside and between imagescan be studied comprehensively, and the morphological and kinematic cor-relations of image sequences can be modeled. The dilated convolutions areadded to avoid the reduction of image resolution caused by the frequent use ofmax-pooling layers. Skip connections are capable of transferring the encodedfeature information to the decoding module directly to ensure the integrity ofthe semantic information. Finally, we creatively extract and concatenate thefeature maps of the same resolution to construct a hybrid pyramid poolingnetwork (HPPN). In consideration of the local key information concerned bythe feature maps of different scales being different, the deep feature mapscontaining high-level semantic information can be obtained by further pool-ing them. The deconvolution operation of these deep feature maps couldreceive more sufficient context information, which is helpful to guide theclassification of pixel points in the process of restoring the image resolution.Therefore, the segmentation of the blood pool and myocardium can be com-pleted more accurately.

The deep learning network is designed to segment the blood pool and my-ocardium in a series of CMR slices S0,S1,S2,...,SN . The segmentation of theobjective regions is conducted by predicting the categories of pixels in eachimage. The categories include three types of {background, bloodpool,myocardium}.The final output feature map has three channels which consists of a back-ground channel x0 and two channels x1,x2 corresponding to the myocardiumand blood pool, with each channel undertaking the task of classifying a typeof target. By doing so, we alleviate the challenge of segmenting multipletargets and simplify the problem to binary classifications. We further em-ploy the binary cross-entropy loss function between each target channel andbackground channel,

L = −2∑

i=1

yilog(pi) + (1− yi)log(1− pi) (2)

where yi represents the ground-truth binary label map for each target and piis calculated as exp(xi)

exp(x0)+exp(xi)for i=1,2.

13

Page 12: An integrated deep learning framework for joint

3.1. Inception module

Our initial feature extraction module is essentially a variation of the in-ception module. Figure 4(a) shows a naive inception module, which mainlycontributes in two aspects. One is to use the 1×1 convolution to raise andlower dimensions, and the other is to carry out convolution and aggregationsimultaneously on multiple scales (Szegedy et al., 2015). As a matter of fact,from the point of view of feature extraction, the convolution of 1×1 seems tobe of little use. However, the convolutional layers of 1×1 invisibly increasethe depth of the network, and improve the nonlinear expression ability of thenetwork under the effect of the rectified linear unit (ReLU) (Xu et al., 2015).Additionally, it is characterized by the addition of a max-pooling branch witha stride of 1, which can achieve effective feature extraction without changingthe image resolution.

Figure 4(b) shows the slight improvements we have made to the naiveinception module. One is an increase in scales. Empirically speaking, theconvolutional layer of a single scale could condense the image into a fea-ture map. However, for a multi-target segmentation task, different targetshave different pixel intensities, which require multiple level feature maps toexpress separate and specific targets. Moreover, multi-scale convolutionallayers are capable of compensating for feature information that cannot becaptured by convolutional layers of a single scale, and all of them have theability to achieve the purpose of discarding unnecessary information. Onthe other hand, after the concatenation of multi-scale feature maps, a max-pooling layer with a stride of 2 is used to condense the coarse feature mapsand make the information more concise.

3.2. Dilated residual network (DRN)

To obtain the representative global CMR image feature representation,we feed the initial feature maps obtained from the modified inception mod-ule into the dilated residual network (DRN). The network is used to describethe deep appearance and structure features of images and to capture themorphological and kinematic characteristics of multiple objects. As shownin Figure 5, we show the structure and parameter settings of the DRN. Thedesign of the DRN is mainly based on symmetric convolution and deconvolu-tion channels. Given the irregular shape of the blood pool and myocardiumand their scattered small parts that are difficult to locate, we have chosendilated convolution instead of normal convolution to better identify these

14

Page 13: An integrated deep learning framework for joint

Figure 4: Illustration of inception module.

targets (Yu and Koltun, 2015; Chen et al., 2016). Another important designis that a ConvLSTM module is added between the down-sampling and up-sampling modules. While extracting the spatial feature information of theimages, the ConvLSTM module can also capture the dynamic dependenceamong the image frames (Shi et al., 2015), which can effectively improve thesegmentation accuracy of cardiovascular images. The formulas of the dilatedconvolution and ConvLSTM are the following:

y[i] = f(∑

x[i+ r · k]× w[k] + b[k])

(3)

where f denotes the kernel-wise nonlinear transformation of ReLU and batchnormalization (Ioffe and Szegedy, 2015), and r is the dilation rate that isadded to the convolution kernel k. We have

Ft = σ(wxf ∗Xt + whf ∗Ht−1 + wcf � Ct−1 + bf )

It = σ(wxi ∗Xt + whi ∗Ht−1 + wci � Ct−1 + bi)

Ct = Ft � Ct−1 + It � tanh(wxc ∗Xt + whc ∗Ht−1 + bc)

Ot = σ(wxo ∗Xt + who ∗Ht−1 + wco � Ct + bo)

Ht = Ot � tanh(Ct)

(4)

15

Page 14: An integrated deep learning framework for joint

Figure 5: Illustration of DRN, which consists of three convolution blocks and three de-convolution blocks, and a ConvLSTM module is placed between the two types of blocks.Input of our DRN is size of 64×64×64 (height × width × channels) initial feature mapsfrom the inception module, and the number of channels set by all convolutional layers is64.

where Xt and Ht−1 represent the current input and status at the last mo-ment respectively. Ft, It and Ot represent the information flow state of theforgetting gate, input gate, and output gate, respectively. Ht represents thecurrent state of the moment. Ct mainly controls whether the current state isused as the output of the ConvLSTM and represents the current output. �denotes the element-wise product, and ∗ the convolution operator. σ denotesthe activation function, and tanh the hyperbolic tangent nonlinear function.w is the weight matrix of different gate, and b the corresponding bias vector.

Our DRN adopts convolutional layers with convolution kernel sizes of

16

Page 15: An integrated deep learning framework for joint

Table 2: Configurations of skip connections.

Skip Connections Image Resolution

Conv1 (1×1) Deconv Block 3# 64×64Conv Block 1# Deconv Block 2# 32×32Conv Block 2# Deconv Block 1# 16×16Conv Block 3# Conv2 (1×1) 8×8

Table 3: Configurations of dilated convolutions.

Dilated Rate

Conv Block 1# / Deconv Block 3# (2, 2)Conv Block 2# / Deconv Block 2# (2, 2)Conv Block 3# / Deconv Block 1# (3, 3)

1×1 and 3×3, and the pooling window size of the max-pooling layers is2×2. We have set two types of dilation rates, 2×2 and 3×3. Table 2 and3 show the configurations of the skip connections and dilated convolutionsin the DRN, respectively. After each convolutional layer, batch normaliza-tion and activation function ReLU are set to accelerate the training speedand improve the training accuracy of the model. To reduce the overfittingrisk of the network, dropout is added to the end of Conv Block to reduceparameters. Theoretically, the feature representation of an image in imagesequences is largely dependent on the feature representation of the previousimages. Hence, we cleverly put a ConvLSTM module into the middle of ConvBlocks and Deconv Blocks to capture the time correlations between imagesequences.

3.3. Hybrid pyramid pooling network (HPPN)

We know that in most cases, it is not guaranteed that our research ob-jects are in the same visual range, so it is necessary to propose an effectivemethod to solve the scale problem of the input image. (He et al., 2014)proposed spatial pyramid pooling (SPP) network in a study of using a deepneural network (DNN) to deal with visual recognition tasks, aiming to solvethe problem that the input image size of a CNN must be fixed. (Liu andHan, 2016) and (Wang et al., 2018) have proved that feature maps of dif-ferent levels exhibit different salient objects. Low-level features are obtained

17

Page 16: An integrated deep learning framework for joint

Figure 6: Illustration of HPPN, which shows parameter configurations of max-poolinglayers and dense connections of up-sampling layers.

by fewer convolutional layers. It has high resolution and contains significantamount of information, such as location and details, but also contains a lotof noise. In contrast, high-level features have strong semantic informationand poor perception of details because of low resolution. In the course ofdeep convolution, the effective use of high- and low-level features can bettercapture the precise location of the target and highlight the details. On thebasis of solving the problem of input size, the SPP network makes full use ofthe difference of multi-level features.

Based on the above analysis, we creatively apply the advantages of pyra-mid pooling layers to solve the segmentation problem of the blood pool andmyocardium. In our network, due to the DRN losing significant amount offeature information in the process of down-sampling, so we use the skip con-nections to transmit the information directly. This structure can not onlyextend context information but also make full use of residual information.Although the middle feature maps with the same resolution can retain thecommon local information through concatenation, it is also doped with alot of unnecessary noise. Thus, we have extracted the intermediate fusedfeature maps of different scales to construct a hybrid pyramid pooling net-work (HPPN). The important local features may be retained while the noiseis removed by a max-pooling operation. The intermediate feature maps ofdifferent scales focus on different local key areas, and their fusion achievesthe aggregation of global feature information, which is of great help to thesegmentation of multiple targets.

The design of the HPPN is different from that of a traditional SPPnetwork. The specific structure is shown in Figure 6. The input of thismodule is four intermediate feature maps of different scales, and then the

18

Page 17: An integrated deep learning framework for joint

max-pooling operation is carried out on them separately. The max-poolingwindow sizes are 8×8, 4×4, 2×2 and 1×1, respectively. The purpose of doingthis is to retain high-level features of the same scale, but they focus on dif-ferent sensitive areas. Low-level feature maps contain abundant informationabout edge, position and shape, but lack semantic information that is help-ful for segmentation, while high-level feature maps are on the opposite. Thefour high-level feature maps are restored to high-resolution maps throughthe three-step 2×2 up-sampling layers and dense connections, and then theglobal guiding segmentation map is obtained through concatenation.

4. Experimental results and analysis

In this section, we first introduce the dataset, including its source, quan-tity and pre-processing. Then, we introduce the configurations, implementa-tion details, evaluation indicators, and results of the experiment. Next, weconduct a series of ablation studies to systematically analyze the effective-ness of each component of the proposed network and its impacts on overallsegmentation performance. Finally, we apply our method to another datasetand compare it with state-of-the-art methods.

4.1. Dataset and pre-processing

In this study, the experiment was performed on 3D short axial imagesof the HVSMR 2016 challenge dataset, including 10 patients with 3D CMRimages (10 training sets and 10 testing sets), each patient having a differentnumber of images taken. Imaging was performed on the axial view of a 1.5-T scanner (Phillips Achieva) without contrast agents using a steady-statefree precession (SSFP) pulse train. During the subjects’ free breathing, theECG and the respiratory-navigator controls were used to remove the heartand respiratory movements (TR=3.4 ms, TE=1.7 ms, α=60). The imagesize and spacing of different subjects were different. In the full-volume train-ing dataset, the average size are 390×390×165 mm and 0.9×0.9×0.85 mm,respectively. Manual segmentation of the blood pool and ventricular my-ocardium was performed by a trained grader and verified by two clinical ex-perts. (The dataset is available at http:// segchd.csail.mit.edu/index.html).

Before training the network, we have pre-processed the dataset. First, wesliced the 3D volume images to obtain the 2D images (in this experiment,we selected axial slices) and removed all the black labels (no target) of eachpatient and their corresponding original images. Then, we adjusted the size

19

Page 18: An integrated deep learning framework for joint

of all images and labels to 128×128 and normalized the values of all pixelpoints. Owing to the insufficient number of training samples, overfitting ofthe model can easily occur. Therefore, random enhancement and rotationtechniques were used to expand the sample numbers of the training dataset(Yu et al., 2017; Dou et al., 2017). Finally, we were able to use a total of 30samples to train our network.

4.2. Implementation details

Our method was implemented on the Tensorflow platform, Python wasused to write the network architecture with the Keras deep learning library,and MatLab was used to realize the data pre-processing. The entire trainingprocess was performed on a 3.40G CPU and 16G NVIDIA GPU. The net-work was trained using the Adam adaptive optimizer with an initial learningrate of 0.001 and mini-batch size of 3. Binary cross-entropy was used tocalculate the loss of pixel-level classification in each iteration, and our modelwas trained for 15,360 iterations in total. Since the labels of testing sets werenot available, effective evaluation criteria could not be obtained. Thus, thetraining sets were divided into training sets and a validation set. We usedleave-one-out method to evaluate the segmentation performance of our net-work. Starting from the first patient, the images of one patient were selectedas the validation set, and the remaining images used as the training sets.The entire training process was repeated 10 times.

4.3. Evaluation indicators

Six indicators based on area overlap and surface distance were used toevaluate the performance of our method: Dice coefficient (Dice), Jaccard co-efficient (Jac), positive predictive value (PPV), sensitivity (Sens), specificity(Spec), and Hausdorff distance of boundaries (HD). These indicators wereused to calculate the segmentation results of the blood pool and myocardiumalone and their formulas are expressed as follows:

Dice(Aseg, Agt) = 2× Aseg ∩ Agt

Aseg + Agt(5)

Jac(Aseg, Agt) =Aseg ∩ Agt

Aseg ∪ Agt(6)

PPV = TP/(TP + FP ) (7)

20

Page 19: An integrated deep learning framework for joint

Figure 7: Segmentation results of 10 subjects.

Sens = TP/(TP + FN) (8)

Spec = TN/(TN + FP ) (9)

HD = (maxi∈seg (maxi∈gt (d (i, j))) ,maxi∈gt (maxi∈seg (d (i, j)))) (10)

where Aseg represents the generated segmentation result, and Agt the realsegmentation result. The greater the value of Dice and Jaccard coefficients,the better the segmentation accuracy. TP represents a positive class of pixelthat can be correctly predicted, FP a negative class of pixel predicted to bea positive class, FN a negative class of pixel that can be correctly predicted,and TN a positive class of pixel predicted to be a negative class. d(i,j)represents the Euclidean distance, and (i,j) represent the boundary point onthe generated or the real segmentation mask.

4.4. Experimental results

In this section, we have quantitatively and qualitatively evaluated theperformance of our segmentation network in the following aspects: (1) viathe results of our method and other advanced segmentation methods, and (2)via a comparison of our method with state-of-the-art segmentation methodsto verify the effectiveness of our method.

4.4.1. Results of our method and other advanced methods

We show the segmentation results of the blood pool and myocardium ofeach subject in Figure 7. The segmentation accuracy of the blood pool ismuch higher than that of the myocardium, indicating that the myocardiumis difficult to segment, which also conforms to the fact that the myocardiumareas in CMR images are relatively small and scattered. For the Dice co-efficient distribution, the Dice value of the blood pool is between 0.90 and0.96, and the Dice value of the myocardium fluctuates around 0.80. For thefour indicators of Jac, PPV, Sens, and Spec, the results of the blood pool

21

Page 20: An integrated deep learning framework for joint

Figure 8: Performance comparison of different methods.

are much better than those of the myocardium. Regarding the HD value,the value of the blood pool is much smaller than that of the myocardium. Ingeneral, the segmentation results of most subjects are kept at the same level,which indicates that our method has a certain stability.

We have implemented the classical segmentation networks FCN andU-Net (Ronneberger et al., 2015). Meanwhile, we also have implementedthe SSLLN framework proposed by (Duan et al., 2018) and SDNet proposedby (Chartsias et al., 2019), which are the effective multi-task segmentationnetworks based on deep learning. Table 4 shows the results of these fivemethods. As shown in Figure 8, it also can be seen from the column diagramof various evaluation indicators that our method has absolute advantages insegmenting the blood pool. However, for the myocardium, the HD valueobtained by our method is larger than that of U-Net, which indicates thatour method still has some shortcomings in the segmentation of small targets.

Regarding the problem that the dataset is too small to train an effectivesegmentation model, we have adopted the data enhancement technology toincrease the original 10 samples to 30 samples. To specify the influence ofthe sample size on model training, Figure 9 shows the validation loss valuesof these methods. It can be seen from the curves that the model with 30samples converges faster and easily tend to a stable state.

Experimental results show that our network has ability to accuratelysegment the blood pool and myocardium and maintain stable segmentationperformance for different objects. Nevertheless, due to the complexity andvariability of targets, there are some bad segmentation results, especially forsmall targets. As shown in Figure 10, we randomly selected a slice with goodprediction and a slice with bad prediction. The top row shows the good pre-diction, that is, our segmentation mask has a very high degree of coincidence

22

Page 21: An integrated deep learning framework for joint

Tab

le4:

Qu

anti

tati

vese

gm

enta

tion

resu

lts

of

our

met

hod

and

fou

rad

van

ced

met

hod

s.T

he

entr

ies

inb

old

hig

hli

ght

the

bes

tco

mp

arab

lere

sult

s.

Met

hod

Dic

eJac

PP

VSen

sSp

ecH

D

Blo

od

Pool

FC

N0.

912±

0.01

90.

838±

0.03

10.

926±

0.03

70.

899±

0.03

30.

985±

0.00

75.

024±

3.04

8

U-N

et0.

927±

0.01

30.

864±

0.02

30.

927±

0.02

60.

928±

0.02

10.

985±

0.00

57.

993±

6.12

1

SS

LL

N0.

908±

0.01

70.

832±

0.02

90.

909±

0.03

60.

908±

0.03

00.

982±

0.00

719

.656±

9.36

7

SD

Net

0.90

1±0.

021

0.82

0±0.

034

0.89

7±0.

046

0.90

7±0.

035

0.97

9±0.

010

20.9

29±

7.27

8

Ours

0.9

46±

0.0

08

0.8

98±

0.0

14

0.9

48±

0.0

17

0.9

45±

0.0

11

0.9

89±

0.0

03

2.0

82±

1.0

35

Myo

card

ium

FC

N0.

728±

0.04

50.

574±

0.05

50.

785±

0.07

10.

683±

0.05

40.

986±

0.00

512

.000±

3.90

2

U-N

et0.

788±

0.03

40.

651±

0.04

60.

838±

0.04

20.

743±

0.03

10.

990±

0.00

25.6

27±

1.5

42

SS

LL

N0.

745±

0.04

50.

596±

0.05

60.

800±

0.04

10.

699±

0.05

70.

987±

0.00

29.

063±

4.49

2

SD

Net

0.72

2±0.

044

0.56

6±0.

054

0.81

7±0.

061

0.64

9±0.

054

0.98

9±0.

004

13.2

63±

5.12

0

Ours

0.8

24±

0.0

27

0.7

01±

0.0

39

0.8

83±

0.0

20

0.7

72±

0.0

37

0.9

92±

0.0

01

11.2

97±

2.68

2

23

Page 22: An integrated deep learning framework for joint

Figure 9: Validation loss values of different numbers of samples.

Figure 10: Examples given to illustrate good and bad segmentation results.

with the real segmentation mask. The bottom row shows the bad prediction,a part of the red box in the lower left corner is mistaken for the blood poolbecause of the similarity in pixel intensity, but a part of the blue box in upperright corner is the real target area. Perhaps there are two main reasons forthe bad segmentation results: (1) partial slices only contain relatively smalland scattered targets, and (2) there are too few similar samples to fully trainthe model.

We have qualitatively analyzed the effectiveness of our method in seg-menting the blood pool and myocardium from three view slices of 3D images.

24

Page 23: An integrated deep learning framework for joint

Figure 11: Qualitative segmentation results of our method and four advanced methods.Segmentation masks of blood pool and myocardium are shown from three view slices, in-cluding sagittal slice, axial slice, and coronal slice, from left to right. First row shows threeview slices and real segmentation masks (white represents blood pool, and gray representsmyocardium). Second row shows the Ground-truth of blood pool and myocardium sep-arately, and third to last rows show predicted results of our method and other advancedmethods (left column is the blood pool, and right column is the myocardium).

As shown in Figure 11, we selected a subject’s 3D image for analysis. Weshow its three view slices and the real segmentation masks as well as thepredicted masks of other methods. In this study, we do not directly seg-ment the slices of other two viewpoints. Instead, we creatively reconstructedthe segmentation masks of axial slices through 3D reconstruction technologyand then sliced it from the other two viewpoints. To more intuitively demon-strate the effectiveness of our method, we have extracted the blood pool andmyocardium from images for separate analysis. As can be seen from thesefigures, the distribution of the blood pool and myocardium is complex andhighly variable, especially the myocardium areas. Although the sample dis-tribution is seriously unbalanced, the predicted results of our method are,to a large extent, more in line with the Ground-truth than those of other

25

Page 24: An integrated deep learning framework for joint

advanced methods.

4.4.2. Performance comparison of our method and state-of-the-art methods

Table 5 shows the segmentation results of our method and state-of-the-artmethods on the 2016 HVSMR dataset. (Cicek et al., 2016) proposed the 3DU-Net framework, that is, all the 2D operations are replaced by 3D opera-tions, and the best segmentation results are obtained by tuning parameters.We also compared our segmentation results with two similar 3D segmentationnetworks: the VoxResNet network proposed by (Chen et al., 2018) to seg-ment MR images of the brain, and the densely connected VoxResNet networkproposed by (Yu et al., 2017) on the basis of VoxResNet framework. That is,adding more skip connections to the residual network. Moreover, similar toour method, (Wolterink et al., 2016) only used the 2D dilated convolutionalnetwork to segment 2D slices. Furthermore, we compared our method withtwo traditional segmentation methods, namely that of, (Tziritas, 2016), whocombined a 3D Markov random field (MRF) model with substructure track-ing, and that of (Shahzad et al., 2016), who combined multiple atlas andlevel sets to automatically segment cardiovascular images. The results of allcomparative methods are presented in Table 5 for analysis and comparison.

For the segmentation of the blood pool, the Dice and HD results areabsolutely superior to other methods. However, on other four classificationindexes, Jac, PPV, Sens, and Spec, our results are slightly lower than thoseof Wolterink in terms of PPV and Spec. For the segmentation of the my-ocardium, our results are the best in terms of Dice, Jac, PPV, and Spec, whileDenseVoxRet achieved the best result in Sens and Shahzad et al. achievedthe best result in HD. However, considering the simultaneous segmentationof the blood pool and myocardium, our method has an absolute advantagein most evaluation indicators. It is worth noting that the standard deviationof each index is basically smaller than that of other methods, which, to someextent, also indicates that our method has strong stability and generaliza-tion ability. In addition, most of the comparable methods we have chosenare based on 3D CMR volume images while our method selected the axialslices of 3D images, which inevitably leads to performance differences amongmethods.

4.4.3. Segmentation results of challenging images

The segmentation results of the example images in Figure 2 are presentedone by one in Figure 12. Accurate segmentation of challenging images proves

26

Page 25: An integrated deep learning framework for joint

Tab

le5:

Per

form

ance

com

par

ison

ofou

rm

eth

od

and

stat

e-of

-th

e-ar

tm

ethod

s.T

he

entr

ies

inb

old

hig

hli

ght

the

bes

tco

mp

a-

rab

lere

sult

s.

Met

hod

Dic

eJac

PP

VSen

sSp

ecH

D

Blo

od

Pool

(Cic

eket

al.,

2016

)0.

926±

0.01

60.

863±

0.02

80.

940±

0.02

80.

916±

0.04

80.

989±

0.00

58.

628±

3.39

0

(Chen

etal

.,20

18)

0.92

9±0.

013

\\

\\

9.96

6±3.

021

(Yu

etal

.,20

17)

0.93

1±0.

011

\\

\\

9.53

3±4.

194

(Wol

teri

nk

etal

.,20

16)

0.92

6±0.

018

0.86

3±0.

030

0.9

51±

0.0

24

0.90

5±0.

047

0.9

92±

0.0

04

7.06

9±2.

857

(Tzi

rita

s,20

16)

0.86

7±0.

047

0.76

8±0.

068

0.86

1±0.

062

0.88

9±0.

108

0.97

2±0.

014

19.7

23±

4.07

8

(Shah

zad

etal

.,20

16)

0.88

5±0.

028

0.79

5±0.

044

0.90

7±0.

052

0.86

7±0.

046

0.98

4±0.

008

9.40

8±3.

059

Ours

(DR

HP

PN

)0.9

46±

0.0

08

0.8

98±

0.0

14

0.94

8±0.

017

0.9

45±

0.0

11

0.98

9±0.

003

2.0

82±

1.0

35

Myo

card

ium

(Cic

eket

al.,

2016

)0.

694±

0.07

60.

536±

0.08

90.

798±

0.07

60.

618±

0.09

20.

992±

0.00

410

.221±

4.33

9

(Chen

etal

.,20

18)

0.77

4±0.

067

\\

\\

6.57

2±3.

551

(Yu

etal

.,20

17)

0.82

1±0.

041

\\

\\

7.29

4±3.

340

(Wol

teri

nk

etal

.,20

16)

0.80

2±0.

060

0.67

3±0.

084

0.80

2±0.

065

0.8

05±

0.0

76

0.99

0±0.

004

6.12

6±3.

565

(Tzi

rita

s,20

16)

0.61

2±0.

153

0.45

7±0.

149

0.66

6±0.

164

0.57

1±0.

150

0.98

5±0.

008

13.1

99±

6.02

5

(Shah

zad

etal

.,20

16)

0.74

7±0.

075

0.60

2±0.

094

0.76

7±0.

054

0.73

4±0.

108

0.98

9±0.

004

5.0

91±

1.6

58

Ours

(DR

HP

PN

)0.8

24±

0.0

27

0.7

01±

0.0

39

0.8

83±

0.0

20

0.77

2±0.

037

0.9

92±

0.0

01

11.2

97±

2.68

2

27

Page 26: An integrated deep learning framework for joint

Figure 12: Examples of images that are challenging to segment, including the original axialslices, the Ground-truth, and our segmentation results. (For the Ground-truth, whiterepresents blood pool, and gray represents myocardium. For our segmentation results,blue represents blood pool, and green represents myocardium.)

the effectiveness of our method. Although there are a few pixels with inaccu-rate classification, but it does not affect the overall judgment of the targets.

4.5. Ablation studies

The proposed segmentation network is mainly based on the effective in-tegration of multiple basic modules. It is necessary to study the effectsof different modules on the performance of the entire segmentation network

28

Page 27: An integrated deep learning framework for joint

Table 6: Main modules of our network and five ablation networks.

ModuleMethod

Ours AS 1 AS 2 AS 3 AS 4 AS 5

Inception module√ × √ √ √ √

Dilated convolution√ √ × √ √ √

ConvLSTM√ √ √ × √ √

Skip connection√ √ √ √ × √

HPPN√ √ √ √ √ ×

Figure 13: Performance comparison of our network and ablation networks.

through detailed ablation experiments. As shown in Table 6, we explored theeffects of general structures such as inception module, dilated convolution,ConvLSTM, and skip connection. In addition, compared with the backboneof our segmentation network, the HPPN can supplement and strengthen thesegmentation map in some respects. Our intention in constructing such aframework is to cope with the challenges posed by the complexity of images,so we investigated the independent contribution of these infrastructures inaddressing challenges (”

√” means to retain the specific module; ”×” means

to remove or replace the specific module).Table 7 and Figure 13 show the segmentation results of our network

and five ablation networks that lack a specific module. Although the seg-mentation performance of these ablation networks is also good, there is stilla small gap between our results and those of the ablation networks. Atthe same time, we also qualitatively demonstrate the segmentation masks ofthese networks in Figure 14, and it is not difficult to find that the structureof our network is more preferable in detail.

29

Page 28: An integrated deep learning framework for joint

Tab

le7:

Qu

anti

tati

vese

gmen

tati

onre

sult

sof

our

net

wor

kan

dfi

veab

lati

onn

etw

orks.

Th

een

trie

sin

bold

hig

hli

ght

the

bes

tco

mp

arab

lere

sult

s.

Met

hod

Dic

eJac

PP

VSen

sSp

ecH

D

Blo

od

Pool

AS

10.

930±

0.01

30.

870±

0.02

20.

947±

0.02

90.

915±

0.02

40.9

90±

0.0

03

3.46

2±1.

635

AS

20.

935±

0.01

00.

877±

0.01

70.

942±

0.02

20.

929±

0.02

60.

988±

0.00

53.

244±

1.04

2

AS

30.

945±

0.00

90.

897±

0.01

50.9

50±

0.0

18

0.94

3±0.

019

0.99

0±0.

004

1.9

57±

0.9

25

AS

40.

942±

0.00

90.

890±

0.01

60.

943±

0.01

90.

941±

0.01

60.

988±

0.00

52.

782±

1.47

5

AS

50.

944±

0.01

00.

895±

0.01

80.

945±

0.02

20.

944±

0.01

80.

989±

0.00

32.

032±

0.72

4

Ours

0.9

46±

0.0

08

0.8

98±

0.0

14

0.94

8±0.

017

0.9

45±

0.0

11

0.98

9±0.

003

2.08

2±1.

035

Myo

card

ium

AS

10.

789±

0.06

00.

655±

0.07

70.

864±

0.03

70.

729±

0.08

20.

992±

0.00

218

.154±

9.38

7

AS

20.

803±

0.03

50.

672±

0.04

70.

876±

0.03

30.

741±

0.04

00.

992±

0.00

114

.264±

4.28

7

AS

30.

818±

0.03

30.

694±

0.04

60.9

04±

0.0

16

0.74

8±0.

048

0.99

2±0.

002

16.0

35±

4.97

4

AS

40.

812±

0.03

40.

685±

0.04

80.

878±

0.03

00.

756±

0.04

20.

992±

0.00

212

.471±

4.30

3

AS

50.

821±

0.02

50.

697±

0.03

50.

886±

0.02

50.

766±

0.03

20.9

93±

0.0

02

15.1

30±

4.77

1

Ours

0.8

24±

0.0

27

0.7

01±

0.0

39

0.88

3±0.

020

0.7

72±

0.0

37

0.99

2±0.

001

11.2

97±

2.6

82

30

Page 29: An integrated deep learning framework for joint

Figure 14: Visualization of segmentation results of blood pool and myocardium by ournetwork and four ablation networks. First row shows the real segmentation mask. Keyareas are marked in different images with dotted boxes of different colors. Each row ofthe figure below has the same dotted box at the same position, which represents the areaswith inaccurate segmentation (left column is the blood pool, and right column is themyocardium).

4.5.1. Inception module for multi-level feature extraction

Owing to the operation of the 1×1 convolution kernel and the strategyof multi-scale convolutional layer fusion, the inception module has a bet-ter ability in multi-level feature extraction and network nonlinear expressioncompared with an ordinary single convolutional layer. Therefore, in our net-work, a modified inception module is designed to achieve the initial featurerepresentation of cardiovascular images. Meanwhile, we further use experi-ments to explore its role in the proposed network. As shown on the thirdline in Table 7, AS 1 represents the experimental results after replacing theinception module with a single scale convolutional layer. It can be clearlyseen from the table that without the application of the inception module,

31

Page 30: An integrated deep learning framework for joint

Figure 15: Segmentation of marginal targets and local weak/no boundaries.

the segmentation performance has been significantly reduced, especially onDice values.

4.5.2. Dilated convolution for marginal targets capture

Theoretically, the dilated convolution is focused on the marginal targets,because it is capable of increasing the receptive field of the convolution kernelwhile keeping the size of it unchanged. In our multi-target segmentationtask, the diversity of the targets morphology and motion makes it difficult tolocate the marginal regions, so we adopted the dilated convolution to improvethe segmentation performance. As shown in Figure 15(a), AS 2 replacesthe dilated convolution in our network with the ordinary convolution. Thesegmentation masks of these three slices that the segmentation performanceof the dilated convolution on the marginal region (the areas depicted in thered boxes and the white arrows) is significant. The quantitative segmentationresults in Table 7 also show that AS 2 underperforms our method on all theindicators.

4.5.3. ConvLSTM for local weak/no boundaries detection

For sequential images, accurate segmentation relies heavily on contextinformation captured by ConvLSTM. Meanwhile, for some images with lo-

32

Page 31: An integrated deep learning framework for joint

Figure 16: Visualization of intermediate feature maps. First row shows the multi-levelinitial feature maps obtained by inception module (image resolution: 64×64). Second rowshows the preliminary segmentation maps obtained by DRN. Third row shows the featuremaps of different scales of HPPN input (from left to right: 64×64, 32×32, 16×16, 8×8).Fourth row shows the feature maps after the concatenation of the output of HPPN andDRN (64×64).

cal weak/no boundaries (the areas depicted in the red boxes and the whitearrows in Figure 15(b)), it is difficult to detect the boundary in terms ofthe current image. However, we can try to use the detection results of theprevious image to guide the boundary detection of the current image. Inthis way, sufficient context information is of great importance to perform-ing segmentation. As shown in Figure 15(b), our method has the ability toachieve accurate detection of blurry boundaries, while AS 3 (the ConvLSTMmodule is removed) does not have that effect. Although Table 7 shows thatAS 3 performs better than our method on the value of PPV, HD, and Spec,ConvLSTM module has still its own independent contribution.

4.5.4. HPPN module for the solution of inconsistent appearance intensity

Due to some images still existing the problem of inconsistent appearanceintensity (highly inconsistent appearance intensity of the same targets; thehigh similarity of appearance intensity of different targets), while the DRNcannot completely solve this problem. As can be seen from the white arrowsin the second row in Figure 16(a), the DRN may ignore the target area where

33

Page 32: An integrated deep learning framework for joint

the appearance intensity does not match. And in Figure 16(b), it seems toweaken the appearance intensity of target area consistent with the Ground-truth. Meanwhile, there is also an incorrect appearance display shown (thenon-target region is incorrectly shown as the target region) as shown in thered box.

To better deal with these problems, we added the HPPN module forfurther processing. After the HPPN module, we have obtained the featuremaps shown in the fourth row in Figure 16. The feature maps of fourthrow show that the target areas can be strengthened and the non-target areascan be weakened. The quantitative results in Table 7 also indicate thatcompared with our network, AS 4 does not achieve that good segmentationperformance due to the lack of HPPN module.

4.5.5. Effectiveness of the inception module and the dense connections

We have evaluated the segmentation performance difference of the in-ception module and the ordinary multi-scale convolution to prove that theformer is more advanced than the latter. Furthermore, the purpose of con-structing the HPPN is to make use of its max-pooling layers of different scalesand realize more accurate location through parallel up-sampling paths, so wealso evaluated the effect of adding dense connections to the structure on seg-mentation performance. Table 8 shows the influence of common multi-scaleconvolution and the inception module on segmentation performance. Table9 shows the effect of the addition of dense connections on segmentation per-formance. Quantitative analysis proves that the module design of our modelcertainly improves the overall segmentation performance.

4.6. Application to another dataset

To investigate the effectiveness and generalization ability of the proposedmethod, we have applied our method to the public challenge dataset: MIC-CAI 2017 Multi-Modality Whole Heart Segmentation (MM-WHS). The goalof the challenge is to segment seven whole-heart substructures in multi-modality (MR/CT) images. Datasets are obtained in clinics using differentscanners, resulting in different image quality, resolution and voxel spacing.In line with our main work, we used the MRI volumes (20 volumes withcorresponding manual segmentation) and segmented its five substructures,including the left ventricle blood cavity (LV), the right ventricle blood cav-ity (RV), the left atrium blood cavity (LA), the right atrium blood cavity(RA) and the myocardium of the left ventricle (LV Myo). High variations

34

Page 33: An integrated deep learning framework for joint

Tab

le8:

Seg

men

tati

on

per

form

ance

ofin

cep

tion

mod

ule

and

ord

inar

ym

ult

i-sc

ale

convol

uti

on.

Th

een

trie

sin

bold

hig

hli

ght

the

bes

tco

mp

arab

lere

sult

s.

Dic

eJac

PP

VSen

sSp

ecH

D

Blo

od

Pool

Ince

pti

onm

od

ule

0.9

46±

0.0

08

0.8

98±

0.0

14

0.94

8±0.

017

0.9

45±

0.0

11

0.98

9±0.

003

2.0

82±

1.0

35

Mu

lti-

scal

eC

onv

0.94

4±0.

008

0.89

4±0.

014

0.9

52±

0.0

18

0.93

6±0.

015

0.9

91±

0.0

03

2.30

3±1.

201

Myo

card

ium

Ince

pti

onm

od

ule

0.8

24±

0.0

27

0.7

01±

0.0

39

0.8

83±

0.0

20

0.7

72±

0.0

37

0.99

2±0.

001

11.2

97±

2.6

82

Mu

lti-

scal

eC

onv

0.82

2±0.

028

0.69

9±0.

041

0.88

2±0.

037

0.77

1±0.

033

0.9

93±

0.0

01

13.4

55±

4.91

9

35

Page 34: An integrated deep learning framework for joint

Tab

le9:

Seg

men

tati

onp

erfo

rman

cew

ith

(w/)

orw

ith

out

(w/o

)ad

din

gd

ense

con

nec

tion

s.T

he

entr

ies

inb

old

hig

hli

ght

the

bes

tco

mp

arab

lere

sult

s.

Den

seco

n-

nec

tion

sD

ice

Jac

PP

VSen

sSp

ecH

D

Blo

od

Pool

w/

0.9

46±

0.0

08

0.8

98±

0.0

14

0.9

48±

0.0

17

0.9

45±

0.0

11

0.98

9±0.

003

2.0

82±

1.0

35

w/o

0.94

4±0.

008

0.89

4±0.

014

0.94

8±0.

020

0.94

0±0.

019

0.9

90±

0.0

04

2.26

2±0.

732

Myo

card

ium

w/

0.8

24±

0.0

27

0.7

01±

0.0

39

0.88

3±0.

020

0.7

72±

0.0

37

0.99

2±0.

001

11.2

97±

2.6

82

w/o

0.81

6±0.

030

0.69

1±0.

043

0.8

97±

0.0

22

0.75

0±0.

041

0.9

94±

0.0

01

13.6

73±

4.28

7

36

Page 35: An integrated deep learning framework for joint

Table 10: Quantitative segmentation results of our method and four advanced methods.The entries in bold highlight the best comparable results.

Method LV RV LA RA LV Myo

FCN 0.884 0.815 0.813 0.838 0.771

U-Net 0.903 0.830 0.833 0.835 0.806

SSLLN 0.893 0.863 0.835 0.860 0.800

SDNet 0.910 0.840 0.794 0.839 0.809

Ours 0.933 0.902 0.905 0.910 0.858

of anatomy and signal intensity bring great difficulties to the whole heartsegmentation, so that this dataset was adopted as a supplementary experi-ment to verify that our method could be applied to different types of imagingdatasets.

The challenge provides 20 patients with original training images (MRI)and label images that vary in sizes and numbers. We have employed thesame data augmentation (Shi et al., 2018; Yang et al., 2017) and image pre-processing techniques as with the CMR images, and the network architectureand implementation details were consistent. In terms of evaluation indicator,the Dice metric is adopted to evaluate the segmentation performance of ourmethod for the sake of facilitating the comparison with other state-of-the-artmethods. Since the labels of the test set are not public, we also adoptedthe leave-one-out method to train the network. Similarly, four advanced seg-mentation networks are applied to this dataset, and the quantitative andqualitative segmentation results are shown in Table 10 and Figure 17. Werandomly selected four slices of different subjects to show the segmentationresults of our method and four advanced methods on MM-WHS dataset. Ascan be seen in Figure 17, the segmentation results of our model are highlyconsistent with the real segmentation mask.

Different from the blood pool and myocardium in the cardiovascular im-ages, the whole heart structures have relatively regular and uniform externalshape, but the continuous movement of the heart still brings great challengeto the segmentation of the whole heart substructures. Regarding the MM-WHS dataset, many researchers have proposed several effective segmentation

37

Page 36: An integrated deep learning framework for joint

Figure 17: Qualitative segmentation results of our method and four advanced methods.

methods. For example, (Mortazi et al., 2017) proposed a method based onmulti-planar deep CNN with an adaptive fusion strategy, and achieved thebest result in that year’ challenge. (Heinrich and Oster, 2018) won the firstplace in the challenge by using discrete nonlinear registration and fast non-local fusion. (Payer et al., 2018) proposed a pipeline of two fully convolu-tional networks, the first CNN is used to locate the center of the boundingbox containing all cardiac substructures while the second CNN is used topredict the categories of pixels in the ROI. Moreover, a deep voxel-wise di-lated residual network based on Bayesian probability proposed by (Shi et al.,2018) also achieved impressive results. These methods are all based on deepCNN, but some non-CNN methods have also been proposed. For example,(Galisot et al., 2018) proposed a method based on local probability atlasand posterior correction. Although there is only a small gap between ourresults and those of Mortazi et al. on LV, we have obtained remarkable seg-mentation results in terms of RV, LA, RA, and LV Myo. In particular, our

38

Page 37: An integrated deep learning framework for joint

Table 11: Performance comparison of our method and state-of-the-art methods. Theentries in bold highlight the best comparable results.

Method LV RV LA RA LV Myo

(Galisot et al., 2018) 0.897 0.819 0.765 0.808 0.763

(Yang et al., 2017) 0.858 0.819 0.787 0.820 0.742

(Shi et al., 2018) 0.914 0.880 0.856 0.873 0.811

(Payer et al., 2018) 0.916 0.868 0.855 0.881 0.778

(Heinrich and Oster, 2018) 0.918 0.871 0.886 0.873 0.781

(Mortazi et al., 2017) 0.932 0.884 0.887 0.874 0.825

Ours (DRHPPN) 0.933 0.902 0.905 0.910 0.858

model achieves significant advantages over other state-of-the-art methods inthis multi-objective segmentation task. On the whole, our method is quiteeffective in whole heart segmentation and it is also proved that our methodis appropriate for this dataset.

5. Conclusion

In this paper, we have proposed a multi-task deep learning method basedon the dilated residual and hybrid pyramid pooling network (DRHPPN).The aim of this framework is to simultaneously and automatically segmentthe blood pool and myocardium in cardiovascular images. Our method isto integrate the CNN-based cardiovascular image representation many timesand creatively analyze and utilize the feature maps extracted in the deepframework. The dynamic relationship of multi-object image sequences ismodeled while learning multi-object morphological changes. Although ourwork is only based on axial slices of 3D CMR images, the stability of ournetwork is verified from the perspectives of both sagittal and coronal slices.Numerous ablation studies and comparative experiments also prove that theproposed method is advanced and effective in segmenting the blood pooland myocardium and has great potential in clinical application, especially inthe pre-operative planning and guidance of CHD. Furthermore, our method

39

Page 38: An integrated deep learning framework for joint

was applied to the MM-WHS dataset and still achieved good segmentationresults, which indicates that our method has a certain generalization ability.

We will continue to study the following key points in the future work: (1)more advanced network will be designed to realize multi-view segmentation of3D CMR images, so as to achieve better segmentation performance; (2) tryingto do more theoretical analysis and applying the theories to the experiments;(3) validating the performance of our method on more datasets.

6. Acknowledgments

This work was supported in part by the National Science Foundation ofChina under Grant 61673020, and in part by the Anhui Provincial NaturalScience Foundation under Grant 1708085QF143.

7. Conflict of Interest

The authors declare on conflict of interest.

References

Afshin, M., Ayed, I. B., Islam, A., Goela, A., Li, S., 2012. Global assess-ment of cardiac function using image statistics in mri. In: Medical ImageComputing and Computer-Assisted Intervention (MICCAI). pp. 535–543.

Chartsias, A., Joyce, T., Papanastasiou, G., Williams, M., Newby, D., Dhar-makumar, R., Tsaftaris, S., 2019. Factorised representation learning incardiac image analysis. Medical Image Analysis 58, 101535.

Chen, H., Dou, Q., Yu, L., Qin, J., Heng, P.-A., 2018. Voxresnet: Deepvoxelwise residual networks for brain segmentation from 3d mr images.Neuroimage 170 (15), 446–455.

Chen, L.-C., Papandreou, G., Kokkinos, I., Murphy, K., Yuille, A. L.,2016. Deeplab: Semantic image segmentation with deep convolutional nets,atrous convolution, and fully connected crfs. IEEE Transactions on PatternAnalysis & Machine Intelligence (TPAMI) 40 (4), 834–848.

Cicek, O., Abdulkadir, A., Lienkamp, S. S., Brox, T., Ronneberger, O., 2016.3d u-net: Learning dense volumetric segmentation from sparse annotation.In: Medical Image Computing and Computer-Assisted Intervention (MIC-CAI). pp. 424–432.

40

Page 39: An integrated deep learning framework for joint

Dou, Q., Yu, L., Chen, H., Jin, Y., Yang, X., Qin, J., Heng, P.-A., 10 2017.3d deeply supervised network for automated segmentation of volumetricmedical images. Vol. 41. pp. 40–54.

Duan, J., Bello, G., Schlemper, J., Bai, W., Dawes, T. J. W., Biffi, C.,de Marvao, A., Doumou, G., O’Regan, D. P., Rueckert, D., 2018. Auto-matic 3d bi-ventricular segmentation of cardiac images by a shape-refinedmulti-task deep learning approach. IEEE Transactions on Medical Imaging(TMI) 38, 2151–2164.

Galisot, G., Brouard, T., Ramel, J.-Y., 2018. Local probabilistic atlases anda posteriori correction for the segmentation of heart images. In: StatisticalAtlases and Computational Models of the Heart ACDC and MMWHSChallenges. pp. 207–214.

Grosgeorge, D., Petitjean, C., Dacher, J.-N., Ruan, S., 2013. Graph cut seg-mentation with a statistical shape model in cardiac mri. Computer Visionand Image Understanding (Cviu) 117 (9), 1027–1035.

He, K., Zhang, X., Ren, S., Sun, J., 2014. Spatial pyramid pooling in deepconvolutional networks for visual recognition. IEEE Transactions on Pat-tern Analysis & Machine Intelligence (TPAMI) 37 (9), 1904–1916.

Heinrich, M. P., Oster, J., 2018. Mri whole heart segmentation using discretenonlinear registration and fast non-local fusion. In: Statistical Atlases andComputational Models of the Heart ACDC and MMWHS Challenges. pp.233–241.

Ioffe, S., Szegedy, C., 2015. Batch normalization: Accelerating deep networktraining by reducing internal covariate shift. In: International Conferenceon Machine Learning. Vol. 37. pp. 448–456.

Liu, J., Xie, H., Zhang, S., Gu, L., 2019. Multi-sequence myocardium seg-mentation with cross-constrained shape and neural network-based initial-ization. Computerized Medical Imaging and Graphics 71, 49–57.

Liu, N., Han, J., 2016. Dhsnet: Deep hierarchical saliency network for salientobject detection. In: 2016 IEEE Conference on Computer Vision and Pat-tern Recognition (CVPR). pp. 678–686.

41

Page 40: An integrated deep learning framework for joint

Long, J., Shelhamer, E., Darrell, T., 2014. Fully convolutional networks forsemantic segmentation. IEEE Transactions on Pattern Analysis and Ma-chine Intelligence (TPAMI) 39 (4), 640–651.

Lorenzo-Valds, M., Sanchez-Ortiz, G. I., Elkington, A. G., Mohiaddin, R. H.,Rueckert, D., 2004. Segmentation of 4d cardiac mr images using a proba-bilistic atlas and the em algorithm. Medical Image Analysis 8 (3), 255–265.

Mahapatra, D., Buhmann, J. M., 2013. Automatic cardiac rv segmentationusing semantic information with graph cuts. In: IEEE 10th InternationalSymposium on Biomedical Imaging (ISBI). pp. 1106–1109.

Maier, O. M. O., Jimenez, D., Santos, A., Ledesma-Carbayo, M. J., 2012.Segmentation of rv in 4d cardiac mr volumes using region-merging graphcuts. In: Computing in Cardiology. pp. 697–700.

Mitchell, S. C., Lelieveldt, B. P. F., Van, der Geest, R. J., Bosch, H. G.,Reiber, J. H. C., Sonka, M., 2001. Multistage hybrid active appearancemodel matching: Segmentation of left and right ventricles in cardiac mrimages. IEEE Transactions on Medical Imaging (TMI) 20 (5), 415–423.

Moolan-Feroze, O., Mirmehdi, M., Hamilton, M., Bucciarelli-Ducci, C., 2014.Segmentation of the right ventricle using diffusion maps and markov ran-dom fields. In: Medical Image Computing and Computer-Assisted Inter-vention (MICCAI). Vol. 17. pp. 682–689.

Mortazi, A., Burt, J., Bagci, U., 2017. Multi-planar deep segmentation net-works for cardiac substructures from mri and ct. In: Statistical Atlasesand Computational Models of the Heart ACDC and MMWHS Challenges.pp. 199–206.

Mukhopadhyay, A., 2016. Total variation random forest: Fully automaticmri segmentation in congenital heart diseases. In: International Workshopon Reconstruction and Analysis of Moving Body Organs. pp. 165–171.

Ordovas, K. G., Higgins, C. B., 2011. Delayed contrast enhancement on mrimages of myocardium: Past, present, future. Radiology 261 (2), 358–374.

Pace, D., Dalca, A., Geva, T., Powell, A., Hedjazi Moghari, M., Golland,P., 2015. Interactive whole-heart segmentation in congenital heart disease.

42

Page 41: An integrated deep learning framework for joint

Medical Image Computing and Computer-Assisted Intervention (MIC-CAI), 80–88.

Payer, C., SternHorst, a., Urschler, B., 2018. Multi-label whole heart seg-mentation using cnns and anatomical label configurations. In: StatisticalAtlases and Computational Models of the Heart ACDC and MMWHSChallenges. pp. 190–198.

Petitjean, C., Dacher, J.-N., 2011. A review of segmentation methods in shortaxis cardiac mr images. Medical Image Analysis 15 (2), 169–184.

Qi, X., Zhang, L., Chen, Y., Pi, Y., Chen, Y., Lv, Q., Yi, Z., 2019. Automateddiagnosis of breast ultrasonography images using deep neural networks.Medical Image Analysis 52, 185–198.

Qin, C., Bai, W., Schlemper, J., Petersen, S. E., Piechnik, S. K., Neubauer,S., Rueckert, D., 2018. Joint learning of motion estimation and segmenta-tion for cardiac mr image sequences. In: Medical Image Computing andComputer-Assisted Intervention (MICCAI). pp. 472–480.

Ronneberger, O., Fischer, P., Brox, T., 2015. U-net: Convolutional networksfor biomedical image segmentation. In: Medical Image Computing andComputer-Assisted Intervention (MICCAI). Vol. 9351. pp. 234–241.

Shahzad, R., Shan, G., Qian, T., Dzyubachyk, O., Geest, R. V. D., 2016.Automated cardiovascular segmentation in patients with congenital heartdisease from 3d cmr scans: Combining multi-atlases and level-sets, 147–155.

Shi, X., Chen, Z., Wang, H., Yeung, D. Y., Wong, W. K., Woo, W. C., 2015.Convolutional lstm network: A machine learning approach for precipita-tion nowcasting. In: Advances in Neural Information Processing Systems28. pp. 802–810.

Shi, Z., Zeng, G., Le, Z., Zhuang, X., Li, L., Yang, G., Zheng, G., 2018.Bayesian voxdrn: A probabilistic deep voxelwise dilated residual networkfor whole heart segmentation from 3d mr images. In: Medical Image Com-puting and Computer-Assisted Intervention (MICCAI). pp. 569–577.

Song, X., Wang, Y., Feng, Q., Wang, Q., 2019. Improved graph cut modelwith features of superpixels and neighborhood patches for myocardium

43

Page 42: An integrated deep learning framework for joint

segmentation from ultrasound image. Mathematical Biosciences and Engi-neering 16 (3), 1115–1137.

Su, H., Liu, F., Xie, Y., Xing, F., Lin, Y., 2015. Region segmentation inhistopathological breast cancer images using deep convolutional neuralnetwork. In: IEEE 12th International Symposium on Biomedical Imag-ing (ISBI). pp. 55–58.

Suinesiaputra, A., Bluemke, D. A., Cowan, B. R., Friedrich, M. G., Kramer,C. M., Kwong, R. K., Plein, S., Schulzmenger, J., Westenberg, J. J. M.,Young, A. A., 2015. Quantification of lv function and mass by cardiovas-cular magnetic resonance: multi-center variability and consensus contours.IEEE Journal of Cardiovascular Magnetic Resonance 17 (1), 63.

Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan,D., Vanhoucke, V., Rabinovich, A., 2015. Going deeper with convolutions.In: 2015 IEEE Conference on Computer Vision and Pattern Recognition(CVPR).

Taha, A., Lo, P., Li, J., Zhao, T., 2018. Kid-net: Convolution networks forkidney vessels segmentation from ct-volumes. In: Medical Image Comput-ing and Computer-Assisted Intervention (MICCAI). pp. 463–467.

Tziritas, G., 2016. Fully-automatic segmentation of cardiac images using3d mrf model optimization and substructures tracking. In: InternationalWorkshop on Reconstruction and Analysis of Moving Body Organs. pp.129–136.

Wang, T., Zhang, L., Wang, S., Lu, H., Yang, G., Ruan, X., Borji, A., 2018.Detect globally, refine locally: A novel approach to saliency detection.In: 2018 IEEE Conference on Computer Vision and Pattern Recognition(CVPR). pp. 3127–3135.

Wolterink, J. M., Leiner, T., Viergever, M. A., Ivana, I. I., 2016. Dilatedconvolutional neural networks for cardiovascular mr segmentation in con-genital heart disease. In: International Workshop on Reconstruction andAnalysis of Moving Body Organs. pp. 95–102.

Woo, J., Slomka, P. J., Kuo, C.-C. J., Hong, B.-W., 2013. Multiphase seg-mentation using an implicit dual shape prior: Application to detection of

44

Page 43: An integrated deep learning framework for joint

left ventricle in cardiac mri. Computer Vision and Image Understanding(Cviu) 117 (9), 1084–1094.

Xu, B., Wang, N., Chen, T., Li, M., 2015. Empirical evaluation of rectifiedactivations in convolutional network. Computer Science.

Xu, C., Xu, L., Gao, Z., Shen, Z., Zhang, H., Zhang, Y., Du, X., Zhao, S.,Ghista, D., Li, S., 2017. Direct detection of pixel-level myocardial infarc-tion areas via a deep-learning algorithm. In: Medical Image Computingand Computer-Assisted Intervention (MICCAI). Vol. 10435. pp. 240–249.

Xue, W., Lum, A., Mercado, A., Landis, M., Warrington, J., Li, S., 2017a.Full quantification of left ventricle via deep multitask learning networkrespecting intra- and inter-task relatedness. In: Medical Image Computingand Computer-Assisted Intervention (MICCAI). pp. 276–284.

Xue, W., Nachum, I. B., Pandey, S., Warrington, J., Leung, S., Li, S., 2017b.Direct estimation of regional wall thicknesses via residual recurrent neu-ral network. In: International Conference on Information Processing inMedical Imaging. pp. 505–516.

Yang, X., Bian, C., Yu, L., Dong, N., Heng, P. A., 2017. Hybrid loss guidedconvolutional networks for whole heart parsing. In: Statistical Atlases andComputational Models of the Heart ACDC and MMWHS Challenges. pp.215–223.

Yu, F., Koltun, V., 2015. Multi-scale context aggregation by dilated convo-lutions. In: Computer Science.

Yu, L., Cheng, J., Dou, Q., Yang, X., Chen, H., Qin, J., Heng, P. A., 2017.Automatic 3d cardiovascular mr segmentation with densely-connected vol-umetric convnets. In: Medical Image Computing and Computer-AssistedIntervention (MICCAI). pp. 287–295.

Zhao, S., Gao, Z., Zhang, H., Xie, Y., Luo, J., Ghista, D., Wei, Z., Bi,X., Xiong, H., Xu, C., Li, S., 2017. Robust segmentation of intimamediaborders with different morphologies and dynamics during the cardiac cycle.IEEE Journal of Biomedical and Health Informatics 22 (5), 1571–1582.

45