Despite its high sensitivity, MR imaging has low specificity and high false positive issues. Therefore, automatic breast lesion detection algorithms are necessary. To this end, we propose a new network, dense attention network (DANet), for breast lesion segmentation in MR images. In DANet, we designed a feature fusion and selection mechanism. Features from the corresponding encoder layer and from all previous decoder layers are fused by concatenation. To highlight the rich-informative channels, a channel attention module is introduced. DANet showed better segmentation results compared to commonly applied segmentation networks on our 2D contrast-enhanced T1-weighted breast MR dataset.
Introduction
Breast cancer is one of the most frequent cancer types in the female population. Early diagnosis is important for the reduction of breast cancer mortality rate. Among all the available breast imaging techniques, MR imaging possesses the highest sensitivity1. However, MR imaging is subject to relatively high false positive rate and low specificity2. And manual reading may prone to errors3. Computer-aided detection/diagnostic (CAD) systems that can automatically and accurately detect breast lesions in MR images is thereby, important and necessary. Current available CAD systems mainly use manual defined features to locate and classify breast lesions4. A few studies applying deep learning methods, such as UNet, to breast cancer detection5. The fast updating models in computer vision have yet to be introduced to breast MR image processing. In this study, we propose a new network, dense attention network (DANet), for the segmentation of breast lesions in MR images. Specifically, our model investigated the importance of feature fusion and selection for our task. Features from the corresponding encoder layer and from all previous decoder layers are fused by concatenation. To highlight the rich-informative channels, a channel attention module is introduced. Results show that our method achieves the best segmentation results compared to three commonly used fully convolutional networks (FCNs), UNet6, FCDenseNet1037, and RefineNet1018.Methods and Experimental Setup
The architecture of the proposed network is shown in Figure 1. Our network adopted the classical FCN architecture. There are five basic blocks in the encoder pathway and the first four are followed by maxpooling operations, which downsample the input images to 1/16 in total. Bilinear upsampling is applied to restore the image resolution. To improve the segmentation performance, we fuse the features from high- and low-levels by concatenation. Feature adaptation is included to clean and prepare the features from the encoder pathway before fusion. Channel attention is introduced right after the fusion to highlight the rich-informative channels. Moreover, we add dense connections between the decoder blocks to fully utilize the different layer features and to avoid possible errors caused by the upsampling process.
We evaluated the performance of the proposed DANet as well as the three comparison FCNs using contrast-enhanced T1-weighted (T1C) breast MR scans of 314 women. The MR images were collected using an Achieva 1.5T system (Philips Healthcare, Best, The Netherlands) with a four-channel phased array breast coil. All acquisitions were done from 2011 to 2017. Axial T1C images with fat suppression were obtained after the intravenous injection of 0.3 mL/kg of gadodiamide (BeiLu Healthcare, Beijing, China) (TR/TE = 5.2 ms/2.3 ms, FOV = 300 mm x 320 mm, section thickness = 1 mm, and flip angle = 15o). Although 3D images were taken, only the central slices with the largest cross-section areas were labelled by two experienced radiologists to be further processed in this study. 5-fold cross-validation experiments were conducted with 80% of the images for training and 20% for validation. The models were implemented with PyTorch on a NVIDIA TITAN Xp GPU (12G). Three independent experiments were done, and the results are presented as (mean ± sd).
Results and Discussion
The performance of the proposed DANet was compared to UNet, FCDenseNet103, and RefineNet101, which were quantified by Dice similarity coefficient (DSC), sensitivity, and relative area difference (DA) (Figure 1). All the three evaluation metrics indicate that our proposed DANet achieved the best segmentation results by obtaining the highest average DSC, the highest average sensitivity, and the lowest average DA. Unpaired t-test was performed, and significant differences were shown between the DSC of DANet and all the other three FCNs (p < 0.05). Segmentation results of two examples were shown in Figure 3. It could be observed that our proposed DANet was better at finding those irregular lesion boundaries (example 1) and at decreasing the false positive incidences (example 2).Conclusion
In this study, we propose a new network, DANet, for the segmentation of breast lesions in 2D MR images. Both quantitative and qualitative results show that DANet achieved the best segmentation performance when compared to UNet, FCDenseNet103, and RefineNet101.