Enhanced 3D brain tumor segmentation using assorted precision training

Adwaitt Pandya1*, Ozioma Collins Oguine2*, Harita Bhargava1* and Shrikant Zade1*

*Correspondence:
Adwaitt Pandya,
adwaitt1999@gmail.com
Ozioma Collins Oguine,
oziomaoguine007@gmail.com
Harita Bhargava,
haritabhargava28@gmail.com
Shrikant Zade,
cdzshrikant@gmail.com

Received: 28 October 2022; Accepted: 08 November 2022; Published: 21 December 2022.

A brain tumor is a medical disorder faced by individuals of all demographics. Medically, it is described as the spread of non-essential cells close to or throughout the brain. Symptoms of this ailment include headaches, seizures, and sensory changes. This research explores two main categories of brain tumors: benign and malignant. Benign spreads steadily, and malignant express growth makes it dangerous. Early identification of brain tumors is a crucial factor for the survival of patients. This research provides a state-of-the-art approach to the early identification of tumors within the brain. We implemented the SegResNet architecture, a widely adopted architecture for three-dimensional segmentation, and trained it using the automatic multi-precision method. We incorporated the dice loss function and dice metric for evaluating the model. We got a dice score of 0.84. For the tumor core, we got a dice score of 0.84; for the whole tumor, 0.90; and for the enhanced tumor, we got a score of 0.79.

Keywords: brain tumor, 3D segmentation, brain tumor segmentation, 3D convolutional neural network, fully convolutional neural network

1. Introduction

A tumor is an abnormal growth of cells in the brain that may or may not be cancerous. Tumors are generally classified into two classes: benign and malignant. Benign tumors are considered non-cancerous, i.e., they grow locally and do not spread to other tissues. They can be fatal if they develop near vital organs like the brain, even after being non-cancerous. Malignant tumors are considered cancerous. New cells are constantly produced in our body to replace the old ones; sometimes DNA gets damaged during this renewal process, so the new cells develop abnormally. These cells continue to multiply faster, thus forming a tumor. Malignant tumors can spread and affect other tissues. Tumors that affect the central nervous system are known as gliomas. Following are the constituents of gliomas:

Edema: Finger-like projection, an agglomerate of fluid or water. FLAIR and T2-weighted sequences produce the best results.

Necrosis: Collection of dead cells. Best seen in the T1 post-contrast sequence.

Enhancing tumor: Indicates breakdown of the blood–brain barrier. Seen in T1c post-contrast sequence.

Non-enhancing tumor: Seen in regions not included in edema, necrosis, or enhancing tumor.

There are many reasons why a person can have a brain tumor-like growth of cells uncontrollably in the brain due to mutation or defect in a gene. This is a reason for causing a brain tumor. Exposure to large amounts of X-rays is also an environmental cause which leads to the development of brain tumors. A tumor’s effects on your body are evaluated by its size, location, and growth rate. General symptoms include changes in the pattern of headaches, nausea or vomiting, vision problems like blurred vision, tiredness, speech and hearing difficulties, and memory problems. Early diagnosis of a tumor provides the patient with the best chance of successful treatment. There are fewer chances of survival, a high cost of treatment, and many more problems arise if the care is delayed.

FIGURE 1
www.bohrpub.com

Figure 1. (A) Two randomly selected images from the T1wCE modality. (B) Intensities of the two images before normalization. (C) Intensities of the two images after normalization. It is imperative to note that certain types of tumors are best seen in different modalities, like edema, which is best seen in T2-weighted sequences and FLAIR images. Necrosis is best visible in the T1 post-contrast sequence, and an enhancing tumor is best seen in the T1c post-contrast sequence (13).

Early diagnosis improves the outcomes of treatment. Diagnosing a brain tumor generally begins with a magnetic resonance imaging (MRI) scan. MRI is an imaging technique that utilizes magnetic fields and radio waves to create detailed images of the organs and tissues. MRI can be used to measure the tumor’s size. For diagnosing a tumor, the accuracy of conventional MRI is generally satisfactory, but we should not rely heavily on it. One of the important and efficient techniques for diagnosing tumors is brain tumor segmentation. The technique of separating tumors from other brain parts in an MRI scan of the brain is called brain tumor segmentation. It separates tumors from normal brain tissues. Tumor segmentation helps in correctly identifying the spatial location of a tumor. Brain tumor segmentation proves to be useful for diagnosis and treatment planning. However, sometimes it is hard to segment the tumor because of irregular boundaries in MRI scans. If a tumor is detected on time due to segmentation, it will prove to be very convenient from the doctor’s perspective to commence treatment planning as soon as possible.

Our dataset consists of a neuroimaging informatics technology initiative (NIFTI) file format. About 10 years ago, the NIFTI file format was envisioned as a replacement for the 7.5 file format for analysis. In image informatics for neuroscience and even neuroradiology research, NIFTI files are frequently employed. For our project, we used the brain tumor segmentation (BraTS) dataset. The Radiological Association of North America (RSNA), the American Society of Neuroradiology (ASNR), and the Medical Image Computing and Computer-Assisted Interventions (MICCAI) society are working together to arrange the BraTS challenge. The model used by us for segmentation is “SegResNet,” and we have trained it on the BraTS 2021 (15) (Task 1) dataset. The following work contains a detailed description of the dataset, proposed methodology, comparative analysis, and results.

2. Theoretical background

2.1. Technology stack

We used the Kaggle notebooks for training, validation, and testing our model. By default, Kaggle provides P100 GPUs for all notebooks with 16 GB of RAM and 16 GB of storage space. We used PyTorch version 1.10.0 and MONAI 0.7.0 for coding. We also used intensity normalization (6) version 2.1.1 for normalizing intensities.

2.2. Dataset

We used the BraTS 2021 task-1 dataset. The collection includes segmentation masks, Native (T1), T1-weighted (T1Gd), T2-weighted (T2), and T2-Fluid Attenuated Inversion Recovery (Recovery) NIFTI volumes for 1,251 individuals from various sources and in axial, sagittal, and coronal orientation. All of the files were NIFTI volumes containing 240 image slices. Each image was 240 × 155 in size.

FIGURE 2
www.bohrpub.com

Figure 2. Augmented samples of different modalities. Each “image channel” corresponds to a different modality.

FIGURE 3
www.bohrpub.com

Figure 3. Brain tumor segmentation masks.

FIGURE 4
www.bohrpub.com

Figure 4. (A) Average loss (left) and average dice score (right) per epoch. (B) Dice score for tumor core (left), whole tumor (center), and enhanced tumor (right).

2.3. Literature survey

For the purpose of segmenting brain tumors, Havaei et al. proposed a CNN architecture that not only utilized local and global features concurrently. They obtained dice scores of 0.85 for segmenting the entire tumor, 0.78 for segmenting the tumor core, and 0.73 for improving tumor segmentation (7).

Pereira et al. (8) explored a way to counter large spatial and structural variability by incorporating small 3 × 3 kernels in their proposed architecture. They attained a dice score of 0.88 on the whole tumor segmentation; for tumor core segmentation, they were able to get a score of 0.83, and for enhancing tumor, they got 0.77. Myronenko et al. (9) described an encoder-decoder-like architecture and a variational autoencoder branch. Their model yielded a dice score of 0.81, 0.90, and 0.86 on enhancing tumor, whole tumor, and tumor core segmentation.

FIGURE 5
www.bohrpub.com

Figure 5. The model’s output (top) is compared to the actual tumor (bottom). The yellow region is the enhanced tumor, the yellow and green regions constitute the tumor core, and the yellow, red, and green regions constitute the whole tumor.

An anisotropic and dilated convolution filter-based cascade of a fully convolutional neural network was proposed by Wang et al. (10). They broke the issue down into a series of binary categorization issues. On the entire tumor, tumor core, and improving tumor segmentation, their model received scores of 0.78, 0.87, and 0.77, respectively.

Mohammadreza et al. (11) extracted text-on-descriptor features such as histograms and first-order intensities, which were then fed into a Random Forest Classifier. They scored 0.84 on the whole tumor, 0.82 on the enhancing tumor, and 0.78 on the tumor core segmentation.

Lyu et al. (12) used a 2-stage model; for stage 1, they used an encoder-decoder-like architecture and variation autoencoder regularization. In stage 2, the network uses attention gates and is trained on a dataset formed by stage 1 output. Their dice scores for the whole tumor, tumor core, and enhancing tumor were 0.87, 0.83, and 0.82, respectively.

3. Methodology

Preprocessing: The dataset comes from various sources. We used the intensity normalization technique described by Shinohara et al. (6) over every image of every modality (FLAIR, T1w, T1wCE, and T2w) to solve the difference in intensities. Separate normalizers were used for each modality and were trained on the images belonging to their respective modalities. Since the images were in different orientations, we oriented all of them in the RAS orientation.

Image Augmentations: We cropped the image, keeping the region of interest of 224 × 224 × 144. We then randomly flipped the images across all three axes and randomly scaled and shifted the intensities. We used no augmentations for validation or testing.

Loss Functions and Metric: We used the dice loss function (14) and metric, a region-based loss function, to calculate similarities. Mathematically, the dice coefficient can be expressed as follows:

D = 2 i N p i g i i N p i 2 + i N g i 2

where D is the dice coefficient, p and g are pairs of corresponding pixel values.(15) For training, we added a small constant to the denominator to tackle the scenarios in which the denominator becomes less than 0.

Hyperparameters: We initialized the model with 16 filters, keeping the number of input channels equal to 4 and yielding an output of 3 channels, corresponding to the three classes discussed earlier. We also applied a dropout with a dropout rate of 0.2. Due to memory constraints, we kept the batch size to 1. We used the Adam optimizer for training with a learning rate of 0.0001. We applied L2 regularization with a regularization coefficient set to 0.00001 and trained the model for 10 epochs.

Model Architecture: We used the SegResNet architecture with the number of downsampling blocks in each layer being 1, 2, 2, and 4, respectively. The number of upsampling blocks in each layer is 1, 1, and 1, respectively.

Training: We trained our model for 10 epochs and saved the best-performing model. We modified the traditional training technique and used the mixed-precision (16) training method, which enhances performance and efficiency and reduces memory requirements. We used automatic mixed precision for both the training and validation tasks.

TABLE 1
www.bohrpub.com

Table 1. Comparison of dice scores obtained by different methods.

4. Result analysis

We converged our loss to 0.11 and got an average dice score of 0.84 on the validation set. As for the separate classes, our dice scores were as follows (on the validation set):

TC = 0.84

WT = 0.90

ET = 0.79

On the test set, we got a mean dice score of 0.86; in the TC class, the dice score was 0.86; in the WT class, it was 0.92; and finally, on ET, it was 0.81.

5. Discussion

Brain tumor segmentation proves to be an effective tool for accurately diagnosing the tumor and its constituents. Our model can segment a tumor from an MRI image efficiently. The model was trained on 750 NIFTI volumes and validated and tested on 250 NIFTI volumes. We were able to get our loss down to 0.11 and got a mean dice score of 0.84. We believe that more data could significantly improve the model’s performance for future research.

6. Conclusion

In this study, we used the SegResNet architecture to segment the brain tumor. Our model produces great results on 200 test cases. Our model’s best score produced a dice score of 0.86 on TC, 0.92 on WT, and 0.81 on ET. In contrast, the mean dice score was 0.84.

Disclosure

The authors declare that they have no competing interests with anyone in publishing this manuscript.

Author contributions

All authors made substantial contributions in conscripting the manuscript and revising it critically for important intellectual content, agreeing to submit it to the current journal, and final approval of the version.

References

1. Baid U, Ghodasara S, Mohan S, Bilello M, Calabrese E, Colak E, et al. The RSNA-ASNR-MICCAI BraTS 2021 benchmark on brain tumor segmentation and radiogenomic classification. arXiv [Preprint]. (2021).

Google Scholar

2. Menze BH, Jakab A, Bauer S, Kalpathy-Cramer J, Farahani K, Kirby J, et al. The multimodal brain tumor image segmentation benchmark (BRATS). IEEE Trans Med Imaging. (2015) 34:1993–2024. doi: 10.1109/TMI.2014.2377694

CrossRef Full Text | Google Scholar

3. Bakas S, Akbari H, Sotiras A, Bilello M, Rozycki M, Kirby JS, et al. Advancing the cancer genome atlas glioma MRI collections with expert segmentation labels and radiomic features. Nat Sci Data. (2017) 4:170117. doi: 10.1038/sdata.2017.117

CrossRef Full Text | Google Scholar

4. Bakas S, Akbari H, Sotiras A, Bilello M, Rozycki M, Kirby J, et al. Segmentation labels and radiomic features for the pre-operative scans of the TCGA-GBM collection. Cancer Imaging Arch. (2017) 286. doi: 10.7937/K9/TCIA.2017.KLXWJJ1Q

CrossRef Full Text | Google Scholar

5. Bakas S, Akbari H, Sotiras A, Bilello M, Rozycki M, Kirby J, et al. Segmentation labels and radiomic features for the pre-operative scans of the TCGA-LGG collection. Cancer Imaging Arch. (2017). doi: 10.7937/K9/TCIA.2017.KLXWJJ1Q

CrossRef Full Text | Google Scholar

6. Shinohara RT, Sweeney EM, Goldsmith J, Shiee N, Mateen FJ, Calabresi PA, et al. Statistical normalization techniques for magnetic resonance imaging. Neuroimage Clin. (2014) 6:9–19. doi: 10.1016/j.nicl.2014.08.008

CrossRef Full Text | Google Scholar

7. Havaei M, Davy A, Warde-Farley D, Biard A, Courville A, Bengio Y, et al. ’Brain tumor segmentation with deep neural networks. arXiv [Preprint]. (2015).

Google Scholar

8. Pereira S, Pinto A, Alves V, Silva C. Brain tumor segmentation using convolutional neural networks in MRI images. IEEE Trans Med Imaging. (2016) 35:1240–51. doi: 10.1109/TMI.2016.2538465

CrossRef Full Text | Google Scholar

9. Myronenko A. 3D MRI brain tumor segmentation using autoencoder regularization. arXiv [Preprint]. (2018).

Google Scholar

10. Wang G, Li W, Ourselin S, Vercauteren T. Automatic brain tumor segmentation based on cascaded convolutional neural networks with uncertainty estimation. Front Comput Neurosci. (2019) 13:56. doi: 10.3389/fncom.2019.00056

CrossRef Full Text | Google Scholar

11. Soltaninejad M, Yang G, Lambrou T, Allinson N, Jones TL, Barrick TR, et al. Supervised learning-based multimodal MRI brain tumor segmentation using texture features from supervoxels. Comput. Methods Prog. Biomed. (2018) 157:69–84. doi: 10.1016/j.cmpb.2018.01.003

CrossRef Full Text | Google Scholar

12. Lyu C, Shu H. A two-stage cascade model with variational autoencoders and attention gates for MRI brain tumor segmentation. Brainlesion. (2021):435–47. doi: 10.1007/978-3-030-72084-1_39

CrossRef Full Text | Google Scholar

13. Iitm N. Segmentation of brain tumors from MRI using deep learning. (2019). Available online at: https://www.youtube.com/watch?v=PcNqAVNCZrE&t=331s

Google Scholar

14. Jadon S. A survey of loss functions for semantic segmentation. In 2020 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB). Proceedings of the 2020 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB). Piscataway, NJ: IEEE (2020). doi: 10.1109/cibcb48159.2020.9277638

CrossRef Full Text | Google Scholar

15. Tiu E. Metrics to evaluate your semantic segmentation model. [online] Medium. (2021). Available online at: https://www.youtube.com/watch?v=PcNqAVNCZrE&t=331s (accessed November 13, 2021).

Google Scholar

16. Micikevicius P, Narang S, Alben J, Gregory FD, Elsen E, Garcia D, et al. Mixed precision training. arXiv [Preprint]. (2017).

Google Scholar

17. Spyridon B, Mauricio R, Andras J, Tefan Bauer S, Markus R, Alessandro C, et al. Identifying the best machine learning algorithms for brain tumor segmentation, progression assessment, and overall survival prediction in the BRATS challenge. Ithaca, NY: Cornell University (2019).

Google Scholar

18. Badrinarayanan V, Kendall A, Cipolla R. SegNet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans Pattern Anal Mach Intell. (2017) 39:2481–95. doi: 10.1109/tpami.2016.2644615

CrossRef Full Text | Google Scholar