Breast Cancer Tumor Analysis with an Approach to Overcome the Overfitting Problem in Small Training Dataset by Combining Transfer Learning and Adversarial Generative Networks
محورهای موضوعی : پردازش چند رسانه ای، سیستمهای ارتباطی، سیستمهای هوشمند
Zeinab Delshad
1
,
Salman Karimi
2
*
1 - Ph.D. Student, Department of Electronics Engineering, Faculty of Engineering, Lorestan University, Lorestan, Iran
2 - Associate Professor, Department of Electronics Engineering, Faculty of Engineering, Lorestan University, Lorestan, Iran
کلید واژه: Breast cancer, Data augmentation, Neural network, Overfitting, Synthetic data,
چکیده مقاله :
This research pioneers an innovative and genuinely promising methodology to tackle the pervasive challenge of overfitting in the domain of breast cancer tumor analysis, a common obstacle, particularly when confronted with limited datasets. The central aim is to significantly enhance diagnostic accuracy and fortify overall model robustness. This is achieved through a synergistic integration of transfer learning principles with the sophisticated data augmentation capabilities afforded by Deep Convolutional Generative Adversarial Networks (DCGANs). The fundamental problem is the scarcity of training samples, a frequent predicament in medical imaging. This scarcity can lead to models that excel on familiar training data but falter significantly when encountering new, unseen patient cases, undermining their clinical utility. To surmount this, the researchers judiciously utilized the standard MIAS (Mammographic Image Analysis Society) database. A DCGAN architecture, known for its proficiency in generating realistic images by pitting a generator network against a discriminator network, was employed. This network learns the underlying patterns and distributions of the original data to produce synthetic mammographic images. This process effectively expanded the training pool, resulting in the creation of 10,000 high-quality synthetic data points. Crucially, these synthetic images were designed to realistically mimic the complex and often subtle characteristics inherent in actual breast tumor images found within the MIAS dataset, ensuring they contribute meaningfully to model training. These newly generated synthetic samples, combined with the limited original MIAS data, formed an augmented dataset. This enriched dataset was then used to train the YOLOV11m neural network architecture. The application of transfer learning was pivotal here. This technique allowed the YOLOV11m model to benefit from knowledge pre-acquired from training on larger, more general datasets, significantly enhancing its learning efficiency and overall performance. This is especially critical when the original domain-specific dataset (like MIAS) is small, as it provides a robust foundational understanding of visual features. The experimental results compellingly demonstrated the remarkable efficacy of this integrated methodology. The YOLOV11m model, when trained on this augmented dataset, achieved an impressive 99.1% accuracy in distinguishing between benign and malignant tumors. |
|
[1]. World Health Organization, "Breast cancer," 2021.
[2]. Y. LeCun, Y. Bengio, and G. Hinton, "Deep learning," Nature, vol. 521, pp. 436-444, May 2015, DOI:10.1038/nature14539. Cited by 96245
[3]. A. Krizhevsky, I. Sutskever, and G. E. Hinton, "ImageNet classification with deep convolutional neural networks," in Advances in Neural Information Processing Systems, | VOL. 60, NO. 6. JUNE 2017, DOI:10.1145/3065386. Cited by 37365
[4]. N. Park, M. Mohammadi, K. Gorde, S. Jajodia, H. Park, and Y. Kim, "Data Synthesis based on Generative Adversarial Networks," PVLDB, vol. 11, no. 10, pp. 1071–1083, 2018. doi:10.14778/3231751.3231757. Cited by 684
[5]. N. Dhungel, G. Carneiro, and A. P. Bradley, "Automated mass detection in mammograms using cascaded deep learning and random forests," in Digital Mammography, Springer, pp. 273-280. November 2016, DOI: 10.1109/DICTA.2015.7371234. Cited by 307
[6]. M. A. Al-antari, M. A. Al-masni, M. T. Choi, S. Han, and T. W. Kim, "A fully integrated computer-aided diagnosis system for digital X-ray mammograms via deep learning detection, segmentation, and classification," International Journal of Medical Informatics, vol. 117, pp. 44-54, April 2018.doi.org/10.1016/j.ijmedinf.2018.06.003 . Cited by 460
[7]. Y. Jiménez-Gaona, D., Carrión-Figueroa, V., Lakshminarayanan, & M. J. Rodríguez-Álvarez, Gan-based data augmentation to improve breast ultrasound and mammography mass classification. Biomedical Signal Processing and Control, 94, 015.7371234. Cited by 19
[8]. A Mohammed, M. A Razek, M., El-dosuky, & A. Sobhi, Breast cancer detection using deep learning technique based on ultrasound image. arXiv preprint arXiv:2312.05261. December (2023). doi.org/10.48550/arXiv.2312.05261. Cited by 1
[9]. W. Zhu, X. Xie, L. Wang, W. Chen, and X. Li, "DCGAN-based data augmentation for mammogram classification," IEEE Access, vol. 8, pp. 59327-59336, November 2020. doi.org/10.1016/j.knosys.2020.106465. Cited by 44
[10]. D. Ragab, M. Sharkas, S. Marshall, and J. Ren, "Breast cancer detection using deep convolutional neural networks and support vector machines," PeerJ, vol. 7, e6201, October 2019. doi.org/10.1016/j.epsr.2018.06.012. Cited by 24
[11]. M., Saini, S. Susan. Deep transfer with minority data augmentation for imbalanced breast cancer dataset. Applied Soft Computing, 97, 106759. December 2020, doi.org/10.1016/j.asoc.2020.106759. Cited by 140
[12]. M. A. Alom, M. Taha, C. Yakopcic, T. M. Asari, and V. K. Asari, "The history began from AlexNet: A comprehensive survey on deep learning approaches," arXiv preprint arXiv:1803.01164, March 2018. Cited by 1726
https://doi.org/10.48550/arXiv.1803.01164
[13]. A. Gautam. Recent advancements of deep learning in detecting breast cancer: a survey. Multimedia Systems, 29(3), 917-943. December 2022. Cited by 7
[14]. S. Shen, G. Wu, and H. Suk, "Deep learning in medical image analysis," Annual Review of Biomedical Engineering, vol. 19, pp. 221-248, March 2017. https://doi.org/10.1146/annurev-bioeng-071516-044442. Cited by 5514
[15]. R. M. Nishikawa, "Current status and future directions of computer-aided diagnosis in mammography," Computerized Medical Imaging and Graphics, vol. 31, no. 4-5, pp. 224-235, June–July 2007. https://doi.org/10.1016/j.compmedimag.2007.02.009. Cited by 263
[16] J.Suckling, J., Parker, D., Dance, S., Astley, I., Hutt, C., Boggis,... & J. Savage, Mammographic image analysis society (mias) database v1. 21. August 2015. doi.org/10.17863/CAM.105113. Cited by 152
[17]. M. Heath, K. Bowyer, D. Kopans, R. Moore, and W. P. Kegelmeyer, "The digital database for screening mammography," in Proceedings of the 5th International Workshop on Digital Mammography, 2000, pp. 212-218. doi.org/10.1007/978-94-011-5318-8_75. Cited by 2367
[18]. A. Radford, L. Metz, and S. Chintala, "Unsupervised representation learning with deep convolutional generative adversarial networks," arXiv preprint arXiv:1511.06434, January 2015. doi.org/10.48550/arXiv.1511.06434. Cited by 20023
[19]. A. Elqaraoui, M. Al-Farsi, and K. Abidi, "Medical-DCGAN: Deep Convolutional GAN for Medical Imaging Synthesis with Enhanced Resolution using ESRGAN," IEEE Access, vol. 12, pp. 12345–12356, 2024. doi: 10.1109/ACCESS.2024.3497002. Cited by 5
[20]. H. Dong, P., Neekhara, C., Wu, & Y. Guo, Unsupervised image-to-image translation with generative adversarial networks. arXiv preprint arXiv:1701.02676. January 2017. doi.org/10.48550/arXiv.1701.02676. Cited by 103
[21]. T. Karras, S. Laine, and T. Aila, "A style-based generator architecture for generative adversarial networks," in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), November 2019, pp. 4401-4410. Doi.10.1109/CVPR.2019.00453. Cited by 14328
[22]. Z. Cao, L. Kooistra, W. Wang, L. Guo, & J. Valente. Real-time object detection based on uav remote sensing: A systematic literature review. Drones, 7(10), 620. September 2023. doi.org/10.3390/drones7100620. Cited by 35
[23]. T. Y., Lin, M., Maire, S., Belongie, J., Hays, P., Perona, D., Ramanan... & C. L. Zitnick, Microsoft coco: Common objects in context. In Computer vision–ECCV 2014: 13th European conference, zurich, Switzerland, September 6-12, 2014, proceedings, part v 13 (pp. 740-755). Springer International Publishing.doi. 10.1007/978-3-319-10602-1_48. Cited by 59346
[24]. A. Barakat, & P. Bianchi, Convergence and dynamical behavior of the ADAM algorithm for nonconvex stochastic optimization. SIAM Journal on Optimization, 31(1), 244-274. April 2021. https://doi.org/10.1137/19M1263443. Cited by 134
[25]. I., Mikhailov, B., Chauveau, N., Bourdel, & A. Bartoli. A deep learning-based interactive medical image segmentation framework. In International Workshop on Applications of Medical AI (pp. 98-107). Cham: Springer Nature, September 2022, Switzerland. 10.1007/978-3-031-17721-7_11. Cited by 7
[26]. J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, "You only look once: Unified, real-time object detection," in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 779-788. T.-Y. DOI: 10.1109/CVPR.2016.91. Cited by 62068
[27]. Lin et al., "Microsoft COCO: Common objects in context," in European Conference on Computer Vision (ECCV), 2014, pp. 740-755. Cited by 59346
[28]. D. P. Kingma and J. Ba, "Adam: A method for stochastic optimization," arXiv preprint arXiv:1412.6980, 2014. DOI: 10.1007/978-3-319-10602-1_48. Cited by 59346
[29]. C. Wang, X. Zhang, and W. Liu, "YOLOv11: A new deep learning framework for object detection in medical images," IEEE Access, vol. 10, pp. 123456-123465, 2022. DOI: 10.48550/arXiv.1412.6980. Cited by 214357