Please use this identifier to cite or link to this item: https://hdl.handle.net/2445/220790
Title: From hand-crafted radiomics to deep learning: evaluating breast cancer classification methods in mammograms
Author: Guzman Requena, Alejandro
Márquez Vara, Noah
Díaz, Oliver
Keywords: Aprenentatge automàtic
Mamografia
Càncer de mama
Xarxes neuronals convolucionals
Machine learning
Mammography
Breast cancer
Convolutional neural networks
Issue Date: 10-Apr-2025
Publisher: SPIE
Citation: Alejandro Guzman, Noah Márquez, Oliver Díaz, "From hand-crafted radiomics to deep learning: evaluating breast cancer classification methods in mammograms," Proc. SPIE 13411, Medical Imaging 2025: Imaging Informatics, 134110U (10 April 2025); https://doi.org/10.1117/12.3046672
Series/Report no: Proceedings SPIE
13411
Abstract: This study evaluates the performance of some machine learning (ML) and deep learning (DL) models for breast cancer tumor classification in mammography (MG) images, by training them on the BCDR dataset. The study compares the use of radiomics-based features in ML models, including Random Forest, Support Vector Machines, and XGBoost, with two deep learning approaches using Convolutional Neural Networks (CNNs) and Vision Transformers (ViTs). Radiomics features were extracted from segmented regions of interest (ROIs) and used to train the ML models, performing hyperparameter tuning and cross-validation to optimize the results. CNN and ViT models were trained using the MG location present in the ROI segmentations, to explore the impact of tumor region localization assistance on classification performance. To examine and verify the performance of the ML models and the ViT, the area under the receiver operating characteristic curve (AUC-ROC) and the training execution time of all experiments (performed on the same device) were used. The results indicate that, while all methods achieve good performance on the training dataset (mean AUCROC scores around 0.9), they exhibit substantial performance drops when tested on external data. Among the evaluated models, ViT achieves the highest overall AUC-ROC in both internal (0.93) and external (0.68) validation, surpassing CNNs and radiomics-based ML models. However, ViT also incurs the highest computational cost, highlighting a trade-off between accuracy and training time. These findings underscore the need for multicenter, multi-vendor data to improve model generalization and reliability, as well as for continued refinement of advanced architectures, such as transformers, to optimize breast cancer lesion classification in clinical settings.
Note: Versió postprint de la comunicació publicada a: https://doi.org/10.1117/12.3046672
It is part of: Comunicació a: Proc. SPIE 13411, Medical Imaging 2025: Imaging Informatics; 134110U (10 April 2025)
URI: https://hdl.handle.net/2445/220790
Related resource: https://doi.org/10.1117/12.3046672
Appears in Collections:Comunicacions a congressos (Matemàtiques i Informàtica)

Files in This Item:
File Description SizeFormat 
UTF-82025 Guzman SPIE.pdfGuzman SPIE3.21 MBAdobe PDFView/Open


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.