2025
Autores
Gouveia, M; Mendes, T; Rodrigues, EM; Oliveira, HP; Pereira, T;
Publicação
APPLIED SCIENCES-BASEL
Abstract
Lung cancer stands as the most prevalent and deadliest type of cancer, with adenocarcinoma being the most common subtype. Computed Tomography (CT) is widely used for detecting tumours and their phenotype characteristics, for an early and accurate diagnosis that impacts patient outcomes. Machine learning algorithms have already shown the potential to recognize patterns in CT scans to classify the cancer subtype. In this work, two distinct pipelines were employed to perform binary classification between adenocarcinoma and non-adenocarcinoma. Firstly, radiomic features were classified by Random Forest and eXtreme Gradient Boosting classifiers. Next, a deep learning approach, based on a Residual Neural Network and a Transformer-based architecture, was utilised. Both 2D and 3D CT data were initially explored, with the Lung-PET-CT-Dx dataset being employed for training and the NSCLC-Radiomics and NSCLC-Radiogenomics datasets used for external evaluation. Overall, the 3D models outperformed the 2D ones, with the best result being achieved by the Hybrid Vision Transformer, with an AUC of 0.869 and a balanced accuracy of 0.816 on the internal test set. However, a lack of generalization capability was observed across all models, with the performances decreasing on the external test sets, a limitation that should be studied and addressed in future work.
2024
Autores
Zolfagharnasab, MH; Freitas, N; Gonçalves, T; Bonci, E; Mavioso, C; Cardoso, MJ; Oliveira, HP; Cardoso, JS;
Publicação
Artificial Intelligence and Imaging for Diagnostic and Treatment Challenges in Breast Care - First Deep Breast Workshop, Deep-Breath 2024, Held in Conjunction with MICCAI 2024, Marrakesh, Morocco, October 10, 2024, Proceedings
Abstract
Breast cancer treatments often affect patients’ body image, making aesthetic outcome predictions vital. This study introduces a Deep Learning (DL) multimodal retrieval pipeline using a dataset of 2,193 instances combining clinical attributes and RGB images of patients’ upper torsos. We evaluate four retrieval techniques: Weighted Euclidean Distance (WED) with various configurations and shallow Artificial Neural Network (ANN) for tabular data, pre-trained and fine-tuned Convolutional Neural Networks (CNNs) and Vision Transformers (ViTs), and a multimodal approach combining both data types. The dataset, categorised into Excellent/Good and Fair/Poor outcomes, is organised into over 20K triplets for training and testing. Results show fine-tuned multimodal ViTs notably enhance performance, achieving up to 73.85% accuracy and 80.62% Adjusted Discounted Cumulative Gain (ADCG). This framework not only aids in managing patient expectations by retrieving the most relevant post-surgical images but also promises broad applications in medical image analysis and retrieval. The main contributions of this paper are the development of a multimodal retrieval system for breast cancer patients based on post-surgery aesthetic outcome and the evaluation of different models on a new dataset annotated by clinicians for image retrieval. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2025.
2024
Autores
Freitas, N; Veloso, C; Mavioso, C; Cardoso, MJ; Oliveira, HP; Cardoso, JS;
Publicação
Artificial Intelligence and Imaging for Diagnostic and Treatment Challenges in Breast Care - First Deep Breast Workshop, Deep-Breath 2024, Held in Conjunction with MICCAI 2024, Marrakesh, Morocco, October 10, 2024, Proceedings
Abstract
Breast cancer is the most common type of cancer in women worldwide. Because of high survival rates, there has been an increased interest in patient Quality of Life after treatment. Aesthetic results play an important role in this aspect, as these treatments can leave a mark on a patient’s self-image. Despite that, there are no standard ways of assessing aesthetic outcomes. Commonly used software such as BCCT.core or BAT require the manual annotation of keypoints, which makes them time-consuming for clinical use and can lead to result variability depending on the user. Recently, there have been attempts to leverage both traditional and Deep Learning algorithms to detect keypoints automatically. In this paper, we compare several methods for the detection of Breast Endpoints across two datasets. Furthermore, we present an extended evaluation of using these models as input for full contour prediction and aesthetic evaluation using the BCCT.core software. Overall, the YOLOv9 model, fine-tuned for this task, presents the best results considering both accuracy and usability, making this architecture the best choice for this application. The main contribution of this paper is the development of a pipeline for full breast contour prediction, which reduces clinician workload and user variability for automatic aesthetic assessment. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2025.
2025
Autores
Rodrigues, M; Gouveia, M; Oliveira, P; Pereira, T;
Publicação
IEEE Access
Abstract
Deep learning techniques have demonstrated significant potential in computer-assisted diagnosis based on medical imaging. However, their integration into clinical workflows remains limited, largely due to concerns about interpretability. To address this challenge, we propose Efficient-Proto-Caps, a lightweight and inherently interpretable model that combines capsule networks with prototype learning for lung nodule characterization. Additionally, an innovative Davies-Bouldin Index with multiple centroids per cluster is employed as a loss function to promote clustering of lung nodule visual attribute representations. When evaluated on the LIDC-IDRI dataset, the most widely recognized benchmark for lung cancer prediction, our model achieved an overall accuracy of 89.7 % in predicting lung nodule malignancy and associated visual attributes. This performance is statistically comparable to that of the baseline model, while utilizing a backbone with only approximately 2 % of the parameters of the baseline model’s backbone. State-of-the-art models achieved better performance in lung nodule malignancy prediction; however, our approach relies on multiclass malignancy predictions and provides a decision rationale aligned with globally accepted clinical guidelines. These results underscore the potential of our approach, as the integration of lightweight and less complex designs into accurate and inherently interpretable models represents a significant advancement toward more transparent and clinically viable computer-assisted diagnostic systems. Furthermore, these findings highlight the model’s potential for broader applicability, extending beyond medicine to other domains where final classifications are grounded in concept-based or example-based attributes. © 2013 IEEE.
2025
Autores
Castro, IAA; Oliveira, HP; Correia, R; Hayes-Gill, B; Morgan, SP; Korposh, S; Gomez, D; Pereira, T;
Publicação
PHYSIOLOGICAL MEASUREMENT
Abstract
Objective.The detection of arterial pulsating signals at the skin periphery with Photoplethysmography (PPG) are easily distorted by motion artifacts. This work explores the alternatives to the aid of PPG reconstruction with movement sensors (accelerometer and/or gyroscope) which to date have demonstrated the best pulsating signal reconstruction. Approach. A generative adversarial network with fully connected layers is proposed for the reconstruction of distorted PPG signals. Artificial corruption was performed to the clean selected signals from the BIDMC Heart Rate dataset, processed from the larger MIMIC II waveform database to create the training, validation and testing sets. Main results. The heart rate (HR) of this dataset was further extracted to evaluate the performance of the model obtaining a mean absolute error of 1.31 bpm comparing the HR of the target and reconstructed PPG signals with HR between 70 and 115 bpm. Significance. The model architecture is effective at reconstructing noisy PPG signals regardless the length and amplitude of the corruption introduced. The performance over a range of HR (70-115 bpm), indicates a promising approach for real-time PPG signal reconstruction without the aid of acceleration or angular velocity inputs.
The access to the final selection minute is only available to applicants.
Please check the confirmation e-mail of your application to obtain the access code.