Cookies Policy
The website need some cookies and similar means to function. If you permit us, we will use those means to collect data on your visits for aggregated statistics to improve our service. Find out More
Accept Reject
  • Menu
Publications

Publications by CTM

2025

Deep Learning-Driven Integration of Multimodal Data for Material Property Predictions

Authors
Costa, V; Oliveira, JM; Ramos, P;

Publication
COMPUTATION

Abstract
Advancements in deep learning have revolutionized materials discovery by enabling predictive modeling of complex material properties. However, single-modal approaches often fail to capture the intricate interplay of compositional, structural, and morphological characteristics. This study introduces a novel multimodal deep learning framework for enhanced material property prediction, integrating textual (chemical compositions), tabular (structural descriptors), and image-based (2D crystal structure visualizations) modalities. Utilizing the Alexandriadatabase, we construct a comprehensive multimodal dataset of 10,000 materials with symmetry-resolved crystallographic data. Specialized neural architectures, such as FT-Transformer for tabular data, Hugging Face Electra-based model for text, and TIMM-based MetaFormer for images, generate modality-specific embeddings, fused through a hybrid strategy into a unified latent space. The framework predicts seven critical material properties, including electronic (band gap, density of states), thermodynamic (formation energy, energy above hull, total energy), magnetic (magnetic moment per volume), and volumetric (volume per atom) features, many governed by crystallographic symmetry. Experimental results demonstrated that multimodal fusion significantly outperforms unimodal baselines. Notably, the bimodal integration of image and text data showed significant gains, reducing the Mean Absolute Error for band gap by approximately 22.7% and for volume per atom by 22.4% compared to the average unimodal models. This combination also achieved a 28.4% reduction in Root Mean Squared Error for formation energy. The full trimodal model (tabular + images + text) yielded competitive, and in several cases the lowest, error metrics, particularly for band gap, magnetic moment per volume and density of states per atom, confirming the value of integrating all three modalities. This scalable, modular framework advances materials informatics, offering a powerful tool for data-driven materials discovery and design.

2025

Conditional Generative Adversarial Network for Predicting the Aesthetic Outcomes of Breast Cancer Treatment

Authors
Montenegro, H; Cardoso, MJ; Cardoso, JS;

Publication
2025 47TH ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY (EMBC)

Abstract
The alterations to the visual appearance of patients' breasts that occur due to breast cancer locoregional treatment can impact the self-esteem and satisfaction of the patients, affecting quality-of-life after treatment. As such, it is imperative that the patients are adequately informed of the potential aesthetic outcomes of treatment, to facilitate the choice of treatment and promote realistic expectations. As breast asymmetries are among the most notable effects of treatment, we propose a conditional generative adversarial network for manipulating the breast shape in torso images, applying it to simulate how the breasts' shape may change through surgical interventions. Experiments on a private breast dataset suggest that the proposed model outperforms the state-of-the-art in the realistic reconstruction of the torso of the patient while effectively manipulating the breasts.

2025

Fusion Strategies for Breast Cancer Characterization Using Traditional and Deep Learning Models

Authors
Lima, PV; Cardoso, JS; Oliveira, HP;

Publication
BIBE

Abstract
Breast cancer remains one of the most prevalent and deadly cancers worldwide, making accurate evaluation of molecular markers important for effective disease management. Biomarkers such as ER, PR, and HER2 are typically assessed because they help inform prognosis and guide treatment decisions. Predicting these characteristics from imaging can support earlier clinical intervention, reduce reliance on invasive procedures, and contribute to more personalized care. While radiomics and deep learning approaches have demonstrated potential, comprehensive comparisons across these methods are still limited. This study evaluated handcrafted features, deep features, and end-to-end deep learning models for predicting ER, PR, and HER2 status from DCE-MRI. Each feature type was first assessed individually and then combined using early and late fusion. Handcrafted and deep features were processed through a pipeline that included resampling, dimensionality reduction, and model selection, while end-to-end models were trained using different initialization strategies and loss functions. The best models achieved AUCs of 0.659 for ER, 0.679 for PR, and 0.686 for HER2. Although late fusion generally improved performance, bias toward the majority classes persisted. Overall, the results suggest that combining different modeling strategies may enhance robustness in breast cancer characterization. © 2025 IEEE.

2025

HER2match dataset

Authors
Klöckner, P; Teixeira, J; Montezuma, D; Cardoso, JS; Horlings, HM; de Oliveira, SP;

Publication

Abstract

2025

BreLoAI - A Scalable Web Application for Breast Cancer Locoregional Treatment Approaches

Authors
Miguel M Romariz; Tiago F Gonçalves; Eduard Bonci; Hélder Oliveira; Carlos Mavioso; Maria J Cardoso; Jaime Cardoso;

Publication
Cureus Journal of Computer Science.

Abstract

2025

End-to-End Occluded Person Re-Identification With Artificial Occlusion Generation

Authors
Capozzi, L; Cardoso, JS; Rebelo, A;

Publication
IEEE ACCESS

Abstract
In recent years, the task of person re-identification (Re-ID) has improved considerably with the advances in deep learning methodologies. However, occluded person Re-ID remains a challenging task, as parts of the body of the individual are frequently hidden by various objects, obstacles, or other people, making the identification process more difficult. To address these issues, we introduce a novel data augmentation strategy using artificial occlusions, consisting of random shapes and objects from a small image dataset that was created. We also propose an end-to-end methodology for occluded person Re-ID, which consists of three branches: a global branch, a feature dropping branch, and an occlusion detection branch. Experimental results show that the use of random shape occlusions is superior to random erasing using our architecture. Results on six datasets consisting of three tasks (holistic, partial and occluded person Re-ID) demonstrate that our method performs favourably against state-of-the-art methodologies.

  • 9
  • 401