2022
Autores
Gonçalves, T; Torto, IR; Teixeira, LF; Cardoso, JS;
Publicação
CoRR
Abstract
2024
Autores
Rio-Torto, I; Gonçalves, T; Cardoso, JS; Teixeira, LF;
Publicação
IEEE INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING, ISBI 2024
Abstract
In fields that rely on high-stakes decisions, such as medicine, interpretability plays a key role in promoting trust and facilitating the adoption of deep learning models by the clinical communities. In the medical image analysis domain, gradient-based class activation maps are the most widely used explanation methods and the field lacks a more in depth investigation into inherently interpretable models that focus on integrating knowledge that ensures the model is learning the correct rules. A new approach, B-cos networks, for increasing the interpretability of deep neural networks by inducing weight-input alignment during training showed promising results on natural image classification. In this work, we study the suitability of these B-cos networks to the medical domain by testing them on different use cases (skin lesions, diabetic retinopathy, cervical cytology, and chest X-rays) and conducting a thorough evaluation of several explanation quality assessment metrics. We find that, just like in natural image classification, B-cos explanations yield more localised maps, but it is not clear that they are better than other methods' explanations when considering more explanation properties.
2024
Autores
Neto, PC; Mamede, RM; Albuquerque, C; Gonçalves, T; Sequeira, AF;
Publicação
2024 IEEE 18TH INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION, FG 2024
Abstract
Face recognition applications have grown in parallel with the size of datasets, complexity of deep learning models and computational power. However, while deep learning models evolve to become more capable and computational power keeps increasing, the datasets available are being retracted and removed from public access. Privacy and ethical concerns are relevant topics within these domains. Through generative artificial intelligence, researchers have put efforts into the development of completely synthetic datasets that can be used to train face recognition systems. Nonetheless, the recent advances have not been sufficient to achieve performance comparable to the state-of-the-art models trained on real data. To study the drift between the performance of models trained on real and synthetic datasets, we leverage a massive attribute classifier (MAC) to create annotations for four datasets: two real and two synthetic. From these annotations, we conduct studies on the distribution of each attribute within all four datasets. Additionally, we further inspect the differences between real and synthetic datasets on the attribute set. When comparing through the Kullback-Leibler divergence we have found differences between real and synthetic samples. Interestingly enough, we have verified that while real samples suffice to explain the synthetic distribution, the opposite could not be further from being true.
2023
Autores
Montenegro, H; Neto, PC; Patrício, C; Torto, IR; Gonçalves, T; Teixeira, LF;
Publicação
Working Notes of the Conference and Labs of the Evaluation Forum (CLEF 2023), Thessaloniki, Greece, September 18th to 21st, 2023.
Abstract
This paper presents the main contributions of the VCMI Team to the ImageCLEFmedical GANs 2023 task. This task aims to evaluate whether synthetic medical images generated using Generative Adversarial Networks (GANs) contain identifiable characteristics of the training data. We propose various approaches to classify a set of real images as having been used or not used in the training of the model that generated a set of synthetic images. We use similarity-based approaches to classify the real images based on their similarity to the generated ones. We develop autoencoders to classify the images through outlier detection techniques. Finally, we develop patch-based methods that operate on patches extracted from real and generated images to measure their similarity. On the development dataset, we attained an F1-score of 0.846 and an accuracy of 0.850 using an autoencoder-based method. On the test dataset, a similarity-based approach achieved the best results, with an F1-score of 0.801 and an accuracy of 0.810. The empirical results support the hypothesis that medical data generated using deep generative models trained without privacy constraints threatens the privacy of patients in the training data. © 2023 Copyright for this paper by its authors.
2023
Autores
Torto, IR; Patrício, C; Montenegro, H; Gonçalves, T; Cardoso, JS;
Publicação
Working Notes of the Conference and Labs of the Evaluation Forum (CLEF 2023), Thessaloniki, Greece, September 18th to 21st, 2023.
Abstract
This paper presents the main contributions of the VCMI Team to the ImageCLEFmedical Caption 2023 task. We addressed both the concept detection and caption prediction tasks. Regarding concept detection, our team employed different approaches to assign concepts to medical images: multi-label classification, adversarial training, autoregressive modelling, image retrieval, and concept retrieval. We also developed three model ensembles merging the results of some of the proposed methods. Our best submission obtained an F1-score of 0.4998, ranking 3rd among nine teams. Regarding the caption prediction task, our team explored two main approaches based on image retrieval and language generation. The language generation approaches, based on a vision model as the encoder and a language model as the decoder, yielded the best results, allowing us to rank 5th among thirteen teams, with a BERTScore of 0.6147. © 2023 Copyright for this paper by its authors.
2023
Autores
Kaidar Person, O; Antunes, M; Cardoso, S; Ciani, O; Cruz, H; Di Micco, R; Gentilini, D; Gonçalves, T; Gouveia, P; Heil, J; Kabata, P; Lopes, D; Martinho, M; Martins, H; Mavioso, C; Mika, M; Montenegro, H; Oliveira, P; Pfob, A; Rotmensz, N; Schinköthe, T; Silva, G; Tarricone, R; Cardoso, M;
Publicação
PLOS ONE
Abstract
BackgroundBreast cancer therapy improved significantly, allowing for different surgical approaches for the same disease stage, therefore offering patients different aesthetic outcomes with similar locoregional control. The purpose of the CINDERELLA trial is to evaluate an artificial-intelligence (AI) cloud-based platform (CINDERELLA platform) vs the standard approach for patient education prior to therapy. MethodsA prospective randomized international multicentre trial comparing two methods for patient education prior to therapy. After institutional ethics approval and a written informed consent, patients planned for locoregional treatment will be randomized to the intervention (CINDERELLA platform) or controls. The patients in the intervention arm will use the newly designed web-application (CINDERELLA platform, CINDERELLA APProach) to access the information related to surgery and/or radiotherapy. Using an AI system, the platform will provide the patient with a picture of her own aesthetic outcome resulting from the surgical procedure she chooses, and an objective evaluation of this aesthetic outcome (e.g., good/fair). The control group will have access to the standard approach. The primary objectives of the trial will be i) to examine the differences between the treatment arms with regards to patients' pre-treatment expectations and the final aesthetic outcomes and ii) in the experimental arm only, the agreement of the pre-treatment AI-evaluation (output) and patient's post-therapy self-evaluation. DiscussionThe project aims to develop an easy-to-use cost-effective AI-powered tool that improves shared decision-making processes. We assume that the CINDERELLA APProach will lead to higher satisfaction, better psychosocial status, and wellbeing of breast cancer patients, and reduce the need for additional surgeries to improve aesthetic outcome.
The access to the final selection minute is only available to applicants.
Please check the confirmation e-mail of your application to obtain the access code.