Cookies Policy
The website need some cookies and similar means to function. If you permit us, we will use those means to collect data on your visits for aggregated statistics to improve our service. Find out More
Accept Reject
  • Menu
About

About

Aurélio Campilho is Emeritus Professor of the University of Porto, Jubilee Full Professor in the Department of Electrical and Computer Engineering, Faculty of Engineering, University of Porto, Portugal. He is a Fellow from EAMBES (European Alliance of Medical and Biological Engineering and Science). He is a Senior Member of the IEEE – The Institute of Electrical and Electronics Engineers. He is coordinator of the Center for Biomedical Engineering Research (C-BER) and develops research at the Biomedical Imaging Lab from C-BER from INESC TEC – Institute for Systems and Computer Engineering, Technology and Science. His current research interests include the areas of biomedical engineering, medical image analysis, image processing and computer vision, particularly in Computer-aided Diagnosis applied in several imaging modalities, including ophthalmic images, carotid ultrasound imaging and computed tomography of the lung.

He is the author of one book (with two editions), co-edited 20 books and published more than 250 articles in international journals and conferences. Organized several special issues of magazines and conferences. He was Associate Editor of the journals IEEE Transactions on Biomedical Engineering and the Machine Vision Applications Journal. From 2004 to 2020, he was General Chair of the International Conferences on Image Analysis and Recognition (ICIAR) conference series.


Interest
Topics
Details

Details

  • Name

    Aurélio Campilho
  • Role

    Affiliated Researcher
  • Since

    01st January 2014
  • Nationality

    Portugal
  • Contacts

    +351222094106
    aurelio.campilho@inesctec.pt
005
Publications

2025

Anatomically-Guided Inpainting for Local Synthesis of Normal Chest Radiographs

Authors
Pedrosa, J; Pereira, SC; Silva, J; Mendonça, AM; Campilho, A;

Publication
DEEP GENERATIVE MODELS, DGM4MICCAI 2024

Abstract
Chest radiography (CXR) is one of the most used medical imaging modalities. Nevertheless, the interpretation of CXR images is time-consuming and subject to variability. As such, automated systems for pathology detection have been proposed and promising results have been obtained, particularly using deep learning. However, these tools suffer from poor explainability, which represents a major hurdle for their adoption in clinical practice. One proposed explainability method in CXR is through contrastive examples, i.e. by showing an alternative version of the CXR except without the lesion being investigated. While image-level normal/healthy image synthesis has been explored in literature, normal patch synthesis via inpainting has received little attention. In this work, a method to synthesize contrastive examples in CXR based on local synthesis of normal CXR patches is proposed. Based on a contextual attention inpainting network (CAttNet), an anatomically-guided inpainting network (AnaCAttNet) is proposed that leverages anatomical information of the original CXR through segmentation to guide the inpainting for a more realistic reconstruction. A quantitative evaluation of the inpainting is performed, showing that AnaCAttNet outperforms CAttNet (FID of 0.0125 and 0.0132 respectively). Qualitative evaluation by three readers also showed that AnaCAttNet delivers superior reconstruction quality and anatomical realism. In conclusion, the proposed anatomical segmentation module for inpainting is shown to improve inpainting performance.

2025

Grad-CAM: The impact of large receptive fields and other caveats

Authors
Santos, R; Pedrosa, J; Mendonça, AM; Campilho, A;

Publication
COMPUTER VISION AND IMAGE UNDERSTANDING

Abstract
The increase in complexity of deep learning models demands explanations that can be obtained with methods like Grad-CAM. This method computes an importance map for the last convolutional layer relative to a specific class, which is then upsampled to match the size of the input. However, this final step assumes that there is a spatial correspondence between the last feature map and the input, which may not be the case. We hypothesize that, for models with large receptive fields, the feature spatial organization is not kept during the forward pass, which may render the explanations devoid of meaning. To test this hypothesis, common architectures were applied to a medical scenario on the public VinDr-CXR dataset, to a subset of ImageNet and to datasets derived from MNIST. The results show a significant dispersion of the spatial information, which goes against the assumption of Grad-CAM, and that explainability maps are affected by this dispersion. Furthermore, we discuss several other caveats regarding Grad-CAM, such as feature map rectification, empty maps and the impact of global average pooling or flatten layers. Altogether, this work addresses some key limitations of Grad-CAM which may go unnoticed for common users, taking one step further in the pursuit for more reliable explainability methods.

2025

Multitask Learning Approach for Foveal Avascular Zone Segmentation in OCTA Images

Authors
Melo, T; Carneiro, A; Campilho, A; Mendonça, AM;

Publication
Pattern Recognition and Image Analysis - 12th Iberian Conference, IbPRIA 2025, Coimbra, Portugal, June 30 - July 3, 2025, Proceedings, Part II

Abstract
The segmentation of the foveal avascular zone (FAZ) in optical coherence tomography angiography (OCTA) images plays a crucial role in diagnosing and monitoring ocular diseases such as diabetic retinopathy (DR) and age-related macular degeneration (AMD). However, accurate FAZ segmentation remains challenging due to image quality and variability. This paper provides a comprehensive review of FAZ segmentation techniques, including traditional image processing methods and recent deep learning-based approaches. We propose two novel deep learning methodologies: a multitask learning framework that integrates vessel and FAZ segmentation, and a conditionally trained network that employs vessel-aware loss functions. The performance of the proposed methods was evaluated on the OCTA-500 dataset using the Dice coefficient, Jaccard index, 95% Hausdorff distance, and average symmetric surface distance. Experimental results demonstrate that the multitask segmentation framework outperforms existing state-of-the-art methods, achieving superior FAZ boundary delineation and segmentation accuracy. The conditionally trained network also improves upon standard U-Net-based approaches but exhibits limitations in refining the FAZ contours. © 2025 Elsevier B.V., All rights reserved.

2024

STERN: Attention-driven Spatial Transformer Network for abnormality detection in chest X-ray images

Authors
Rocha, J; Pereira, SC; Pedrosa, J; Campilho, A; Mendonça, AM;

Publication
ARTIFICIAL INTELLIGENCE IN MEDICINE

Abstract
Chest X-ray scans are frequently requested to detect the presence of abnormalities, due to their low-cost and non-invasive nature. The interpretation of these images can be automated to prioritize more urgent exams through deep learning models, but the presence of image artifacts, e.g. lettering, often generates a harmful bias in the classifiers and an increase of false positive results. Consequently, healthcare would benefit from a system that selects the thoracic region of interest prior to deciding whether an image is possibly pathologic. The current work tackles this binary classification exercise, in which an image is either normal or abnormal, using an attention-driven and spatially unsupervised Spatial Transformer Network (STERN), that takes advantage of a novel domain-specific loss to better frame the region of interest. Unlike the state of the art, in which this type of networks is usually employed for image alignment, this work proposes a spatial transformer module that is used specifically for attention, as an alternative to the standard object detection models that typically precede the classifier to crop out the region of interest. In sum, the proposed end-to-end architecture dynamically scales and aligns the input images to maximize the classifier's performance, by selecting the thorax with translation and non-isotropic scaling transformations, and thus eliminating artifacts. Additionally, this paper provides an extensive and objective analysis of the selected regions of interest, by proposing a set of mathematical evaluation metrics. The results indicate that the STERN achieves similar results to using YOLO-cropped images, with reduced computational cost and without the need for localization labels. More specifically, the system is able to distinguish abnormal frontal images from the CheXpert dataset, with a mean AUC of 85.67% -a 2.55% improvement vs. the 0.98% improvement achieved by the YOLO-based counterpart in comparison to a standard baseline classifier. At the same time, the STERN approach requires less than 2/3 of the training parameters, while increasing the inference time per batch in less than 2 ms. Code available via GitHub.

2024

Automated image label extraction from radiology reports - A review

Authors
Pereira, SC; Mendonca, AM; Campilho, A; Sousa, P; Lopes, CT;

Publication
ARTIFICIAL INTELLIGENCE IN MEDICINE

Abstract
Machine Learning models need large amounts of annotated data for training. In the field of medical imaging, labeled data is especially difficult to obtain because the annotations have to be performed by qualified physicians. Natural Language Processing (NLP) tools can be applied to radiology reports to extract labels for medical images automatically. Compared to manual labeling, this approach requires smaller annotation efforts and can therefore facilitate the creation of labeled medical image data sets. In this article, we summarize the literature on this topic spanning from 2013 to 2023, starting with a meta-analysis of the included articles, followed by a qualitative and quantitative systematization of the results. Overall, we found four types of studies on the extraction of labels from radiology reports: those describing systems based on symbolic NLP, statistical NLP, neural NLP, and those describing systems combining or comparing two or more of the latter. Despite the large variety of existing approaches, there is still room for further improvement. This work can contribute to the development of new techniques or the improvement of existing ones.

Supervised
thesis

2022

Artificial Intelligence-based Decision Support Models for COVID-19 Detection

Author
Sofia Perestrelo de Vasconcelos Cardoso Pereira

Institution
UP-FEUP

2022

Collaborative Tools for Lung Cancer Diagnosis in Computed Tomography

Author
Carlos Alexandre Nunes Ferreira

Institution
UP-FEUP

2022

Explainable Artificial Medical Intelligence for Automated Thoracic Pathology Screening

Author
Joana Maria Neves da Rocha

Institution
UP-FEUP

2022

content based image retrieval as a computer aided diagnosis tool for radiologists

Author
José Ricardo Ferreira de Castro Ramos

Institution
UP-FEUP

2022

Computer-aided diagnosis and follow-up of prevalent eye diseases using OCT/OCTA images

Author
Tânia Filipa Fernandes de Melo

Institution
UP-FEUP