Cookies
O website necessita de alguns cookies e outros recursos semelhantes para funcionar. Caso o permita, o INESC TEC irá utilizar cookies para recolher dados sobre as suas visitas, contribuindo, assim, para estatísticas agregadas que permitem melhorar o nosso serviço. Ver mais
Aceitar Rejeitar
  • Menu
Publicações

Publicações por CTM

2025

FX-MAD: Frequency-domain Explainability and Explainability-driven Unsupervised Detection of Face Morphing Attacks

Autores
Huber, M; Neto, PC; Sequeira, AF; Damer, N;

Publicação
2025 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WORKSHOPS, WACVW

Abstract
Face recognition (FR) systems are vulnerable to morphing attacks, which refer to face images created by morphing the facial features of two different identities into one face image to create an image that can match both identities, allowing serious security breaches. In this work, we apply a frequency-based explanation method from the area of explainable face recognition to shine a light on how FR models behave when processing a bona fide or attack pair from a frequency perspective. In extensive experiments, we used two different state-of-the-art FR models and six different morphing attacks to investigate possible differences in behavior. Our results show that FR models rely differently on different frequency bands when making decisions for bona fide pairs and morphing attacks. In the following step, we show that this behavioral difference can be used to detect morphing attacks in an unsupervised setup solely based on the observed frequency-importance differences in a generalizable manner.

2025

An Integrated and User-Friendly Platform for the Deployment of Explainable Artificial Intelligence Methods Applied to Face Recognition

Autores
Albuquerque, C; Neto, PC; Gonc, T; Sequeira, AF;

Publicação
HCI FOR CYBERSECURITY, PRIVACY AND TRUST, HCI-CPT 2025, PT II

Abstract
Face recognition technology, despite its advancements and increasing accuracy, still presents significant challenges in explainability and ethical concerns, especially when applied in sensitive domains such as surveillance, law enforcement, and access control. The opaque nature of deep learning models jeopardises transparency, bias, and user trust. Concurrently, the proliferation of web applications presents a unique opportunity to develop accessible and interactive tools for demonstrating and analysing these complex systems. These tools can facilitate model decision exploration with various images, aiding in bias mitigation or enhancing users' trust by allowing them to see the model in action and understand its reasoning. We propose an explainable face recognition web application designed to support enrolment, identification, authentication, and verification while providing visual explanations through pixel-wise importance maps to clarify the model's decision-making process. The system is built in compliance with the European Union General Data Protection Regulation, ensuring data privacy and user control over personal information. The application is also designed for scalability, capable of efficiently managing large datasets. Load tests conducted on databases containing up to 1,000,000 images confirm its efficiency. This scalability ensures robust performance and a seamless user experience even with database growth.

2025

How Knowledge Distillation Mitigates the Synthetic Gap in Fair Face Recognition

Autores
Neto, PC; Colakovic, I; Karakatic, S; Sequeira, AF;

Publicação
COMPUTER VISION-ECCV 2024 WORKSHOPS, PT XX

Abstract
Leveraging the capabilities of Knowledge Distillation (KD) strategies, we devise a strategy to fight the recent retraction of face recognition datasets. Given a pretrained Teacher model trained on a real dataset, we show that carefully utilising synthetic datasets, or a mix between real and synthetic datasets to distil knowledge from this teacher to smaller students can yield surprising results. In this sense, we trained 33 different models with and without KD, on different datasets, with different architectures and losses. And our findings are consistent, using KD leads to performance gains across all ethnicities and decreased bias. In addition, it helps to mitigate the performance gap between real and synthetic datasets. This approach addresses the limitations of synthetic data training, improving both the accuracy and fairness of face recognition models.

2025

Fairness Under Cover: Evaluating the Impact of Occlusions on Demographic Bias in Facial Recognition

Autores
Mamede, RM; Neto, PC; Sequeira, AF;

Publicação
COMPUTER VISION-ECCV 2024 WORKSHOPS, PT XXI

Abstract
This study investigates the effects of occlusions on the fairness of face recognition systems, particularly focusing on demographic biases. Using the Racial Faces in the Wild (RFW) dataset and synthetically added realistic occlusions, we evaluate their effect on the performance of face recognition models trained on the BUPT-Balanced and BUPT-GlobalFace datasets. We note increases in the dispersion of FMR, FNMR, and accuracy alongside decreases in fairness according to Equalized Odds, Demographic Parity, STD of Accuracy, and Fairness Discrepancy Rate. Additionally, we utilize a pixel attribution method to understand the importance of occlusions in model predictions, proposing a new metric, Face Occlusion Impact Ratio (FOIR), that quantifies the extent to which occlusions affect model performance across different demographic groups. Our results indicate that occlusions exacerbate existing demographic biases, with models placing higher importance on occlusions in an unequal fashion across demographics.

2025

Model compression techniques in biometrics applications: A survey

Autores
Caldeira, E; Neto, PC; Huber, M; Damer, N; Sequeira, AF;

Publicação
INFORMATION FUSION

Abstract
The development of deep learning algorithms has extensively empowered humanity's task automatization capacity. However, the huge improvement in the performance of these models is highly correlated with their increasing level of complexity, limiting their usefulness in human-oriented applications, which are usually deployed in resource-constrained devices. This led to the development of compression techniques that drastically reduce the computational and memory costs of deep learning models without significant performance degradation. These compressed models are especially essential when implementing multi-model fusion solutions where multiple models are required to operate simultaneously. This paper aims to systematize the current literature on this topic by presenting a comprehensive survey of model compression techniques in biometrics applications, namely quantization, knowledge distillation and pruning. We conduct a critical analysis of the comparative value of these techniques, focusing on their advantages and disadvantages and presenting suggestions for future work directions that can potentially improve the current methods. Additionally, we discuss and analyze the link between model bias and model compression, highlighting the need to direct compression research toward model fairness in future works.

2025

SHAPing Latent Spaces in Facial Attribute Classification Models

Autores
Ferreira, Leonardo; Gonçalves, Tiago; Neto, Pedro C.; Sequeira, Ana; Mamede, Rafael; Oliveira, Mafalda;

Publicação

Abstract
This study investigates the use of SHAP (SHapley Additive exPlanations) values as an explainable artificial intelligence (xAI) technique applied on a facial attribute classification task. We analyse the consistency of SHAP value distributions across diverse classifier architectures that share the same feature extractor, revealing that key features driving attribute classification remain stable regardless of classifier architecture. Our findings highlight the challenges in interpreting SHAP values at the individual sample level, as their reliability depends on the model’s ability to learn distinct class-specific features; models exploiting inter-class correlations yield less representative SHAP explanations. Furthermore, pixel-level SHAP analysis reveals that superior classification accuracy does not necessarily equate to meaningful semantic understanding; notably, despite FaceNet exhibiting lower performance than CLIP, it demonstrated a more nuanced grasp of the underlying class attributes. Finally, we address the computational scalability of SHAP, demonstrating that KernelExplainer becomes infeasible for high-dimensional pixel data, whereas DeepExplainer and GradientExplainer offer more practical alternatives with trade-offs. Our results suggest that SHAP is most effective for small to medium feature sets or tabular data, providing interpretable and computationally manageable explanations.

  • 24
  • 402