Cookies Policy
The website need some cookies and similar means to function. If you permit us, we will use those means to collect data on your visits for aggregated statistics to improve our service. Find out More
Accept Reject
  • Menu
Publications

Publications by LIAAD

2025

Simulating Biases for Interpretable Fairness in Offline and Online Classifiers

Authors
Inácio, R; Kokkinogenis, Z; Cerqueira, V; Soares, C;

Publication
CoRR

Abstract

2025

Generating Large Semi-Synthetic Graphs of Any Size

Authors
Tuna, R; Soares, C;

Publication
CoRR

Abstract

2025

MASTFM: Meta-learning and Data Augmentation to Stress Test Forecasting Models

Authors
Inácio, R; Cerqueira, V; Barandas, M; Soares, C;

Publication
Machine Learning and Knowledge Discovery in Databases. Applied Data Science Track and Demo Track - European Conference, ECML PKDD 2025, Porto, Portugal, September 15-19, 2025, Proceedings, Part X

Abstract

2025

Mast: interpretable stress testing via meta-learning for forecasting model robustness evaluation

Authors
Inácio, R; Cerqueira, V; Barandas, M; Soares, C;

Publication
Mach. Learn.

Abstract

2025

SPATA: Systematic Pattern Analysis for Detailed and Transparent Data Cards

Authors
Vitorino, J; Maia, E; Praça, I; Soares, C;

Publication
CoRR

Abstract

2025

Evaluating Transfer Learning Methods on Real-World Data Streams: A Case Study in Financial Fraud Detection

Authors
Pereira, RR; Bono, J; Ferreira, HM; Ribeiro, P; Soares, C; Bizarro, P;

Publication
Machine Learning and Knowledge Discovery in Databases. Applied Data Science Track - European Conference, ECML PKDD 2025, Porto, Portugal, September 15-19, 2025, Proceedings, Part IX

Abstract
When the available data for a target domain is limited, transfer learning (TL) methods leverage related data-rich source domains to train and evaluate models, before deploying them on the target domain. However, most TL methods assume fixed levels of labeled and unlabeled target data, which contrasts with real-world scenarios where both data and labels arrive progressively over time. As a result, evaluations based on these static assumptions may not reflect how methods perform in practice. To support a more realistic assessment of TL methods in dynamic settings, we propose an evaluation framework that (1) simulates varying data availability over time, (2) creates multiple domains via resampling of a given dataset and (3) introduces inter-domain variability through controlled transformations, e.g., including time-dependent covariate and concept shifts. These capabilities enable the systematic simulation of a large number of variants of the experiments, providing deeper insights into how algorithms may behave when deployed. We demonstrate the usefulness of the proposed framework by performing a case study on a proprietary real-world suite of card payment datasets. To support reproducibility, we also apply the framework on the publicly available Bank Account Fraud (BAF) dataset. By providing a methodology for evaluating TL methods over time and in different data availability conditions, our framework supports a better understanding of model behavior in real-world environments, which enables more informed decisions when deploying models in new domains. © 2025 Elsevier B.V., All rights reserved.

  • 11
  • 514