Cookies
O website necessita de alguns cookies e outros recursos semelhantes para funcionar. Caso o permita, o INESC TEC irá utilizar cookies para recolher dados sobre as suas visitas, contribuindo, assim, para estatísticas agregadas que permitem melhorar o nosso serviço. Ver mais
Aceitar Rejeitar
  • Menu
Publicações

Publicações por LIAAD

2024

Pre-trained language models: What do they know?

Autores
Guimaraes, N; Campos, R; Jorge, A;

Publicação
WILEY INTERDISCIPLINARY REVIEWS-DATA MINING AND KNOWLEDGE DISCOVERY

Abstract
Large language models (LLMs) have substantially pushed artificial intelligence (AI) research and applications in the last few years. They are currently able to achieve high effectiveness in different natural language processing (NLP) tasks, such as machine translation, named entity recognition, text classification, question answering, or text summarization. Recently, significant attention has been drawn to OpenAI's GPT models' capabilities and extremely accessible interface. LLMs are nowadays routinely used and studied for downstream tasks and specific applications with great success, pushing forward the state of the art in almost all of them. However, they also exhibit impressive inference capabilities when used off the shelf without further training. In this paper, we aim to study the behavior of pre-trained language models (PLMs) in some inference tasks they were not initially trained for. Therefore, we focus our attention on very recent research works related to the inference capabilities of PLMs in some selected tasks such as factual probing and common-sense reasoning. We highlight relevant achievements made by these models, as well as some of their current limitations that open opportunities for further research.This article is categorized under:Fundamental Concepts of Data and Knowledge > Key Design Issues in DataMiningTechnologies > Artificial Intelligence

2024

Predictive Maintenance for Industry 4.0 & 5.0

Autores
Ribeiro, RP;

Publicação
Proceedings of the 1st International Conference on Explainable AI for Neural and Symbolic Methods, EXPLAINS 2024, Porto, Portugal, November 20-22, 2024.

Abstract

2024

Artificial Intelligence Approaches for Predictive Maintenance in the Steel Industry: A Survey

Autores
Jakubowski, J; Strzelecka, NW; Ribeiro, RP; Pashami, S; Bobek, S; Gama, J; Nalepa, GJ;

Publicação
CoRR

Abstract

2024

Aequitas Flow: Streamlining Fair ML Experimentation

Autores
Jesus, SM; Saleiro, P; Silva, IOe; Jorge, BM; Ribeiro, RP; Gama, J; Bizarro, P; Ghani, R;

Publicação
CoRR

Abstract
Aequitas Flow is an open-source framework and toolkit for end-to-end Fair Machine Learning (ML) experimentation, and benchmarking in Python. This package fills integration gaps that exist in other fair ML packages. In addition to the existing audit capabilities in Aequitas, the Aequitas Flow module provides a pipeline for fairness-aware model training, hyperparameter optimization, and evaluation, enabling easy-to-use and rapid experiments and analysis of results. Aimed at ML practitioners and researchers, the framework offers implementations of methods, datasets, metrics, and standard interfaces for these components to improve extensibility. By facilitating the development of fair ML practices, Aequitas Flow hopes to enhance the incorporation of fairness concepts in AI systems making AI systems more robust and fair. © 2025 Elsevier B.V., All rights reserved.

2024

More (Enough) Is Better: Towards Few-Shot Illegal Landfill Waste Segmentation

Autores
Molina, M; Veloso, B; Ferreira, CA; Ribeiro, RP; Gama, J;

Publicação
ECAI 2024

Abstract
Image segmentation for detecting illegal landfill waste in aerial images is essential for environmental crime monitoring. Despite advancements in segmentation models, the primary challenge in this domain is the lack of annotated data due to the unknown locations of illegal waste disposals. This work mainly focuses on evaluating segmentation models for identifying individual illegal landfill waste segments using limited annotations. This research seeks to lay the groundwork for a comprehensive model evaluation to contribute to environmental crime monitoring and sustainability efforts by proposing to harness the combination of agnostic segmentation and supervised classification approaches. We mainly explore different metrics and combinations to better understand how to measure the quality of this applied segmentation problem.

2024

Towards Evaluation of Explainable Artificial Intelligence in Streaming Data

Autores
Mozolewski, M; Bobek, S; Ribeiro, RP; Nalepa, GJ; Gama, J;

Publicação
EXPLAINABLE ARTIFICIAL INTELLIGENCE, XAI 2024, PT IV

Abstract
This study introduces a method to assess the quality of Explainable Artificial Intelligence (XAI) algorithms in dynamic data streams, concentrating on the fidelity and stability of feature-importance and rule-based explanations. We employ XAI metrics, such as fidelity and Lipschitz Stability, to compare explainers between each other and introduce the Comparative Expert Stability Index (CESI) for benchmarking explainers against domain knowledge. We adopted the aforementioned metrics to the streaming data scenario and tested them in an unsupervised classification scenario with simulated distribution shifts as different classes. The necessity for adaptable explainers in complex scenarios, like failure detection is underscored, stressing the importance of continued research into versatile explanation techniques to enhance XAI system robustness and interpretability.

  • 39
  • 516