2024
Authors
Fernandes, L; Pereira, T; Oliveira, HP;
Publication
2024 IEEE 37TH INTERNATIONAL SYMPOSIUM ON COMPUTER-BASED MEDICAL SYSTEMS, CBMS 2024
Abstract
Currently, lung cancer is one of the deadliest diseases that affects millions of people globally. However, Artificial Intelligence is being increasingly integrated with healthcare practices, with the goal to aid in the early diagnosis of lung cancer. Although such methods have shown very promising results, they still lack transparency to the user, which consequently could make their generalised adoption a challenging task. Therefore, in this work we explore the use of post-hoc explainable methods, to better understand the inner-workings of an already established multitasking framework that executes the segmentation and the classification task of lung nodules simultaneously. The idea behind such study is to understand how a multitasking approach impacts the model's performance in the lung nodule classification task when compared to single-task models. Our results show that the multitasking approach works as an attention mechanism by aiding the model to learn more meaningful features. Furthermore, the multitasking framework was able to achieve a better performance in regard to the explainability metric, with an increase of 7% when compared to our baseline, and also during the classification and segmentation task, with an increase of 4.84% and 15.03%; for each task respectively, when also compared to the studied baselines.
2025
Authors
Nogueira, AFR; Oliveira, HP; Teixeira, LF;
Publication
IMAGE AND VISION COMPUTING
Abstract
3D human pose estimation aims to reconstruct the human skeleton of all the individuals in a scene by detecting several body joints. The creation of accurate and efficient methods is required for several real-world applications including animation, human-robot interaction, surveillance systems or sports, among many others. However, several obstacles such as occlusions, random camera perspectives, or the scarcity of 3D labelled data, have been hampering the models' performance and limiting their deployment in real-world scenarios. The higher availability of cameras has led researchers to explore multi-view solutions due to the advantage of being able to exploit different perspectives to reconstruct the pose. Most existing reviews focus mainly on monocular 3D human pose estimation and a comprehensive survey only on multi-view approaches to determine the 3D pose has been missing since 2012. Thus, the goal of this survey is to fill that gap and present an overview of the methodologies related to 3D pose estimation in multi-view settings, understand what were the strategies found to address the various challenges and also, identify their limitations. According to the reviewed articles, it was possible to find that most methods are fully-supervised approaches based on geometric constraints. Nonetheless, most of the methods suffer from 2D pose mismatches, to which the incorporation of temporal consistency and depth information have been suggested to reduce the impact of this limitation, besides working directly with 3D features can completely surpass this problem but at the expense of higher computational complexity. Models with lower supervision levels were identified to overcome some of the issues related to 3D pose, particularly the scarcity of labelled datasets. Therefore, no method is yet capable of solving all the challenges associated with the reconstruction of the 3D pose. Due to the existing trade-off between complexity and performance, the best method depends on the application scenario. Therefore, further research is still required to develop an approach capable of quickly inferring a highly accurate 3D pose with bearable computation cost. To this goal, techniques such as active learning, methods that learn with a low level of supervision, the incorporation of temporal consistency, view selection, estimation of depth information and multi-modal approaches might be interesting strategies to keep in mind when developing a new methodology to solve this task.
2024
Authors
Teixeira, FB; Ricardo, M; Coelho, A; Oliveira, HP; Viana, P; Paulino, N; Fontes, H; Marques, P; Campos, R; Pessoa, LM;
Publication
2024 JOINT EUROPEAN CONFERENCE ON NETWORKS AND COMMUNICATIONS & 6G SUMMIT, EUCNC/6G SUMMIT 2024
Abstract
Telecommunications and computer vision have evolved separately so far. Yet, with the shift to sub-terahertz (sub-THz) and terahertz (THz) radio communications, there is an opportunity to explore computer vision technologies together with radio communications, considering the dependency of both technologies on Line of Sight. The combination of radio sensing and computer vision can address challenges such as obstructions and poor lighting. Also, machine learning algorithms, capable of processing multimodal data, play a crucial role in deriving insights from raw and low-level sensing data, offering a new level of abstraction that can enhance various applications and use cases such as beamforming and terminal handovers. This paper introduces CONVERGE, a pioneering vision-radio paradigm that bridges this gap by leveraging Integrated Sensing and Communication (ISAC) to facilitate a dual View-to-Communicate, Communicate-to-View approach. CONVERGE offers tools that merge wireless communications and computer vision, establishing a novel Research Infrastructure (RI) that will be open to the scientific community and capable of providing open datasets. This new infrastructure will support future research in 6G and beyond concerning multiple verticals, such as telecommunications, automotive, manufacturing, media, and health.
2025
Authors
Silva, F; Oliveira, HP; Pereira, T;
Publication
ACM COMPUTING SURVEYS
Abstract
The large gap between the generalization level of state-of-the-art machine learning and human learning systems calls for the development of artificial intelligence (AI) models that are truly inspired by human cognition. In tasks related to image analysis, searching for pixel-level regularities has reached a power of information extraction still far from what humans capture with image-based observations. This leads to poor generalization when even small shifts occur at the level of the observations. We explore a perspective on this problem that is directed to learning the generative process with causality-related foundations, using models capable of combining symbolic manipulation, probabilistic reasoning, and pattern recognition abilities. We briefly review and explore connections of research from machine learning, cognitive science, and related fields of human behavior to support our perspective for the direction to more robust and human-like artificial learning systems.
2024
Authors
Vale, P; Boer, J; Oliveira, HP; Pereira, T;
Publication
2024 IEEE 37TH INTERNATIONAL SYMPOSIUM ON COMPUTER-BASED MEDICAL SYSTEMS, CBMS 2024
Abstract
The early and accurate detection and the grading characterization of brain cancer will generate a positive impact on the treatment plan of those patients. AI-based models can help analyze the Magnetic Resonance Imaging (MRI) to make an initial assessment of the tumor grading. The objective of this work was to develop an Al-based model to classify the grading of the tumor using the MRI. Two regions of interest were explored, with several levels of complexity for the neural network architecture, and Iwo strategies to deal with Unbalanced data. The best results were obtained for the most complex architecture (Resnet50) with a combination of weighted random sampler and data augmentation achieving a balanced accuracy of 62.26%. This work confirmed that complex problems required a more dense neural network and strategies to deal with the unbalanced data.
2024
Authors
Teixeira, M; Silva, F; Ferreira, RM; Pereira, T; Figueiredo, C; Oliveira, HP;
Publication
NPJ PRECISION ONCOLOGY
Abstract
Recent studies have shown that the microbiome can impact cancer development, progression, and response to therapies suggesting microbiome-based approaches for cancer characterization. As cancer-related signatures are complex and implicate many taxa, their discovery often requires Machine Learning approaches. This review discusses Machine Learning methods for cancer characterization from microbiome data. It focuses on the implications of choices undertaken during sample collection, feature selection and pre-processing. It also discusses ML model selection, guiding how to choose an ML model, and model validation. Finally, it enumerates current limitations and how these may be surpassed. Proposed methods, often based on Random Forests, show promising results, however insufficient for widespread clinical usage. Studies often report conflicting results mainly due to ML models with poor generalizability. We expect that evaluating models with expanded, hold-out datasets, removing technical artifacts, exploring representations of the microbiome other than taxonomical profiles, leveraging advances in deep learning, and developing ML models better adapted to the characteristics of microbiome data will improve the performance and generalizability of models and enable their usage in the clinic.
The access to the final selection minute is only available to applicants.
Please check the confirmation e-mail of your application to obtain the access code.