Cookies
O website necessita de alguns cookies e outros recursos semelhantes para funcionar. Caso o permita, o INESC TEC irá utilizar cookies para recolher dados sobre as suas visitas, contribuindo, assim, para estatísticas agregadas que permitem melhorar o nosso serviço. Ver mais
Aceitar Rejeitar
  • Menu
Publicações

Publicações por Ana Maria Rebelo

2014

MobBIO: A Multimodal Database Captured with a Portable Handheld Device

Autores
Sequeira, AF; Monteiro, JC; Rebelo, A; Oliveira, HP;

Publicação
PROCEEDINGS OF THE 2014 9TH INTERNATIONAL CONFERENCE ON COMPUTER VISION, THEORY AND APPLICATIONS (VISAPP 2014), VOL 3

Abstract
Biometrics represents a return to a natural way of identification: testing someone by what (s) he is, instead of relying on something (s) he owns or knows seems likely to be the way forward. Biometric systems that include multiple sources of information are known as multimodal. Such systems are generally regarded as an alternative to fight a variety of problems all unimodal systems stumble upon. One of the main challenges found in the development of biometric recognition systems is the shortage of publicly available databases acquired under real unconstrained working conditions. Motivated by such need the MobBIO database was created using an Asus EeePad Transformer tablet, with mobile biometric systems in mind. The proposed database is composed by three modalities: iris, face and voice.

2017

Multimodal Learning for Sign Language Recognition

Autores
Ferreira, PM; Cardoso, JS; Rebelo, A;

Publicação
PATTERN RECOGNITION AND IMAGE ANALYSIS (IBPRIA 2017)

Abstract
Sign Language Recognition (SLR) has becoming one of the most important research areas in the field of human computer interaction. SLR systems are meant to automatically translate sign language into text or speech, in order to reduce the communicational gap between deaf and hearing people. The aim of this paper is to exploit multimodal learning techniques for an accurate SLR, making use of data provided by Kinect and Leap Motion. In this regard, single-modality approaches as well as different multimodal methods, mainly based on convolutional neural networks, are proposed. Experimental results demonstrate that multimodal learning yields an overall improvement in the sign recognition performance.

2018

Physiological Inspired Deep Neural Networks for Emotion Recognition

Autores
Ferreira, PM; Marques, F; Cardoso, JS; Rebelo, A;

Publicação
IEEE ACCESS

Abstract
Facial expression recognition (FER) is currently one of the most active research topics due to its wide range of applications in the human-computer interaction field. An important part of the recent success of automatic FER was achieved thanks to the emergence of deep learning approaches. However, training deep networks for FER is still a very challenging task, since most of the available FER data sets are relatively small. Although transfer learning can partially alleviate the issue, the performance of deep models is still below of its full potential as deep features may contain redundant information from the pre-trained domain. Instead, we propose a novel end-to-end neural network architecture along with a well-designed loss function based on the strong prior knowledge that facial expressions are the result of the motions of some facial muscles and components. The loss function is defined to regularize the entire learning process so that the proposed neural network is able to explicitly learn expression-specific features. Experimental results demonstrate the effectiveness of the proposed model in both lab-controlled and wild environments. In particular, the proposed neural network provides quite promising results, outperforming in most cases the current state-of-the-art methods.

2018

Are Deep Learning Methods Ready for Prime Time in Fingerprints Minutiae Extraction?

Autores
Rebelo, A; Oliveira, T; Correia, ME; Cardoso, JS;

Publicação
Progress in Pattern Recognition, Image Analysis, Computer Vision, and Applications - 23rd Iberoamerican Congress, CIARP 2018, Madrid, Spain, November 19-22, 2018, Proceedings

Abstract
Currently the breakthroughs in most computer vision problems have been achieved by applying deep learning methods. The traditional methodologies that used to successfully discriminate the data features appear to be overwhelmed by the capabilities of learning of the deep network architectures. Nevertheless, many recent works choose to integrate the old handcrafted features into the deep convolutional networks to increase even more their impressive performance. In fingerprint recognition, the minutiae are specific points used to identify individuals and their extraction is a crucial module in a fingerprint recognition system. This can only be emphasized by the fact that the US Federal Bureau of Investigation (FBI) sets as a threshold for a positive identification a number of 8 common minutiae. Deep neural networks have been used to learn possible representations of fingerprint minutiae but, however surprisingly, in this paper it is shown that for now the best choice for an automatic minutiae extraction system is still the traditional road map. A comparison study was conducted with state-of-the-art methods and the best results were achieved by handcraft features. © Springer Nature Switzerland AG 2019.

2019

On the role of multimodal learning in the recognition of sign language

Autores
Ferreira, PM; Cardoso, JS; Rebelo, A;

Publicação
MULTIMEDIA TOOLS AND APPLICATIONS

Abstract
Sign Language Recognition (SLR) has become one of the most important research areas in the field of human computer interaction. SLR systems are meant to automatically translate sign language into text or speech, in order to reduce the communicational gap between deaf and hearing people. The aim of this paper is to exploit multimodal learning techniques for an accurate SLR, making use of data provided by Kinect and Leap Motion. In this regard, single-modality approaches as well as different multimodal methods, mainly based on convolutional neural networks, are proposed. Our main contribution is a novel multimodal end-to-end neural network that explicitly models private feature representations that are specific to each modality and shared feature representations that are similar between modalities. By imposing such regularization in the learning process, the underlying idea is to increase the discriminative ability of the learned features and, hence, improve the generalization capability of the model. Experimental results demonstrate that multimodal learning yields an overall improvement in the sign recognition performance. In particular, the novel neural network architecture outperforms the current state-of-the-art methods for the SLR task.

2021

DeSIRe: Deep Signer-Invariant Representations for Sign Language Recognition

Autores
Ferreira, PM; Pernes, D; Rebelo, A; Cardoso, JS;

Publicação
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS

Abstract
As a key technology to help bridging the gap between deaf and hearing people, sign language recognition (SLR) has become one of the most active research topics in the human-computer interaction field. Although several SLR methodologies have been proposed, the development of a real-world SLR system is still a very challenging task. One of the main challenges is related to the large intersigner variability that exists in the manual signing process of sign languages. To address this problem, we propose a novel end-to-end deep neural network that explicitly models highly discriminative signer-independent latent representations from the input data. The key idea of our model is to learn a distribution over latent representations, conditionally independent of signer identity. Accordingly, the learned latent representations will preserve as much information as possible about the signs, and discard signer-specific traits that are irrelevant for recognition. By imposing such regularization in the representation space, the result is a truly signer-independent model which is robust to different and new test signers. The experimental results demonstrate the effectiveness of the proposed model in several SLR databases.

  • 2
  • 5