Cookies Policy
The website need some cookies and similar means to function. If you permit us, we will use those means to collect data on your visits for aggregated statistics to improve our service. Find out More
Accept Reject
  • Menu
Publications

Publications by Ricardo Pereira Cruz

2021

Ordinal losses for classification of cervical cancer risk

Authors
Albuquerque, T; Cruz, R; Cardoso, JS;

Publication
PEERJ COMPUTER SCIENCE

Abstract
Cervical cancer is the fourth leading cause of cancer-related deaths in women, especially in low to middle-income countries. Despite the outburst of recent scientific advances, there is no totally effective treatment, especially when diagnosed in an advanced stage. Screening tests, such as cytology or colposcopy, have been responsible for a substantial decrease in cervical cancer deaths. Cervical cancer automatic screening via Pap smear is a highly valuable cell imaging-based detection tool, where cells must be classified as being within one of a multitude of ordinal classes, ranging from abnormal to normal. Current approaches to ordinal inference for neural networks are found to not sufficiently take advantage of the ordinal problem or to be too uncompromising. A non-parametric ordinal loss for neuronal networks is proposed that promotes the output probabilities to follow a unimodal distribution. This is done by imposing a set of different constraints over all pairs of consecutive labels which allows for a more flexible decision boundary relative to approaches from the literature. Our proposed loss is contrasted against other methods from the literature by using a plethora of deep architectures. A first conclusion is the benefit of using non-parametric ordinal losses against parametric losses in cervical cancer risk prediction. Additionally, the proposed loss is found to be the top-performer in several cases. The best performing model scores an accuracy of 75.6% for seven classes and 81.3% for four classes.

2021

Background Invariance by Adversarial Learning

Authors
Cruz, R; Prates, RM; Simas, EF; Costa, JFP; Cardoso, JS;

Publication
2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR)

Abstract
Convolutional neural networks are shown to be vulnerable to changes in the background. The proposed method is an end-to-end method that augments the training set by introducing new backgrounds during the training process. These backgrounds are created by a generative network that is trained as an adversary to the model. A case study is explored based on overhead power line insulators detection using a drone - a training set is prepared from photographs taken inside a laboratory and then evaluated using photographs that are harder to collect from outside the laboratory. The proposed method improves performance by over 20% for this case study.

2022

Quasi-Unimodal Distributions for Ordinal Classification

Authors
Albuquerque, T; Cruz, R; Cardoso, JS;

Publication
MATHEMATICS

Abstract
Ordinal classification tasks are present in a large number of different domains. However, common losses for deep neural networks, such as cross-entropy, do not properly weight the relative ordering between classes. For that reason, many losses have been proposed in the literature, which model the output probabilities as following a unimodal distribution. This manuscript reviews many of these losses on three different datasets and suggests a potential improvement that focuses the unimodal constraint on the neighborhood around the true class, allowing for a more flexible distribution, aptly called quasi-unimodal loss. For this purpose, two constraints are proposed: A first constraint concerns the relative order of the top-three probabilities, and a second constraint ensures that the remaining output probabilities are not higher than the top three. Therefore, gradient descent focuses on improving the decision boundary around the true class in detriment to the more distant classes. The proposed loss is found to be competitive in several cases.

2023

Rethinking low-cost microscopy workflow: Image enhancement using deep based Extended Depth of Field methods

Authors
Albuquerque, T; Rosado, L; Cruz, RPM; Vasconcelos, MJM; Oliveira, T; Cardoso, JS;

Publication
Intell. Syst. Appl.

Abstract
Microscopic techniques in low-to-middle income countries are constrained by the lack of adequate equipment and trained operators. Since light microscopy delivers crucial methods for the diagnosis and screening of numerous diseases, several efforts have been made by the scientific community to develop low-cost devices such as 3D-printed portable microscopes. Nevertheless, these devices present some drawbacks that directly affect image quality: the capture of the samples is done via mobile phones; more affordable lenses are usually used, leading to poorer physical properties and images with lower depth of field; misalignments in the microscopic set-up regarding optical, mechanical, and illumination components are frequent, causing image distortions such as chromatic aberrations. This work investigates several pre-processing methods to tackle the presented issues and proposed a new workflow for low-cost microscopy. Additionally, two new deep learning models based on Convolutional Neural Networks are also proposed (EDoF-CNN-Fast and EDoF-CNN-Pairwise) to generate Extended Depth of Field (EDoF) images, and compared against state-of-the-art approaches. The models were tested using two different datasets of cytology microscopic images: public Cervix93 and a new dataset that has been made publicly available containing images captured with µSmartScope. Experimental results demonstrate that the proposed workflow can achieve state-of-the-art performance when generating EDoF images from low-cost microscopes. © 2022 The Author(s)

2023

Two-Stage Framework for Faster Semantic Segmentation

Authors
Cruz, R; Silva, DTE; Goncalves, T; Carneiro, D; Cardoso, JS;

Publication
SENSORS

Abstract
Semantic segmentation consists of classifying each pixel according to a set of classes. Conventional models spend as much effort classifying easy-to-segment pixels as they do classifying hard-to-segment pixels. This is inefficient, especially when deploying to situations with computational constraints. In this work, we propose a framework wherein the model first produces a rough segmentation of the image, and then patches of the image estimated as hard to segment are refined. The framework is evaluated in four datasets (autonomous driving and biomedical), across four state-of-the-art architectures. Our method accelerates inference time by four, with additional gains for training time, at the cost of some output quality.

2023

Two-Stage Semantic Segmentation in Neural Networks

Authors
Silva, DTE; Cruz, R; Goncalves, T; Carneiro, D;

Publication
FIFTEENTH INTERNATIONAL CONFERENCE ON MACHINE VISION, ICMV 2022

Abstract
Semantic segmentation consists of classifying each pixel according to a set of classes. This process is particularly slow for high-resolution images, which are present in many applications, ranging from biomedicine to the automotive industry. In this work, we propose an algorithm targeted to segment high-resolution images based on two stages. During stage 1, a lower-resolution interpolation of the image is the input of a first neural network, whose low-resolution output is resized to the original resolution. Then, in stage 2, the probabilities resulting from stage 1 are divided into contiguous patches, with less confident ones being collected and refined by a second neural network. The main novelty of this algorithm is the aggregation of the low-resolution result from stage 1 with the high-resolution patches from stage 2. We propose the U-Net architecture segmentation, evaluated in six databases. Our method shows similar results to the baseline regarding the Dice coefficient, with fewer arithmetic operations.

  • 3
  • 5