Cookies
O website necessita de alguns cookies e outros recursos semelhantes para funcionar. Caso o permita, o INESC TEC irá utilizar cookies para recolher dados sobre as suas visitas, contribuindo, assim, para estatísticas agregadas que permitem melhorar o nosso serviço. Ver mais
Aceitar Rejeitar
  • Menu
Publicações

2024

Enhancing Object Detection in Maritime Environments Using Metadata

Autores
Fernandes, DS; Bispo, J; Bento, LC; Figueiredo, M;

Publicação
PROGRESS IN PATTERN RECOGNITION, IMAGE ANALYSIS, COMPUTER VISION, AND APPLICATIONS, CIARP 2023, PT II

Abstract
Over the years, many solutions have been suggested in order to improve object detection in maritime environments. However, none of these approaches uses flight information, such as altitude, camera angle, time of the day, and atmospheric conditions, to improve detection accuracy and network robustness, even though this information is often available and captured by the UAV. This work aims to develop a network unaffected by image-capturing conditions, such as altitude and angle. To achieve this, metadata was integrated into the neural network, and an adversarial learning training approach was employed. This was built on top of the YOLOv7, which is a state-of-the-art realtime object detector. To evaluate the effectiveness of this methodology, comprehensive experiments and analyses were conducted. Findings reveal that the improvements achieved by this approach are minimal when trying to create networks that generalize more across these specific domains. The YOLOv7 mosaic augmentation was identified as one potential responsible for this minimal impact because it also enhances the model's ability to become invariant to these image-capturing conditions. Another potential cause is the fact that the domains considered (altitude and angle) are not orthogonal with respect to their impact on captured images. Further experiments should be conducted using datasets that offer more diverse metadata, such as adverse weather and sea conditions, which may be more representative of real maritime surveillance conditions. The source code of this work is publicly available at https://git hub.com/ipleiria-robotics/maritime-metadata-adaptation.

2024

Image Transfer over MQTT in IoT: Message Segmentation and Encryption for Remote Indicator Panels

Autores
Valente, D; Brito, T; Correia, M; Carvalho, JA; Lima, J;

Publicação
OPTIMIZATION, LEARNING ALGORITHMS AND APPLICATIONS, PT I, OL2A 2023

Abstract
The Internet of Things (IoT) has revolutionized how objects and devices interact, creating new possibilities for seamless connectivity and data exchange. This paper presents a unique and effective method for transferring images via the Message Queuing Telemetry Transport (MQTT) protocol in an encrypted manner. The image is split into multiple messages, with each carrying a segment of the image, and employ top-notch encryption techniques to ensure secure communication. Applying this process, the message payload is split into smaller segments, and consequently, it minimizes the network bandwidth impact while mitigating potential of packet loss or latency issues. Furthermore, by applying encryption techniques, we guarantee the confidentiality and integrity of the image data during transmission, safeguarding against unauthorized access or tampering. Our experiments in a real-world scenario involving remote indicator panels with LEDs verify the effectiveness of our approach. By using our proposed method, we successfully transmit images over MQTT, achieving secure and reliable data transfer while ensuring the integrity of the image content. Our results demonstrate the feasibility and effectiveness of the proposed approach for image transfer in IoT applications. The combination of message segmentation, MQTT protocol, and encryption techniques offers a practical solution for transmitting images in resource-constrained IoT networks while maintaining data security. This approach can be applied in different applications.

2024

Automating the Annotation of Medical Images in Capsule Endoscopy Through Convolutional Neural Networks and CBIR

Autores
Fernandes, R; Salgado, M; Paçal, I; Cunha, A;

Publicação
Lecture Notes of the Institute for Computer Sciences, Social-Informatics and Telecommunications Engineering, LNICST

Abstract
This research addresses the significant challenge of automating the annotation of medical images, with a focus on capsule endoscopy videos. The study introduces a novel approach that synergistically combines Deep Learning and Content-Based Image Retrieval (CBIR) techniques to streamline the annotation process. Two pre-trained Convolutional Neural Networks (CNNs), MobileNet and VGG16, were employed to extract and compare visual features from medical images. The methodology underwent rigorous validation using various performance metrics such as accuracy, AUC, precision, and recall. The MobileNet model demonstrated exceptional performance with a test accuracy of 98.4%, an AUC of 99.9%, a precision of 98.2%, and a recall of 98.6%. On the other hand, the VGG16 model achieved a test accuracy of 95.4%, an AUC of 99.2%, a precision of 97.3%, and a recall of 93.5%. These results indicate the high efficacy of the proposed method in the automated annotation of medical images, establishing it as a promising tool for medical applications. The study also highlights potential avenues for future research, including expanding the image retrieval scope to encompass entire endoscopy video databases. © ICST Institute for Computer Sciences, Social Informatics and Telecommunications Engineering 2024.

2024

Preface

Autores
Mamede, S; Santos, A;

Publicação
Creating Learning Organizations Through Digital Transformation

Abstract
[No abstract available]

2024

Innovating in Nursing Education: A Game Prototype for Bridging the Gap in Family-Centered Care

Autores
de Oliveira, JF; Campos, J; Martins, T; Fernandes, CS; Ferreira, MC;

Publicação
PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON UBIQUITOUS COMPUTING AND AMBIENT INTELLIGENCE, UCAMI 2024

Abstract
In recent years, there has been an increasing trend towards innovative and interactive learning approaches. Serious games have emerged as a promising solution in health education, offering engaging and immersive learning experiences. This article presents the development steps of a mobile application to promote knowledge of nursing assessment and intervention in the family. A prototype was developed for Android devices using React Native technology and Firebase database, incorporating gamification elements. It was then evaluated by potential users. The results showed that the proposed solution successfully enhanced nurses' learning about family issues and dynamics, receiving positive feedback from users regarding its effectiveness and usability. By leveraging the power of mobile technology and gamification, this researchwork seeks to bridge an existing gap, contributing to the advancement of game-based educational approaches in the health field.

2024

A Vision Transformer Approach to Fundus Image Classification

Autores
Leite, D; Camara, J; Rodrigues, J; Cunha, A;

Publicação
Lecture Notes of the Institute for Computer Sciences, Social-Informatics and Telecommunications Engineering, LNICST

Abstract
Glaucoma is a condition that affects the optic nerve, with loss of retinal nerve fibers, increased excavation of the optic nerve, and a progressive decrease in the visual field. It is the leading cause of irreversible blindness in the world. Manual classification of glaucoma is a complex and time-consuming process that requires assessing a variety of ocular features by experienced clinicians. Automated detection can assist the specialist in early diagnosis and effective treatment of glaucoma and prevent vision loss. This study developed a deep learning model based on vision transformers, called ViT-BRSET, to detect patients with increased excavation of the optic nerve automatically. ViT-BRSET is a neural network architecture that is particularly effective for computer vision tasks. The results of this study were promising, with an accuracy of 0.94, an F1-score of 0.91, and a recall of 0.94. The model was trained on a new dataset called BRSET, which consists of 16,112 fundus images of patients with increased excavation of the optic nerve. The results of this study suggest that ViT-BRSET has the potential to improve early diagnosis through early detection of optic nerve excavation, one of the main signs of glaucomatous disease. ViT-BRSET can be used to mass-screen patients, identifying those who need further examination by a doctor. © ICST Institute for Computer Sciences, Social Informatics and Telecommunications Engineering 2024.

  • 41
  • 4076