Cookies Policy
The website need some cookies and similar means to function. If you permit us, we will use those means to collect data on your visits for aggregated statistics to improve our service. Find out More
Accept Reject
  • Menu
About

About

I am Associate Professor with Habilitation at the Informatics Department of University of Minho, where I teach Fault-Tolerant Distributed Systems. I serve as a member of the Board of Directors of INESC TEC, director of the Minho Advanced Computing Centre, and co-director of the UT Austin Portugal program. I am the national representative to the EuroHPC JU.

I received my PhD. degree in 2000 from the École Polytechnique Fédérale de Lausanne. My main research contributions have been in the field of fault-tolerant distributed agreement and epidemic communication algorithms and in the conception, development and assessment of dependable database systems. I have coordinated the H2020 SafeCloud project on secure processing in the Cloud and the FP6 GORDA project on open database replication, two previous national projects on scalable dependable databases, ESCADA and StrongRep, and the U. Minho team in the FP7 CumuloNimbo and LeanBigdata projects. I published over 100 research papers on large scale and dependable distributed systems, and served on the programme committee of several highly reputed conferences. I chaired the PC of IFIP DAIS and IEEE SRDS, and served as general chair of SRDS and ACM Eurosys.

Interest
Topics
Details

Details

  • Name

    Rui Carlos Oliveira
  • Role

    Diretor
  • Since

    01st November 2011
017
Publications

2025

Uma extensão de Raft com propagação epidémica

Authors
Gonçalves, A; Alonso, AN; Pereira, J; Oliveira, R;

Publication
CoRR

Abstract

2025

Machine Learning Regression-Based Prediction for Improving Performance and Energy Consumption in HPC Platforms

Authors
Coelho, M; Ocana, K; Pereira, A; Porto, A; Cardoso, DO; Lorenzon, A; Oliveira, R; Navaux, POA; Osthoff, C;

Publication
HIGH PERFORMANCE COMPUTING, CARLA 2024

Abstract
High-performance computing is pivotal for processing large datasets and executing complex simulations, ensuring faster and more accurate results. Improving the performance of software and scientific workflows in such environments requires careful analysis of their computational behavior and energy consumption. Therefore, maximizing computational throughput in these environments, through adequate software configuration and resource allocation, is essential for improving performance. The work presented in this paper focuses on leveraging regression-based machine learning and decision trees to analyze and optimize resource allocation in high-performance computing environments based on application's performance and energy metrics. Applied to a bioinformatics case study, these models enable informed decision-making by selecting the appropriate computing resources to enhance the performance of a phylogenomics software. Our contribution is to better explore and understand the efficient resource management of supercomputers, namely Santos Dumont. We show that the predictions for application's execution time using the proposed method are accurate for various amounts of computing nodes, while energy consumption predictions are less precise. The application parameters most relevant for this work are identified and the relative importance of each application parameter to the accuracy of the prediction is analysed.

2023

Privacy-Preserving Machine Learning on Apache Spark

Authors
Brito, CV; Ferreira, PG; Portela, BL; Oliveira, RC; Paulo, JT;

Publication
IEEE ACCESS

Abstract
The adoption of third-party machine learning (ML) cloud services is highly dependent on the security guarantees and the performance penalty they incur on workloads for model training and inference. This paper explores security/performance trade-offs for the distributed Apache Spark framework and its ML library. Concretely, we build upon a key insight: in specific deployment settings, one can reveal carefully chosen non-sensitive operations (e.g. statistical calculations). This allows us to considerably improve the performance of privacy-preserving solutions without exposing the protocol to pervasive ML attacks. In more detail, we propose Soteria, a system for distributed privacy-preserving ML that leverages Trusted Execution Environments (e.g. Intel SGX) to run computations over sensitive information in isolated containers (enclaves). Unlike previous work, where all ML-related computation is performed at trusted enclaves, we introduce a hybrid scheme, combining computation done inside and outside these enclaves. The experimental evaluation validates that our approach reduces the runtime of ML algorithms by up to 41% when compared to previous related work. Our protocol is accompanied by a security proof and a discussion regarding resilience against a wide spectrum of ML attacks.

2023

Toward a Practical and Timely Diagnosis of Application's I/O Behavior

Authors
Esteves, T; Macedo, R; Oliveira, R; Paulo, J;

Publication
IEEE ACCESS

Abstract
We present DIO, a generic tool for observing inefficient and erroneous I/O interactions between applications and in-kernel storage backends that lead to performance, dependability, and correctness issues. DIO eases the analysis and enables near real-time visualization of complex I/O patterns for data-intensive applications generating millions of storage requests. This is achieved by non-intrusively intercepting system calls, enriching collected data with relevant context, and providing timely analysis and visualization for traced events. We demonstrate its usefulness by analyzing four production-level applications. Results show that DIO enables diagnosing inefficient I/O patterns that lead to poor application performance, unexpected and redundant I/O calls caused by high-level libraries, resource contention in multithreaded I/O that leads to high tail latency, and erroneous file accesses that cause data loss. Moreover, through a detailed evaluation, we show that, when comparing DIO's inline diagnosis pipeline with a similar state-of-the-art solution, our system captures up to 28x more events while keeping tracing performance overhead between 14% and 51%.

2023

LOOM: A Closed-Box Disaggregated Database System

Authors
Coelho, F; Alonso, AN; Ferreira, L; Pereira, J; Oliveira, R;

Publication
PROCEEDINGS OF12TH LATIN-AMERICAN SYMPOSIUM ON DEPENDABLE AND SECURE COMPUTING, LADC 2023

Abstract
Cloud native database systems provide highly available and scalable services as part of cloud platforms by transparently replicating and partitioning data across automatically managed resources. Some systems, such as Google Spanner, are designed and implemented from scratch. Others, such as Amazon Aurora, derive from traditional database systems for better compatibility but disaggregate storage to cloud services. Unfortunately, because they follow an open-box approach and fork the original code base, they are difficult to implement and maintain. We address this problem with Loom, a replicated and partitioned database system built on top of PostgreSQL that delegates durable storage to a distributed log native to the cloud. Unlike previous disaggregation proposals, Loom is a closed-box approach that uses the original server through existing interfaces to simplify implementation and improve robustness and maintainability. Experimental evaluation achieves 6x higher throughput and 5x lower response time than standard replication and competes with the state of the art in cloud and HPC hardware.