2014
Authors
Lau, N; Moreira, AP; Ventura, R; Faria, BM;
Publication
ICARSC
Abstract
2019
Authors
Almeida, L; Reis, LP; Moreira, AP;
Publication
ICARSC
Abstract
2021
Authors
Amorim, A; Guimares, D; Mendona, T; Neto, P; Costa, P; Moreira, AP;
Publication
ROBOTICS AND COMPUTER-INTEGRATED MANUFACTURING
Abstract
Robots are increasingly present in our lives, sharing the workspace and tasks with human co-workers. However, existing interfaces for human-robot interaction / cooperation (HRI/C) have limited levels of intuitiveness to use and safety is a major concern when humans and robots share the same workspace. Many times, this is due to the lack of a reliable estimation of the human pose in space which is the primary input to calculate the human-robot minimum distance (required for safety and collision avoidance) and HRI/C featuring machine learning algorithms classifying human behaviours / gestures. Each sensor type has its own characteristics resulting in problems such as occlusions (vision) and drift (inertial) when used in an isolated fashion. In this paper, it is proposed a combined system that merges the human tracking provided by a 3D vision sensor with the pose estimation provided by a set of inertial measurement units (IMUs) placed in human body limbs. The IMUs compensate the gaps in occluded areas to have tracking continuity. To mitigate the lingering effects of the IMU offset we propose a continuous online calculation of the offset value. Experimental tests were designed to simulate human motion in a human-robot collaborative environment where the robot moves away to avoid unexpected collisions with de human. Results indicate that our approach is able to capture the human's position, for example the forearm, with a precision in the millimetre range and robustness to occlusions.
2021
Authors
de Souza, JPC; Costa, CM; Rocha, LF; Arrais, R; Moreira, AP; Pires, EJS; Boaventura Cunha, J;
Publication
ROBOTICS AND COMPUTER-INTEGRATED MANUFACTURING
Abstract
Several approaches with interesting results have been proposed over the years for robot grasp planning. However, the industry suffers from the lack of an intuitive and reliable system able to automatically estimate grasp poses while also allowing the integration of grasp information from the accumulated knowledge of the end user. In the presented paper it is proposed a non-object-agnostic grasping pipeline motivated by picking use cases from the aerospace industry. The planning system extends the functionality of the simulated annealing optimization algorithm for allowing its application within an industrial use case. Therefore, this paper addresses the first step of the design of a reconfigurable and modular grasping pipeline. The key idea is the creation of an intuitive and functional grasping framework for being used by factory floor operators according to the task demands. This software pipeline is capable of generating grasp solutions in an offline phase, and later on, in the robot operation phase, can choose the best grasp pose by taking into consideration a set of heuristics that try to achieve a successful grasp while also requiring the least effort for the robotic arm. The results are presented in a simulated and a real factory environment, relying on a mobile platform developed for intralogistic tasks. With this architecture, new state-of-art methodologies can be integrated in the future for growing the grasping pipeline and make it more robust and applicable to a wider range of use cases.
2021
Authors
Baltazar, AR; Petry, MR; Silva, MF; Moreira, AP;
Publication
SN APPLIED SCIENCES
Abstract
The transport of patients from the inpatient service to the operating room is a recurrent task in a hospital routine. This task is repetitive, non-ergonomic, time consuming, and requires the labor of patient transporters. In this paper is presented a system, named Connected Driverless Wheelchair, that can receive transportation requests directly from the hospital information management system, pick up patients at their beds, navigate autonomously through different floors, avoid obstacles, communicate with elevators, and drop patients off at the designated operating room. As a result, a prototype capable of transporting patients autonomously in hospital environments was obtained. Although it was impossible to test the final developed system at the hospital as planned, due to the COVID-19 pandemic, the extensive tests conducted at the robotics laboratory facilities, and our previous experience in integrating mobile robots in hospitals, allowed to conclude that it is perfectly prepared for this integration to be carried out.The achieved results are relevant since this is a system that may be applied to support these types of tasks in the future, making the transport of patients more efficient (both from a cost and time perspective), without unpredictable delays and, in some cases, safer.
2021
Authors
Sousa, RB; Petry, MR; Moreira, AP;
Publication
CONTROLO 2020
Abstract
Data acquisition is a critical task for localisation and perception of mobile robots. It is necessary to compute the relative pose between onboard sensors to process the data in a common frame. Thus, extrinsic calibration computes the sensor’s relative pose improving data consistency between them. This paper performs a literature review on extrinsic sensor calibration methods prioritising the most recent ones. The sensors types considered were laser scanners, cameras and IMUs. It was found methods for robot–laser, laser–laser, laser–camera, robot–camera, camera–camera, camera–IMU, IMU–IMU and laser–IMU calibration. The analysed methods allow the full calibration of a sensory system composed of lasers, cameras and IMUs. © The Editor(s) (if applicable) and The Author(s), under exclusive license to Springer Nature Switzerland AG 2021.
The access to the final selection minute is only available to applicants.
Please check the confirmation e-mail of your application to obtain the access code.