| |
Last updated on October 19, 2022. This conference program is tentative and subject to change
Technical Program for Tuesday October 25, 2022
|
TuA-1 |
Rm1 (Room A) |
Award Session V |
Regular session |
Chair: Matsuno, Fumitoshi | Kyoto University |
Co-Chair: Harada, Kensuke | Osaka University |
|
10:00-10:15, Paper TuA-1.1 | |
Robot-Assisted Nuclear Disaster Response: Report and Insights from a Field Exercise (Finalist for IROS Best Paper Award on Safety, Security, and Rescue Robotics in Memory of Motohiro Kisoi Sponsored by IRSI) |
|
Chiou, Manolis | University of Birmingham |
Epsimos, Georgios-Theofanis | University of Birmingham |
Nikolaou, Grigoris | University of West Attica |
Pappas, Pantelis | Univercity of West Attica |
Petousakis, Giannis | University of Birmingham |
Muhl, Stefan | Kerntechnische Hilfdienst GmbH (KHG) |
Stolkin, Rustam | University of Birmingham |
Keywords: Robotics in Hazardous Fields, Search and Rescue Robots, Human Factors and Human-in-the-Loop
Abstract: This paper reports on insights by robotics researchers that participated in a 5-day robot-assisted nuclear disaster response field exercise conducted by Kerntechnische Hilfdienst GmbH (KHG) in Karlsruhe, Germany. The German nuclear industry established KHG to provide a robot-assisted emergency response capability for nuclear accidents. We present a systematic description of the equipment used; the robot operators' training program; the field exercise and robot tasks; and the protocols followed during the exercise. Additionally, we provide insights and suggestions for advancing disaster response robotics based on these observations. Specifically, the main degradation in performance comes from the cognitive and attentional demands on the operator. Furthermore, robotic platforms and modules should aim to be robust and reliable in addition to their ease of use. Last, as emergency response stakeholders are often skeptical about using autonomous systems, we suggest adopting a variable autonomy paradigm to integrate autonomous robotic capabilities with the human-in-the-loop gradually. This middle ground between teleoperation and autonomy can increase end-user acceptance while directly alleviating some of the operator's robot control burden and maintaining the resilience of the human-in-the-loop.
|
|
10:15-10:30, Paper TuA-1.2 | |
Power-Based Safety Layer for Aerial Vehicles in Physical Interaction Using Lyapunov Exponents (Finalist for IROS Best Paper Award on Safety, Security, and Rescue Robotics in Memory of Motohiro Kisoi Sponsored by IRSI) |
|
Cuniato, Eugenio | ETH Zurich |
Lawrance, Nicholas Robert Jonathon | ETH Zürich |
Tognon, Marco | ETH Zurich |
Siegwart, Roland | ETH Zurich |
Keywords: Aerial Systems: Mechanics and Control, Robot Safety, Force Control
Abstract: As the performance of autonomous systems increases, safety concerns arise, especially when operating in non-structured environments. To deal with these concerns, this work presents a safety layer for mechanical systems that detects and responds to unstable dynamics caused by external disturbances. The safety layer is implemented independently and on top of already present nominal controllers, like pose or wrench tracking, and limits power flow when the system's response would lead to instability. This approach is based on the computation of the Largest Lyapunov Exponent (LLE) of the system’s error dynamics, which represent a measure of the dynamics’ divergence or convergence rate. By actively computing this metric, divergent and possibly dangerous system behaviors can be promptly detected. The LLE is then used in combination with Control Barrier Functions (CBFs) to impose power limit constraints on a jerk controlled system. The proposed architecture is experimentally validated on an Omnidirectional Micro Aerial Vehicle (OMAV) both in free flight and interaction tasks.
|
|
10:30-10:45, Paper TuA-1.3 | |
Risk-Aware Motion Planning for Collision-Tolerant Aerial Robots Subject to Localization Uncertainty (Finalist for IROS Best Paper Award on Safety, Security, and Rescue Robotics in Memory of Motohiro Kisoi Sponsored by IRSI) |
|
De Petris, Paolo | NTNU - Norwegian University of Science and Technology |
Dharmadhikari, Mihir Rahul | NTNU - Norwegian University of Science and Technology |
Nguyen, Huan | NTNU - Norwegian University of Science and Technology |
Alexis, Kostas | NTNU - Norwegian University of Science and Technology |
Keywords: Aerial Systems: Perception and Autonomy, Motion and Path Planning
Abstract: This paper contributes a novel strategy towards risk-aware motion planning for collision-tolerant aerial robots subject to localization uncertainty. Attuned to the fact that micro aerial vehicles are often tasked to navigate within GPS-denied, possibly unknown, confined and obstacle-filled environments the proposed method exploits collision-tolerance at the robot design level to mitigate the risks of collisions especially as their likelihood increases with growing uncertainty. Accounting for the maximum kinetic energy with which an impact is considered safe, alongside the robot dynamics, the planner builds a set of admissible uncertainty-aware and collision-inclusive paths over a horizon involving multiple motion steps. The first step of the best path is executed by the robot, while the procedure is then repeated in a receding horizon manner. Evaluated in extensive simulation studies and experimental results with a collision-tolerant flying robot, the planner successfully considers the interplay between uncertainty and the likelihood of a collision, balances the risks of possible impacts and enables to navigate safely within highly cluttered environments.
|
|
10:45-11:00, Paper TuA-1.4 | |
A Planning-And-Control Framework for Aerial Manipulation of Articulated Objects (Finalist for IROS Best Paper Award on Mobile Manipulation Sponsored by OMRON Sinic X Corp.) |
|
Brunner, Maximilian | ETH Zurich |
Rizzi, Giuseppe Maria | ETH Zurich |
Studiger, Matthias | ETH Zürich |
Siegwart, Roland | ETH Zurich |
Tognon, Marco | ETH Zurich |
Keywords: Aerial Systems: Applications, Aerial Systems: Perception and Autonomy, Manipulation Planning
Abstract: While the variety of applications for Aerial Manipulators (AMs) has increased over the last years, they are mostly limited to push-and-slide tasks. More complex manipulations of dynamic environments are poorly addressed and still require handcrafted designs of hardware, control, and trajectory planning. In this paper we focus on the active manipulation of articulated objects with AMs. We present a novel planning and control approach that allows the AM to execute complex interaction maneuvers with as little as possible priors given by the operator. Our framework combines sampling-based predictive control to generate pose trajectories with an impedance controller for compliant behaviours, applied to a fully-actuated flying platform. The framework leverages a physics engine to simulate the dynamics of the platform and the environment in order to find optimal motions to execute manipulation tasks. Experiments on two selected examples of pulling open a door and of turning a valve show the feasibility of the proposed approach.
|
|
11:00-11:15, Paper TuA-1.5 | |
Safe Drone Flight with Time-Varying Backup Controllers (Finalist for IROS Best Paper Award on Safety, Security, and Rescue Robotics in Memory of Motohiro Kisoi Sponsored by IRSI) |
|
Singletary, Andrew | California Institute of Technology |
Swann, Aiden | Caltech |
Jimenez Rodriguez, Ivan Dario | California Institute of Technology |
Ames, Aaron | Caltech |
Keywords: Robot Safety, Aerial Systems: Mechanics and Control, Multi-Robot Systems
Abstract: The weight, space, and power limitations of small aerial vehicles often prevent the application of modern control techniques without significant model simplifications. Moreover, high-speed agile behavior, such as that exhibited in drone racing, make these simplified models too unreliable for safety-critical control. In this work, we introduce the concept of time-varying backup controllers (TBCs): user-specified maneuvers combined with backup controllers that generate reference trajectories which guarantee the safety of nonlinear systems. TBCs reduce conservatism when compared to traditional backup controllers and can be directly applied to multi-agent coordination to guarantee safety. Theoretically, we provide conditions under which TBCs strictly reduce conservatism, describe how to switch between several TBC's and show how to embed TBCs in a multi-agent setting. Experimentally, we verify that TBCs safely increase operational freedom when filtering a pilot's actions and demonstrate robustness and computational efficiency when applied to decentralized safety filtering of two quadrotors.
|
|
11:15-11:30, Paper TuA-1.6 | |
Mobile Manipulation Leveraging Multiple Views (Finalist for IROS Best Paper Award on Mobile Manipulation Sponsored by OMRON Sinic X Corp.) |
|
Watkins-Valls, David | Columbia University |
Allen, Peter | Columbia University |
Maia, Henrique | Columbia University |
Seshadri, Madhavan | Columbia University in the City of New York |
Sanabria, Jonathan | Columbia University in the City of New York |
Waytowich, Nicholas | US. Army Research Laboratory |
Varley, Jacob | Google |
Keywords: Mobile Manipulation, Vision-Based Navigation, Machine Learning for Robot Control
Abstract: While both navigation and manipulation are challenging topics in isolation, many tasks require the ability to both navigate and manipulate in concert. To this end, we propose a mobile manipulation system that leverages novel navigation and shape completion methods to manipulate an object with a mobile robot. Our system utilizes uncertainty in the initial estimation of a manipulation target to calculate a predicted next-best-view. Without the need of localization, the robot then uses the predicted panoramic view at the next-best-view location to navigate to the desired location, capture a second view of the object, create a new model that predicts the shape of object more accurately than a single image alone, and uses this model for grasp planning. We show that the system is highly effective for mobile manipulation tasks through simulation experiments using real world data, as well as ablations on each component of our system.
|
|
TuA-2 |
Rm2 (Room B-1) |
Award Session VI |
Regular session |
Chair: Oh, Sehoon | DGIST |
Co-Chair: von Drigalski, Felix Wolf Hans Erich | Mujin Inc |
|
10:00-10:15, Paper TuA-2.1 | |
Safety-Critical Manipulation for Collision-Free Food Preparation (Finalist for IROS Best Paper Award for Industrial Robotics Research with Real-World Applications Sponsored by Mujin Inc.) |
|
Singletary, Andrew | California Institute of Technology |
Guffey, William | Miso Robotics |
Molnar, Tamas G. | California Institute of Technology |
Sinnet, Ryan | Miso Robotics, Inc |
Ames, Aaron | Caltech |
Keywords: Robot Safety, Manipulation Planning, Integrated Planning and Control
Abstract: Recent advances allow for the automation of food preparation in high-throughput environments, yet the successful deployment of these robots requires the planning and execution of quick, robust, and ultimately collision-free behaviors. In this work, we showcase a novel framework for modifying previously generated trajectories of robotic manipulators in highly detailed and dynamic collision environments using Control Barrier Functions (CBFs). This method dynamically re-plans previously validated behaviors in the presence of changing environments---and does so in a computationally efficient manner. Moreover, the approach provides rigorous safety guarantees of the resulting trajectories, factoring in the true underlying dynamics of the manipulator. This methodology is extensively validated on a full-scale robotic manipulator in a real-world cooking environment, and has resulted in substantial improvements in computation time and robustness over re-planning.
|
|
10:15-10:30, Paper TuA-2.2 | |
Impedance Control on Arbitrary Surfaces for Ultrasound Scanning Using Discrete Differential Geometry (Finalist for IROS Best Paper Award for Industrial Robotics Research with Real-World Applications Sponsored by Mujin Inc.) |
|
Dyck, Michael | Institute of Robotics and Mechatronics, German Aerospace Center |
Sachtler, Arne | Technical University of Munich (TUM) |
Klodmann, Julian | German Aerospace Center |
Albu-Schäffer, Alin | DLR - German Aerospace Center |
Keywords: Compliance and Impedance Control, Computational Geometry, Medical Robots and Systems
Abstract: We propose an approach to robotic ultrasound scanning systems using a passivity-based impedance control scheme on arbitrary surfaces. First, we introduce task coordinates depending on the geometry of the surface, which enables hands-on guidance of the robot along the surface, as well as teleoperated and autonomous ultrasound image acquisition. Our coordinates allow controlling the signed distance of the robot to the surface and alignment of the tool to the surface normal using classical impedance control. This corresponds to implicitly obtaining a foliation of parallel surfaces. By setting the desired signed distance negative, i.e. into the surface, we obtain passive contact forces. We extend the approach to also incorporate coordinates that allow controlling the specific point on the surface and, likewise, on all parallel surfaces. Finally, we demonstrate the performance of the controller on the seven degrees of freedom lightweight robot DLR MIRO. In the experiments the robot can track complex trajectories while keeping the distance error below 1mm and applying an almost constant contact force.
|
|
10:30-10:45, Paper TuA-2.3 | |
Soft Tissue Characterisation Using a Novel Robotic Medical Percussion Device with Acoustic Analysis and Neural Networks (Finalist for IROS Best Application Paper Award Sponsored by ICROS) |
|
Zhang Qiu, Pilar | Imperial College London |
Tan, Yongxuan | Imperial College London |
Thompson, Oliver | Imperial College London |
Cobley, Bennet | Imperial College London |
Nanayakkara, Thrishantha | Imperial College London |
Keywords: Medical Robots and Systems, AI-Based Methods, Mechanism Design
Abstract: Medical percussion is a common manual examination procedure used by physicians to determine the state of underlying tissues from their acoustic responses. Although it has been used for centuries, there is a limited quantitative understanding of its dynamics, leading to subjectivity and a lack of detailed standardisation. This paper presents a novel compliant two-degree-of-freedom robotic device inspired by the human percussion action, and validates its performance in two tissue characterisation experiments. In Experiment 1, spectrotemporal analysis using 1-D Continuous Wavelet Transform (CWT) proved the potential of the device to identify hard nodules, mimicking lipomas, embedded in silicone phantoms representing a patient’s abdominal region. In Experiment 2, Gaussian Mixture Modelling (GMM) and Neural Network (NN) predictive models were implemented to classify composite phantom tissues of varying density and thickness. The proposed device and methods showed up to 97.5% accuracy in the classification of phantoms, proving the potential of robotic solutions to standardise and improve the accuracy of percussion diagnostic procedures.
|
|
10:45-11:00, Paper TuA-2.4 | |
Multi-Directional Bicycle Robot for Bridge Inspection with Steel Defect Detection System (Finalist for IROS Best Application Paper Award Sponsored by ICROS) |
|
Ahmed, Habib | University of Nevada Reno |
Nguyen, Son | University of Nevada, Reno |
La, Duc | University of Nevada, Reno |
Le, Chuong | University of Oklahoma |
La, Hung | University of Nevada at Reno |
Keywords: Climbing Robots, Field Robots, Robotics and Automation in Construction
Abstract: This paper presents a novel design of a multi-directional bicycle robot, which is developed for the inspection of steel structures, in particular, steel-reinforced bridges. The locomotion concept is based on arranging two magnetic wheels in a bicycle-like configuration with two independent steering actuators. This configuration allows the robot to possess multi-directional mobility. An additional free joint helps the robot adapt naturally to non-flat and complex steel structures. The robot’s design provides the advantage of being mechanically simple and providing high-level mobility across diverse steel structures. In addition, a visual sensor is equipped that allows the data collection for steel defect detection with offline training and validation. The paper also provides a novel pipeline for Steel Defect Detection, which utilizes multiple datasets (one for training and one for validation) from real bridges. The quantitative results have been reported for three Deep Encoder-Decoder Networks (i.e., LinkNet, UNet, DeepLab) with their corresponding Encoder modules (i.e., ResNet-18, ResNet-34, RegNet-X2, EfficientNet-B0, and EfficientNet-B2). Due to space concerns, the qualitative results have been outlined in Appendix, with a link in Fig. 11 caption to access the result provided.
|
|
11:00-11:15, Paper TuA-2.5 | |
Tactile-Sensitive NewtonianVAE for High-Accuracy Industrial Connector Insertion (Finalist for IROS Best Application Paper Award Sponsored by ICROS) |
|
Okumura, Ryo | Panasonic Holdings Corporation |
Nishio, Nobuki | Panasonic Corp |
Taniguchi, Tadahiro | Ritsumeikan University |
Keywords: Computer Vision for Automation, Force and Tactile Sensing, Machine Learning for Robot Control
Abstract: An industrial connector insertion task requires submillimeter positioning and grasp pose compensation for a plug. Thus, highly accurate estimation of the relative pose between a plug and socket is fundamental for achieving the task. World models are promising technologies for visuomotor control because they obtain appropriate state representation to jointly optimize feature extraction and latent dynamics model. Recent studies show that the NewtonianVAE, a type of the world model, acquires latent space equivalent to mapping from images to physical coordinates. Proportional control can be achieved in the latent space of NewtonianVAE. However, applying NewtonianVAE to high-accuracy industrial tasks in physical environments is an open problem. Moreover, the existing framework does not consider the grasp pose compensation in the obtained latent space. In this work, we proposed tactile-sensitive NewtonianVAE and applied it to a USB connector insertion with grasp pose variation in the physical environments. We adopted a GelSight-type tactile sensor and estimated the insertion position compensated by the grasp pose of the plug. Our method trains the latent space in an end-to-end manner, and no additional engineering and annotation are required. Simple proportional control is available in the obtained latent space. Moreover, we showed that the original NewtonianVAE fails in some situations, and demonstrated that domain knowledge induction improves model accuracy. This domain knowledge can be easily obtained using robot specification and grasp pose error measurement. We demonstrated that our proposed method achieved a 100% success rate and 0.3 mm positioning accuracy in the USB connector insertion task in the physical environment. It outperformed SOTA CNN-based two-stage goal pose regression with grasp pose compensation using coordinate transformation.
|
|
TuA-3 |
Rm3 (Room B-2) |
Grasping 4 |
Regular session |
Chair: Watanabe, Tetsuyou | Kanazawa University |
Co-Chair: Yu, Haoyong | National University of Singapore |
|
10:00-10:10, Paper TuA-3.1 | |
Enabling Massage Actions: An Interactive Parallel Robot with Compliant Joints |
|
Dong, Huixu | Zhejiang University |
Feng, Yue | Nanyang Technological University |
Qiu, Chen | Nanyang Technological University |
Pan, Ye | Senior Post Doctoral Research Associate, Computer Science, Uiver |
He, Miao | Chongqing University of Technology |
Chen, I-Ming | Nanyang Technological University |
Keywords: Grippers and Other End-Effectors, Compliant Joints and Mechanisms, Compliance and Impedance Control
Abstract: We propose a parallel massage robot with compliant joints based on the series elastic actuator (SEA), offering a unified force-position control approach. First, the kinematic and static force models are established for obtaining the corresponding control variables. Then, a novel force-position control strategy is proposed to separately control the force-position along the normal direction of the surface and another two-direction displacement, without the requirement of a robotic dynamics model. To evaluate its performance, we implement a series of robotic massage experiments. The results demonstrate that the proposed massage manipulator can successfully achieve desired forces and motion patterns of massage tasks, arriving at a high-score user experience.
|
|
10:10-10:20, Paper TuA-3.2 | |
Designing Underactuated Graspers with Dynamically Variable Geometry Using Potential Energy Map Based Analysis |
|
Yako, Connor | Stanford University |
Yuan, Shenli | Stanford University |
Salisbury, Kenneth | Stanford University |
Keywords: Grippers and Other End-Effectors, Underactuated Robots, Methods and Tools for Robot System Design
Abstract: In this paper we present a potential energy map based approach that provides a framework for the design and control of a robotic grasper. Unlike other potential energy map approaches, our framework considers friction for a more realistic perspective on grasper performance. Our analysis establishes the importance of considering dynamically variable geometry in grasper design, namely palm width, link lengths, and transmission ratios, which are assumed to be able to change in real-time. Our analysis assumes a two-phalanx tendon-pulley underactuated grasper, but it can be extended to other underactuated mechanisms. We demonstrate the utility of these novel potential energy maps and the method used to generate them in order by showing how various design parameters impact the grasping and in-hand manipulation performance of a particular design across a range of object sizes and friction coefficients. Optimal grasping designs have palms that scale with object size and transmission ratios that scale with the coefficient of friction. Using a custom in-hand manipulation metric, we compared the in-hand manipulation capabilities of a grasper that only dynamically varied its palm size, link lengths, and transmission ratios to a grasper with a variable palm and controllable actuation efforts. The analysis revealed the advantage of dynamically variable geometry; by varying only its palm size, link lengths, and transmission ratios in real-time, safe, caged in-hand manipulation of a wide range of objects could be performed.
|
|
10:20-10:30, Paper TuA-3.3 | |
A Novel Human-Safe Robotic Gripper: An Application of a Programmable Permanent Magnet Actuator |
|
Ulagaoozhian, Chandramouly | Ecole De Technologie Superieure |
Duchaine, Vincent | Ecole De Technologie Superieure |
Keywords: Grippers and Other End-Effectors, Human-Robot Collaboration, Safety in HRI
Abstract: While collaborative robotic arms offer significant safety benefits, safety of the overall manipulator system cannot be guaranteed unless equally strict safety requirements are satisfied by the accompanying end-effector. Current robot grippers are not made in a way that fulfills such a requirement, resulting in collaborative robots needing to operate in a protected environment. This paper presents a novel permanent magnet actuator inside of a conventional industrial electric gripper which results in an end-effector that has an unmatched force range of 1-2N to 43N and exhibits interesting characteristics suited to the requirements of a safe gripper such as torque holding without power, variable stiffness and force sensing.
|
|
10:30-10:40, Paper TuA-3.4 | |
The Good Grasp, the Bad Grasp, and the Plateau in Tactile Based Grasp Stability Prediction |
|
Kwiatkowski, Jennifer | École De Technologie Supérieure |
Jolaei, Mohammad | École De Technologie Supérieure (ÉTS) |
Bernier, Alexandre | École De Technologie Supérieure |
Duchaine, Vincent | Ecole De Technologie Superieure |
Keywords: Deep Learning in Grasping and Manipulation, Perception for Grasping and Manipulation, Force and Tactile Sensing
Abstract: Research around tactile sensing for grasp stability prediction in robotic manipulators continues to be popular, however few works are able to achieve a high classification accuracy. Due to simulation complexity, data-driven methods are often forced to rely on experimental data, yielding small, often unbalanced, data sets. In this work, the authors use a 3972 sample data set to explore the effects of the data set composition on the performance of a classifier. While maintaining a similar overall accuracy, the ability to recognize a grasp failure was significantly impacted by the composition of the data set. The authors propose an autonomous pipeline designed to generate more diverse failure grasps. On failure-rich data, a tactile-based classifier with a balanced training set achieved a classification accuracy of 84.68% while maintaining a recall of the grasp failure class of 76%. This represents a 71.79% improvement in recall over a model trained on a larger but unbalanced data set.
|
|
10:40-10:50, Paper TuA-3.5 | |
Extrinsic Dexterous Manipulation with a Direct-Drive Hand: A Case Study |
|
Gupta, Arnav | Carnegie Mellon University |
Mao, Yuemin | Carnegie Mellon University |
Bhatia, Ankit | Carnegie Mellon University |
Cheng, Xianyi | Carnegie Mellon University |
King, Jonathan | The Robotics Institute, Carnegie Mellon University |
Hou, Yifan | Carnegie Mellon University |
Mason, Matthew T. | Carnegie Mellon University |
Keywords: Grippers and Other End-Effectors, Dexterous Manipulation, Industrial Robots
Abstract: This paper explores a novel approach to dexterous manipulation, aimed at levels of speed, precision, robustness, and simplicity suitable for practical deployment. The enabling technology is a Direct-drive Hand (DDHand) comprising two fingers, two DOFs each, that exhibit high speed and a light touch. The test application is the dexterous manipulation of three small and irregular parts, moving them to a grasp suitable for a subsequent assembly operation, regardless of initial presentation. We employed four primitive behaviors that use ground contact as a “third finger”, prior to or during the grasp process: pushing, pivoting, toppling, and squeeze- grasping. In our experiments, each part was presented from 30 to 90 times randomly positioned in each stable pose. Success rates varied from 83% to 100%. The time to manipulate and grasp was 6.32 seconds on average, varying from 2.07 to 16 seconds. In some cases, performance was robust, precise, and fast enough for practical applications, but in other cases, pose uncertainty required time-consuming vision and arm motions. The paper concludes with a discussion of further improvements required to make the primitives robust, eliminate uncertainty, and reduce this dependence on vision and arm motion.
|
|
10:50-11:00, Paper TuA-3.6 | |
GTac-Gripper: A Reconfigurable Under-Actuated Four-Fingered Robotic Gripper with Tactile Sensing |
|
Lu, Zeyu | National University of Singapore |
Guo, Haotian | National University of Singapore |
Zhang, Wensi | National University of Singapore |
Yu, Haoyong | National University of Singapore |
Keywords: Grippers and Other End-Effectors, Multifingered Hands, Force and Tactile Sensing
Abstract: Humans can use different grasping poses and forces for everyday objects of different shapes and sizes. Grasping and manipulating everyday objects have been longstanding challenges in robotics. Performing multiple grasping configurations is difficult for robotic end-effectors with limited degrees of freedom (DOF). Integrating tactile sensing into robotic grippers will facilitate grasping and manipulating a wider range of objects. In this letter, we present a robotic gripper with a reconfigurable mechanism and tactile sensors (GTac) integrated into the fingers and palm. Each finger consists of two phalanges with a 2 DOF underactuated design and a metacarpophalangeal (MCP) joint. Our gripper with four adaptive fingers can perform 5 grasping configurations and obtain 228 tactile feedback signals (normal and shear forces) at 150 Hz. Our results show that the gripper can grasp various everyday objects and achieve in-hand manipulation including translation and rotation with closed-loop control. In the YCB benchmark assessment, the gripper achieved a score of 93% (round objects), 0% (flat objects), 78% (tools), 90% (articulated objects), and 65% in total This research provides a new hardware design and could be beneficial to various robotic applications in the domestic and industrial fields.
|
|
11:00-11:10, Paper TuA-3.7 | |
Single-Fingered Reconfigurable Robotic Gripper with a Folding Mechanism for Narrow Working Spaces |
|
Nishimura, Toshihiro | Kanazawa University |
Muryoe, Tsubasa | Kanazawa University |
Asama, Yoshitatsu | Panasonic Corporation |
Ikeuchi, Hiroki | Panasonic Corporation |
Toshima, Ryo | Panasonic Corporation |
Watanabe, Tetsuyou | Kanazawa University |
Keywords: Grippers and Other End-Effectors, Mechanism Design, Grasping
Abstract: This paper proposes a novel single-fingered reconfigurable robotic gripper for grasping objects in narrow working spaces. The finger of the developed gripper realizes two configurations, namely, the insertion and grasping modes, using only a single motor. In the insertion mode, the finger assumes a thin shape such that it can insert its tip into a narrow space. The grasping mode of the finger is activated through a folding mechanism. Mode switching can be achieved in two ways: switching the mode actively by a motor, or combining passive rotation of the fingertip through contact with the support surface and active motorized construction of the claw. The latter approach is effective when it is unclear how much finger insertion is required for a specific task. The structure provides a simple control scheme. The performance of the proposed robotic gripper design and control methodology was experimentally evaluated. The minimum width of the insertion space required to grasp an object is 4 mm (1 mm, when using a strategy).
|
|
11:10-11:20, Paper TuA-3.8 | |
SEED: Series Elastic End Effectors in 6D for Visuotactile Tool Use |
|
Suh, Hyung Ju Terry | Massachusetts Institute of Technology |
Kuppuswamy, Naveen | Toyota Research Institute |
Pang, Tao | Massachusetts Institute of Technology |
Mitiguy, Paul | Stanford University |
Alspach, Alex | Toyota Research Institute |
Tedrake, Russ | Massachusetts Institute of Technology |
Keywords: Grippers and Other End-Effectors, Soft Robot Applications, Force Control
Abstract: We propose the framework of Series Elastic End Effectors in 6D (SEED), which combines a spatially compliant element with visuotactile sensing to grasp and manipulate tools in the wild. Our framework generalizes the benefits of series elasticity to 6-dof, while providing an abstraction of control using visuotactile sensing. We propose an algorithm for relative pose estimation from visuotactile sensing, and a spatial hybrid force-position controller capable of achieving stable force interaction with the environment. We demonstrate the effectiveness of our framework on tools that require regulation of spatial forces. Video link: https://youtu.be/2-YuIfspDrk
|
|
11:20-11:30, Paper TuA-3.9 | |
Elongatable Gripper Fingers with Integrated Stretchable Tactile Sensors for Underactuated Grasping and Dexterous Manipulation (I) |
|
Yoon, Sohee John | Seoul National University |
Choi, Minsik | Seoul National University |
Jeong, Bomin | Seoul National University |
Park, Yong-Lae | Seoul National University |
Keywords: Multifingered Hands, Underactuated Robots, Soft Sensors and Actuators
Abstract: The ability to grasp a wider range of objects in size and shape directly relates to the performance of robotic grippers. Adapting to complex geometries of objects requires large degrees of freedom to allow complex configurations. However, complexity in controlling many individual joints leads to introduction of underactuated mechanisms, in which traditional finger designs composed of revolute joints allow only flexion/extension motions. In this article, we propose a length-adjustable linkage mechanism in the underactuated finger controlled by an antagonistic tendon pair. The resulting gripper can elongate the fingers for an increased task space or shorten them for a finer spatial resolution. For tactile sensing, hyperelastic soft sensors are used to stretch with finger elongation. Contact pressures measured by the soft sensors are used in force-feedback control for which either the joint angles or the link lengths are adjusted. Lastly, a multimodal control scheme that combines elongation and flexion modes is demonstrated with tasks of dexterous manipulation.
|
|
TuA-4 |
Rm4 (Room C-1) |
Manipulation Systems 4 |
Regular session |
Chair: Sanfeliu, Alberto | Universitat Politècnica De Cataluyna |
Co-Chair: Klee, David | Northeastern University |
|
10:00-10:10, Paper TuA-4.1 | |
Challenges and Outlook in Robotic Manipulation of Deformable Objects (I) |
|
Zhu, Jihong | TU Delft |
Cherubini, Andrea | LIRMM - Universite De Montpellier CNRS |
Dune, Claire | Université De Toulon |
Navarro-Alarcon, David | The Hong Kong Polytechnic University |
Alambeigi, Farshid | University of Texas at Austin |
Berenson, Dmitry | University of Michigan |
Ficuciello, Fanny | Università Di Napoli Federico II |
Harada, Kensuke | Osaka University |
Kober, Jens | TU Delft |
Li, Xiang | Tsinghua University |
Pan, Jia | University of Hong Kong |
Yuan, Wenzhen | Carnegie Mellon University |
Gienger, Michael | Honda Research Institute Europe |
Keywords: Sensor-based Control, Dexterous Manipulation, Perception for Grasping and Manipulation
Abstract: Deformable object manipulation (DOM) is an emerging research problem in robotics. The ability to manipulate deformable objects endows robots with higher autonomy and promises new applications in the industrial, services, and healthcare sectors. However, compared to rigid object manipulation, the manipulation of deformable objects is considerably more complex, and is still an open research problem. Addressing DOM challenges demand breakthroughs in almost all aspects of robotics, namely, hardware design, sensing, (deformation) modeling, planning, and control. In this article, we review recent advances and highlight the main challenges when considering deformation in each sub-field. A particular focus of our paper lies in the discussions of these challenges and proposing future directions of research.
|
|
10:10-10:20, Paper TuA-4.2 | |
Robust Robotic 3-D Drawing Using Closed-Loop Planning and Online Picked Pens (I) |
|
Liu, Ruishuang | Osaka University |
Wan, Weiwei | Osaka University |
Koyama, Keisuke | Osaka University |
Harada, Kensuke | Osaka University |
Keywords: Manipulation Planning, Task and Motion Planning, Task Planning
Abstract: This paper develops a flexible and robust robotic system for autonomously drawing on 3D surfaces. The system takes 2D drawing strokes and a 3D target surface (mesh or point clouds) as input. It maps the 2D strokes onto the 3D surface and generates a robot motion to draw the mapped strokes using visual recognition, grasp pose reasoning, and motion planning. The system is flexible compared to conventional robotic drawing systems as we do not fix drawing tools to the end of a robot arm. Instead, a robot recognizes and picks up pens online and holds the pens to draw 3D strokes. Meanwhile, the system has high robustness thanks to the following crafts: First, a high-quality mapping method is developed to minimize deformation in the strokes. Second, visual detection is used to re-estimate the drawing tool's pose before executing each drawing motion. Third, force control is employed to compensate for noisy visual detection and calibration and ensure a firm touch between the pen tip and the surface. Fourth, error detection and recovery are implemented to deal with slippage and other anomalies. The planning and executions are performed in a closed-loop manner until the strokes are successfully drawn. We evaluate the system and analyze the necessity of the various crafts using different real-world tasks. The results show that the proposed system is flexible and robust to generate robotic motion that picks up the pens and successfully draws 3D strokes on given surfaces.
|
|
10:20-10:30, Paper TuA-4.3 | |
6D Robotic Assembly Based on RGB-Only Object Pose Estimation |
|
Fu, Bowen | Tsinghua University |
Leong, Sek Kun | Tsinghua University |
Lian, Xiaocong | Tsinghua University |
Ji, Xiangyang | Tsinghua University |
Keywords: Deep Learning in Grasping and Manipulation, Assembly, Deep Learning for Visual Perception
Abstract: Vision-based robotic assembly is a crucial yet challenging task as the interaction with multiple objects requires high levels of precision. In this paper, we propose an integrated 6D robotic system to perceive, grasp, manipulate and assemble blocks with tight tolerances. Aiming to provide an off-the-shelf RGB-only solution, our system is built upon a monocular 6D object pose estimation network trained solely with synthetic images leveraging physically-based rendering. Subsequently, pose-guided 6D transformation along with collision-free assembly is proposed to construct any designed structure with arbitrary initial poses. Our novel 3-axis calibration operation further enhances the precision and robustness by disentangling 6D pose estimation and robotic assembly. Both quantitative and qualitative results demonstrate the effectiveness of our proposed 6D robotic assembly system.
|
|
10:30-10:40, Paper TuA-4.4 | |
Context and Intention Aware 3D Human Body Motion Prediction Using an Attention Deep Learning Model in Handover Tasks |
|
Laplaza, Javier | Universitat Politècnica De Catalunya |
Moreno-Noguer, Francesc | CSIC |
Sanfeliu, Alberto | Universitat Politècnica De Cataluyna |
Keywords: Deep Learning in Grasping and Manipulation, Human-Robot Collaboration, Human-Aware Motion Planning
Abstract: This work explores how contextual information and human intention affect the motion prediction of humans during a handover operation with a social robot. By classifying human intention in four different classes, we developed a model able to generate a different motion for each intention class. Furthermore, the model uses a multi-headed attention architecture to add contextual information to the pipeline, such as the position of the robot end effector (REE) or the position of obstacles in the interaction scene. We generate predictions up to two and half seconds in the future given an input sequence of one second containing the previous motion of the human. The results show an improvement of the prediction accuracy, both for the full skeleton prediction and the human hand used for the delivery. The model also allows to generate different sequences with the desired human intention.
|
|
10:40-10:50, Paper TuA-4.5 | |
Learning a State Estimator for Tactile In-Hand Manipulation |
|
Röstel, Lennart | German Aerospace Center (DLR) |
Sievers, Leon | German Aerospace Center |
Pitz, Johannes | German Aerospace Center |
Bäuml, Berthold | German Aerospace Center (DLR) |
Keywords: Deep Learning in Grasping and Manipulation, Probability and Statistical Methods, In-Hand Manipulation
Abstract: We study the problem of estimating the pose of an object which is being manipulated by a multi-fingered robotic hand by only using proprioceptive feedback. To address this challenging problem, we propose a novel variant of differentiable particle filters, which combines two key extensions. First, our learned proposal distribution incorporates recent measurements in a way that mitigates weight degeneracy. Second, the particle update works on non-euclidean manifolds like Lie-groups, enabling learning-based pose estimation in 3D on SE(3). We show that the method can represent the rich and often multi-modal distributions over poses that arise in tactile state estimation. The models are trained in simulation, but by using domain randomization, we obtain state estimators that can be employed for pose estimation on a real robotic hand (equipped with joint torque sensors). Moreover, the estimator runs fast, allowing for online usage with update rates of more than 100Hz on a single CPU core. We quantitatively evaluate our method and benchmark it against other approaches in simulation. We also show qualitative experiments on the real torque-controlled DLR-Hand II.
|
|
10:50-11:00, Paper TuA-4.6 | |
A Two-Stage Learning Architecture That Generates High-Quality Grasps for a Multi-Fingered Hand |
|
Winkelbauer, Dominik | DLR |
Bäuml, Berthold | German Aerospace Center (DLR) |
Humt, Matthias | German Aerospace Center (DLR), Technical University Munich (TUM) |
Thuerey, Nils | Technical University Munich |
Triebel, Rudolph | German Aerospace Center (DLR) |
Keywords: Deep Learning in Grasping and Manipulation, Multifingered Hands, Grasping
Abstract: In this work, we investigate the problem of planning stable grasps for object manipulations using an 18-DOF robotic hand with four fingers. The main challenge here is the high-dimensional search space, and we address this problem using a novel two-stage learning process. In the first stage, we train an autoregressive network called the hand-pose-generator, which learns to generate a distribution of valid 6D poses of the palm for a given volumetric object representation. In the second stage, we employ a network that regresses 12D finger positions and scalar grasp qualities from given object representations and palm poses. To train our networks, we use synthetic training data generated by a novel grasp planning algorithm, which also proceeds stage-wise: first the palm pose, then the finger positions. Here, we devise a Bayesian Optimization scheme for the palm pose and a physics-based grasp pose metric to rate stable grasps. In experiments on the YCB benchmark data set, we show a grasp success rate of over 83%, as well as qualitative results on real scenarios of grasping unknown objects.
|
|
11:00-11:10, Paper TuA-4.7 | |
Graph-Structured Policy Learning for Multi-Goal Manipulation Tasks |
|
Klee, David | Northeastern University |
Biza, Ondrej | Czech Technical University in Prague |
Platt, Robert | Northeastern University |
Keywords: Deep Learning in Grasping and Manipulation, Reinforcement Learning, Transfer Learning
Abstract: Multi-goal policy learning for robotic manipulation is challenging. Prior successes have used state-based representations of the objects or provided demonstration data to facilitate learning. In this paper, by hand-coding a high-level discrete representation of the domain, we show that policies to reach dozens of goals can be learned with a single network using Q-learning from pixels. The agent focuses learning on simpler, local policies which are sequenced together by planning in the abstract space. We compare our method against standard multi-goal RL baselines, as well as other methods that leverage the discrete representation, on a challenging block construction domain. We find that our method can build more than a hundred different block structures, and demonstrate forward transfer to structures with novel objects. Lastly, we deploy the policy learned in simulation on a real robot.
|
|
11:10-11:20, Paper TuA-4.8 | |
Scene Editing As Teleoperation: A Case Study in 6DoF Kit Assembly |
|
Li, Yulong | Columbia University |
Agrawal, Shubham | Columbia University |
Liu, Jen-Shuo | Columbia University |
Feiner, Steven | Columbia University |
Song, Shuran | Columbia University |
Keywords: Deep Learning in Grasping and Manipulation, Deep Learning for Visual Perception, Telerobotics and Teleoperation
Abstract: Studies in robot teleoperation have been centered around action specifications -- from continuous joint control to discrete end-effector pose control. However, these "robot-centric" interfaces often require skilled operators with extensive robotics expertise. To make teleoperation accessible to non-expert users, we propose the framework "Scene Editing as Teleoperation" (SEaT), where the key idea is to transform the traditional "robot-centric" interface into a "scene-centric" interface -- instead of controlling the robot, users focus on specifying the task's goal by manipulating digital twins of the real-world objects. As a result, a user can perform teleoperation without any expert knowledge of the robot hardware. To achieve this goal, we utilize a category-agnostic scene-completion algorithm that translates the real-world workspace (with unknown objects) into a manipulable virtual scene representation and an action-snapping algorithm that refines the user input before generating the robot's action plan. To train the algorithms, we procedurely generated a large-scale, diverse kit-assembly dataset that contains object-kit pairs that mimic real-world object-kitting tasks. Our experiments in simulation and on a real-world system demonstrate that our framework improves both the efficiency and success rate for 6DoF kit-assembly tasks. A user study demonstrates that SEaT framework participants achieve a higher task success rate and report a lower subjective workload compared to an alternative robot-centric interface.
|
|
11:20-11:30, Paper TuA-4.9 | |
On the Importance of Label Encoding and Uncertainty Estimation for Robotic Grasp Detection |
|
Stephan, Benedict | Ilmenau University of Technology |
Aganian, Dustin | University of Technology Ilmenau |
Hinneburg, Lars | Ilmenau University of Technology |
Eisenbach, Markus | Ilmenau University of Technology |
Mueller, Steffen | Ilmenau University of Technology |
Gross, Horst-Michael | Ilmenau University of Technology |
Keywords: Deep Learning in Grasping and Manipulation, Grasping
Abstract: Automated grasping of arbitrary objects is an essential skill for many applications such as smart manufacturing and human robot interaction. This makes grasp detection a vital skill for automated robotic systems. Recent work in model-free grasp detection uses point cloud data as input and typically outperforms the earlier work on RGB(D)-based methods. We show that RGB(D)-based methods are being underestimated due to suboptimal label encodings used for training. Using the evaluation pipeline of the GraspNet-1Billion dataset, we investigate different encodings and propose a novel encoding that significantly improves grasp detection on depth images. Additionally, we show shortcomings of the 2D rectangle grasps supplied by the GraspNet-1Billion dataset and propose a filtering scheme by which the ground truth labels can be improved significantly. Furthermore, we apply established methods for uncertainty estimation on our trained models since knowing when we can trust the model's decisions provides an advantage for real-world application. By doing so, we are the first to directly estimate uncertainties of detected grasps. We also investigate the applicability of the estimated aleatoric and epistemic uncertainties based on their theoretical properties. Additionally, we demonstrate the correlation between estimated uncertainties and grasp quality, thus improving selection of high quality grasp detections. By all these modifications, our approach using only depth images can compete with point-cloud-based approaches for grasp detection despite the lower degree of freedom for grasp poses in 2D image space.
|
|
TuA-5 |
Rm5 (Room C-2) |
Navigation Systems 3 |
Regular session |
Chair: Nagatani, Keiji | The University of Tokyo |
Co-Chair: Ohno, Kazunori | Tohoku University |
|
10:00-10:10, Paper TuA-5.1 | |
A Unified MPC Design Approach for AGV Path Following |
|
Kokot, Mirko | University of Zagreb |
Miklic, Damjan | Romb Technologies |
Petrovic, Tamara | Univ. of Zagreb |
Keywords: Autonomous Vehicle Navigation, Motion and Path Planning, Service Robotics
Abstract: This paper presents a unified approach to the design of Model Predictive Controllers (MPC), custom-tailored for path following by Automated Guided Vehicles (AGVs). The approach can be applied in a unified manner to several relevant AGV kinematic configurations, including tricycle, differential, and double steer-drive. By leveraging Linear Parameter Varying (LPV) MPC, it provides maximum maneuverability and industrial-grade positioning accuracy. We incorporate state-of-the-art optimized velocity planning, to maximize vehicle utilization. Experimental validation is performed on three different kinematic configurations, including a real forklift with tricycle configuration, using industrially-relevant positioning maneuvers.
|
|
10:10-10:20, Paper TuA-5.2 | |
GPU-Parallelized Iterative LQR with Input Constraints for Fast Collision Avoidance of Autonomous Vehicles |
|
Lee, Yeongseok | Korea Advanced Institute of Science and Technology |
Cho, Minsu | Korea Advanced Institute of Science and Techonology |
Kim, Kyung-Soo | KAIST(Korea Advanced Institute of Science and Technology) |
Keywords: Autonomous Vehicle Navigation, Motion and Path Planning, Control Architectures and Programming
Abstract: Collision avoidance in emergency situations is a crucial and challenging task in motion planning for autonomous vehicles. Especially in the field of optimization-based planning using nonlinear model predictive control, many efforts to achieve real-time performance are still ongoing. Among various approaches, the iterative linear quadratic regulator (iLQR) is known as an efficient means of nonlinear optimization. Additionally, parallel computing architectures, such as GPUs, are more widely applied in autonomous vehicles. In this paper, we propose 1) a parallel computing framework for iLQR with input constraints considering the characteristics of the problem and 2) a proper environmental formulation that is suitable for lower-precision GPU computations. The GPU-accelerated framework was tested on a real-time simulation-in-the-loop system using CarMaker and ROS at a 20 Hz sampling rate on a low-performance mobile computer and was compared against the same framework realized with a CPU.
|
|
10:20-10:30, Paper TuA-5.3 | |
RIANet: Road Graph and Image Attention Network for Urban Autonomous Driving |
|
Ha, Timothy | Seoul National University |
Oh, Jeongwoo | Seoul National University |
Chung, Hojun | Seoul National University |
Lee, Gunmin | Seoul National University |
Oh, Songhwai | Seoul National University |
Keywords: Autonomous Vehicle Navigation, Imitation Learning, Sensor Fusion
Abstract: In this paper, we present a novel autonomous driving framework, called a road graph and image attention network (RIANet), which computes the attention scores of objects in the image using the road graph feature. The process of the proposed method is as follows: First, the feature encoder module encodes the road graph, image, and additional features of the scene. The attention network module then incorporates the encoded features and computes the scene context feature via the attention mechanism. Finally, the low-level controller module drives the ego-vehicle based on the scene context feature. In the experiments, we use an urban scene driving simulator named CARLA to train and test the proposed method. The results show that the proposed method outperforms existing autonomous driving methods.
|
|
10:30-10:40, Paper TuA-5.4 | |
Sem-Aug: Improving Camera-LiDAR Feature Fusion with Semantic Augmentation for 3D Vehicle Detection |
|
Zhao, Lin | School of Automation, Beijing Institute of Technology |
Wang, Meiling | Beijing Institute of Technology |
Yue, Yufeng | Beijing Institute of Technology |
Keywords: Computer Vision for Transportation, Intelligent Transportation Systems, Object Detection, Segmentation and Categorization
Abstract: Camera-LiDAR fusion provides precise distance measurements and fine-grained textures, making it a promising option for 3D vehicle detection in autonomous driving scenarios. Previous camera-LiDAR based 3D vehicle detection approaches mainly focused on employing image-based pre-trained models to fetch semantic features. However, these methods may perform inferior to the LiDAR-based ones when lacking semantic segmentation labels in autonomous driving tasks. Motivated by this observation, we propose a novel semantic augmentation method, namely Sem-Aug, to guide high-confidence camera-LiDAR fusion feature generation and boost the performance of multimodal 3D vehicle detection. The key novelty of semantic augmentation lies in the 2D segmentation mask auto-labeling, which provides supervision for semantic segmentation sub-network to mitigate the poor generalization performance of camera-LiDAR fusion. Using semantic-augmentation-guided camera-LiDAR fusion features, Sem-Aug achieves remarkable performance on the representative autonomous driving KITTI dataset compared to both the LiDAR-based baseline and previous multimodal 3D vehicle detectors. Qualitative and quantitative experiments demonstrate that Sem-Aug provides significant improvements in challenging Hard detection scenarios caused by occlusion and truncation.
|
|
10:40-10:50, Paper TuA-5.5 | |
P2EG: Prediction and Planning Integrated Robust Decision-Making for Automated Vehicle Negotiating in Narrow Lane with Explorative Game |
|
Zhang, Qianyi | Nankai University |
Li, Xiao | Meituan Group |
He, Ethan | Meituan Group |
Ding, Shuguang | Meituan Group |
Wang, Naizheng | Meituan Group |
Liu, Jingtai | Nankai University |
Keywords: Autonomous Vehicle Navigation, Planning under Uncertainty, Motion and Path Planning
Abstract: In the narrow lane scene of autonomous driving, it is critical for the ego car to recognize the intentions of social vehicles and cooperate with them. However, cooperating with social vehicles is challenging due to insufficient information. This paper proposes an Explorative Game that adopts Participant Game and Perfect Bayesian Equilibrium to exploratively perform some aggressive actions to obtain additional information, thus the autonomous vehicle can cooperate robustly and efficiently. Explorative Game assumes each vehicle maintains a unique belief about the current situation and attributes insecurity and instability to the conflict of various Perfect Bayesian Equilibriums formed by various beliefs. Aggressive actions enable the ego car to proactively guide social vehicles to cooperate as it expects and encourage them to express their intentions as quickly and clearly as possible so that the equilibriums can converge and the conflict can be eliminated. Additional information reduces the error between the actual intentions of social vehicles and the estimated intentions from the ego car, helping rationally prune potential interactions and update parameters of the reward function. We demonstrate our algorithm on recorded data as well as virtual environments with manually controlled social vehicles to prove the efficiency of cooperation and the robustness of decision-making. And it has been running for more than 20 kilometers in the real world.
|
|
10:50-11:00, Paper TuA-5.6 | |
Visual Mapping and Localization System Based on Compact Instance-Level Road Markings with Spatial Uncertainty |
|
Zhou, Yuxuan | Wuhan University |
Li, Xingxing | Wuhan University |
Li, Shengyu | Wuhan University |
Wang, Xuanbin | Wuhan University |
Keywords: Autonomous Vehicle Navigation, Mapping, Localization
Abstract: High-definition (HD) map is crucial for intelligent vehicles to perform high-level localization and navigation. To improve the availability and usability of HD map, it is meaningful to investigate crowd-sourced mapping solutions and low-cost map-aided localization schemes which don't rely on high-end sensors. In this paper, we propose a novel vision-based mapping and localization system, which could generate compact instance-level road maps automatically and provide high-availability map-aided localization. The spatial uncertainties of the map elements are taken into consideration by analyzing the inverse perspective mapping (IPM) model, which enables more flexible map usages in both mapping and localization phases of the system. Besides, a pose graph optimization framework is developed for accurate pose estimation by fusing global positioning (GNSS), local navigation (odometry) and map matching information together. Real-world experiments in urban environment were conducted to validate different phases of the system, including on-vehicle mapping, multi-source map merging and map-aided localization.
|
|
11:00-11:10, Paper TuA-5.7 | |
Motion Planning for HyTAQs: A Topology-Guided Unified NMPC Approach |
|
Wu, Tong | Harbin Institute of Technology |
Zhu, Yimin | Harbin Institute of Technology |
Zhang, Lixian | Harbin Institute of Technology |
Yang, Jianan | Harbin Institute of Technology |
Ding, Yihang | Harbin Institute of Technology |
Keywords: Autonomous Vehicle Navigation, Motion and Path Planning, Optimization and Optimal Control
Abstract: In this study, a topology-guided unified nonlinear model predictive control (NMPC) approach is proposed for autonomous navigation of a class of Hybrid Terrestrial and Aerial Quadrotors (HyTAQs) in unknown environments. The approach can fully exploit the hybrid terrestrial-aerial locomotion of the vehicle and as such ensure a high navigation efficiency. A unified terrestrial-aerial NMPC is first formulated with a type of complementarity constraints involving the hybrid dynamics, together with the collision avoidance constraints for safety. Further, a topological roadmap with both terrestrial and aerial paths is leveraged to guide the kinodynamic path searching and thus the unified NMPC. Then, a complete and distinctive navigation framework is established and validated on our self-developed HyTAQ. Compared with the existing unified terrestrial-aerial planning methods, ours takes the vehicle dynamics into account for the first attempt and achieves a more reasonable decision of modes switching. Experimental results are presented to demonstrate the effectiveness and superiority of the proposed approach.
|
|
11:10-11:20, Paper TuA-5.8 | |
SEAN 2.0: Formalizing and Generating Social Situations for Robot Navigation |
|
Tsoi, Nathan | Yale University |
Xiang, Alec | Yale University |
Yu, Peter | Yale University |
Sohn, Samuel | Rutgers University |
Schwartz, Greg | Yale University |
Ramesh, Subashri | Yale University |
Hussein, Mohamed | Rutgers University-Camden |
Gupta, Anjali | Yale University |
Kapadia, Mubbasir | Rutgers University |
Vázquez, Marynel | Yale University |
Keywords: Software Tools for Benchmarking and Reproducibility, Social HRI, Modeling and Simulating Humans
Abstract: We present SEAN 2.0, an open-source system designed to advance social navigation via the training and benchmarking of navigation policies in varied social contexts. A key limitation of current social navigation research is that policies are often trained and evaluated considering only a few social contexts, which are fragmented across prior work. Inspired by work in psychology, we describe navigation context based on social situations, which encompass the robot task and environmental factors, and propose logic-based classifiers for five common examples. SEAN 2.0 allows a robot to experience these social situations via different methods for specifying and generating pedestrian motion, including a novel Behavior Graph method. Our experiments show that when data collected using the Behavior Graph method is used to learn a robot navigation policy, that policy outperforms others trained using alternative methods for pedestrian control. Also, social situations were found to be useful for understanding performance across social contexts. Other components of SEAN 2.0 include vision and depth sensors, several physical environments, different means of specifying robot tasks, and a range of evaluation metrics for social robot navigation. User feedback for SEAN 2.0 indicated that the system was "easier to navigate and more user friendly" than our prior work, SEAN 1.0.
|
|
TuA-6 |
Rm6 (Room D) |
SLAM 4 |
Regular session |
Chair: Mangelson, Joshua | Brigham Young University |
Co-Chair: Pomerleau, Francois | Université Laval |
|
10:00-10:10, Paper TuA-6.1 | |
Group-K Consistent Measurement Set Maximization for Robust Outlier Detection |
|
Forsgren, Brendon | Brigham Young University |
Vasudevan, Ram | University of Michigan |
Kaess, Michael | Carnegie Mellon University |
McLain, T.W. | Brigham Young University |
Mangelson, Joshua | Brigham Young University |
Keywords: SLAM, Range Sensing
Abstract: This paper presents a method for the robust selection of measurements in a simultaneous localization and mapping (SLAM) framework. Existing methods check consistency or compatibility on a pairwise basis, however many measurement types are not sufficiently constrained in a pairwise scenario to determine if either measurement is inconsistent with the other. This paper presents group-k consistency maximization (GkCM) that estimates the largest set of measurements that is internally group-k consistent. Solving for the largest set of group-k consistent measurements can be formulated as an instance of the maximum clique problem on generalized graphs and can be solved by adapting current methods. This paper evaluates the performance of GkCM using simulated data and compares it to pairwise consistency maximization (PCM) presented in previous work.
|
|
10:10-10:20, Paper TuA-6.2 | |
Floorplan-Aware Camera Poses Refinement |
|
Sokolova, Anna | Samsung AI Center, Moscow |
Nikitin, Filipp | Samsung Research |
Vorontsova, Anna | Samsung AI Center, Moscow |
Konushin, Anton | Samsung AI Center Moscow |
Keywords: SLAM, Mapping, Semantic Scene Understanding
Abstract: Processing large indoor scenes is a challenging task, as scan registration and camera trajectory estimation methods accumulate errors across time. As a result, the quality of reconstructed scans is insufficient for some applications, such as visual-based localization and navigation, where the correct position of walls is crucial. For many indoor scenes, there exists an image of a technical floorplan that contains information about the geometry and main structural elements of the scene, such as walls, partitions, and doors. We argue that such a floorplan is a useful source of spatial information, which can guide a 3D model optimization. The standard RGB-D 3D reconstruction pipeline consists of a tracking module applied to an RGB-D sequence and a bundle adjustment (BA) module that takes the posed RGB-D sequence and corrects the camera poses to improve consistency. We propose a novel optimization algorithm expanding conventional BA that leverages the prior knowledge about the scene structure in the form of a floorplan. Our experiments on the Redwood dataset and our self-captured data demonstrate that utilizing floorplan improves accuracy of 3D reconstructions.
|
|
10:20-10:30, Paper TuA-6.3 | |
MOTSLAM: MOT-Assisted Monocular Dynamic SLAM Using Single-View Depth Estimation |
|
Zhang, Hanwei | Kyushu University |
Uchiyama, Hideaki | Nara Institute of Science and Technology |
Ono, Shintaro | The University of Tokyo |
Kawasaki, Hiroshi | Kyushu University |
Keywords: SLAM, Visual Tracking, Deep Learning Methods
Abstract: Visual SLAM systems targeting static scenes have been developed with satisfactory accuracy and robustness. Dynamic 3D object tracking has then become a significant capability in visual SLAM with the requirement of understanding dynamic surroundings in various scenarios including autonomous driving, augmented and virtual reality. However, performing dynamic SLAM solely with monocular images remains a challenging problem due to the difficulty of associating dynamic features and estimating their positions. In this paper, we present MOTSLAM, a dynamic visual SLAM system with the monocular configuration that tracks both poses and bounding boxes of dynamic objects. MOTSLAM first performs multiple object tracking (MOT) with associated both 2D and 3D bounding box detection to create initial 3D objects. Then, neural-network-based monocular depth estimation is applied to fetch the depth of dynamic features. Finally, camera poses, object poses, and both static, as well as dynamic map points, are jointly optimized using a novel bundle adjustment. Our experiments on the KITTI dataset demonstrate that our system has reached best performance on both camera ego-motion and object tracking on monocular dynamic SLAM.
|
|
10:30-10:40, Paper TuA-6.4 | |
Gravity-Constrained Point Cloud Registration |
|
Kubelka, Vladimir | Örebro University |
Vaidis, Maxime | Université Laval |
Pomerleau, Francois | Université Laval |
Keywords: SLAM, Sensor Fusion, Range Sensing
Abstract: Visual and lidar Simultaneous Localization and Mapping (SLAM) algorithms benefit from the Inertial Measurement Unit (IMU) modality. The high-rate inertial data complement the other lower-rate modalities. Moreover, in the absence of constant acceleration, the gravity vector makes two attitude angles out of three observable in the global coordinate frame. In visual odometry, this is already being used to reduce the 6-Degrees Of Freedom (DOF) pose estimation problem to 4-DOF. In lidar SLAM, the gravity measurements are often used as a penalty in the back-end global map optimization to prevent map deformations. In this work, we propose an Iterative Closest Point (ICP)-based front-end which exploits the observable DOF and provides pose estimates aligned with the gravity vector. We believe that this front-end has the potential to support the loop closure identification, thus speeding up convergences of global map optimizations. The presented approach has been extensively tested against accurate ground-truth localization in large-scale outdoor environments as well as in the Subterranean Challenge organized by Defense Advanced Research Projects Agency (DARPA). We show that it can reduce the localization drift by 30% when compared to the standard 6-DOF ICP. Moreover, the code is readily available to the community as a part of the libpointmatcher library.
|
|
10:40-10:50, Paper TuA-6.5 | |
When Geometry Is Not Enough: Using Reflector Markers in Lidar SLAM |
|
Kurz, Gerhard | Robert Bosch GmbH |
Scherer, Sebastian Andreas | Robert Bosch GmbH |
Biber, Peter | Robert Bosch GmbH |
Fleer, David | Bosch Rexroth AG |
Keywords: SLAM, Mapping, Localization
Abstract: Lidar-based SLAM systems perform well in a wide range of circumstances by relying on the geometry of the environment. However, even mature and reliable approaches struggle when the environment contains structureless areas such as long hallways. To allow the use of lidar-based SLAM in such environments, we propose to add reflector markers in specific locations that would otherwise be difficult. We present an algorithm to reliably detect these markers and two approaches to fuse the detected markers with geometry-based scan matching. The performance of the proposed methods is demonstrated on real-world datasets from several industrial environments.
|
|
10:50-11:00, Paper TuA-6.6 | |
LOCUS 2.0: Robust and Computationally Efficient LiDAR Odometry for Real-Time 3D Mapping |
|
Reinke, Andrzej | University of Bonn |
Palieri, Matteo | Polytechnic University of Bari |
Morrell, Benjamin | Jet Propulsion Laboratory, California Institute of Technology |
Chang, Yun | MIT |
Ebadi, Kamak | NASA Jet Propulsion Laboratory |
Carlone, Luca | Massachusetts Institute of Technology |
Agha-mohammadi, Ali-akbar | NASA-JPL, Caltech |
Keywords: SLAM, Sensor Fusion, Data Sets for SLAM
Abstract: LiDAR odometry has attracted considerable attention as a robust localization method for autonomous robots operating in complex GNSS-denied environments. However, achieving reliable and efficient performance on heterogeneous platforms in large-scale environments remains an open challenge due to the limitations of onboard computation and memory resources needed for autonomous operation. In this work, we present LOCUS 2.0, a robust and computationally-efficient LiDAR odometry system for real-time underground 3D mapping. LOCUS 2.0 includes a novel normals-based GICP formulation that reduces the computation time of point cloud alignment, an adaptive voxelization strategy that maintains the desired computation load regardless of the environment’s geometry, and a sliding-window map approach that bounds the memory consumption and management. The proposed approach is shown to be suitable to be deployed on heterogeneous robotic platforms involved in large-scale explorations under severe computation and memory constraints. We demonstrate LOCUS 2.0, a key element of the CoSTAR team’s entry in the DARPA Subterranean Challenge, across various underground scenarios. We release LOCUS 2.0 as an open-source library and also release a LiDAR-based odometry dataset in challenging and largescale underground environments. The dataset features a legged and wheeled platform in multiple environments including fog, dust, darkness, and geometrically degenerate surroundings with a total of 11 h of operations and 16 km of distance traveled.
|
|
11:00-11:10, Paper TuA-6.7 | |
The Hilti SLAM Challenge Dataset |
|
Helmberger, Michael | HILTI AG |
Morin, Kristian | HILTI AG |
Berner, Beda | Hilti AG |
Kumar, Nitish | ETH Zurich |
Cioffi, Giovanni | University of Zurich |
Scaramuzza, Davide | University of Zurich |
Keywords: SLAM, Sensor Fusion
Abstract: Research in Simultaneous Localization and Mapping (SLAM) has made outstanding progress over the past years. SLAM systems are nowadays transitioning from academic to real world applications. However, this transition has posed new demanding challenges in terms of accuracy and robustness. To develop new SLAM systems that can address these challenges, new datasets containing cutting-edge hardware and realistic scenarios are required. We propose the Hilti SLAM Challenge Dataset. Our dataset contains indoor sequences of offices, labs, and construction environments and outdoor sequences of construction sites and parking areas. All these sequences are characterized by featureless areas and varying illumination conditions that are typical in real-world scenarios and pose great challenges to SLAM algorithms that have been developed in confined lab environments. Accurate sparse ground truth, at millimeter level, is provided for each sequence. The sensor platform used to record the data includes a number of visual, lidar, and inertial sensors, which are spatially and temporally calibrated. The purpose of this dataset is to foster the research in sensor fusion to develop SLAM algorithms that can be deployed in tasks where high accuracy and robustness are required, e.g., in construction environments. Many academic and industrial groups tested their SLAM systems on the proposed dataset in the Hilti SLAM Challenge. The results of the challenge, which are summarized in this paper, show that the proposed dataset is an important asset in the development of new SLAM algorithms that are ready to be deployed in the real-world.
|
|
11:10-11:20, Paper TuA-6.8 | |
Photometric Single-View Dense 3D Reconstruction in Endoscopy |
|
Batlle, Victor M. | I3A. Universidad De Zaragoza |
Montiel, J.M.M | I3A. Universidad De Zaragoza |
Tardos, Juan D. | Universidad De Zaragoza |
Keywords: SLAM, Computer Vision for Medical Robotics
Abstract: Visual SLAM inside the human body will open the way to computer-assisted navigation in endoscopy. However, due to space limitations, medical endoscopes only provide monocular images, leading to systems lacking true scale. In this paper we exploit the controlled lighting in colonoscopy to achieve the first in-vivo 3D reconstruction of the human colon using photometric stereo on a calibrated monocular endoscope. Our method works in a real medical environment, providing both a suitable in-place calibration procedure and a depth estimation technique adapted to the colon's tubular geometry. We validate our method on simulated colonoscopies, obtaining a mean error of 7% in depth estimation, which is below 3 mm on average. Our qualitative results on the EndoMapper dataset show that the method is able to correctly estimate the colon shape in real human colonoscopies, paving the ground for true-scale monocular SLAM in endoscopy.
|
|
11:20-11:30, Paper TuA-6.9 | |
Continuous-Time Stereo-Inertial Odometry |
|
Hug, David | ETH Zurich |
Bänninger, Philipp | ETH Zurich |
Alzugaray, Ignacio | ETH Zürich |
Chli, Margarita | ETH Zurich |
Keywords: SLAM, Sensor Fusion
Abstract: The emerging paradigm of Continuous-Time Simultaneous Localization And Mapping (CTSLAM) has become a competitive alternative to conventional discrete-time approaches in recent times and holds the additional promise of fusing multi-modal sensor setups in a truly generic manner, rendering its importance to robotic navigation and manipulation seminal. In this spirit, this work expands upon continuous-time concepts, evaluates their suitability in common stereo and stereo-inertial online configurations, and provides an extensible, generic, robust, and modular open-source implementation to the community. The presented experimental analysis records the performance of our approach in these setups against the state-of-the-art in discrete-time Simultaneous Localization And Mapping (SLAM) on established datasets, achieving competitive results, and provides a direct comparison between online discrete- and continuous-time approaches for the first time. Targeting the absence of open-sourced, continuous-time pipelines and their associated, oftentimes prohibitive, initial developmental overhead, our implementation is made public.
|
|
TuA-7 |
Rm7 (Room E) |
Virtual Reality and Interfaces |
Regular session |
Chair: Kovac, Mirko | Imperial College London |
Co-Chair: Fukui, Rui | The University of Tokyo |
|
10:00-10:10, Paper TuA-7.1 | |
Immersive View and Interface Design for Teleoperated Aerial Manipulation |
|
Kocer, Basaran Bahadir | Imperial College London |
Stedman, Harvey | University College London |
Caves, Izaak | Imperial College London |
Van Zalk, Nejra | Imperial College London |
Pawar, Vijay Manohar | University College London |
Kovac, Mirko | Imperial College London |
Keywords: Virtual Reality and Interfaces, Telerobotics and Teleoperation, Human-Centered Robotics
Abstract: The recent momentum in aerial manipulation has led to an interest in developing virtual reality interfaces for aerial physical interaction tasks with simple, intuitive, and reliable control and perception. However, this requires the use of expensive subsystems and there is still a research gap between interface design, user evaluations and the effect on aerial manipulation tasks. Here, we present a methodology for low-cost available drone systems with a Unity-based interface for immersive FPV teleoperation. We applied our approach in a flight track where a cluttered environment is used to simulate a demanding aerial manipulation task inspired by forestry drones and canopy sampling. Through objective measures of teleoperation performance and subjective questionnaires, we found that operators performed worse using the FPV interface and had higher perceived levels of cognitive load when compared to traditional interface design. Additional analysis of physiological measures highlighted that objective stress levels and cognitive load were also influenced by task duration and perceived performance, providing an insight into what interfaces could target to support teleoperator requirements during aerial manipulation tasks.
|
|
10:10-10:20, Paper TuA-7.2 | |
WFH-VR: Teleoperating a Robot Arm to Set a Dining Table across the Globe Via Virtual Reality |
|
Yim, Lai Sum | National Yang Ming Chiao Tung University |
Vo, Quang | George Mason University |
Huang, Ching-I | National Chiao Tung University |
Wang, Chu-Ruei | National Yang Ming Chiao Tung University |
McQueary, Wren | George Mason University |
Wang, Hsueh-Cheng | National Yang Ming Chiao Tung University, Taiwan |
Huang, Haikun | George Mason University |
Yu, Lap-Fai | George Mason University |
Keywords: Virtual Reality and Interfaces, Telerobotics and Teleoperation, Physical Human-Robot Interaction
Abstract: This paper presents an easy-to-deploy, virtual reality-based teleoperation system for controlling a robot arm. The proposed system is based on a consumer-grade virtual reality device (Oculus Quest 2) with a low-cost robot arm (a LoCoBot) to allow easy replication and set up. The proposed Work-from-Home Virtual Reality (WFH-VR) system allows the user to feel an intimate connection with the real remote robot arm. Virtual representations of the robot and objects to be manipulated in the real-world are presented in VR by streaming data pertaining to orientation and poses. The user studies suggest that 1) the proposed telerobotic system is effective under conditions both with and without network latency, whereas a method that simply streams video does not. This design enables the system implemented at an arbitrary distance from the actual work site. 2) The proposed system allows novices to perform manipulation tasks requiring higher dexterity than traditional keyboard controls can support, such as setting tableware. All results, hardware settings, and questionnaire feedback can be obtained at https://arg-nctu.github.io/projects/vr-robot-arm.html.
|
|
10:20-10:30, Paper TuA-7.3 | |
A Deep Learning Technique As a Sensor Fusion for Enhancing the Position in a Virtual Reality Micro-Environment |
|
Prieto Prada, John | DGIST |
Luna, Miguel | DGIST |
Park, Sang Hyun | DGIST |
Song, Cheol | DGIST |
Keywords: Virtual Reality and Interfaces, Sensor Fusion, AI-Based Methods
Abstract: Most virtual reality (VR) applications use a commercial controller for interaction. However, a typical virtual reality controller (VRC) lacks positional precision and accuracy in millimeter-scale scenarios. This lack of precision and accuracy is caused by built-in sensors’ drift. Therefore, the tracking performance of a VRC needs to be enhanced for millimeter-scale scenarios. Herein, we introduce a novel way of enhancing the tracking performance of a commercial VRC in a millimeter-scale environment using a deep learning (DL) algorithm. Specifically, we use a long short-term memory (LSTM) model trained with data collected from a linear motor, an IMU sensor, and a VRC. We integrate the virtual environment developed in Unity software with the LSTM model running in Python. We designed three experimental conditions: the VRC, Kalman filter (KF), and LSTM modes. Furthermore, we evaluate tracking performances in the three conditions and two other experimental scenarios, namely stationary and dynamic. In the stationary experimental scenario, the system is left motionless for 10 s. By contrast, in the dynamic experimental scenarios, the linear stage moves the system by 12 mm along the X, Y, and Z axes. The experimental results indicate that the deep learning model outperforms the standard controller’s positional performance by 85.69 % and 92.14 % in static and dynamic situations, respectively.
|
|
10:30-10:40, Paper TuA-7.4 | |
A Wearable Multi-Joint Wrist Contour Measuring Device for Hand Shape Recognition |
|
Mitani, Tatsuro | The University of Tokyo |
Okishiba, Shunsuke | The University of Tokyo |
Tateyama, Naoki | The University of Tokyo |
Yamanojo, Koshi | The University of Tokyo |
Warisawa, Shin'ichi | The University of Tokyo |
Fukui, Rui | The University of Tokyo |
Keywords: Virtual Reality and Interfaces, Sensor Fusion
Abstract: Recently, various types of hand shape recognition systems have been developed for human-machine interfaces. However, most wearable recognition systems cannot robustly handle the variations in attachment positions of the devices. Thus, we propose a hand shape recognition system using a wearable multi-joint wrist contour measuring device to realize robust and effective recognition of hand shape, regardless of attachment position variations. In particular, this device can measure the wrist contour and band flexion data to recognize hand shape. The wrist contour data are measured using photo-reflectors mounted inside the device, and the band flexion data are measured using photo-interrupters installed at the device joints. Additionally, the attachment position information is extracted from the wrist contour or band flexion data using dimensional compression or attachment position recognition to achieve robustness against the position variations. Subsequently, the extracted information is incorporated into the hand shape recognizer. The results of the recognition experiment demonstrate that the attachment position information extracted from the band flexion data using dimensional compression could effectively realize robust hand shape recognition, considering variations in the device attachment position.
|
|
10:40-10:50, Paper TuA-7.5 | |
A Torque Controlled Approach for Virtual Remote Centre of Motion Implementation |
|
Minelli, Marco | University of Modena and Reggio Emilia |
Secchi, Cristian | Univ. of Modena & Reggio Emilia |
Keywords: Motion Control, Medical Robots and Systems
Abstract: In this paper, we propose a novel torque controller for the implementation virtual remote center of motion. The controller allows the system to implement the required behavior and guarantees the satisfaction of the remote center of motion constraint. Exploiting the Udwadia-Kalaba equation for constrained dynamic systems, the controller is synthesized considering the dynamic effect the constraint produces on the manipulator, achieving more effective control with respect to kinematic strategies, and allowing the implementation of compliance behaviors. Simulations and experimental validation with a KUKA LWR 4+ with 7 degrees of freedom has been performed to check the performances of the proposed controller. Results show the effectiveness of the proposed controller with different control action, and the capability to interact with the environment by implementing compliant motion control.
|
|
10:50-11:00, Paper TuA-7.6 | |
Semi-Automatic Infrared Calibration for Augmented Reality Systems in Surgery |
|
Iqbal, Hisham | Imperial College London |
Rodriguez y Baena, Ferdinando | Imperial College, London, UK |
Keywords: Virtual Reality and Interfaces, Human Performance Augmentation, Computer Vision for Medical Robotics
Abstract: Augmented reality (AR) has the potential to improve the immersion and efficiency of computer-assisted orthopaedic surgery (CAOS) by allowing surgeons to maintain focus on the operating site rather than external displays in the operating theatre. Successful deployment of AR to CAOS requires a calibration that can accurately calculate the spatial relationship between real and holographic objects. Several studies attempt this calibration through manual alignment or with additional fiducial markers in the surgical scene. We propose a calibration system that offers a direct method for the calibration of AR head-mounted displays (HMDs) with CAOS systems, by using infrared-reflective marker-arrays widely used in CAOS. In our fast, user-agnostic setup, a HoloLens 2 detected the pose of marker arrays using infrared response and time-of-flight depth obtained through sensors onboard the HMD. Registration with a commercially available CAOS system was achieved when an IR marker-array was visible to both devices. Study tests found relative-tracking mean errors of 2.03 mm and 1.12 degrees when calculating the relative pose between two static marker-arrays at short ranges. When using the calibration result to provide in-situ holographic guidance for a simulated wire-insertion task, a pre-clinical test reported mean errors of 2.07 mm and 1.54 degrees when compared to a pre-planned trajectory.
|
|
11:00-11:10, Paper TuA-7.7 | |
Detecting Touch and Grasp Gestures Using a Wrist-Worn Optical and Inertial Sensing Network |
|
Cofer, Savannah | Stanford University |
Chen, Tyler | Stanford University |
Yang, Jackie (Junrui) | Stanford University |
Follmer, Sean | Stanford University |
Keywords: Virtual Reality and Interfaces, Human Detection and Tracking, Touch in HRI
Abstract: Freehand gesture based interaction promises to enable rich interaction in applications such as augmented reality (AR), virtual reality (VR), Human-Robot Interaction (HRI), and Robotic Prosthetic devices. However, current sensing approaches are limited; camera-based solutions are constrained by optical occlusion, and devices that interpret muscle activity are unreliable. This work presents a novel wrist-worn sensing device that combines near-infrared (NIR) sensing through 20 active LED-photodiode pairs and 6 DOF inertial measurement unit (IMU) sensing to enable high-accuracy detection of surface touch and grasp interactions for applications in AR and robotic prosthetic devices. Two convolutional neural networks are used to map device inputs to detect touch events, and subsequently classify them by gesture type and direction. We evaluate the accuracy and temporal precision of our system for event detection and classification. Results from an in-lab user study of 12 participants show an average of 97% touch detection accuracy and 98% grasp detection accuracy.
|
|
11:10-11:20, Paper TuA-7.8 | |
Towards Reproducible Evaluations for Flying Drone Controllers in Virtual Environments |
|
Li, Zheng | Fudan University |
Huang, Yiming | HKUST |
Yau, Yui-Pan | The Hong Kong University of Science and Technology |
Hui, Pan | Hong Kong University of Science and Technology |
Lee, Lik-Hang | KAIST |
Keywords: Virtual Reality and Interfaces, Force and Tactile Sensing, Design and Human Factors
Abstract: Research attention on natural user interfaces (NUIs) for drone flights are rising. Nevertheless, NUIs are highly diversified, and primarily evaluated by different physical environments leading to hard-to-compare performance between such solutions. We propose a virtual environment, namely VRFlightSim, enabling comparative evaluations with enriched drone flight details to address this issue. We first replicated a state-of-the-art (SOTA) interface and designed two tasks (crossing and pointing) in our virtual environment. Then, two user studies with 13 participants demonstrate the necessity of VRFlightSim and further highlight the potential of open-data interface designs.
|
|
11:20-11:30, Paper TuA-7.9 | |
Learning with Yourself: A Tangible Twin Robot System to Promote STEM Education |
|
Gao, Jiasi | Tsinghua University |
Gong, Jiangtao | Lenovo Research |
Zhou, Guyue | Tsinghua University |
Guo, Haole | Tsinghua University |
Qi, Tong | Tsinghua University |
Keywords: Education Robotics, Physical Human-Robot Interaction, Design and Human Factors
Abstract: This paper presents a customized programmable robotic system, TanTwin (Tangible Twin), designed to promote STEM education for K-12 children. Firstly, TanTwin is implemented based on a wheel-robot with standard LEGO bricks. With several deep neural networks, a child can convert a captured portrait of himself/herself into standard LEGO bricks, therefore he/she can build a tangible twin robot of himself/herself automatically. Besides, to adapt to the customized appearance, the corresponding visual element and content of the robotic system were also changed by a rule-based adaption algorithm. To demonstrate the effectiveness of TanTwin and to investigate whether tangible twin robots could contribute to children’s learning, we conducted a controlled experimental study to compare learning with a TanTwin and with a standard robot system through measuring students’ cognitive learning outcomes. The pre-/post- knowledge test results indicated that learning with a tangible twin robot leads to significantly better learning outcomes. Given the results, we validate our system and customization technology can promote STEM education.
|
|
TuA-8 |
Rm8 (Room F) |
Tendon Driven Mechanisms |
Regular session |
Chair: Tahara, Kenji | Kyushu University |
Co-Chair: Hawkes, Elliot Wright | University of California, Santa Barbara |
|
10:00-10:10, Paper TuA-8.1 | |
Kinematics-Inertial Fusion for Localization of a 4-Cable Underactuated Suspended Robot Considering Cable Sag |
|
Allak, Eren | University of Klagenfurt |
Khorrambakht, Rooholla | K. N. Toosi University of Technology |
Brommer, Christian | University of Klagenfurt |
Weiss, Stephan | Universität Klagenfurt |
Keywords: Tendon/Wire Mechanism, Sensor Fusion, Parallel Robots
Abstract: Suspended Cable-Driven Parallel Robots (SCDPR) have intriguing capabilities on large scales but still have open challenges in precisely estimating the end-effector pose. The cables exhibit a downward curved shape, also known as cable sag which needs to be accounted for in the pose estimation. The catenary equations can accurately describe this phenomenon but are only accurate in equilibrium conditions. Thus, pose estimation for large-scale SCDPR in dynamic motion is an open challenge. This work proposes a real-time pose estimation algorithm for dynamic trajectories of SCDPRs, which is accurate over large areas. We present a novel approach that considers cable sag to reduce the estimation error for large scales while also employing an Inertial Measurement Unit (IMU) to improve estimation accuracy for dynamic motion. Our approach reduces the RMSE to less than a third compared to standard methods not considering cable sag. Similarly, the inclusion of the IMU reduces the RMSE in dynamic situations by 40% compared to non-IMU aided approaches considering cable sag. Furthermore, we evaluate our Extended Kalman Filter (EKF) based algorithm on a real system with ground truth pose information.
|
|
10:10-10:20, Paper TuA-8.2 | |
End-Point Stiffness and Joint Viscosity Control of Musculoskeletal Robotic Arm Using Muscle Redundancy |
|
Tsuboi, Shoki | Kyushu University |
Kino, Hitoshi | Chukyo University |
Tahara, Kenji | Kyushu University |
Keywords: Tendon/Wire Mechanism, Redundant Robots, Compliance and Impedance Control
Abstract: This study focuses on replicating the musculoskeletal system of human arms for mimicking its movement. Muscle redundancy is critical for regulating the mechanical impedance of arms and legs. However, when implementing muscle redundancy on robots, making an ill-posed problem that cannot determine the muscle forces uniquely. In this paper, first, a method for controlling end-point stiffness in the muscle space for the joint and muscle redundant system is described. Next, the muscle model imitating the nonlinear viscosity characteristic of human muscles is introduced. Then, a method to control the joint viscosity by adjusting the internal forces of muscles adequately without affecting the stiffness control directly is proposed. Finally, numerical simulations are performed to investigate the effectiveness of the proposed method.
|
|
10:20-10:30, Paper TuA-8.3 | |
Data-Driven Kinematic Control Scheme for Cable-Driven Parallel Robots Allowing Collisions |
|
Zou, Yongwei | Harbin Institute of Technology, Shenzhen |
Hu, Yusheng | Harbin Institute of Technology, Shenzhen |
Cao, Huanhui | Harbin Institute of Technology |
Xu, Yuchen | Harbin Institute of Technology(Shenzhen) |
Yu, Yuanjie | Harbin Institute of Technology, Shenzhen |
Lu, Wenjie | Harbin Institute of Technology (Shenzhen) |
Xiong, Hao | Harbin Institute of Technology, Shenzhen |
Keywords: Tendon/Wire Mechanism, Parallel Robots
Abstract: Cable-Driven Parallel Robots (CDPRs) have been proposed for a variety of applications such as material handling, rehabilitation, and instrumentation. However, the collision-free constraint of CDPRs limits the workspace of CDPRs and the feasible position of anchor points. To address the collision-free constraint of CDPRs, a data-driven kinematic control scheme is developed for CDPRs, enabling a CDPR to control its pose even if suffering collisions between a cable and the base or the end-effector. To deal with the collisions, the data-driven kinematic control scheme utilizes a motion model obtained based on data samples of the motion of the CDPR, rather than the Jacobian matrix of the CDPR, to map a control law in the task space to the time derivative of the length of cables in the joint space. To evaluate the effectiveness of the developed data-driven kinematic control scheme, experiments of controlling a suspended CDPR with two cables allowing collisions are conducted.
|
|
10:30-10:40, Paper TuA-8.4 | |
Miniature, Lightweight, High-Force, Capstan Winch for Mobile Robots |
|
Heap, William | University of California, Santa Barbara |
Keeley, Christopher | University of California, Santa Barbara |
Yao, Elvy | University of California, Santa Barbara |
Naclerio, Nicholas | University of California, Santa Barbara |
Hawkes, Elliot Wright | University of California, Santa Barbara |
Keywords: Tendon/Wire Mechanism, Mechanism Design, Actuation and Joint Mechanisms
Abstract: Actuators that apply tension forces are widely applicable in robotics. In many applications of tensile actuators, a large stroke length, high force, and small, light device are important. For these requirements, the best current solution is a winch, which uses a rotating shaft to pull lightweight cable. However, most winches accumulate cable in a spool on their shaft which limits maximum stroke length and force at a miniature scale. An alternative is a capstan winch, in which the cable wraps around the shaft in a single-layered spiral before passing off the shaft. Although high-force and high- stroke versions exist, miniaturization has not been successfully demonstrated. We present the design, modeling, and characterization of a miniaturized capstan winch. The 16 g winch is capable of lifting 4.5 kg (280x body weight) a distance of 4.3 m (67x body length). We also demonstrate it actuating a jumping robot and pulling a remote-controlled car out of a ditch. Through its miniature design and high-force, high-stroke performance, our winch expands the potential capabilities of small-scale robots.
|
|
10:40-10:50, Paper TuA-8.5 | |
RAMIEL: A Parallel-Wire Driven Monopedal Robot for High and Continuous Jumping |
|
Suzuki, Temma | The University of Tokyo |
Toshimitsu, Yasunori | University of Tokyo |
Nagamatsu, Yuya | The University of Tokyo |
Kawaharazuka, Kento | The University of Tokyo |
Miki, Akihiro | The University of Tokyo |
Ribayashi, Yoshimoto | JSK Robotics Lab |
Bando, Masahiro | The University of Tokyo |
Kojima, Kunio | The University of Tokyo |
Kakiuchi, Yohei | Toyohashi University of Technology |
Okada, Kei | The University of Tokyo |
Inaba, Masayuki | The University of Tokyo |
Keywords: Tendon/Wire Mechanism, Legged Robots, Mechanism Design
Abstract: Legged robots with high locomotive performance have been extensively studied, and various leg structures have been proposed. Especially, a leg structure that can achieve both continuous and high jumps is advantageous for moving around in a three-dimensional environment. In this study, we propose a parallel wire-driven leg structure, which has one DoF of linear motion and two DoFs of rotation and is controlled by six wires, as a structure that can achieve both continuous jumping and high jumping. The proposed structure can simultaneously achieve high controllability on each DoF, long acceleration distance and high power required for jumping. In order to verify the jumping performance of the parallel wire-driven leg structure, we have developed a parallel wire-driven monopedal robot, RAMIEL. RAMIEL is equipped with quasi-direct drive, high power wire winding mechanisms and a lightweight leg, and can achieve a maximum jumping height of 1.6 m and a maximum of seven continuous jumps.
|
|
10:50-11:00, Paper TuA-8.6 | |
Workspace-Based Model Predictive Control for Cable-Driven Robots (I) |
|
Song, Chen | The Chinese University of Hong Kong |
Lau, Darwin | The Chinese University of Hong Kong |
Keywords: Optimization and Optimal Control, Motion Control, Tendon/Wire Mechanism
Abstract: The control of cable-driven robots is challenging due to the system’s non-linearity, actuation redundancy and the unilaterally bounded actuation constraints. To solve this problem, a workspace-based model predictive control (W-MPC) scheme is proposed which combines the online model predictive control with offline workspace analysis. Using the workspace, a set of convex constraints can be generated for a given reference trajectory. This can then be used to formulate a convex optimization problem for the online W-MPC. Meanwhile strict recursive feasibility and stability are obtained by taking advantage of the predictive feature of MPC. To demonstrate the effectiveness of the proposed W-MPC, simulation was performed on a 2-link planar cabled-riven robot and a spatial cable-driven parallel robot for both nominal and non-nominal scenarios. Hardware experiment was also carried out using a 3 degree-of-freedom planar cable robot. The results show that the controller is efficient and effective to motion tracking with the cable force constraints satisfied despite the existence of various model uncertainties.
|
|
11:00-11:10, Paper TuA-8.7 | |
Dexterity Analysis and Motion Optimization of In-Situ Torsionally-Steerable Flexible Surgical Robots |
|
Kong, Yixuan | Harbin Institute of Technology, Shenzhen |
Wang, Jiaole | Harbin Institute of Technology, Shenzhen |
Zhang, Ning | Harbin Institute of Technology, Shenzhen |
Song, Shuang | Harbin Institute of Technology (Shenzhen) |
Li, Bing | Harbin Institute of Technology (Shenzhen) |
Keywords: Tendon/Wire Mechanism, Flexible Robotics, Medical Robots and Systems
Abstract: Flexible robots with in-situ torsion can be used in laryngeal endoscopic surgery which can maintain the position and approach vector of the end-effector during the operation. However, the inherent errors would be produced by in-situ torsional motion which are different due to the various configuration of serpentine module in robot. In this paper, the kinematics model is established according to the structure of serpentine module. The dexterity analysis shows that the singular position is reduced and the angular velocity of dexterity is improved comparing with the robot without in-situ torsion function. The theoretical position errors caused by in-situ torsion is quantitatively analyzed by simulation. It is found that the maximum error is 5.19mm at the bending angle of 120°.In addition, the existence of joints in the robot arm also leads to the occurrence of rotation errors. The configuration and number of the joints are optimized to improve the accuracy. Finally, the experiments are carried out to verify the effectiveness of the proposed design and model. The results indicate that the flexible surgical robot have higher motion dexterity. And the inherent error during the in-situ torsion motion can be eliminated by structural optimization.
|
|
TuA-9 |
Rm9 (Room G) |
Planning under Uncertainty |
Regular session |
Chair: Zhang, Yunzhou | Northeastern University |
Co-Chair: Bezzo, Nicola | University of Virginia |
|
10:00-10:10, Paper TuA-9.1 | |
Meta-Learning-Based Proactive Online Planning for UAVs under Degraded Conditions |
|
Gao, Shijie | University of Virginia |
Yel, Esen | Stanford University |
Bezzo, Nicola | University of Virginia |
Keywords: Planning under Uncertainty, Failure Detection and Recovery, Aerial Systems: Applications
Abstract: Changes in model dynamics due to factors like actuator faults, platform aging, and unexpected disturbances can challenge an autonomous robot during real-world operations affecting its intended behavior and safety. Under such circumstances, it becomes critical to improve tracking performance, predict future states of the system, and replan to maintain safety and liveness conditions. In this letter, we propose a meta-learning-based framework to learn a model to predict the future system's states and their uncertainties under unforeseen and untrained conditions. Meta-learning is considered for this problem thanks to its ability to easily adapt to new tasks with a few data points gathered at runtime. We use the predictions from the meta-learned model to detect unsafe situations and proactively replan the system's trajectory when an unsafe situation is detected (e.g., a collision with an object). The proposed framework is applied and validated with both simulations and experiments on a faulty UAV performing an infrastructure inspection mission, demonstrating safety improvements.
|
|
10:10-10:20, Paper TuA-9.2 | |
Path-Tree Optimization in Discrete Partially Observable Environments Using Rapidly-Exploring Belief-Space Graphs |
|
Phiquepal, Camille | University of Stuttgart |
Orthey, Andreas | TU Berlin |
Toussaint, Marc | TU Berlin |
Keywords: Planning under Uncertainty, Motion and Path Planning, Mobile Manipulation
Abstract: Robots often need to solve path planning problems where essential and discrete aspects of the environment are partially observable. This introduces a multi-modality, where the robot must be able to observe and infer the state of its environment. To tackle this problem, we introduce the Path- Tree Optimization (PTO) algorithm which plans a path-tree in belief-space. A path-tree is a tree-like motion with branching points where the robot receives an observation leading to a belief-state update. The robot takes different branches depend- ing on the observation received. The algorithm has three main steps. First, a rapidly-exploring random graph (RRG) on the state space is grown. Second, the RRG is expanded to a belief- space graph by querying the observation model. In a third step, dynamic programming is performed on the belief-space graph to extract a path-tree. The resulting path-tree combines exploration with exploitation i.e. it balances the need for gaining knowledge about the environment with the need for reaching the goal. We demonstrate the algorithm capabilities on navigation and mobile manipulation tasks, and show its advantage over a baseline using a task and motion planning approach (TAMP) both in terms of optimality and runtime.
|
|
10:20-10:30, Paper TuA-9.3 | |
Object-Aware SLAM Based on Efficient Quadric Initialization and Joint Data Association |
|
Cao, Zhenzhong | Northeastern University |
Zhang, Yunzhou | Northeastern University |
Tian, Rui | Northeastern University |
Ma, Rong | Beijing Simulation Center |
Hu, Xinggang | Northeastern University |
Coleman, Sonya | University of Ulster |
Kerr, Dermot | University of Ulster |
Keywords: SLAM, Localization, Mapping
Abstract: Semantic simultaneous localization and mapping (SLAM) is a popular technology enabling indoor mobile robots to sufficiently perceive and interact with the environment. In this paper, we propose an object-aware semantic SLAM system, which consists of a quadric initialization method, an object-level data association method, and a multi-constraint optimization factor graph. To overcome the limitation of multi-view observations and the requirement of dense point clouds for objects, an efficient quadric initialization method based on object detection and surfel construction is proposed, which can efficiently initialize quadrics within fewer frames and with small viewing angles. The robust object-level joint data association method and the tightly coupled multi-constraint factor graph for quadrics optimization and joint bundle adjustment enable the accurate estimation of constructed quadrics and camera poses. Extensive experiments using public datasets show that the proposed system achieves competitive performance with respect to accuracy and robustness of object quadric estimation and camera localization compared with stateof-the-art methods.
|
|
10:30-10:40, Paper TuA-9.4 | |
Accelerated Reinforcement Learning for Temporal Logic Control Objectives |
|
Kantaros, Yiannis | Washington University in St. Louis |
Keywords: Planning under Uncertainty, Reinforcement Learning, Integrated Planning and Learning
Abstract: This paper addresses the problem of learning control policies for mobile robots, modeled as unknown Markov Decision Processes (MDPs), that are tasked with temporal logic missions, such as sequencing, coverage, or surveillance. The MDP captures uncertainty in the workspace structure and the outcomes of control decisions. The control objective is to synthesize a control policy that maximizes the probability of accomplishing a high-level task, specified as a Linear Temporal Logic (LTL) formula. To address this problem, we propose a novel accelerated model-based reinforcement learning (RL) algorithm for LTL control objectives that is capable of learning control policies significantly faster than related approaches. Its sample-efficiency relies on biasing exploration towards directions that may contribute to task satisfaction. This is accomplished by leveraging an automaton representation of the LTL task as well as a continuously learned MDP model. Finally, we provide comparative experiments that demonstrate the sample efficiency of the proposed method against recent RL methods for LTL objectives.
|
|
10:40-10:50, Paper TuA-9.5 | |
Smooth Model Predictive Path Integral Control without Smoothing |
|
Kim, Taekyung | Agency for Defense Development |
Park, Gyuhyun | Agency for Defense Development(ADD) |
Kwak, Kiho | Agency for Defense Development |
Bae, Jihwan | Agency for Defense Development |
Lee, Wonsuk | Agency for Defense Development |
Keywords: Planning under Uncertainty, Optimization and Optimal Control, Model Learning for Control
Abstract: We present a sampling-based control approach that can generate smooth actions for general nonlinear systems without external smoothing algorithms. Model Predictive Path Integral (MPPI) control has been utilized in numerous robotic applications due to its appealing characteristics to solve non-convex optimization problems. However, the stochastic nature of sampling-based methods can cause significant chattering in the resulting commands. Chattering becomes more prominent in cases where the environment changes rapidly, possibly even causing the MPPI to diverge. To address this issue, we propose a method that seamlessly combines MPPI with an input-lifting strategy. In addition, we introduce a new action cost to smooth control sequence during trajectory rollouts while preserving the information theoretic interpretation of MPPI, which was derived from non-affine dynamics. We validate our method in two nonlinear control tasks with neural network dynamics: a pendulum swing-up task and a challenging autonomous driving task. The experimental results demonstrate that our method outperforms the MPPI baselines with additionally applied smoothing algorithms.
|
|
10:50-11:00, Paper TuA-9.6 | |
Monte-Carlo Robot Path Planning |
|
Dam, Tuan | Technische Universität Darmastadt |
Chalvatzaki, Georgia | Technische Universität Darmastadt, Intelligent Robotic Systems |
Peters, Jan | Technische Universität Darmstadt |
Pajarinen, Joni | Aalto University |
Keywords: Planning under Uncertainty, Motion and Path Planning, Planning, Scheduling and Coordination
Abstract: Path planning is a crucial algorithmic approach for designing robot behaviors. Sampling-based approaches, like rapidly exploring random trees (RRTs) or probabilistic roadmaps, are prominent algorithmic solutions for path planning problems. Despite its exponential convergence rate, RRT can only find suboptimal paths. On the other hand, RRT*, a widely-used extension to RRT, guarantees probabilistic completeness for finding optimal paths but suffers in practice from slow convergence in complex environments. Furthermore, real-world robotic environments are often partially observable or with poorly described dynamics, casting the application of RRT* in complex tasks suboptimal. This paper studies a novel algorithmic formulation of the popular Monte-Carlo tree search (MCTS) algorithm for robot path planning. Notably, we study Monte-Carlo Path Planning (MCPP) by analyzing and proving, on the one part, its exponential convergence rate to the optimal path in fully observable Markov decision processes (MDPs), and on the other part, its probabilistic completeness for finding feasible paths in partially observable MDPs (POMDP) assuming limited distance observability (proof sketch). Our algorithmic contribution allows us to employ recently proposed variants of MCTS with different exploration strategies for robot path planning. Experimental evaluation in simulated 2D and 3D environments with a 7 degrees of freedom (DOF) manipulator, as well as in a real-world robot path planning task, demonstrate the superiority of MCPP in POMDP tasks.
|
|
11:00-11:10, Paper TuA-9.7 | |
Qualitative Belief Space Planning Via Compositions |
|
Zilberman, Itai | Technion |
Indelman, Vadim | Technion - Israel Institute of Technology |
Keywords: Planning under Uncertainty
Abstract: Planning under uncertainty is a fundamental problem in robotics. Classical approaches rely on a metrical representation of the world and robot's states to infer the next course of action. While these approaches are considered accurate, they are often susceptible to metric errors and tend to be costly regarding memory and time consumption. However, in some cases, relying on qualitative geometric information alone is sufficient. Hence, the issues described above become an unnecessary burden. This work presents a novel qualitative Belief Space Planning (BSP) approach, highly suitable for platforms with low-cost sensors and particularly appealing in sparse environment scenarios. Our algorithm generalizes its predecessors by avoiding any deterministic assumptions. Moreover, it smoothly incorporates spatial information propagation techniques, known as compositions. We demonstrate our algorithm in simulations and the advantage of using compositions in particular.
|
|
11:10-11:20, Paper TuA-9.8 | |
Task and Motion Informed Trees (TMIT*): Almost-Surely Asymptotically Optimal Integrated Task and Motion Planning |
|
Thomason, William | Rice University |
Strub, Marlin Polo | NASA Jet Propulsion Laboratory |
Gammell, Jonathan | University of Oxford |
Keywords: Task and Motion Planning, Motion and Path Planning, Manipulation Planning
Abstract: High-level autonomy requires discrete and continuous reasoning to decide both what actions to take and how to execute them. Integrated Task and Motion Planning (TMP) algorithms solve these hybrid problems jointly to consider constraints between the discrete symbolic actions (i.e., the task plan) and their continuous geometric realization (i.e., motion plans). This joint approach solves more difficult problems than approaches that address the task and motion subproblems independently. TMP algorithms combine and extend results from both task and motion planning. TMP has mainly focused on computational performance and completeness and less on solution optimality. Optimal TMP is difficult because the independent optima of the subproblems may not be the optimal integrated solution, which can only be found by jointly optimizing both plans. This paper presents Task and Motion Informed Trees (TMIT*), an optimal TMP algorithm that combines results from makespan-optimal task planning and almost-surely asymptotically optimal motion planning. TMIT* interleaves asymmetric forward and reverse searches to delay computationally expensive operations until necessary and perform an efficient informed search directly in the problem’s hybrid state space. This allows it to solve problems quickly and then converge towards the optimal solution with additional computational time, as demonstrated on the evaluated robotic-manipulation benchmark problems.
|
|
11:20-11:30, Paper TuA-9.9 | |
Generalizable Task Planning through Representation Pretraining |
|
Wang, Chen | Stanford University |
Xu, Danfei | Stanford Univesity |
Fei-Fei, Li | Stanford University |
Keywords: Integrated Planning and Learning, Task and Motion Planning, Representation Learning
Abstract: The ability to plan for multi-step manipulation tasks in unseen situations is crucial for future home robots. But collecting sufficient experience data for end-to-end learning is often infeasible in the real world, as deploying robots in many environments can be prohibitively expensive. On the other hand, large-scale scene understanding datasets contain diverse and rich semantic and geometric information. But how to leverage such information for manipulation remains an open problem. In this paper, we propose a learning-to-plan method that can generalize to new object instances by leveraging object-level representations extracted from a synthetic scene understanding dataset. We evaluate our method with a suite of challenging multi-step manipulation tasks inspired by household activities and show that our model achieves measurably better success rate than state-of-the-art end-to-end approaches.
|
|
TuA-10 |
Rm10 (Room H) |
Robot Safety |
Regular session |
Chair: Elara, Mohan Rajesh | Singapore University of Technology and Design |
Co-Chair: Biswas, Swagata | TCS Research |
|
10:00-10:10, Paper TuA-10.1 | |
An Analytical Study of Motion of Autonomous Vehicles under Imperfect Sensing |
|
Biswas, Swagata | TCS Research |
Paul, Himadri Sekhar | TCS Research |
Bagchi, Saurabh | Purdue University |
Keywords: Robot Safety, Autonomous Vehicle Navigation, Sensor-based Control
Abstract: A fully tested autonomous system works predictably under ideal or assumed environment. However, its behavior is not fully defined when some components malfunction or fail. In this paper, we consider automated guided vehicle (AGV), equipped with multiple sensors, executing a traversal task in a static unknown environment. We have analytically studied the system, computed a set of performance and safety metrics, and validated it with simulation results in Webots. We have also analyzed the effect on system performance under independent and correlated sensing errors. We have also performed sensitivity analysis to identify the most critical components in any given system; and this can be utilized to increase the reliability of the system and its conformance to safety objectives.
|
|
10:10-10:20, Paper TuA-10.2 | |
Adhesion Risk Assessment of an Aircraft Inspection Robot for Improving Operator Awareness |
|
Muthugala Arachchige, Viraj Jagathpriya Muthugala | Singapore University of Technology and Design |
Vega-Heredia, Manuel | Singapore University of Technology and Design & Universidad Autó |
Nay, Htet Lin | Department of Engineering Product Development, Singapore Univers |
Samarakoon Mudiyanselage, Bhagya Prasangi Samarakoon | Singapore University of Technology and Design |
Elara, Mohan Rajesh | Singapore University of Technology and Design |
Keywords: Robot Safety, Human-Robot Collaboration, Service Robotics
Abstract: Vacuum-adhesion-based climbing robots have been developed to cater to the demands in the cleaning and inspection work of airplanes. A robot intended to clean and inspect an airplane faces a Risk of Adhesion (RoA) based on the robot and the surface conditions, such as worn-out suction cups. These sorts of underlying conditions are not easily noticeable for an operator of a robot and might lead to catastrophic events. Therefore, the ability of a robot to self-assess the RoA in a scenario and notify the operator is crucial for ensuring safety. Particularly, an aircraft inspection robot should have adhesion awareness. This paper proposes a novel method to self-assess the RoA of a vacuum-adhesion-based robot intended to clean and inspect airplanes. The RoA is assessed by a fuzzy inference system that analyzes the present pressure difference and the current duty setting of the vacuum pump of a robot. The robot operator can collaborate with the robot to take precautions based on the assessed RoA to ensure safety. The outcomes of the experiment conducted on an airplane skin validate the ability of the proposed method to assess the RoA associated with heterogeneous operating conditions effectively. Thus, the utilization of the proposed method would improve the safety of a vacuum-adhesion-based robot intended to clean and inspect airplanes.
|
|
10:20-10:30, Paper TuA-10.3 | |
How Do We Fail? Stress Testing Perception in Autonomous Vehicles |
|
Delecki, Harrison | Stanford University |
Itkina, Masha | Stanford University |
Lange, Bernard | Stanford University |
Senanayake, Ransalu | Stanford University |
Kochenderfer, Mykel | Stanford University |
Keywords: Robot Safety, Intelligent Transportation Systems
Abstract: Autonomous vehicles (AVs) rely on environment perception and behavior prediction to reason about agents in their surroundings. These perception systems must be robust to adverse weather such as rain, fog, and snow. However, validation of these systems is challenging due to their complexity and dependence on observation histories. This paper presents a method for characterizing failures of LiDAR-based perception systems for AVs in adverse weather conditions. We develop a methodology based in reinforcement learning to find likely failures in object tracking and trajectory prediction due to sequences of disturbances. We apply disturbances using a physics-based data augmentation technique for simulating LiDAR point clouds in adverse weather conditions. Experiments performed across a wide range of driving scenarios from a real-world driving dataset show that our proposed approach finds high likelihood failures with smaller input disturbances compared to baselines while remaining computationally tractable. Identified failures can inform future development of robust perception systems for AVs.
|
|
10:30-10:40, Paper TuA-10.4 | |
HiddenGems: Efficient Safety Boundary Detection with Active Learning |
|
Petrov, Aleksandar | University of Oxford |
Fang, Carter | ETH Zurich |
Pham, Minh Khang | Motional |
Eng, You Hong | Future Urban Mobility, Singapore-MIT Alliance for Research and T |
Fu, James Guo Ming | NuTonomy |
Pendleton, Scott Drew | Motional |
Keywords: Robot Safety, Methods and Tools for Robot System Design, Intelligent Transportation Systems
Abstract: Evaluating safety performance in a resource-efficient way is crucial for the development of autonomous systems. Simulation of parameterized scenarios is a popular testing strategy but parameter sweeps can be prohibitively expensive. To address this, we propose HiddenGems: a sample-efficient method for discovering the boundary between compliant and non-compliant behavior via active learning. Given a parameterized scenario, one or more compliance metrics, and a simulation oracle, HiddenGems maps the compliant and non-compliant domains of the scenario. The methodology enables critical test case identification, comparative analysis of different versions of the system under test, as well as verification of design objectives. We evaluate HiddenGems on a scenario with a jaywalker crossing in front of an autonomous vehicle and obtain compliance boundary estimates for collision, lane keep, and acceleration metrics individually and in combination, with 6 times fewer simulations than a parameter sweep. We also show how HiddenGems can be used to detect and rectify a failure mode for an unprotected turn with 86% fewer simulations.
|
|
10:40-10:50, Paper TuA-10.5 | |
Enpheeph: A Fault Injection Framework for Spiking and Compressed Deep Neural Networks |
|
Colucci, Alessio | TU Wien |
Steininger, Andreas | TU Wien |
Shafique, Muhammad | New York University Abu Dhabi |
Keywords: Robot Safety, Object Detection, Segmentation and Categorization, Deep Learning Methods
Abstract: Research on Deep Neural Networks (DNNs) has focused on improving performance and accuracy for real-world deployments, leading to new models, such as Spiking Neural Networks (SNNs), and optimization techniques, e.g., quantization and pruning for compressed networks. However, the deployment of these innovative models and optimization techniques introduces possible reliability issues, which is a pillar for DNNs to be widely used in safety-critical applications, e.g., autonomous driving. Moreover, scaling technology nodes have the associated risk of multiple faults happening at the same time, a possibility not addressed in state-of-the-art resiliency analyses. Towards better reliability analysis for DNNs, we present enpheeph, a Fault Injection Framework for Spiking and Compressed DNNs. The enpheeph framework enables optimized execution on specialized hardware devices, e.g., GPUs, while providing complete customizability to investigate different fault models, emulating various reliability constraints and use-cases. Hence, the faults can be executed on SNNs as well as compressed networks with minimal-to-none modifications to the underlying code, a feat that is not achievable by other state-of-the-art tools. To evaluate our enpheeph framework, we analyze the resiliency of different DNN and SNN models, with different compression techniques. By injecting a random and increasing number of faults, we show that DNNs can show a reduction in accuracy with a fault rate as low as 7 x 10 ^ (-7) faults per parameter, with an accuracy drop higher than 40%. Run-time overhead when executing enpheeph is less than 20% of the baseline execution time when executing 100 000 faults concurrently, at least 10x lower than state-of-the-art frameworks, making enpheeph future-proof for complex fault injection scenarios. We release the source code of our enpheeph framework under an open-source license at https://github.com/Alexei95/enpheeph.
|
|
10:50-11:00, Paper TuA-10.6 | |
Continuous Safety Control of a Mobile Robot in Cluttered Environments |
|
Wu, Si | Northeastern University, China |
Liu, Tengfei | Northeastern University, China |
Niu, Qinglin | Northeastern University, China |
Jiang, Zhong-Ping | New York University |
Keywords: Robot Safety, Optimization and Optimal Control
Abstract: This paper studies the safety control problem for mobile robots working in cluttered environments. A compact set is employed to represent the obstacles, and a direction-distance function is used to describe the obstacle-measurement model. The major contribution is a nontrivial modification of the quadratic programming (QP) approach for continuous safety control of integrator-modeled mobile robots. In particular, a refinement of the Moreau-Yosida method is proposed to regularize the measurement model while retaining feasibility and safety. The second contribution is the development of a new feasible set shaping technique with a positive basis for a QP-based continuous safety controller. Physical experiments are employed to verify the proposed method.
|
|
11:00-11:10, Paper TuA-10.7 | |
Dependability Analysis of Deep Reinforcement Learning Based Robotics and Autonomous Systems through Probabilistic Model Checking |
|
Dong, Yi | University of Liverpool |
Zhao, Xingyu | University of Liverpool |
Huang, Xiaowei | University of Liverpool |
Keywords: Robot Safety, Probability and Statistical Methods, Reinforcement Learning
Abstract: While Deep Reinforcement Learning (DRL) provides transformational capabilities to the control of Robotics and Autonomous Systems (RAS), the black-box nature of DRL and uncertain deployment environments of RAS pose new challenges on its dependability. Although existing works impose constraints on the DRL policy to ensure successful completion of the mission, it is far from adequate to assess the DRL-driven RAS in a holistic way considering all dependability properties. In this paper, we formally define a set of dependability properties in temporal logic and construct a Discrete-Time Markov Chain (DTMC) to model the dynamics of risk/failures of a DRL-driven RAS interacting with the stochastic environment. We then conduct Probabilistic Model Checking (PMC) on the designed DTMC to verify those properties. Our experimental results show that the proposed method is effective as a holistic assessment framework while uncovering conflicts between the properties that may need trade-offs in training. Moreover, we find that the standard DRL training cannot improve dependability properties, thus requiring bespoke optimisation objectives. Finally, our method offers sensitivity analysis of dependability properties to disturbance levels from environments, providing insights for the assurance of real RAS.
|
|
11:10-11:20, Paper TuA-10.8 | |
On Safety Testing, Validation, and Characterization with Scenario-Sampling: A Case Study of Legged Robots |
|
Weng, Bowen | The Ohio State University |
Castillo, Guillermo | The Ohio State University |
Zhang, Wei | Southern University of Science and Technology |
Hereid, Ayonga | Ohio State University |
Keywords: Robot Safety, Performance Evaluation and Benchmarking, Legged Robots
Abstract: The dynamic response of the legged robot locomotion is non-Lipschitz and can be stochastic due to environmental uncertainties. To test, validate, and characterize the safety performance of legged robots, existing solutions on observed and inferred risk can be incomplete and sampling inefficient. Some formal verification methods suffer from the model precision and other surrogate assumptions. In this paper, we propose a scenario sampling based testing framework that characterizes the overall safety performance of a legged robot by specifying (i) where (in terms of a set of states) the robot is potentially safe, and (ii) how safe the robot is within the specified set. The framework can also help certify the commercial deployment of the legged robot in real-world environment along with human and compare safety performance among legged robots with different mechanical structures and dynamic properties. The proposed framework is further deployed to evaluate a group of state-of-the-art legged robot locomotion controllers from various model-based, deep neural network involved, and reinforcement learning based methods in the literature. Among a series of intended work domains of the studied legged robots (e.g. tracking speed on sloped surface, with abrupt changes on demanded velocity, and against adversarial push-over disturbances), we show that the method can adequately capture the overall safety characterization and the subtle performance insights. Many of the observed safety outcomes, to the best of our knowledge, have never been reported by the existing work in the legged robot literature.
|
|
11:20-11:30, Paper TuA-10.9 | |
Semi-Perspective Decoupled Heatmaps for 3D Robot Pose Estimation from Depth Maps |
|
Simoni, Alessandro | University of Modena and Reggio Emilia |
Pini, Stefano | University of Modena and Reggio Emilia |
Borghi, Guido | University of Bologna |
Vezzani, Roberto | University of Modena and Reggio Emilia |
Keywords: Gesture, Posture and Facial Expressions, Computer Vision for Automation, RGB-D Perception
Abstract: Knowing the exact 3D location of workers and robots in a collaborative environment enables several real applications, such as the detection of unsafe situations or the study of mutual interactions for statistical and social purposes. In this paper, we propose a non-invasive and light-invariant framework based on depth devices and deep neural networks to estimate the 3D pose of robots from an external camera. The method can be applied to any robot without requiring hardware access to the internal states. We introduce a novel representation of the predicted pose, namely Semi-Perspective Decoupled Heatmaps (SPDH), to accurately compute 3D joint locations in world coordinates adapting efficient deep networks designed for the 2D Human Pose Estimation. The proposed approach, which takes as input a depth representation based on XYZ coordinates, can be trained on synthetic depth data and applied to real-world settings without the need for domain adaptation techniques. To this end, we present the SimBa dataset, based on both synthetic and real depth images, and use it for the experimental evaluation. Results show that the proposed approach, made of a specific depth map representation and the SPDH, overcomes the current state of the art.
|
|
TuA-11 |
Rm11 (Room I) |
Human and Humanoid Motion Analysis and Synthesis |
Regular session |
Chair: Asfour, Tamim | Karlsruhe Institute of Technology (KIT) |
Co-Chair: Tanaka, Takayuki | Hokkaido University |
|
10:00-10:10, Paper TuA-11.1 | |
Understanding Spatio-Temporal Relations in Human-Object Interaction Using Pyramid Graph Convolutional Network |
|
Xing, Hao | Technical University of Munich (TUM) |
Burschka, Darius | Technische Universitaet Muenchen |
Keywords: Human and Humanoid Motion Analysis and Synthesis, Human Detection and Tracking, Learning from Demonstration
Abstract: Human activities recognition is an important task for an intelligent robot, especially in the field of human-robot collaboration, it requires not only the label of sub-activities but also the temporal structure of the activity. In order to automatically recognize both the label and the temporal structure in sequence of human-object interaction, we propose a novel Pyramid Graph Convolutional Network (PGCN), which employs a pyramidal encoder-decoder architecture consisting of an attention based graph convolution network and a temporal pyramid pooling module for downsampling and upsampling interaction sequence on the temporal axis, respectively. The system represents the 2D or 3D spatial relation of human and objects from the detection results in video data as a graph. To learn the human-object relations, a new attention graph convolutional network is trained to extract condensed information from the graph representation. To segment action into sub-actions, a novel temporal pyramid pooling module is proposed, which upsamples compressed features back to the original time scale and classifies actions per frame. We explore various attention layers, namely spatial attention, temporal attention and channel attention, and combine different upsampling decoders to test the performance on action recognition and segmentation. We evaluate our model on two challenging datasets in the field of human-object interaction recognition, i.e. Bimanual Actions and IKEA Assembly datasets. We demonstrate that our classifier significantly improves both framewise action recognition and segmentation, e.g., F1 micro and F1@50 scores on Bimanual Actions dataset are improved by 4.3% and 8.5% respectively.
|
|
10:10-10:20, Paper TuA-11.2 | |
Gastrocnemius and Power Amplifier Soleus Spring-Tendons Achieve Fast Human-Like Walking in a Bipedal Robot |
|
Kiss, Bernadett | Max Planck Institute for Intelligent Systems |
Gonen, Emre Cemal | Max Planck Institute for Intelligent Systems |
Mo, An | MPI IS Stuttgart |
Buchmann, Alexandra | Technical University of Munich |
Renjewski, Daniel | Technische Universität München |
Badri-Spröwitz, Alexander | Max Planck Institute for Intelligent Systems |
Keywords: Humanoid and Bipedal Locomotion, Humanoid Robot Systems, Legged Robots
Abstract: Legged locomotion in humans is governed by natural dynamics of the human body and neural control. One mechanism that is assumed to contribute to the high efficiency of human walking is the impulsive ankle push-off, which potentially powers the swing leg catapult. However, the mechanics of the human lower leg with its complex muscle-tendon units spanning over single and multiple joints is not yet understood. Legged robots allow testing the interaction between complex leg mechanics, control, and environment in real-world walking gait. We developed a 0.49 m tall, 2.2 kg anthropomorphic bipedal robot with Soleus and Gastrocnemius muscle-tendon units represented by linear springs, acting as mono- and biarticular elastic structures around the robot's ankle and knee joints. We tested the influence of three Soleus and Gastrocnemius spring-tendon configurations on the ankle power curves, the coordination of the ankle and knee joint movements, the total cost of transport, and walking speed. We controlled the robot with a feed-forward central pattern generator, leading to walking speeds between 0.35 m/s and 0.57 m/s at 1.0 Hz locomotion frequency, at 0.35 m leg length. We found differences between all three configurations; the Soleus spring-tendon modulates the robot's speed and energy efficiency likely by ankle power amplification, while the Gastrocnemius spring-tendon changes the movement coordination between ankle and knee joints during push-off.
|
|
10:20-10:30, Paper TuA-11.3 | |
A Riemannian Take on Human Motion Analysis and Retargeting |
|
Klein, Holger | Karlsruhe Institute of Technology |
Jaquier, Noémie | Karlsruhe Institute of Technology |
Meixner, Andre | Karlsruhe Institute of Technology (KIT) |
Asfour, Tamim | Karlsruhe Institute of Technology (KIT) |
Keywords: Human and Humanoid Motion Analysis and Synthesis, Dynamics
Abstract: Dynamic motions of humans and robots are widely driven by posture-dependent nonlinear interactions between their degrees of freedom. However, these dynamical effects remain mostly overlooked when studying the mechanisms of human movement generation. Inspired by recent works, we hypothesize that human motions are planned as sequences of geodesic synergies, and thus correspond to coordinated joint movements achieved with piecewise minimum energy. The underlying computational model is built on Riemannian geometry to account for the inertial characteristics of the body. Through the analysis of various human arm motions, we find that our model segments motions into geodesic synergies, and successfully predicts observed arm postures, hand trajectories, as well as their respective velocity profiles. Moreover, we show that our analysis can further be exploited to transfer arm motions to robots by reproducing individual human synergies as geodesic paths in the robot configuration space.
|
|
10:30-10:40, Paper TuA-11.4 | |
Human-To-Robot Manipulability Domain Adaptation with Parallel Transport and Manifold-Aware ICP |
|
Reithmeir, Anna | Technical University of Munich |
Figueredo, Luis Felipe Cruz | Technical University of Munich (TUM) |
Haddadin, Sami | Technical University of Munich |
Keywords: Human and Humanoid Motion Analysis and Synthesis, Learning from Demonstration, Human-Robot Collaboration
Abstract: Manipulability ellipsoids efficiently capture the human pose and reveal information about the task at hand. Their use in task-dependent robot teaching – particularly their transfer from a teacher to a learner – can advance emulation of human-like motion. Although in recent literature focus is shifted towards manipulability transfer between two robots, the adaptation to the capabilities of the other kinematic system is to date not addressed and research in transfer from human to robot is still in its infancy. This work presents a novel manipulability domain adaptation method for the transfer of manipulability information to the domain of another kinematic system. As manipulability matrices/ellipsoids are symmetric positive-definite (SPD) they can be viewed as points on the Riemannian manifold of SPD matrices. We are the first to address the problem of manipulability transfer from the perspective of point cloud registration. We propose a manifold-aware Iterative Closest Point algorithm (ICP) with parallel transport initialization. Furthermore, we introduce a correspondence matching heuristic for manipulability ellipsoids based on inherent geometric features. We confirm our method in simulation experiments with 2-DoF manipulators as well as 7-DoF models representing the human-arm kinematics.
|
|
10:40-10:50, Paper TuA-11.5 | |
From Human Walking to Bipedal Robot Locomotion: Reflex Inspired Compensation on Planned and Unplanned Downsteps |
|
Verhagen, Joris Petrus Martinus | Delft University of Technology |
Xiong, Xiaobin | California Institute of Technology |
Ames, Aaron | Caltech |
Seth, Ajay | Delft University of Technology |
Keywords: Humanoid and Bipedal Locomotion, Biomimetics, Legged Robots
Abstract: Humans are able to negotiate downstep behaviors---both planned and unplanned---with remarkable agility and ease. The goal of this paper is to systematically study the translation of this human behavior to bipedal walking robots, even if the morphology is inherently different. Concretely, we begin with human data wherein planned and unplanned downsteps are taken. We analyze this data from the perspective of reduced-order modelling of the human, encoding the center of mass (CoM) kinematics and contact forces, which allows for the translation of these behaviors into the corresponding reduced-order model of a bipedal robot. We embed the resulting behaviors into the full-order dynamics of a bipedal robot via nonlinear optimization-based controllers. The end result is the demonstration of planned and unplanned downsteps in simulation on an underactuated walking robot.
|
|
10:50-11:00, Paper TuA-11.6 | |
End-To-End from Human Hand Synergies to Robot Hand Tendon Routing |
|
Hidalgo Carvajal, Diego Xavier | Technical University of Munich |
Herneth, Christopher | Technical University Munich |
Naceri, Abdeldjallil | TUM |
Haddadin, Sami | Technical University of Munich |
Keywords: Grasping, Multifingered Hands, Deep Learning in Grasping and Manipulation
Abstract: The human hand capabilities are paramount for highly dexterous manipulation interactions. Unfortunately, the limitations of current technologies make replicating such capabilities unfeasible. Although several works have focused on directly attempting to create robot hands able to mimic human ones closely, few of them have attempted to create generalizable platforms, where robotic hand mechanisms can be iteratively selected and customized to different tasks. In order to build highly dexterous robotic hands in the future, it is crucial to understand not only human manipulation, but also develop methods to leverage robotic mechanisms limitations to mimic human hand interactions accurately. In this letter, we propose an end-to-end framework capable of generating underactuated tendon routings that allow a generic robot hand model to reproduce desired observed human grasp motion synergies accurately. Our contributions are threefold: (1) an end to end framework to generate task-oriented robot hand tendon routings, with the potential to implement desired synergies, (2) a novel grammar based representation of robot hand tendon routings, and (3) a schematic visualization of robot hand tendon routings. The latter two contributions have the potential to embed and compare properties among robot hands. Our results in simulation show that the proposed method produces tendon routing mechanisms that are able to closely mimic the joint trajectories of human subjects performing the same experimental tasks, while achieving dynamically stable grasping postures.
|
|
11:00-11:10, Paper TuA-11.7 | |
A Centaur System for Assisting Human Walking with Load Carriage |
|
Yang, Ping | Southern University of Science and Technology |
Yan, Haoyun | Southern University of Science and Technology |
Yang, Bowen | Southern University of Science and Technology |
Li, Jianquan | Department of Mechanical and Energy Engineering, Southern Univer |
Li, Kailin | Department of Mechanical and Energy Engineering, Southern Univer |
Leng, Yuquan | Southern University of Science and Technology |
Fu, Chenglong | Southern University of Science and Technology (SUSTech) |
Keywords: Human Performance Augmentation, Physically Assistive Devices, Human-Robot Collaboration
Abstract: Walking with load is a common task in daily life and disaster rescue. Long-term load carriage may cause irreversible damage to the human body. Although remarkable progress has been made in the field of wearable robots, it is still far from avoiding interference to human legs, which will lead to energy consumption. In this paper, a novel wearable robot, Centaur, for assisting load carriage has been proposed. The Centaur system consists of two rigid robotic legs of two degrees-of-freedom (DOFs) to transfer load weight to the ground. Different from exoskeletons, the robotic legs of the Centaur are placed behind the human rather than attached to human limbs, which can provide a larger support polygon and avoid additional interference to the wearer. Additionally, the Centaur can attain the locomotion stability of the quadruped while maintaining the motion agility of the biped itself. This paper also presents an interactive motion control strategy based on the human-robot interaction force. This control strategy incorporates legged robotics walking controller and real-time walking trajectory planning to realize the cooperative walking with human beings. Finally, experiments of human walking with load carriage have been conducted on flat terrain to verify the concept of the Centaur system. The result demonstrates that the Centaur system can effectively reduce 70.03% of load weight during the single stance phase, which indicates that the Centaur system provides a new solution for assisting human walking with load-carriage.
|
|
11:10-11:20, Paper TuA-11.8 | |
Construction of a Simulator to Reproduce the Changes of Running by Motion Strategy with Spring-Loaded Inverted Pendulum Model |
|
Kitagawa, Masaki | Hokkaido University |
Tanaka, Takayuki | Hokkaido University |
Murai, Akihiko | The National Institute of Advanced Industrial Science and Techno |
Keywords: Human Performance Augmentation, Modeling and Simulating Humans, Optimization and Optimal Control
Abstract: This study aims to construct a running simulator based on a motion generation and control system that enables the description of motion strategies using the spring-loaded inverted pendulum (SLIP) model. The problems of stability and robustness encountered in the running simulation with the SLIP model are elucidated, and stable running is achieved by controlling the stiffness and the attitude angle dynamically at touchdown, as well as human energy adjustment that is introduced to consider the active motion strategy. As a result, passive and active control by humans can be expressed, and a framework that can express the changes in running due to motion strategies is constructed. Finally, we discuss the possibility of describing and elucidating the motion strategies.
|
|
11:20-11:30, Paper TuA-11.9 | |
Koopman Pose Predictions for Temporally Consistent Human Walking Estimations |
|
Mitjans, Marc | Boston University |
Levine, David Michael | Brigham and Women's Hospital |
Awad, Louis | Harvard University |
Tron, Roberto | Boston University |
Keywords: Human Detection and Tracking, Sensor Fusion, RGB-D Perception
Abstract: We tackle the problem of tracking the human lower body as an initial step toward an automatic motion assessment system for clinical mobility evaluation, using a multimodal system that combines Inertial Measurement Unit (IMU) data, RGB images, and point cloud depth measurements. This system applies the factor graph representation to an optimization problem that provides 3-D skeleton joint estimations. In this paper, we focus on improving the temporal consistency of the estimated human trajectories to greatly extend the range of operability of the depth sensor. More specifically, we introduce a new factor graph factor based on Koopman theory that embeds the nonlinear dynamics of several lower-limb movement activities. This factor performs a two-step process: first, a custom activity recognition module based on spatial temporal graph convolutional networks recognizes the walking activity; then, a Koopman pose prediction of the subsequent skeleton is used as an a priori estimation to drive the optimization problem toward more consistent results. We tested the performance of this module on datasets composed of multiple clinical lower-limb mobility tests, and we show that our approach reduces outliers on the skeleton form by almost 1 m, while preserving natural walking trajectories at depths up to more than 10 m.
|
|
TuA-12 |
Rm12 (Room J) |
Vision |
Regular session |
Chair: Yamashita, Atsushi | The University of Tokyo |
Co-Chair: Wang, Xiaolong | UC San Diego |
|
10:00-10:10, Paper TuA-12.1 | |
LapSeg3D: Weakly Supervised Semantic Segmentation of Point Clouds Representing Laparoscopic Scenes |
|
Alt, Benjamin | ArtiMinds Robotics |
Kunz, Christian | Karlsruhe Institute of Technology |
Katic, Darko | Karlsruhe Institute for Technology (KIT) |
Younis, Rayan | Heidelberg University |
Jäkel, Rainer | Karlsruhe Institute of Technology |
Müller-Stich, Beat | University Hospital Heidelberg, Department of Surgery |
Wagner, Martin | Heidelberg University Hospital |
Mathis-Ullrich, Franziska | Karlsruhe Institute of Technology |
Keywords: Computer Vision for Medical Robotics, Surgical Robotics: Laparoscopy, RGB-D Perception
Abstract: The semantic segmentation of surgical scenes is a prerequisite for task automation in robot assisted interventions. We propose LapSeg3D, a novel DNN-based approach for the voxel-wise annotation of point clouds representing surgical scenes. As the manual annotation of training data is highly time consuming, we introduce a semi-autonomous clustering-based pipeline for the annotation of the gallbladder, which is used to generate segmented labels for the DNN. When evaluated against manually annotated data, LapSeg3D achieves an F1 score of 0.94 for gallbladder segmentation on various datasets of ex-vivo porcine livers. We show LapSeg3D to generalize accurately across different gallbladders and datasets recorded with different RGB-D camera systems.
|
|
10:10-10:20, Paper TuA-12.2 | |
Ego+X: An Egocentric Vision System for Global 3D Human Pose Estimation and Social Interaction Characterization |
|
Liu, Yuxuan | Shanghai Jiao Tong University |
Yang, Jianxin | Shanghai Jiao Tong University |
Gu, Xiao | Imperial College London |
Guo, Yao | Shanghai Jiao Tong University |
Yang, Guang-Zhong | Shanghai Jiao Tong University |
Keywords: Computer Vision for Medical Robotics, Social HRI, Human-Centered Robotics
Abstract: Egocentric vision is an emerging topic, which has demonstrated great potential in assistive healthcare scenarios, ranging from human-centric behavior analysis to personal social assistance. Within this field, due to the heterogeneity of visual perception from first-person views, egocentric pose estimation is one of the most significant prerequisites for enabling various downstream applications. However, existing methods for egocentric pose estimation mainly focus on predicting the pose represented in the camera coordinates from a single image, which ignores the latent cues in the temporal domain and results in less accuracy. In this paper, we propose Ego+X, an egocentric vision based system for 3D canonical pose estimation and human-centric social interaction characterization. Our system is composed of two head-mounted egocentric cameras, where one is faced downwards and the other looks outwards. By leveraging the global context provided by visual SLAM, we first propose Ego-Glo for spatial-accurate and temporal-consistent egocentric 3D pose estimation in the canonical coordinate system. With the help of an egocentric camera looking outwards, we then propose Ego-Soc by extending Ego-Glo to various social interaction tasks, e.g., object detection and human-human interaction. Quantitative and qualitative experiments have been conducted to demonstrate the effectiveness of our proposed Ego+X.
|
|
10:20-10:30, Paper TuA-12.3 | |
Tracking Monocular Camera Pose and Deformation for SLAM Inside the Human Body |
|
Gomez Rodriguez, Juan Jose | University of Zaragoza |
Montiel, J.M.M | I3A. Universidad De Zaragoza |
Tardos, Juan D. | Universidad De Zaragoza |
Keywords: Computer Vision for Medical Robotics, Localization
Abstract: Monocular SLAM in deformable scenes will open the way to multiple medical applications like computer- assisted navigation in endoscopy, automatic drug delivery or autonomous robotic surgery. In this paper we propose a novel method to simultaneously track the camera pose and the 3D scene deformation, without any assumption about environment topology or shape. The method uses an illumination-invariant photometric method to track image features and estimates camera motion and deformation combining reprojection error with spatial and temporal regularization of deformations. Our results in simulated colonoscopies show the method’s accuracy and robustness in complex scenes under increasing levels of deformation. Our qualitative results in human colonoscopies from Endomapper dataset show that the method is able to successfully cope with the challenges of real endoscopies: deformations, low texture and strong illumination changes. We also compare with previous tracking methods in simpler scenarios from Hamlyn dataset where we obtain competitive performance, without needing any topological assumption.
|
|
10:30-10:40, Paper TuA-12.4 | |
Markerless Suture Needle 6D Pose Tracking with Robust Uncertainty Estimation for Autonomous Minimally Invasive Robotic Surgery |
|
Chiu, Zih-Yun | University of California, San Diego |
Liao, Albert | University of California, San Diego |
Richter, Florian | University of California, San Diego |
Johnson, Bjorn | University of California, San Diego |
Yip, Michael C. | University of California, San Diego |
Keywords: Computer Vision for Medical Robotics, Visual Tracking, Medical Robots and Systems
Abstract: Suture needle localization is necessary for autonomous suturing. Previous approaches in autonomous suturing often relied on fiducial markers rather than markerless detection schemes for localizing a suture needle due to the inconsistency of markerless detections. However, fiducial markers are not practical for real-world applications and can often be occluded from environmental factors in surgery (e.g., blood). Therefore in this work, we present a robust tracking approach for estimating the 6D pose of a suture needle when using inconsistent detections. We define observation models based on suture needles' geometry that captures the uncertainty of the detections and fuse them temporally in a probabilistic fashion. In our experiments, we compare different permutations of the observation models in the suture needle localization task to show their effectiveness. Our proposed method outperforms previous approaches in localizing a suture needle. We also demonstrate the proposed tracking method in an autonomous suture needle regrasping task and ex vivo environments.
|
|
10:40-10:50, Paper TuA-12.5 | |
Monocular Depth Estimation for Equirectangular Videos |
|
Fraser, Helmi | Edinburgh Centre for Robotics, Heriot-Watt University |
Wang, Sen | Edinburgh Centre for Robotics, Heriot-Watt University |
Keywords: Omnidirectional Vision, RGB-D Perception
Abstract: Depth estimation from panoramic imagery has received minimal attention in contrast to standard perspective imagery, which constitutes the majority of the literature on the key research topic. The vast - and frequently complete - field of view provided by such panoramic photographs makes them appealing for a variety of applications, including robots, autonomous vehicles, and virtual reality. Consumer-level camera systems capable of capturing such images are likewise growing more affordable, and may be desirable complements to autonomous systems' sensor packages. They do, however, introduce significant distortions and violate some assumptions regarding perspective view images. Additionally, many state-of-the-art algorithms are not designed for its projection model, and their depth estimation performance tends to degrade when being applied to panoramic imagery. This paper presents a novel technique for adapting view synthesis-based depth estimation models to omnidirectional vision. Specifically, we: 1) integrate a ``virtual'' spherical camera model into the training pipeline, facilitating the model training, 2) exploit spherical convolutional layers to perform convolution operations on equirectangular images, handling the severe distortion, and 3) propose an optical flow-based masking scheme to mitigate the effect of unwanted pixels during training. Our qualitative and quantitative results demonstrate that these simple yet efficient designs result in significantly improved depth estimations when compared to previous approaches.
|
|
10:50-11:00, Paper TuA-12.6 | |
Visual Servoing with Geometrically Interpretable Neural Perception |
|
Paolillo, Antonio | IDSIA USI-SUPSI |
Nava, Mirko | IDSIA |
Piga, Dario | SUPSI-IDSIA |
Giusti, Alessandro | IDSIA Lugano, SUPSI |
Keywords: Visual Servoing, Deep Learning for Visual Perception
Abstract: An increasing number of nonspecialist robotic users demand easy-to-use machines. In the context of visual servoing, the removal of explicit image processing is becoming a trend, allowing an easy application of this technique. This work presents a deep learning approach for solving the perception problem within the visual servoing scheme. An artificial neural network is trained using the supervision coming from the knowledge of the controller and the visual features motion model. In this way, it is possible to give a geometrical interpretation to the estimated visual features, which can be used in the analytical law of the visual servoing. The approach keeps perception and control decoupled, conferring flexibility and interpretability on the whole framework. Simulated and real experiments with a robotic manipulator validate our approach.
|
|
11:00-11:10, Paper TuA-12.7 | |
Online Adaptation for Implicit Object Tracking and Shape Reconstruction in the Wild |
|
Ye, Jianglong | UC San Diego |
Chen, Yuntao | TuSimple |
Wang, Naiyan | TuSimple |
Wang, Xiaolong | UC San Diego |
Keywords: Visual Tracking, Deep Learning for Visual Perception
Abstract: Tracking and reconstructing 3D objects from cluttered scenes are the key components for computer vision, robotics and autonomous driving systems. While recent progress in implicit function has shown encouraging results on high-quality 3D shape reconstruction, it is still very challenging to generalize to cluttered and partially observable LiDAR data. In this paper, we propose to leverage the continuity in video data. We introduce a novel and unified framework which utilizes a neural implicit function to simultaneously track and reconstruct 3D objects in the wild. Our approach adapts the DeepSDF model (i.e., an instantiation of the implicit function) in the video online, iteratively improving the shape reconstruction while in return improving the tracking, and vice versa. We experiment with both Waymo and KITTI datasets, and show significant improvements over state-of-the-art methods for both tracking and shape reconstruction tasks.
|
|
11:10-11:20, Paper TuA-12.8 | |
DeepFusionMOT: A 3D Multi-Object Tracking Framework Based on Camera-LiDAR Fusion with Deep Association |
|
Wang, Xiyang | Chongqing University |
Fu, Chunyun | Chongqing University |
Li, Zhankun | Chongqing University |
Lai, Ying | Guangdong Songshan Polytechnic |
He, Jiawei | Chongqing University |
Keywords: Visual Tracking, Computer Vision for Transportation, Computer Vision for Automation
Abstract: In the recent literature, on the one hand, many 3D multi-object tracking (MOT) works have focused on tracking accuracy and neglected computation speed, commonly by designing rather complex cost functions and feature extractors. On the other hand, some methods have focused too much on computation speed at the expense of tracking accuracy. In view of these issues, this paper proposes a robust and fast camera-LiDAR fusion-based MOT method that achieves a good trade-off between accuracy and speed. Relying on the characteristics of camera and LiDAR sensors, an effective deep association mechanism is designed and embedded in the proposed MOT method. This association mechanism realizes tracking of an object in a 2D domain when the object is far away and only detected by the camera, and updating of the 2D trajectory with 3D information obtained when the object appears in the LiDAR field of view to achieve a smooth fusion of 2D and 3D trajectories. Extensive experiments based on the KITTI dataset indicate that our proposed method presents obvious advantages over the state-of-the-art MOT methods in terms of both tracking accuracy and processing speed. Our code is made publicly available for the benefit of the community.
|
|
11:20-11:30, Paper TuA-12.9 | |
EV-Catcher: High Speed Object Catching Using Low-Latency Event-Based Neural Networks |
|
Wang, Ziyun | University of Pennsylvania |
Cladera Ojeda, Fernando | University of Pennsylvania |
Bisulco, Anthony | Samsung |
Lee, Daewon | Samsung AI Center New York |
Taylor, Camillo Jose | University of Pennsylvania |
Daniilidis, Kostas | University of Pennsylvania |
Hsieh, M. Ani | University of Pennsylvania |
Lee, Daniel | Cornell Tech |
Isler, Volkan | University of Minnesota |
Keywords: Visual Tracking, Sensor-based Control
Abstract: Event-based sensors have recently drawn increasing interest in robotic perception due to their lower latency, higher dynamic range, and lower bandwidth requirements compared to standard CMOS-based imagers. These properties make them ideal tools for real-time perception tasks in highly dynamic environments. In this work, we demonstrate an application where event cameras excel: accurately estimating the impact location of fast-moving objects. We introduce a lightweight event representation called Binary Event History Image (BEHI) to encode event data at low latency, as well as a learning-based approach that allows real-time inference of a confidence-enabled control signal to the robot. To validate our approach, we present an experimental catching system in which we catch fast-flying ping-pong balls. We show that the system is capable of achieving a success rate of 81% in catching balls targeted at different locations, with a velocity of up to 13 m/s even on compute-constrained embedded platforms such as the Nvidia Jetson NX.
|
|
TuA-13 |
Rm13 (Room K) |
Mapping 4 |
Regular session |
Chair: Ghaffari, Maani | University of Michigan |
Co-Chair: Ewen, Parker | University of Michigan |
|
10:00-10:10, Paper TuA-13.1 | |
Voxfield: Non-Projective Signed Distance Fields for Online Planning and 3D Reconstruction |
|
Pan, Yue | ETH Zurich |
Kompis, Yves | ETH Zürich |
Bartolomei, Luca | ETH Zurich |
Mascaro Palliser, Ruben | ETH Zurich |
Stachniss, Cyrill | University of Bonn |
Chli, Margarita | ETH Zurich |
Keywords: Mapping, RGB-D Perception, Motion and Path Planning
Abstract: Creating accurate maps of complex, unknown environments is of utmost importance for truly autonomous navigation robot. However, building these maps online is far from trivial, especially when dealing with large amounts of raw sensor readings on a computation and energy constrained mobile system, such as a small drone. While numerous approaches tackling this problem have emerged in recent years, the mapping accuracy is often sacrificed as systematic approximation errors are tolerated for efficiency's sake. Motivated by these challenges, we propose Voxfield, a mapping framework that can generate maps online with higher accuracy and lower computational burden than the state of the art. Built upon the novel formulation of non-projective truncated signed distance fields (TSDFs), our approach produces more accurate and complete maps, suitable for surface reconstruction. Additionally, it enables efficient generation of euclidean signed distance fields (ESDFs), useful e.g. for path planning, that does not suffer from typical approximation errors. Through a series of experiments with public datasets, both real-world and synthetic, we demonstrate that our method beats the state of the art in map coverage, accuracy and computational time. Moreover, we show that Voxfield can be utilized as a back-end in recent multi-resolution mapping frameworks, producing high quality maps even in large-scale experiments. Finally, we validate our method by running it onboard a quadrotor, showing it can generate accurate ESDF maps usable for real-time path planning and obstacle avoidance.
|
|
10:10-10:20, Paper TuA-13.2 | |
3D Lidar Reconstruction with Probabilistic Depth Completion for Robotic Navigation |
|
Tao, Yifu | University of Oxford |
Popovic, Marija | University of Bonn |
Wang, Yiduo | University of Oxford |
Digumarti, Sundara Tejaswi | University of Oxford |
Chebrolu, Nived | University of Oxford |
Fallon, Maurice | University of Oxford |
Keywords: Mapping, SLAM
Abstract: Safe motion planning in robotics requires planning into space which has been verified to be free of obstacles. However, obtaining such environment representations using lidars is challenging by virtue of the sparsity of their depth mea- surements. We present a learning-aided 3D lidar reconstruction framework that upsamples sparse lidar depth measurements with the aid of overlapping camera images so as to generate denser reconstructions with more definitively free space than can be achieved with the raw lidar measurements alone. We use a neural network with an encoder-decoder structure to predict dense depth images along with depth uncertainty estimates which are fused using a volumetric mapping system. We conduct experiments on real-world outdoor datasets captured using a handheld sensing device and a legged robot. Using input data from a 16-beam lidar mapping a building network, our experiments showed that the amount of estimated free space was increased by more than 40% with our approach. We also show that our approach trained on a synthetic dataset generalises well to real-world outdoor scenes without additional fine-tuning. Finally, we demonstrate how motion planning tasks can benefit from these denser reconstructions.
|
|
10:20-10:30, Paper TuA-13.3 | |
Scalable Fiducial Tag Localization on a 3D Prior Map Via Graph-Theoretic Global Tag-Map Registration |
|
Koide, Kenji | National Institute of Advanced Industrial Science and Technology |
Oishi, Shuji | National Institute of Advanced Industrial Science and Technology |
Yokozuka, Masashi | Nat. Inst. of Advanced Industrial Science and Technology |
Banno, Atsuhiko | National Instisute of Advanced Industrial Science and Technology |
Keywords: Mapping, Vision-Based Navigation, Localization
Abstract: This paper presents an accurate and scalable method for fiducial tag localization on a 3D prior environmental map. The proposed method comprises three steps: 1) visual odometry-based landmark SLAM for estimating the relative poses between fiducial tags, 2) geometrical matching-based global tag-map registration via maximum clique finding, and 3) tag pose refinement based on direct camera-map alignment with normalized information distance. Through simulation-based evaluations, the proposed method achieved a 98 % global tag-map registration success rate and an average tag pose estimation accuracy of a few centimeters. Experimental results in a real environment demonstrated that it enables to localize over 110 fiducial tags placed in an environment in 25 minutes for data recording and post-processing.
|
|
10:30-10:40, Paper TuA-13.4 | |
These Maps Are Made for Walking: Real-Time Terrain Property Estimation for Mobile Robots |
|
Ewen, Parker | University of Michigan |
Li, Adam | University of Michigan |
Chen, Yuxin | University of Michigan |
Hong, Steven | University of Michigan |
Vasudevan, Ram | University of Michigan |
Keywords: Mapping, Semantic Scene Understanding, RGB-D Perception
Abstract: The equations of motion governing mobile robots are dependent on terrain properties such as the coefficient of friction, and contact model parameters. Estimating these properties is thus essential for robotic navigation. Ideally any map estimating terrain properties should run in real time, mitigate sensor noise, and provide probability distributions of the aforementioned properties, thus enabling risk-mitigating navigation and planning. This paper addresses these needs and proposes a Bayesian inference framework for semantic mapping which recursively estimates both the terrain surface profile and a probability distribution for terrain properties using data from a single RGB-D camera. The proposed framework is evaluated in simulation against other semantic mapping methods and is shown to outperform these state-of-the-art methods in terms of correctly estimating simulated ground-truth terrain properties when evaluated using a precision-recall curve and the Kullback-Leibler divergence test. Additionally, the proposed method is deployed on a physical legged robotic platform in both indoor and outdoor environments, and we show our method correctly predicts terrain properties in both cases. The proposed framework runs in real-time and includes a ROS interface for easy integration.
|
|
10:40-10:50, Paper TuA-13.5 | |
MapLite 2.0: Online HD Map Inference Using a Prior SD Map |
|
Ort, Teddy | Massachusetts Institute of Technology |
Walls, Jeffrey | University of Michigan |
Parkison, Steven | University of Michigan |
Gilitschenski, Igor | University of Toronto |
Rus, Daniela | MIT |
Keywords: Mapping, Autonomous Vehicle Navigation, Localization
Abstract: Deploying fully autonomous vehicles has been a subject of intense research in both industry and academia. However, the majority of these efforts have relied heavily on High Definition (HD) prior maps. These are necessary to provide the planning and control modules a rich model of the operating environment. While this approach has shown success, it drastically limits both the scale and scope of these deployments as creating and maintaining HD maps for very large areas can be prohibitive. In this work, we present a new method for building the HD map online by starting with a Standard Definition (SD) prior map such as a navigational road map, and incorporating onboard sensors to infer the local HD map. We evaluate our method extensively on 100 sequences of real-world vehicle data and demonstrate that it can infer a highly structured HD map-like model of the world accurately using only SD prior maps and onboard sensors.
|
|
10:50-11:00, Paper TuA-13.6 | |
Towards High-Definition Maps: A Framework Leveraging Semantic Segmentation to Improve NDT Map Compression and Descriptivity |
|
Manninen, Petri Juhani | Finnish Geospatial Research Institute |
Hyyti, Heikki Sakari | Finnish Geospatial Research Institute FGI, National Land Survey |
Kyrki, Ville | Aalto University |
Maanpää, Jyri Sakari | Department of Remote Sensing and Photogrammetry, Finnish Geospat |
Taher, Josef | Aalto University |
Hyyppä, Juha | Finnish Geospatial Research Institute |
Keywords: Mapping, Localization
Abstract: High-Definition (HD) maps are needed for robust navigation of autonomous vehicles, limited by the on-board storage capacity. To solve this, we propose a novel framework, Environment-Aware Normal Distributions Transform (EA-NDT), that significantly improves compression of standard NDT map representation. The compressed representation of EA-NDT is based on semantic-aided clustering of point clouds resulting in more optimal cells compared to grid cells of standard NDT. To evaluate EA-NDT, we present an open-source implementation that extracts planar and cylindrical primitive features from a point cloud and further divides them into smaller cells to represent the data as an EA-NDT HD map. We collected an open suburban environment dataset and evaluated EA-NDT HD map representation against the standard NDT representation. Compared to the standard NDT, EA-NDT achieved consistently at least 1.5× higher map compression while maintaining the same descriptive capability. Moreover, we showed that EA-NDT is capable of producing maps with significantly higher descriptivity score when using the same number of cells than the standard NDT.
|
|
11:00-11:10, Paper TuA-13.7 | |
Robust Structure Identification and Room Segmentation of Cluttered Indoor Environments from Occupancy Grid Maps |
|
Luperto, Matteo | Università Degli Studi Di Milano |
Kucner, Tomasz Piotr | Aalto University |
Tassi, Andrea | Politecnico Di Milano |
Magnusson, Martin | Örebro University |
Amigoni, Francesco | Politecnico Di Milano |
Keywords: Mapping
Abstract: Identifying the environment's structure, i.e., to detect core components as rooms and walls, can facilitate several tasks fundamental for the successful operation of indoor autonomous mobile robots, including semantic environment understanding. These robots often rely on 2D occupancy maps for core tasks such as localisation, motion and task planning. However, reliable identification of structure and room segmentation from 2D occupancy maps is still an open problem due to clutter (e.g., furniture and movable object), occlusions, and partial coverage. We propose a method for the RObust StructurE identification and ROom SEgmentation (ROSE^2) of 2D occupancy maps, which may be cluttered and incomplete. ROSE^2 identifies the main directions of walls and is resilient to clutter and partial observations, allowing to extract a clean, abstract geometrical floor-plan-like description of the environment, which is used to segment, i.e., to identify rooms in, the original occupancy grid map. ROSE^2 is tested in several real-world publicly-available cluttered maps obtained in different conditions. The results show how it can robustly identify the environment structure in 2D occupancy maps suffering from clutter and partial observations, while significantly improving room segmentation accuracy. Thanks to the combination of clutter removal and robust room segmentation ROSE^2 consistently achieves higher performance than the state-of-the-art methods, against which it is compared.
|
|
11:10-11:20, Paper TuA-13.8 | |
Struct-MDC: Mesh-Refined Unsupervised Depth Completion Leveraging Structural Regularities from Visual SLAM |
|
Jeon, Jinwoo | KAIST |
Lim, Hyunjun | Korea Advanced Institute of Science and Technology |
Seo, Dong-Uk | Korea Advanced Institute of Science and Technology |
Myung, Hyun | KAIST (Korea Advanced Institute of Science and Technology) |
Keywords: Mapping, SLAM, Deep Learning for Visual Perception
Abstract: Feature-based visual simultaneous localization and mapping (SLAM) methods only estimate the depth of extracted features, generating a sparse depth map. To solve this sparsity problem, depth completion tasks that estimate a dense depth from a sparse depth have gained significant importance in recent years. Existing methodologies that use sparse depth from visual SLAM mainly employ point features. However, point features have limitations in preserving structural regularities owing to textureless environments and sparsity problems. To deal with these issues, we perform depth completion with visual SLAM using line features, which can better contain structural regularities than point features. The proposed methodology creates a convex hull region by performing constrained Delaunay triangulation with depth interpolation using line features. However, the generated depth includes low-frequency information and is discontinuous at the convex hull boundary. Therefore, we propose a mesh depth refinement (MDR) module to address this problem. The MDR module effectively transfers the high-frequency details of an input image to the interpolated depth and plays a vital role in bridging the conventional and deep learning-based approaches. The Struct-MDC outperforms other state-of-the-art algorithms on public and our custom datasets, and even outperforms supervised methodologies for some metrics. In addition, the effectiveness of the proposed MDR module is verified by a rigorous ablation study.
|
|
11:20-11:30, Paper TuA-13.9 | |
MotionSC: Data Set and Network for Real-Time Semantic Mapping in Dynamic Environments |
|
Wilson, Joseph | University of Michigan |
Song, Jingyu | University of Michigan |
Fu, Yuewei | University of Michigan |
Zhang, Arthur | University of Michigan |
Capodieci, Andrew | Neya Robotics |
Jayakumar, Paramsothy | U.S. Army DEVCOM Ground Vehicle Systems Center |
Barton, Kira | University of Michigan at Ann Arbor |
Ghaffari, Maani | University of Michigan |
Keywords: Data Sets for Robotic Vision, Deep Learning for Visual Perception, Mapping
Abstract: This work addresses a gap in semantic scene completion (SSC) data by creating a novel outdoor data set with accurate and complete dynamic scenes. Our data set is formed from randomly sampled views of the world at each time step, which supervises generalizability to complete scenes without occlusions or traces. We create SSC baselines from state-of-the-art open source networks and construct a benchmark real-time dense local semantic mapping algorithm, MotionSC, by leveraging recent 3D deep learning architectures to enhance SSC with temporal information. Our network shows that the proposed data set can quantify and supervise accurate scene completion in the presence of dynamic objects, which can lead to the development of improved dynamic mapping algorithms. All software is available at https://github.com/UMich-CURLY/3DMapping.
|
|
TuA-14 |
Rm14 (Room 501) |
Soft Robot Materials and Design 1 |
Regular session |
Chair: Legrand, Julie | VUB |
Co-Chair: Coad, Margaret M. | University of Notre Dame |
|
10:00-10:10, Paper TuA-14.1 | |
Rigid Skeleton Enhanced Dexterous Soft Finger Possessing Proprioception |
|
Zhen, Ruichen | Harbin Institute of Technology |
Jiang, Li | Harbin Institute of Technology |
Keywords: Soft Robot Materials and Design, Biologically-Inspired Robots, Soft Sensors and Actuators
Abstract: This work presents a humanoid soft robotics finger design with rigid skeletons and proprioceptive sensors. This 4-DOFs dexterous finger has soft joints and rigid phalanxes, which is about the size of human hand. To enhance the overall stiffness and for human-like behavior and configuration, rigid-soft actuators which we called quasi-joint is introduced. Although their lengths are shortened in this design, the soft actuators can still bend over 90°, exhibiting joint-like flexion and abduction/adduction. Thus IP joints and MCP joint are realized. EGaIn soft sensors are embedded into the structure for bending detection. In addition, multi-step molding fabrication method is introduced for this complex multi-material structure. This rigid-soft finger is a preliminary work and modular part of a highly dexterous humanoid soft robotic hand.
|
|
10:10-10:20, Paper TuA-14.2 | |
Self-Propelled Soft Everting Toroidal Robot for Navigation and Climbing in Confined Spaces |
|
Badillo, Nelson | University of Notre Dame |
Coad, Margaret M. | University of Notre Dame |
Keywords: Soft Robot Materials and Design, Mechanism Design, Climbing Robots
Abstract: There are many spaces inaccessible to humans where robots could help deliver sensors and equipment. Many of these spaces contain three-dimensional passageways and uneven terrain that pose challenges for robot design and control. Everting toroidal robots, which move via simultaneous eversion and inversion of their body material, are promising for navigation in these types of spaces. We present a novel soft everting toroidal robot that propels itself using a motorized device inside an air-filled membrane. Our robot requires only a single control signal to move, can conform to its environment, and can climb vertically with a motor torque that is independent of the force used to brace the robot against its environment. We derive and validate models of the forces involved in its motion, and we demonstrate the robot's ability to navigate a maze and climb a pipe.
|
|
10:20-10:30, Paper TuA-14.3 | |
A Geometric Design Approach for Continuum Robots by Piecewise Approximation of Free-Form Shapes |
|
Wang, Sicheng | Purdue University |
Blumenschein, Laura | Purdue University |
Keywords: Soft Robot Materials and Design, Kinematics
Abstract: As soft, continuum robots see increasing areas of application, many scenarios have arisen where it is necessary to consider the geometric shape of the robot. The current approaches to robot kinematics, such as the piecewise constant-curvature (PCC) model, are effective in representing simple overall robot geometry and estimating the end-effector state, but they are less intuitive for planing robots that involve complex geometries. In this work, we propose a solution to the geometric design problem by a two-part approach: a free-form spline defines a "shape curve" that describes the overall geometry of the robot, and then a "kinematic curve" composed of shapes that are feasible to replicate with continuum robots is fitted to the shape curve. As an implementation of this approach, we specifically explore the application of piecewise cubic Bezier curves in designing the shape curve of the robot, and pairs of arcs to construct the kinematic curves. Finally, the approach is applied to a tip-extension "vine" robot that is designed and fabricated to "grow" along a designed path and access the top surface of an obstacle.
|
|
10:30-10:40, Paper TuA-14.4 | |
A Multi-Segmented Soft Finger Using Snap-Through Instability of a Soft Valve with a Slit |
|
Park, Tongil | Korea Institute of Medical Microrobotics |
Choi, Eunpyo | Chonnam National University |
Kim, Chang-Sei | Chonnam National University |
Park, Jong-oh | Chonnam National University |
Hong, Ayoung | Chonnam National University |
Keywords: Soft Robot Materials and Design, Grippers and Other End-Effectors, Soft Sensors and Actuators
Abstract: Soft fingers with multiple segments can perform various grasping modes when each segment is individually controlled, although this requires a number of inputs and leads to a complicated structure. In this paper, we propose a multi-segment soft finger capable of generating dual modes only using a single input channel. We use the snap-through behavior of a soft spherical shell with a slit as a flow valve between two finger segments. Experimental results showed that geometrical designs and material properties of the valve determine its critical pressure (i.e., the pressure causing buckling of the shell) and affect the proposed soft fingers' shape deformation and mode transition. We finally characterized an antipodal gripper with two proposed fingers in terms of the acquisition region and grasp robustness. The gripper could achieve not only a broad range of precision grasp by fingers with a passive distal segment but also robust stability of power grasp by fingers with an active distal segment, without adding extra input channels.
|
|
10:40-10:50, Paper TuA-14.5 | |
Deformation-Driven Closed-Chain Soft Mobile Robot Aimed for Rolling and Climbing Locomotion |
|
Johnsen, Luka Paul | Tokyo Institute of Technology |
Tsukagoshi, Hideyuki | Tokyo Institute of Technology |
Keywords: Soft Robot Materials and Design, Hydraulic/Pneumatic Actuators, Soft Robot Applications
Abstract: The inherent compliance of soft mobile robots makes them a safe option for interaction with humans and fragile environments. To expand their possibilities around our daily life, it is desirable for them to have enhanced traversal ability with safe structure and movement. To meet this requirement, we propose a new type of soft mobile robot that moves while selecting rotational, wave, and climbing locomotion. The proposed robot, named DECSO, forms a closed chain of modular segment actuators that individually generate contraction and bidirectional bending motion by applying negative pressure. In this paper, after presenting methods for generating the above three kinds of locomotion, the characteristics and modeling of the segment actuator are illustrated. Furthermore, the experimental results show that a 4-segment prototype achieved smooth rotational movement in the horizontal plane at a velocity of 75 mm/s, while a 6-segment prototype was able not only to move forward by travelling waves, but also to cross over the human body and obstacles of 62% of its own height by rotational movement. These findings suggest a new possibility for soft mobile robots that can safely touch and move around the human body.
|
|
10:50-11:00, Paper TuA-14.6 | |
Soft, Multi-Layer, Disposable, Kirigami Based Robotic Grippers: On Handling of Delicate, Contaminated, and Everyday Objects |
|
Buzzatto, Joao | The University of Auckland |
Shahmohammadi, Mojtaba | University of Auckland |
Liang, JunBang | The University of Auckland |
Sanches, Felipe Padula | University of Auckland |
Matsunaga, Saori | Mitsubishi Electric Corporation |
Haraguchi, Rintaro | MitsubishiElectric Corp |
Mariyama, Toshisada | Mitsubishi Electric Corporation |
MacDonald, Bruce | University of Auckland |
Liarokapis, Minas | The University of Auckland |
Keywords: Soft Robot Materials and Design, Soft Robot Applications
Abstract: Grasping and manipulation are complex and demanding tasks, especially when executed in dynamic and unstructured environments. Typically, such tasks are executed by rigid articulated end-effectors, with a plethora of actuators that need sophisticated sensing and complex control laws to execute them efficiently. Soft robotics offers an alternative that allows for simplified execution of these demanding tasks, enabling the creation of robust, efficient, lightweight, and affordable solutions that are easy to control and operate. In this work, we introduce a new class of soft, kirigami-based robotic grippers, study their post-contact behavior and investigate different cut patterns for their development. We follow an experimental approach in which several designs are proposed and employed in a series of grasping and force exertion tests to compare their capabilities and post-contact behavior. The results of such experiments indicate a clear relationship between degree of reconfiguration and grasping force, and provide key insights into the effect of the cut patterns in the performance of the designs. These findings are then used in the design process of an improved version of multi-layer, disposable kirigami grippers that are fabricated employing simple 3D printed layers or laser cut PET films and silicone rubber using the concept of Hybrid Deposition Manufacturing (HDM). A series of experimental results demonstrate that the proposed design and manufacturing methods can enable the creation of soft, kirigami-based grippers with superior grasping capabilities that can handle delicate, contaminated, and everyday life objects and can even be disposed off in an automated way (e.g., after handling hazardous materials, such as medical waste).
|
|
11:00-11:10, Paper TuA-14.7 | |
Topology Optimized Multi-Material Self-Healing Actuator with Reduced Out of Plane Deformation |
|
Wang, Zhanwei | Vrije Universiteit Brussel |
Terryn, Seppe | Vrije Universiteit Brussel (VUB) |
Legrand, Julie | VUB |
Ferrentino, Pasquale | Vrije Universiteit Brussels |
Kashef Tabrizian, Seyedreza | Vrije Universiteit Brussel |
Brancart, Joost | Vrije Universiteit Brussel (VUB) |
Roels, Ellen | Vrije Universiteit Brussel |
Van Assche, Guy | Vrije Universiteit Brussel (VUB) |
Vanderborght, Bram | Vrije Universiteit Brussel |
Keywords: Soft Robot Materials and Design, Soft Robot Applications, Modeling, Control, and Learning for Soft Robots
Abstract: Recent advances in soft robotics in academia have led to the adoption of soft grippers in industrial settings. Due to their soft bending actuators, these grippers can handle delicate objects with great care. However, due to their flexibility, the actuators are prone to out-of-plane deformations upon asymmetric loading. These undesired deformations lead to reduced grasp performance and may cause instability or failure of the grip. While the state-of-the-art contributions describe complex designs to limit those deformations, this work focuses on a complementary path investigating the material distribution. In this paper, a novel bending actuator is developed with improved out-of-plane deformation resistance by optimizing the material distribution in multi-material designs composed of two polymers with different mechanical properties. This is made possible by the strong interfacial strength of Diels-Alder chemical bonds in the used polymers, which have a self-healing capability. A Solid Isotropic Material with Penalization (SIMP) topology optimization is performed to increase the out-of-plane resistance. The actuator is simulated using FEA COMSOL in which the (hyper) elastic materials are simulated by Mooney-Rivlin models, fitted on experimental uniaxial tensile test data. This multi-material actuator and a reference single material actuator were manufactured and modeled. Via experimental characterization and validation in FEA simulations, it is shown that the actuator performance, characterized by the in-plane performance and out-plane resistance, can be increased by an optimized multi-material composition, without changing the geometrical shape of the actuator.
|
|
11:10-11:20, Paper TuA-14.8 | |
Performance Evaluation for Braided McKibben Pneumatic Actuators in Telescopic Nested Structure |
|
Kadir, Raihan | Student |
Mohd Faudzi, Ahmad `Athif | Universiti Teknologi Malaysia |
Abdul Rahman, Mohd Azizi | Malaysia-Japan International Institute of Technology |
Keywords: Soft Robot Materials and Design, Hydraulic/Pneumatic Actuators, Soft Sensors and Actuators
Abstract: This study presents the usage of Braided Actuators with Nested Structures to improve the stroke characteristics of the McKibben actuators. Thin McKibben actuators can contract 23 % of their original length when pressurized. Usage of braiding and nested structures to improve contraction ratio has proven successful. In this study, we use a combined telescopic Nested Structure and Braided Actuators to increase the contraction ratio of the actuators. The Nested Braided Actuator (NBA) performance was compared to Single Actuator (SA), Braided Actuator (BA), and Nested Actuator (NA). The results at 350 kPa indicate that, NBA had the highest contraction ratio of 45.5 %, followed by NA (39.38 %), BA (29.57 %), and SA (23.41 %). At 350 kPa, the contraction force per actuator NBA was 36.29 N, NA exerted 41.33 N, BA exerted 30.41 N and SA exerted 43.8 N. The actuator's performance showed that it was capable of high stroke with only a 20 – 30 % loss in contraction force compared to the SA.
|
|
11:20-11:30, Paper TuA-14.9 | |
Energy-Efficient Tunable-Stiffness Soft Robots Using Second Moment of Area Actuation |
|
Micklem, Leo | University of Southampton |
Weymouth, Gabriel | University of Southampton |
Thornton, Blair | University of Southampton |
Keywords: Soft Robot Materials and Design, Hydraulic/Pneumatic Actuators, Marine Robotics
Abstract: The optimal stiffness for soft swimming robots depends on swimming speed, which means no single stiffness can maximise efficiency in all swimming conditions. Tunable-stiffness would produce an increased range of high-efficiency swimming speeds for robots with flexible propulsors and enable soft control surfaces for steering underwater vehicles. We propose and demonstrate a method for tunable soft robotic stiffness using inflatable rubber tubes to stiffen a silicone foil through pressure and second moment of area change. We achieved double the effective stiffness of the system for an input pressure change from 0 to 0.8 bar and 2 J energy input. We achieved a resonant amplitude gain of 5 to 7 times the input amplitude and tripled the high-gain frequency range compared to a foil with fixed stiffness. These results show that changing second moment of area is an energy effective approach to tunable-stiffness robots.
|
|
TuA-15 |
Rm15 (Room 509) |
Localization 4 |
Regular session |
Chair: Nelson, Bradley J. | ETH Zurich |
Co-Chair: Suzuki, Taro | Chiba Institute of Technology |
|
10:00-10:10, Paper TuA-15.1 | |
Acoustic Localization and Communication Using a MEMS Microphone for Low-Cost and Low-Power Bio-Inspired Underwater Robots |
|
Hinduja, Akshay | Carnegie Mellon University |
Ohm, Yunsik | Carnegie Mellon Univesity |
Liao, Jiahe | Carnegie Mellon University |
Majidi, Carmel | Carnegie Mellon University |
Kaess, Michael | Carnegie Mellon University |
Keywords: Localization, Marine Robotics, Biologically-Inspired Robots
Abstract: Having accurate localization capabilities is one of the fundamental requirements of autonomous robots. For underwater vehicles, the choices for effective localization are limited due to limitations of GPS use in water and poor environmental visibility that makes camera-based methods ineffective. Popular inertial navigation methods for underwater localization using Doppler-velocity log sensors, sonar, high-end inertial navigation systems, or acoustic positioning systems require bulky expensive hardware which are incompatible with low-cost, bio-inspired underwater robots. In this paper, we introduce an approach for underwater robot localization inspired by GPS methods known as acoustic pseudoranging. Our method allows us to localize multiple bio-inspired robots equipped with commonly available micro electro-mechanical systems microphones. This is achieved by estimating the time difference of arrival of acoustic signals sent simultaneously through four speakers with a known constellation geometry. We also leverage the same acoustic framework to perform one-way communication with multiple robots to execute some primitive motions. To our knowledge, this is the first application of the approach for the on-board localization of small bio-inspired robots in water. Hardware schematics and the accompanying code are released to aid further development in the field.
|
|
10:10-10:20, Paper TuA-15.2 | |
Using Magnetic Fields to Navigate and Simultaneously Localize Catheters in Endoluminal Environments |
|
Fischer, Cedric | ETH Zurich |
Boehler, Quentin | ETH Zurich |
Nelson, Bradley J. | ETH Zurich |
Keywords: Localization, Surgical Robotics: Steerable Catheters/Needles, Surgical Robotics: Planning
Abstract: Remote magnetic navigation offers an ideal platform for automated catheter navigation. Magnetically guided catheters show great dexterity and can reach locations that are otherwise challenging to access. By automating aspects of catheterization procedures, we can simplify and expedite the procedure to allow surgeons to focus on other critical tasks during the surgery. In this article, we describe an automation strategy that is based on the center line of extracted and registered vascular geometries. Position feedback is accomplished with a Hall-effect sensor embedded near the distal end of the catheter. Sensor measurements are compared to the magnetic field predicted by a linear model of the electromagnetic navigation system. By defining specific magnetic field gradients and applying the known vascular geometry, the magnetic fields can be utilized for the simultaneous navigation and localization of the catheter. This eliminates the need for other external, dedicated mapping systems, and the use of fluoroscopy imaging is minimized. The concept is tested in 2d vascular models and the accuracy of the localization is assessed with overhead camera tracking.
|
|
10:20-10:30, Paper TuA-15.3 | |
Robust Slip-Aware Fusion for Mobile Robots State Estimation |
|
Hashemi, Ehsan | University of Alberta |
He, Xingkang | University of Notre Dame |
Johansson, Karl H. | Royal Institute of Technology |
Keywords: Localization, Wheeled Robots, Sensor Fusion
Abstract: A novel robust and slip-aware speed estimation framework is developed and experimentally verified for mobile robot navigation by designing proprioceptive robust observers at each wheel. The observer for each corner is proved to be consistent, in the sense that it can provide an upper bound of the mean square estimation error (MSE) timely. Under proper conditions, the MSE is proved to be uniformly bounded. A covariance intersection fusion method is used to fuse the wheel-level estimates, such that the updated estimate remains consistent. The estimated slips at each wheel are then used for a robust consensus to improve the reliability of speed estimation in harsh and combined-slip scenarios. As confirmed by indoor and outdoor experiments under different surface conditions, the developed framework addresses state estimation challenges for mobile robots that experience uneven torque distribution or large slip. The novel proprioceptive observer can also be integrated with existing tightly-coupled visual-inertial navigation systems.
|
|
10:30-10:40, Paper TuA-15.4 | |
Highly-Efficient Binary Neural Networks for Visual Place Recognition |
|
Ferrarini, Bruno | Universtiy of Essex |
Milford, Michael J | Queensland University of Technology |
McDonald-Maier, Klaus | University of Essex |
Ehsan, Shoaib | University of Essex |
Keywords: Localization, Autonomous Vehicle Navigation
Abstract: VPR is a fundamental task for autonomous navigation as it enables a robot to localize itself in the workspace when a known location is detected. Although accuracy is an essential requirement for a VPR technique, computational and energy efficiency are not less important for real-world applications. CNN-based techniques archive state-of-the-art VPR performance but are computationally intensive and energy demanding. Binary neural networks (BNN) have been recently proposed to address VPR efficiently. Although a typical BNN is an order of magnitude more efficient than a CNN, its processing time and energy usage can be further improved. In a typical BNN, the first convolution is not completely binarized for the sake of accuracy. Consequently, the first layer is the slowest network stage, requiring a large share of the entire computational effort. This paper presents a class of BNNs for VPR that combines depthwise separable factorization and binarization to replace the first convolutional layer to improve computational and energy efficiency. Our best model achieves higher VPR performance while spending considerably less time and energy to process an image than a BNN using a non-binary convolution as a first stage.
|
|
10:40-10:50, Paper TuA-15.5 | |
GNSS Odometry: Precise Trajectory Estimation Based on Carrier Phase Cycle Slip Estimation |
|
Suzuki, Taro | Chiba Institute of Technology |
Keywords: Localization, SLAM
Abstract: This paper proposes a highly accurate trajectory estimation method for outdoor mobile robots using global navigation satellite system (GNSS) time differences of carrier phase (TDCP) measurements. By using GNSS TDCP, the relative 3D position can be estimated with millimeter precision. However, when a phenomenon called cycle slip occurs, wherein the carrier phase measurement jumps and becomes discontinuous, it is impossible to accurately estimate the relative position using TDCP. Although previous studies have eliminated the effect of cycle slip using a robust optimization technique, it was difficult to completely eliminate the effect of outliers. In this paper, we propose a method to detect GNSS carrier phase cycle slip, estimate the amount of cycle slip, and modify the observed TDCP to calculate the relative position using the factor graph optimization framework. The estimated relative position acts as a loop closure in graph optimization and contributes to the reduction in the integration error of the relative position. Experiments with an unmanned aerial vehicle showed that by modifying the cycle slip using the proposed method, the vehicle trajectory could be estimated with an accuracy of 5–30 cm using only a single GNSS receiver, without using any other external data or sensors.
|
|
10:50-11:00, Paper TuA-15.6 | |
Square-Root Robocentric Visual-Inertial Odometry with Online Spatiotemporal Calibration |
|
Huai, Zheng | University of Delaware |
Huang, Guoquan | University of Delaware |
Keywords: Localization, Vision-Based Navigation
Abstract: Robocentric visual-inertial odometry (R-VIO) in our recent work [1] models the probabilistic state estimation problem with respect to a moving local (body) frame, which is contrary to a fixed global (world) frame as in the world-centric formulation, thus avoiding the observability mismatch issue and achieving better estimation consistency. To further improve efficiency and robustness in order to be amenable for the resource-constrained applications, in this paper, we propose a novel information-based estimator, termed R-VIO2. In particular, the numerical stability and computational efficiency are significantly boosted by using the i) square-root expression and ii) incremental QR-based update combined with back substitution. Moreover, the spatial transformation and time offset between visual and inertial sensors are jointly calibrated online to robustify the estimator performance in the presence of unknown parameter errors. The proposed R-VIO2 has been extensively tested on public benchmark dataset as well as in a large-scale real-world experiment, and shown to achieve very competitive accuracy and superior time efficiency against the state-of-the-art visual-inertial navigation methods.
|
|
11:00-11:10, Paper TuA-15.7 | |
Are We Ready for Radar to Replace Lidar in All-Weather Mapping and Localization? |
|
Burnett, Keenan | University of Toronto |
Wu, Yuchen | University of Toronto |
Yoon, David Juny | University of Toronto |
Schoellig, Angela P. | University of Toronto |
Barfoot, Timothy | University of Toronto |
Keywords: Localization, Range Sensing, Intelligent Transportation Systems
Abstract: We present an extensive comparison between three topometric localization systems: radar-only, lidar-only, and a cross-modal radar-to-lidar system across varying seasonal and weather conditions using the Boreas dataset. Contrary to our expectations, our experiments showed that our lidar-only pipeline achieved the best localization accuracy even during a snowstorm. Our results seem to suggest that the sensitivity of lidar localization to moderate precipitation has been exaggerated in prior works. However, our radar-only pipeline was able to achieve competitive accuracy with a much smaller map. Furthermore, radar localization and radar sensors still have room to improve and may yet prove valuable in extreme weather or as a redundant backup system. Code for this project can be found at: https://github.com/utiasASRL/vtr3
|
|
11:10-11:20, Paper TuA-15.8 | |
Continuous-Time Factor Graph Optimization for Trajectory Smoothness of GNSS/INS Navigation in Temporarily GNSS-Denied Environments |
|
Zhang, Haoming | RWTH Aachen University |
Xia, Xiao | Institute of Automatic Control, RWTH Aachen University |
Nitsch, Maximilian | Institute of Automatic Control, RWTH Aachen University |
Abel, Dirk | RWTH Aachen University |
Keywords: Localization, Sensor Fusion, Marine Robotics
Abstract: For autonomous systems, the smoothness of estimated trajectories is essential to ensure robust and performant vehicle control. Unfortunately, approaches for state estimation using Kalman filters are sensitive to measurement outliers that can diverge dramatically. We propose a state-estimation algorithm using factor graph optimization (FGO) in continuous-time for GNSS/INS navigation systems to track this problem. The focus is on the smoothness of the estimated trajectory in environments where the GNSS observations become temporarily unreliable. Therefore, an inland shipping scenario with bridge crossings is selected as an application example. To estimate the trajectory in continuous-time, we integrate a White-Noise-On-Acceleration (WNOA) motion prior factor based on Gaussian process regression between successive states. Our results show a 30% improvement in accuracy and increased smoothness of the estimated trajectory with FGO compared to Kalman filtering.
|
|
11:20-11:30, Paper TuA-15.9 | |
Maintaining Robot Localizability with Bayesian Cramer-Rao Lower Bounds |
|
Cano, Justin | Polytechnique Montréal |
Chauffaut, Corentin | Université De Toulouse |
Chaumette, Eric | University of Toulouse/Isae-Supaero |
Pages, Gaël | ISAE-SUPAERO |
Le Ny, Jerome | Polytechnique Montreal |
Keywords: Localization, Range Sensing, Multi-Robot Systems
Abstract: Accurate and real-time position estimates are crucial for mobile robots. This work focuses on ranging-based positioning systems, which rely on distance measurements between known points, called anchors, and a tag to localize. The topology of the network formed by the anchors strongly influences the tag's localizability, i.e., its ability to be accurately localized. Here, the tag and some anchors are supposed to be carried by robots, which allows enhancing the positioning accuracy by planing the anchors' motions. We leverage Bayesian Cram'er-Rao Lower Bounds (CRLBs) on the estimates' covariance in order to quantify the tag's localizability. This class of CRLBs can capture prior information on the tag's position and take it into account when deploying the anchors. We propose a method to decrease a potential function based on the Bayesian CRLB in order to maintain the localizability of the tag while having some prior knowledge about its position distribution. Then, we present a new experiment highlighting the link between the localizability potential and the precision expected in practice. Finally, two real-time anchor motion planners are demonstrated with ranging measurements in the presence or absence of prior information about the tag's position.
|
|
TuA-16 |
Rm16 (Room 510) |
Reinforcement Learning 3 |
Regular session |
Chair: Patel, Niravkumar | Indian Institute of Technology Madras |
Co-Chair: Kobayashi, Taisuke | National Institute of Informatics |
|
10:00-10:10, Paper TuA-16.1 | |
Safe Reinforcement Learning Using Black-Box Reachability Analysis |
|
Selim, Mahmoud | Ain Shams University |
Alanwar, Amr | KTH Royal Institute of Technology |
Kousik, Shreyas | Stanford University |
Gao, Grace | Stanford University |
Pavone, Marco | Stanford University |
Johansson, Karl H. | Royal Institute of Technology |
Keywords: Reinforcement Learning, Robot Safety, Motion and Path Planning
Abstract: Reinforcement learning (RL) is capable of sophisticated motion planning and control for robots in uncertain environments. However, state-of-the-art deep RL approaches typically lack safety guarantees, especially when the robot and environment models are unknown. To justify widespread deployment, robots must respect safety constraints without sacrificing performance. Thus, we propose a Black-box Reachability-based Safety Layer (BRSL) with three main components: (1) data-driven reachability analysis for a black-box robot model, (2) a trajectory rollout planner that predicts future actions and observations using an ensemble of neural networks trained online, and (3) a differentiable polytope collision check between the reachable set and obstacles that enables correcting unsafe actions. In simulation, BRSL outperforms other state-of-the-art safe RL methods on a Turtlebot 3, a quadrotor, a trajectory-tracking point mass, and a hexarotor in wind with an unsafe set adjacent to the area of highest reward.
|
|
10:10-10:20, Paper TuA-16.2 | |
Hierarchical Primitive Composition: Simultaneous Activation of Skills with Inconsistent Action Dimensions in Multiple Hierarchies |
|
Lee, Jeong-Hoon | Yonsei Univ |
Choi, Jongeun | Yonsei University |
Keywords: Reinforcement Learning, Incremental Learning, Deep Learning in Grasping and Manipulation
Abstract: Deep reinforcement learning has shown its effectiveness in various applications, providing a promising direction for solving tasks with high complexity. However, naively applying classical RL for learning a complex long-horizon task with a single control policy is prohibitively inefficient. Thus, policy modularization tackles this problem by learning a set of modules that are mapped to primitive and properly orchestrating them. In this study, we further expand the discussion by incorporating simultaneous activation of the skills and structuring them into multiple hierarchies in a recursive fashion. Moreover, we sought to devise an algorithm that can properly orchestrate the skills with different action spaces via multiplicative Gaussian distributions, which highly increase the reusability. By exploiting the modularity, interpretability can also be achieved by observing the modules that are used in the new task if each of the skills is known. We demonstrate how the proposed scheme can be employed in practice by solving a pick and place task with a 6 DoF manipulator, and examine the effects of each property from ablation studies.
|
|
10:20-10:30, Paper TuA-16.3 | |
Vision-Guided Quadrupedal Locomotion in the Wild with Multi-Modal Delay Randomization |
|
Imai, Chieko Sarah | University of California San Diego |
Zhang, Minghao | Tsinghua University |
Zhang, Yuchen | University Od California, San Diego |
Kierebinski, Marcin | University of California, San Diego |
Yang, Ruihan | UC San Diego |
Qin, Yuzhe | UCSD |
Wang, Xiaolong | UC San Diego |
Keywords: Reinforcement Learning, Legged Robots
Abstract: Developing robust vision-guided controllers for quadrupedal robots in complex environments, with various obstacles, dynamical surroundings and uneven terrains, is very challenging. While Reinforcement Learning (RL) provides a promising paradigm for agile locomotion skills with vision inputs in simulation, it is still very challenging to deploy the RL policy in the real world. Our key insight is that aside from the discrepancy in the domain gap, in visual appearance between the simulation and the real world, the latency from the control pipeline is also a major cause of difficulty. In this paper, we propose Multi-Modal Delay Randomization (MMDR) to address this issue when training RL agents. Specifically, we simulate the latency of real hardware by using past observations, sampled with randomized periods, for both proprioception and vision. We train the RL policy for end-to-end control in a physical simulator without any predefined controller or reference motion, and directly deploy it on the real A1 quadruped robot running in the wild. We evaluate our method in different outdoor environments with complex terrains and obstacles. We demonstrate the robot can smoothly maneuver at a high speed, avoid the obstacles, and show significant improvement over the baselines.
|
|
10:30-10:40, Paper TuA-16.4 | |
Temporal Logic Guided Meta Q-Learning of Multiple Tasks |
|
Zhang, Hao | University of Science and Technology of China |
Kan, Zhen | University of Science and Technology of China |
Keywords: Reinforcement Learning, Motion and Path Planning, Formal Methods in Robotics and Automation
Abstract: Reinforcement learning (RL) based approaches have enabled robots to perform various tasks. However, most existing RL algorithms focus on learning a particular task, without considering generalization to new tasks. To address this issue, by combining meta learning and reinforcement learning, we develop a meta Q-learning of multi-task (MQMT) framework where the robot effectively learns a meta model from a diverse set of training tasks and then generalizes the learned model to a new set of tasks that have never been encountered during training using only a small amount of additional data. Particularly, the multiple tasks are specified by co-safe linear temporal logic specification. As a semantics-preserving rewriting operation, LTL progression is exploited to decompose training tasks into learnable sub-goals, which not only enables simultaneous learning of multiple tasks, but also facilitates reward design by converting non-Markovian reward process to Markovian ones. Reward shaping is further incorporated into the reward design to relax the sparse reward issue to improve reinforcement learning. The simulation and experiment results demonstrate the effectiveness of the MQMT framework.
|
|
10:40-10:50, Paper TuA-16.5 | |
Model-Free Neural Lyapunov Control for Safe Robot Navigation |
|
Xiong, Zikang | Purdue University |
Eappen, Joe Kurian | Purdue University |
Qureshi, Ahmed H. | Purdue University |
Jagannathan, Suresh | Purdue University |
Keywords: Machine Learning for Robot Control, Reinforcement Learning, Motion and Path Planning
Abstract: Model-free Deep Reinforcement Learning (DRL) controllers have demonstrated promising results on various challenging non-linear control tasks. While a model-free DRL algorithm can solve unknown dynamics and high-dimensional problems, it lacks safety assurance. Although safety constraints can be encoded as part of a reward function, there still exists a large gap between an RL controller trained with this modified reward and a safe controller. In contrast, instead of implicitly encoding safety constraints with rewards, we explicitly co-learn a Twin Neural Lyapunov Function (TNLF) with the control policy in the DRL training loop and use the learned TNLF to build a runtime monitor. Combined with the path generated from a planner, the monitor chooses appropriate waypoints that guide the learned controller to provide collision-free control trajectories. Our approach inherits the scalability advantages from DRL while enhancing safety guarantees. Our experimental evaluation demonstrates the effectiveness of our approach compared to DRL with augmented rewards and constrained DRL methods over a range of high-dimensional safety-sensitive navigation tasks.
|
|
10:50-11:00, Paper TuA-16.6 | |
Efficient Off-Policy Safe Reinforcement Learning Using Trust Region Conditional Value at Risk |
|
Kim, Dohyeong | Seoul National University |
Oh, Songhwai | Seoul National University |
Keywords: Reinforcement Learning, Robot Safety, Collision Avoidance
Abstract: This paper aims to solve a safe reinforcement learning (RL) problem with risk measure-based constraints. As risk measures, such as conditional value at risk (CVaR), focus on the tail distribution of cost signals, constraining risk measures can effectively prevent a failure in the worst case. An on-policy safe RL method, called TRC, deals with a CVaR-constrained RL problem using a trust region method and can generate policies with almost zero constraint violations with high returns. However, to achieve outstanding performance in complex environments and satisfy safety constraints quickly, RL methods are required to be sample efficient. To this end, we propose an off-policy safe RL method with CVaR constraints, called off-policy TRC. If off-policy data from replay buffers is directly used to train TRC, the estimation error caused by the distributional shift results in performance degradation. To resolve this issue, we propose novel surrogate functions, in which the effect of the distributional shift can be reduced, and introduce an adaptive trust-region constraint to ensure a policy not to deviate far from replay buffers. buffer. The proposed method has been evaluated in simulation and real-world environments and satisfied safety constraints within a few steps while achieving high returns even in complex robotic tasks.
|
|
11:00-11:10, Paper TuA-16.7 | |
Variable Impedance Skill Learning for Contact-Rich Manipulation |
|
Yang, Quantao | Orebro University |
Dürr, Alexander | Lund University |
Topp, Elin Anna | Lund University - LTH |
Stork, Johannes A. | Orebro University |
Stoyanov, Todor | Örebro University |
Keywords: Reinforcement Learning, Compliance and Impedance Control, Machine Learning for Robot Control
Abstract: Reinforcement Learning (RL) has the potential of solving complex continuous control tasks, with direct applications to robotics. Nevertheless, current state-of-the-art methods require a vast amount of interaction experience and are generally not safe or feasible to learn directly on a physical robot. We address this challenge by a framework for learning latent action spaces for RL agents from demonstrated trajectories. We extend this framework by connecting it to a variable impedance Cartesian space controller, allowing us to learn contact-rich tasks safely and efficiently. Our method learns from trajectories that incorporate both positional, but also crucially impedance-space information. We evaluate our method on a number of peg-in-hole task variants with a Franka Panda arm and demonstrate that learning variable impedance actions for RL in Cartesian space can be deployed directly on the real robot, without resorting to learning in simulation.
|
|
11:10-11:20, Paper TuA-16.8 | |
Source Term Estimation Using Deep Reinforcement Learning with Gaussian Mixture Model Feature Extraction for Mobile Sensors |
|
Park, Minkyu | UNIST (Ulsan National Institute of Science and Technology) |
Ladosz, Pawel | UNIST |
Oh, Hyondong | UNIST |
Keywords: Reinforcement Learning, Planning under Uncertainty, Aerial Systems: Perception and Autonomy
Abstract: This paper proposes a deep reinforcement learning method for mobile sensors to estimate the properties of the source of the hazardous gas release. The problem of estimating the properties of the released gas is generally termed as the source term estimation (STE) problem. Since the sensor measurements from atmospheric gas dispersion are sparse, intermittent, and time-varying due to the turbulence and the sensor noise, STE is considered to be a challenging problem. The particle filter is adopted to estimate the source term under such stochastic noise conditions. The deep deterministic policy gradient (DDPG) is also employed to find the best source search policy in terms of successful estimation and traveled distance. Through ablation studies, we demonstrate that the use of the Gaussian mixture model, which clusters the potential source positions from the particle filter, as an input to the DDPG and the gated recurrent unit functioning as a memory in DDPG help to improve the STE performance. Besides, simulation results in randomized source term conditions and previously-unseen environments show the superior STE performance of the proposed algorithm compared with the existing information-theoretic STE algorithm.
|
|
11:20-11:30, Paper TuA-16.9 | |
Autonomous Learning of Page Flipping Movements Via Tactile Feedback (I) |
|
Zheng, Yi | University of California, Los Angeles |
Veiga, Filipe Fernandes | MIT |
Peters, Jan | Technische Universität Darmstadt |
Santos, Veronica J. | University of California, Los Angeles |
Keywords: Reinforcement Learning, Perception for Grasping and Manipulation, Force and Tactile Sensing
Abstract: Robotic manipulation is challenging when both the objects being manipulated and the tactile sensors are deformable. In this work, we addressed the interplay between the manipulation of deformable objects, tactile sensing, and model-free reinforcement learning on a real robot. We showed how a real robot can learn to manipulate a deformable, thin-shell object via feedback from deformable, multimodal tactile sensors. We addressed the learning of a page flipping task using a two-stage approach. For the first stage, we learned nominal page flipping trajectories for two page sizes by constructing a reward function that quantifies functional task performance from the perspective of tactile sensing. For the second stage, we learned adapted trajectories using tactile-driven perceptual coupling, with an intuitive assumption that, while the page flipping trajectories for different task contexts (page sizes) might differ, similar tactile feedback should be expected from functional trajectories for each context. We also investigated the quality of information encoded by two different representations of tactile sensing data: one based on the artificial apical tuft of bio-inspired tactile sensors, and another based on PCA eigenvalues. The results and effectiveness of our learning framework were demonstrated on a real 7-DOF robot arm and gripper outfitted with tactile sensors.
|
|
TuA-17 |
Rm17 (Room 553) |
Climbing and Wheeled Robots |
Regular session |
Chair: Nadan, Paul | Carnegie Mellon University |
Co-Chair: Mashimo, Tomoaki | Okayama University |
|
10:00-10:10, Paper TuA-17.1 | |
Magnetic Field Modeling of Linear Halbach Array for Wall-Climbing Robot Based on Radial Basis Function Neural Network |
|
Liu, Xiaofei | Shenzhen Institute of Advanced Technology, Chinese Academy of Sc |
Yi, Zhengkun | Shenzhen Institute of Advanced Technology, Chinese Academy of Sc |
Wu, Xinyu | SIAT |
Shang, Wanfeng | Shenzhen Institutes of Advanced Technology, Chinese Academy of S |
Keywords: AI-Based Methods, Dynamics
Abstract: Aiming at the problem that it is difficult to calculate the force of permanent magnets in the magnetic field, this paper proposes a nonlinear mechanical model of linear array magnetic field based on radial basis function neural network (RBFNN). Combined with the linear Halbach array adsorption module of the wall-climbing robot, the three-dimensional geometric magnetic fields of four typical linear array permanent magnets were constructed, and the theoretical models of the interaction between the magnetic fields were given respectively. Further, the finite element simulation calculation of the magnetic force was carried out using COMSOL Multiphysics software. According to the parametric scanning results of the orthogonal test, a nonlinear intelligent prediction model of the force between magnetic fields with local loss sensitivity is established by using the RBFNN numerical fitting method. The average deviation of the network test set is 1.19, and the standard deviation is 0.80. The intelligent prediction model has strong general performance, faster convergence speed and stronger flexibility, which provides a theoretical basis for the interaction and control of array magnetic fields.
|
|
10:10-10:20, Paper TuA-17.2 | |
OmniWheg: An Omnidirectional Wheel-Leg Transformable Robot |
|
Cao, Ruixiang | Shanghaitech University |
Gu, Jun | ShanghaiTech University |
Yu, Chen | ShanghaiTech University |
Rosendo, Andre | ShanghaiTech University |
Keywords: Wheeled Robots, Legged Robots, Climbing Robots
Abstract: This paper presents the design, analysis, and performance evaluation of an omnidirectional transformable wheel-leg robot called OmniWheg. We design a novel mechanism consisting of a separable omni-wheel and 4-bar linkages, allowing the robot to transform between omni-wheeled and legged modes smoothly. In wheeled mode, the robot can move in all directions and efficiently adjust the relative position of its wheels, while it can overcome common obstacles in legged mode, such as stairs and steps. Unlike other articles studying whegs, this implementation with omnidirectional wheels allows the correction of misalignments between right and left wheels before traversing obstacles, which effectively improves the success rate and simplifies the preparation process before the wheel-leg transformation. We describe the design concept, mechanism, and the dynamic characteristic of the wheel-leg structure. We then evaluate its performance in various scenarios, including passing obstacles, climbing steps of different heights, and turning/moving omnidirectionally. Our results confirm that this mobile platform can overcome common indoor obstacles and move flexibly on the flat ground with the new transformable wheel-leg mechanism, while keeping a high degree of stability.
|
|
10:20-10:30, Paper TuA-17.3 | |
SCALER: A Tough Versatile Quadruped Free-Climber Robot |
|
Tanaka, Yusuke | University of California, Los Angeles |
Shirai, Yuki | University of California, Los Angeles |
Lin, Xuan | UCLA |
Schperberg, Alexander | University of California Los Angeles |
Kato, Hayato | UCLA |
Swerdlow, Alexander | University of California, Los Angeles |
Kumagai, Naoya | University of California, Los Angeles |
Hong, Dennis | UCLA |
Keywords: Climbing Robots, Legged Robots, Grippers and Other End-Effectors
Abstract: This paper introduces SCALER, a quadrupedal robot that demonstrates climbing on bouldering walls, overhangs, ceilings and trotting on the ground. SCALER is one of the first high-degrees of freedom four-limbed robots that can free-climb under the Earth's gravity and one of the most mechanically efficient quadrupeds on the ground. Where other state-of-the-art climbers specialize in climbing, SCALER promises practical free-climbing with payload textit{and} ground locomotion, which realizes true versatile mobility. A new climbing gait, SKATE gait, increases the payload by utilizing the SCALER body linkage mechanism. SCALER achieves a maximum normalized locomotion speed of 1.87 /s, or 0.56 m/s on the ground and 1.0 /min, or 0.35 m/min in bouldering wall climbing. Payload capacity reaches 233 % of the SCALER weight on the ground and 35 % on the vertical wall. Our GOAT gripper, a mechanically adaptable underactuated two-finger gripper, successfully grasps convex and non-convex objects and supports SCALER.
|
|
10:30-10:40, Paper TuA-17.4 | |
Microspine Design for Additive Manufacturing |
|
Nadan, Paul | Carnegie Mellon University |
Patel, Dinesh | Carnegie Mellon University |
Pavlov, Catherine | Carnegie Mellon University |
Backus, Spencer | NASA Jet Propulsion Lab |
Johnson, Aaron M. | Carnegie Mellon University |
Keywords: Climbing Robots, Grippers and Other End-Effectors, Additive Manufacturing
Abstract: Microspine grippers allow robots to ascend steep rocky slopes and cliff faces, enabling scientific exploration of exposed strata on Earth and other solar system bodies. Historically, the Shape Deposition Manufacturing (SDM) process has been used to fabricate multi-material suspensions for load-sharing among multiple microspines. We instead apply the Hybrid Deposition Manufacturing (HDM) process to microspine fabrication, and we further propose a novel 3D-printed microspine suspension design that can be manufactured via Fused Deposition Manufacturing (FDM) alone, using a single flexible material with an embedded fishhook. We use a model of microspine stiffness that allows designers to compensate for order-of-magnitude changes in material tensile modulus by adjusting geometric parameters of the design. The stiffness model and the FDM microspine design are validated through tensile testing, and mechanical properties of the HDM and FDM designs are compared against a standard SDM microspine design. We demonstrate that the FDM process can produce microspines with equivalent normal and axial stiffness and superior maximum load and fatigue response to SDM microspines, and discuss additional advantages of the FDM process for rapid prototyping and broader accessibility.
|
|
10:40-10:50, Paper TuA-17.5 | |
Design, Fabrication, and Characterization of a Hybrid Bionic Spherical Robotics with Multilegged Feedback Mechanism |
|
Xu, Minyi | Dalian Maritime University |
Zheng, Jiaxi | Dalian Maritime University |
Xu, Peng | Dalian Maritime University |
Meng, Zhaochen | Dalian Maritime University |
Liu, Jianhua | Dalian Maritime University |
Wang, Siyuan | Dalian Maritime University |
Wang, Xinyu | Dalian Maritime University |
Xie, Guangming | Peking University |
Tao, Jin | Nankai University |
Keywords: Engineering for Robotic Systems, Simulation and Animation, Motion Control
Abstract: Spherical robots have many desirable traits when designing mass efficient systems interacted with unstructured terrain. In this paper, we propose a hybrid bionic spherical robot based on the morphological properties of sea urchins and the movement characteristics of tumbleweeds. This robot enables it to move freely in harsh terrain with the distributed high-aspect-ratio telescopic units, where our control strategy is based on the a central pattern generator and combines foot pressure measurements and actuator state information. In particular, these data from multiple foot pressure sensors also are used to compute the robots' center of mass, with evaluating locomotion state during rolling maneuvers. The experimental results show the ability of our design to provide reliable pose estimates, also overcoming a challenge in the field of mobile robotics, including switch directions flexibly in harsh terrain due to their anisotropy.
|
|
10:50-11:00, Paper TuA-17.6 | |
The Concept of Rod-Driven Locomotion for Spherical Lunar Exploration Robots |
|
Zevering, Jasper | Julius-Maximilians-Universität Würzburg |
Borrmann, Dorit | Julius-Maximilians-University of Würzburg |
Bredenbeck, Anton | TU Delft |
Nuechter, Andreas | University of Würzburg |
Keywords: Space Robotics and Automation
Abstract: A spherical robotic probe has several advantages in rough environments and has therefore raised interest for application in planetary exploration. A sphere is well-suited to protect high-sensitive payloads, however, the locomotion system for planetary surfaces raises several challenges. This paper presents a novel locomotion system consisting of linear actuators which are usable in a multi-functional fashion. Apart from pushing and bringing leverage for locomotion the extendable rods enable a tripod mode for improved sensing. The developed solutions offer a mathematical-physical system description, simple algorithms for the control of locomotion and balancing as well as general calculations for determining the maximum achievable performance parameters of such a robot. The first built prototype shows the basic suitability of the system and reveals directions for further research.
|
|
11:00-11:10, Paper TuA-17.7 | |
Nonlinear Model Predictive Control with Cost Function Scheduling for a Wheeled Mobile Robot |
|
Lim, Jaehyun | Yonsei University |
Lee, Hyeonwoo | Yonsei University |
Choi, Jongeun | Yonsei University |
Keywords: Wheeled Robots, Optimization and Optimal Control, Motion and Path Planning
Abstract: Designing a cost function for nonlinear model predictive control (MPC) with a sparse/binary stage cost is challenging. This paper proposes a novel MPC approach with a scheduled quadratic stage cost function that approximates the true stage cost in order to optimally control a nonlinear system with a sparse/binary stage cost. The cost function parameter is optimally scheduled by a parameter scheduling policy obtained by solving a Markov decision process (MDP) constructed from sampled trajectories from any nonlinear MPC solver. The proposed approach is implemented into a differential drive wheeled mobile robot (WMR) designed for smart warehousing via the robot operating system (ROS) framework. The simulation and experimental results successfully demonstrate the effectiveness of our MPC approach in cases of the point stabilization problem of a differential drive WMR.
|
|
11:10-11:20, Paper TuA-17.8 | |
Caterpillar-Inspired Insect-Scale Climbing Robot Using Dry Adhesives |
|
Khalil, Mohamed M. | Toyohashi University of Technology |
Mashimo, Tomoaki | Okayama University |
Keywords: Climbing Robots, Micro/Nano Robots
Abstract: Surface-to-surface transitions, vertical and inverted locomotion, and high payload capacity are important requirements for a climbing robot. Despite many previous approaches, the miniaturization of climbing robots that satisfy these requirements is still a big challenge. In this paper, we present an insect-scale wheeled climbing robot that employs a low-cost dry adhesive technology. The design, inspired by caterpillars, consists of 2 main links that are connected by a pivot joint. The prototype robot measures a length of 40 mm and a width of 10 mm, and it weighs 1.7 g. On a horizontal surface, the robot moves with a speed of 12.3 mm/s and can drag a load weight of 10 g (six times its weight) at a speed of 4 mm/s. The high torque-to-weight ratio achieved by two micro-geared ultrasonic motors permits vertical and inverted locomotion while carrying a high payload capacity (120% of the robot’s weight). The locomotion strategy for surface-to-surface transitions, i.e., movements at right angle corners, is substantiated by the functionality of the pivot joint. To satisfy the design requirements, many dry adhesive tapes are evaluated, and the climbing robot with an appropriate adhesion force is demonstrated.
|
|
11:20-11:30, Paper TuA-17.9 | |
Electro-Hydraulic Rolling Soft Wheel: Design, Hybrid Dynamic Modeling, and Model Predictive Control (I) |
|
Ly, Khoi | University of Colorado |
Mayekar, Jatin | University of Colorado Boulder, ABB |
Aguasvivas Manzano, Sarah | University of Colorado Boulder |
Keplinger, Christoph | University of Colorado |
Rentschler, Mark | University of Colorado at Boulder |
Correll, Nikolaus | University of Colorado at Boulder |
Keywords: Product Design, Development and Prototyping, Dynamics, Optimization and Optimal Control
Abstract: Locomotion through rolling is attractive compared to other forms of locomotion thanks to uniform designs, high degree of mobility, dynamic stability, and self-recovery from collision. Despite previous efforts to design rolling soft systems, pneumatic and other soft actuators are often limited in terms of high-speed dynamics, system integration, and/or functionalities. Furthermore, mathematical description of the rolling dynamics for this type of robot and how the models can be used for speed control are often not mentioned. This article introduces a cylindrical-shaped shell-bulging rolling soft wheel that employs an array of 16 folded HASEL actuators as a mean for improved rolling performance. The actuators represent the soft components with discrete forces that propel the wheel, whereas the wheel’s frame is rigid but allows for smooth, continuous change in position and speed. We discuss the interplay between the electrical and mechanical design choices, the modeling of the wheel’s hybrid (continuous and discrete) dynamic behavior, and the implementation of a model predictive controller (MPC) for the robot’s speed. With the balance of several design factors, we show the wheel’s ability to carry integrated hardware with a maximum rolling speed at 0.7 m/s (or 2.2 body lengths per second), despite its total weight of 979 g, allowing the wheel to outperform the existing rolling soft wheels with comparable weights and sizes. We also show that the MPC enables the wheel to accelerate and leverage its inherent braking capability to reach desired speeds—a critical function that did not exist in previous rolling soft systems.
|
|
TuA-18 |
Rm18 (Room 554) |
Motion and Path Planning 4 |
Regular session |
Chair: Julien, Leclerc | University of Houston |
Co-Chair: Muthugala Arachchige, Viraj Jagathpriya Muthugala | Singapore University of Technology and Design |
|
10:00-10:10, Paper TuA-18.1 | |
Linear MPC-Based Motion Planning for Autonomous Surgery |
|
Minelli, Marco | University of Modena and Reggio Emilia |
Sozzi, Alessio | University of Ferrara |
De Rossi, Giacomo | University of Verona |
Ferraguti, Federica | Università Degli Studi Di Modena E Reggio Emilia |
Farsoni, Saverio | University of Ferrara |
Setti, Francesco | University of Verona |
Muradore, Riccardo | University of Verona |
Bonfe, Marcello | University of Ferrara |
Secchi, Cristian | Univ. of Modena & Reggio Emilia |
Keywords: Motion and Path Planning, Medical Robots and Systems, Optimization and Optimal Control
Abstract: Within the context of Robotic Minimally Invasive Surgery (R-MIS), we propose a novel linear model predictive controller formulation for the coordination of multiple autonomous robotic arms. The controller is synthesized by formulating a linear approximation of non-linear constraints, which allows the controller to be both computationally faster and better performing due to the increased prediction horizon allowed within the real-time control requirements for the proposed surgical application. The solution is validated under the expected constraints of a surgical scenario in which multiple laparoscopic tools must move and coordinate in a shared environment.
|
|
10:10-10:20, Paper TuA-18.2 | |
Disk-Graph Probabilistic Roadmap: Biased Distance Sampling for Path Planning in a Partially Unknown Environment |
|
Noël, Thibault | INRIA Rennes |
Kabbour, Salah Eddine | GroupeCreative |
Lehuger, Antoine Lehuger | Groupe Créative |
Marchand, Eric | Univ Rennes, Inria, CNRS, IRISA |
Chaumette, Francois | Inria Rennes-Bretagne Atlantique |
Keywords: Motion and Path Planning, Range Sensing, Autonomous Agents
Abstract: In this paper, we propose a new sampling-based path planning approach, focusing on the challenges linked to autonomous exploration. Our method relies on the definition of a disk graph of free-space bubbles, from which we derive a biased sampling function that expands the graph towards known free space for maximal navigability and frontiers discovery. The proposed method demonstrates an exploratory behavior similar to Rapidly-exploring Random Trees, while retaining the connectivity and flexibility of a graph-based planner. We demonstrate the interest of our method by first comparing its path planning capabilities against state-of-the-art approaches, before discussing exploration-specific aspects, namely replanning capabilities and incremental construction of the graph. A simple frontiers-driven exploration controller derived from our planning method is also demonstrated using the Pioneer platform.
|
|
10:20-10:30, Paper TuA-18.3 | |
Elevation State-Space: Surfel-Based Navigation in Uneven Environments for Mobile Robots |
|
Atas, Fetullah | National Taipei University of Technology |
Cielniak, Grzegorz | University of Lincoln |
Grimstad, Lars | Norwegian University of Life Sciences |
Keywords: Motion and Path Planning, Field Robots
Abstract: This paper introduces a new method for robot motion planning and navigation in uneven environments through a surfel representation of underlying point clouds. The proposed method addresses the shortcomings of state-of-the-art navigation methods by incorporating both kinematic and physical constraints of a robot with standard motion planning algorithms (e.g., those from the Open Motion Planning Library), thus enabling efficient sampling-based planners for challenging uneven terrain navigation on raw point cloud maps. Unlike techniques based on Digital Elevation Maps (DEMs), our novel surfel-based state-space formulation and implementation are based on raw point cloud maps, allowing for the modeling of overlapping surfaces such as bridges, piers, and tunnels. Experimental results demonstrate the robustness of the proposed method for robot navigation in real and simulated unstructured environments. The proposed approach also optimizes planners' performances by boosting their success rates up to 5x for challenging unstructured terrain planning and navigation, thanks to our surfel-based approach's robot constraint-aware sampling strategy. Finally, we provide an open-source implementation of the proposed method to benefit the robotics community.
|
|
10:30-10:40, Paper TuA-18.4 | |
Locomotion Policy Guided Traversability Learning Using Volumetric Representations of Complex Environments |
|
Frey, Jonas | ETH Zurich |
Höller, David | ETH |
Khattak, Shehryar | ETH Zurich |
Hutter, Marco | ETH Zurich |
Keywords: Motion and Path Planning, Legged Robots, Deep Learning Methods
Abstract: Despite the progress in legged robotic locomotion, autonomous navigation in unknown environments remains an open problem. Ideally, the navigation system utilizes the full potential of the robots' locomotion capabilities while operating within safety limits under uncertainty. The robot must sense and analyze the traversability of the surrounding terrain, which depends on the hardware, locomotion control, and terrain properties. It may contain information about the risk, energy, or time consumption needed to traverse the terrain. To avoid hand-crafted traversability cost functions we propose to collect traversability information about the robot and locomotion policy by simulating the traversal over randomly generated terrains using a physics simulator. Thousand of robots are simulated in parallel controlled by the same locomotion policy used in reality to acquire 57 years of real-world locomotion experience equivalent. For deployment on the real robot, a sparse convolutional network is trained to predict the simulated traversability cost, which is tailored to the deployed locomotion policy, from an entirely geometric representation of the environment in the form of a 3D voxel-occupancy map. This representation avoids the need for commonly used elevation maps, which are error-prone in the presence of overhanging obstacles and multi-floor or low-ceiling scenarios. The effectiveness of the proposed traversability prediction network is demonstrated for path planning for the legged robot ANYmal in various indoor and natural environments.
|
|
10:40-10:50, Paper TuA-18.5 | |
Jerk-Continuous Online Trajectory Generation for Robot Manipulator with Arbitrary Initial State and Kinematic Constraints |
|
Zhao, Haoran | University of Houston |
Abdurahiman, Nihal | Hamad Medical Corporation |
Navkar, Nikhil | Hamad Medical Corporation |
Julien, Leclerc | University of Houston |
Becker, Aaron | University of Houston |
Keywords: Motion and Path Planning, Industrial Robots, Manipulation Planning
Abstract: This work presents an online trajectory generation algorithm using a sinusoidal jerk profile. The generator takes initial acceleration, velocity and position as input, and plans a multi-segment trajectory to a goal position under jerk, acceleration, and velocity limits. By analyzing the critical constraints and conditions, the corresponding closed-form solution for the time factors and trajectory profiles are derived. The proposed algorithm was first derived in Mathematica and then converted into a C++ implementation. Finally, the algorithm was utilized and demonstrated in ROS & Gazebo using a UR3 robot. Both the Mathematica and C++ implementations can be accessed at https://github.com/Haoran-Zhao/Jerk-continuous-online-trajectory-generator-with-constraints.git
|
|
10:50-11:00, Paper TuA-18.6 | |
Imitation Learning and Model Integrated Excavator Trajectory Planning |
|
Guo, Qiangqiang | Washington University |
Ye, Zhixian | Baidu |
Wang, Liyang | Baidu USA |
Zhang, Liangjun | Baidu |
Keywords: Motion and Path Planning, Robotics and Automation in Construction, Imitation Learning
Abstract: Automated excavation is promising to improve the safety and efficiency of excavators, and trajectory planning is one of the most important techniques. In this paper, we propose a two-stage method that integrates data-driven imitation learning and model-based trajectory optimization to generate optimal trajectories for autonomous excavators. We firstly train a deep neural network using demonstration data to mimic the operation patterns of human experts under various terrain states including their geometry shape and material type. Then, for a particular terrain, we use a stochastic trajectory optimization method to improve the trajectory generated by the neural network to guarantee kinematics feasibility, improve smoothness, satisfy hard constraints, and achieve desired excavation volumes. We test the proposed algorithm on a Franka robot arm. The experimental results show that the proposed two-stage algorithm by combing expert knowledge and model optimization can increase the excavation weights by up to 24.77% meanwhile with low variance.
|
|
11:00-11:10, Paper TuA-18.7 | |
Online Complete Coverage Path Planning of a Reconfigurable Robot Using Glasius Bio-Inspired Neural Network and Genetic Algorithm |
|
Samarakoon Mudiyanselage, Bhagya Prasangi Samarakoon | Singapore University of Technology and Design |
Muthugala Arachchige, Viraj Jagathpriya Muthugala | Singapore University of Technology and Design |
Elara, Mohan Rajesh | Singapore University of Technology and Design |
Keywords: Motion and Path Planning, Service Robotics, AI-Based Methods
Abstract: Area coverage is crucial for robotics applications such as cleaning, painting, exploration, and inspections. Hinged reconfigurable robots have been introduced for these application domains to improve the area coverage performance. However, the existing coverage algorithms of hinged reconfigurable robots require improvements in the aspects; consideration of beyond a limited set of reconfigurable shapes, coordinated reconfiguration and navigation, and online decision-making. Therefore, this paper proposes a novel online Complete Coverage Path Planning (CCPP) method for a hinged reconfigurable robot. The proposed CCPP method is designed with two sub-methods, the Global Coverage Path Planning (GCPP) and Local Coverage Path Planning (LCPP). The GCPP method has been implemented, adapting a Glasius Bio-inspired Neural Network (GBNN) that performs online path planning considering a fixed shape for the robot. Obstacle regions that the GCPP would not adequately cover due to access constraints are covered by the LCPP method that considers concurrent reconfiguration and navigation of the robot. A genetic algorithm determines the reconfiguration parameters that ascertain collision-free coverage and access of obstacle regions. Experimental results validate that the proposed online CCPP method is effective in ascertaining the complete area coverage in heterogeneous environments, including dynamic workspaces. Furthermore, the deployment of the LCPP method can considerably improve the coverage.
|
|
11:10-11:20, Paper TuA-18.8 | |
Quantity Over Quality: Training an AV Motion Planner with Large Scale Commodity Vision Data |
|
Platinsky, Lukas | Lyft |
Naseer, Tayyab | Woven Planet |
Chen, Hui | Woven Planet |
Haines, Ben | Woven Planet |
Zhu, Haoyue | Woven Planet |
Grimmett, Hugo | Lyft |
Del Pero, Luca | Lyft |
Keywords: Motion and Path Planning, Autonomous Vehicle Navigation, Deep Learning Methods
Abstract: With the Autonomous Vehicle (AV) industry shifting towards machine-learned approaches for motion planning, the performance of self-driving systems is starting to rely heavily on large quantities of expert driving demonstrations. However, collecting this demonstration data typically involves expensive HD sensor suites (LiDAR + RADAR + cameras), which quickly becomes financially infeasible at the scales required. This motivates the use of commodity sensors like cameras for data collection, which are an order of magnitude cheaper than HD sensor suites, but offer lower fidelity. Leveraging these sensors for training an AV motion planner opens a financially viable path to observe the 'long tail' of driving events. As our main contribution we show it is possible to train a high-performance motion planner using commodity vision data which outperforms planners trained on HD-sensor data for a fraction of the cost. To the best of our knowledge, we are the first to demonstrate this using real-world data. We compare the performance of the autonomy system on these two different sensor configurations, and show that we can compensate for the lower sensor fidelity by means of increased quantity: a planner trained on 100h of commodity vision data outperforms the one with 25h of expensive HD data. We also share the engineering challenges we had to tackle to make this work.
|
|
11:20-11:30, Paper TuA-18.9 | |
TIGRIS: An Informed Sampling-Based Algorithm for Informative Path Planning |
|
Moon, Brady | Carnegie Mellon University |
Chatterjee, Satrajit | Carnegie Mellon University |
Scherer, Sebastian | Carnegie Mellon University |
Keywords: Motion and Path Planning, Field Robots
Abstract: Informative path planning is an important and challenging problem in robotics that remains to be solved in a manner that allows for wide-spread implementation and real-world practical adoption. Among various reasons for this, one is the lack of approaches that allow for informative path planning in high-dimensional spaces and non-trivial sensor constraints. In this work we present a sampling-based approach that allows us to tackle the challenges of large and high-dimensional search spaces. This is done by performing informed sampling in the high-dimensional continuous space and incorporating potential information gain along edges in the reward estimation. This method rapidly generates a global path that maximizes information gain for the given path budget constraints. We discuss the details of our implementation for an example use case of searching for multiple objects of interest in a large search space using a fixed-wing UAV with a forward-facing camera. We compare our approach to a sampling-based planner baseline and demonstrate how our contributions allow our approach to consistently out-perform the baseline by 18.0%. With this we thus present a practical and generalizable informative path planning framework that can be used for very large environments, limited budgets, and high dimensional search spaces, such as robots with motion constraints or high-dimensional configuration spaces.
|
|
TuA-19 |
Rm19 (Room 555) |
Biologically-Inspired Robots 4 |
Regular session |
Chair: Sartoretti, Guillaume Adrien | National University of Singapore (NUS) |
Co-Chair: Nakamura, Taro | Chuo University |
|
10:00-10:10, Paper TuA-19.1 | |
Online Learning Feedback Control Method Considering Hysteresis for Musculoskeletal Structures |
|
Kawaharazuka, Kento | The University of Tokyo |
Okada, Kei | The University of Tokyo |
Inaba, Masayuki | The University of Tokyo |
Keywords: Biomimetics, Learning from Experience, Tendon/Wire Mechanism
Abstract: While the musculoskeletal humanoid has various biomimetic benefits, its complex modeling is difficult, and many learning control methods have been developed. However, for the actual robot, the hysteresis of its joint angle tracking is still an obstacle, and realizing target posture quickly and accurately has been difficult. Therefore, we develop a feedback control method considering the hysteresis. To solve the problem in feedback controls caused by the closed-link structure of the musculoskeletal body, we update a neural network representing the relationship between the error of joint angles and the change in target muscle lengths online, and realize target joint angles accurately in a few trials. We compare the performance of several configurations with various network structures and loss definitions, and verify the effectiveness of this study on an actual musculoskeletal humanoid, Musashi.
|
|
10:10-10:20, Paper TuA-19.2 | |
Realization of Seated Walk by a Musculoskeletal Humanoid with Buttock-Contact Sensors from Human Constrained Teaching |
|
Kawaharazuka, Kento | The University of Tokyo |
Okada, Kei | The University of Tokyo |
Inaba, Masayuki | The University of Tokyo |
Keywords: Biomimetics, Soft Sensors and Actuators, Tendon/Wire Mechanism
Abstract: In this study, seated walk, a movement of walking while sitting on a chair with casters, is realized on a musculoskeletal humanoid from human teaching. The body is balanced by using buttock-contact sensors implemented on the planar interskeletal structure of the human mimetic musculoskeletal robot. Also, we develop a constrained teaching method in which one-dimensional control command, its transition, and a transition condition are described for each state in advance, and a threshold value for each transition condition such as joint angles and foot contact sensor values is determined based on human teaching. Complex behaviors can be easily generated from simple inputs. In the musculoskeletal humanoid MusashiOLegs, forward, backward, and rotational movements of seated walk are realized.
|
|
10:20-10:30, Paper TuA-19.3 | |
A Control Architecture of a Distributed Actuator System for a Bio-Inspired Spine |
|
Ku, Bonhyun | University of Illinois at Urbana-Champaign |
Banerjee, Arijit | University of Illinois at Urbana-Champaign |
Keywords: Biologically-Inspired Robots, Biomimetics, Actuation and Joint Mechanisms
Abstract: Control of an articulated spine is important for humanoids' dynamic and balanced motion. Although there have been many spinal structures for humanoids, their actuation is still limited due to the usage of geared-motors for joints. This paper introduces position control of a distributed electromechanical spine in a vertical plane. The spine dynamics model is approximated as an open chain. Gravitational and spring torques are compensated for the control. Moreover, torque-to-current conversion for the actuator is developed. Experimental results show the implemented control of the electromechanical spine for undulatory motions.
|
|
10:30-10:40, Paper TuA-19.4 | |
Development of a Conveyor-Type Object Release Mechanism for a Parallel Gripper with a Mushroom-Shaped Gecko-Inspired Surface |
|
Nagahama, Shunsuke | Waseda University |
Nakao, Atsushi | Waseda University |
Sugano, Shigeki | Waseda University |
Keywords: Grasping, Biologically-Inspired Robots, Actuation and Joint Mechanisms
Abstract: It is known that the surface microstructure, which mimics the surface of a gecko's foot, exerts a large gripping force with a tiny contact force. By applying this structure to the fingertips of a two-fingered parallel gripper, stable grasping can be achieved independent of the wetting and frictional state of the contact surface. However, when releasing an object, the adhesive force of the microstructure is large, which hinders the release of the object. In this study, we developed a releasing mechanism using a conveyor mechanism, focusing on the characteristic of the micro-protrusion structure that easily peels off in the direction of rotation. This mechanism is driven in conjunction with the gripper's grasping and releasing motions, and experiments have confirmed that it can stably release the object.
|
|
10:40-10:50, Paper TuA-19.5 | |
Instantaneous Force Generation Mechanism Based on the Striking Motion of Mantis shrimp—Design and Control Method of Cavitation by Simulation and Experiment |
|
Ito, Fumio | Chuo University |
Ishii, Yusuke | Chuo University |
Kurumaya, Shunichi | Chuo University |
Kagaya, Katsushi | The University of Tokyo |
Nakamura, Taro | Chuo University |
Keywords: Biomimetics, Soft Robot Applications, Dynamics
Abstract: In this paper, we describe simulations and experiments on the occurrence of cavitation with an instantaneous force generation mechanism based on the motion of the mantis shrimp. The aim of this study is to strengthen the striking force of the developed mechanism by realizing the mechanism of cavitation occurrence. This paper presents a motion and cavitation model for the developed mechanism and provides the results of the simulations and experiments considering the occurrence of cavitation. In previous studies, mantis shrimp have been shown to have two consecutive impacts on a hard object with cavitation in water. However, no robot can imitate the striking mechanism of the mantis shrimp and replicate two consecutive impacts. In this study, the mechanism was designed and developed based on a mantis shrimp model. From the experiment, it was observed that the developed mechanism caused cavitation only at the point of impact, rather than during arm swing. This result suggested that the striking force was strengthened, providing insight into the striking mechanism of the mantis shrimp.
|
|
10:50-11:00, Paper TuA-19.6 | |
Upside-Down Brachiation Robot Using Elastic Energy Stored through Soft Body Deformation |
|
Nonoyama, Kisuke | Shinshu University |
Shimizu, Masahiro | Osaka University |
Umedachi, Takuya | Shinshu University |
Keywords: Soft Robot Applications, Soft Robot Materials and Design, Flexible Robotics
Abstract: Storing elastic energy in a soft body and releasing it instantly enables ultrafast movement beyond muscle capability, which small animals (mainly arthropods) realize. We applied this mechanism to a soft robot to achieve locomotion on top of a tubular surface such as a branch (i.e., upside-down brachiation). To achieve arboreal locomotion, the robot must have strong grippers to support its body and minimize the duration of time when one gripper is off the branch. By using a simulation model and prototype, we demonstrated that storing and releasing elastic energy in a soft body satisfies these requirements, allowing us to fabricate a lightweight robot. The prototype completed one step of the locomotion process in 0.22 secs, making it 2.3 times faster than the original speed of the mounted motor. In addition, we confirmed that the robot climbs 0- to 90-degree branches.
|
|
11:00-11:10, Paper TuA-19.7 | |
Learning of Balance Controller Considering Changes in Body State for Musculoskeletal Humanoids |
|
Kawaharazuka, Kento | The University of Tokyo |
Ribayashi, Yoshimoto | JSK Robotics Lab |
Miki, Akihiro | The University of Tokyo |
Toshimitsu, Yasunori | University of Tokyo |
Suzuki, Temma | The University of Tokyo |
Okada, Kei | The University of Tokyo |
Inaba, Masayuki | The University of Tokyo |
Keywords: Biomimetics, Learning from Experience, Bioinspired Robot Learning
Abstract: The musculoskeletal humanoid is difficult to modelize due to the flexibility and redundancy of its body, whose state can change over time, and so balance control of its legs is challenging. There are some cases where ordinary PID controls may cause instability. In this study, to solve these problems, we propose a method of learning a correlation model among the joint angle, muscle tension, and muscle length of the ankle and the zero moment point to perform balance control. In addition, information on the changing body state is embedded in the model using parametric bias, and the model estimates and adapts to the current body state by learning this information online. This makes it possible to adapt to changes in upper body posture that are not directly taken into account in the model, since it is difficult to learn the complete dynamics of the whole body considering the amount of data and computation. The model can also adapt to changes in body state, such as the change in footwear and change in the joint origin due to recalibration. The effectiveness of this method is verified by a simulation and by using an actual musculoskeletal humanoid, Musashi.
|
|
11:10-11:20, Paper TuA-19.8 | |
Locomotion Via Active Suction in a Sea Star-Inspired Soft Robot |
|
Ishida, Michael | University of California, San Diego |
Sandoval, Jessica | University of California San Diego |
Lee, Sebastian | University of California Berkeley |
Huen, Sidney | University of California, San Diego |
Tolley, Michael T. | University of California, San Diego |
Keywords: Biologically-Inspired Robots, Soft Robot Applications, Hydraulic/Pneumatic Actuators
Abstract: Some marine animals, such as sea stars, have developed versatile adhesive appendages that can be used for a variety of behaviors including sticking to surfaces, manipulating objects, and locomoting. These appendages couple reversible adhesion capabilities with muscular structures that can do work on the world while still being soft enough to conform to various surfaces and to squeeze through tight spaces. We took inspiration from the hydraulic tube feet of sea star to create modules consisting of soft active suction discs and soft pneumatic linear actuators. Tube feet convert fluid motion into linear actuation using soft tubes that are constrained radially. In this work, we study the tradeoff between the stiffness of such radial constraints, and the ability of the linear actuators to apply axial forces. The adhesive force of the suction disc is dependent on both the pressure differential between the suction cavity and the environment and the deformation within the body of the disc while it is being loaded. We show that the tube foot modules are capable of creating locomotion over flat surfaces, up small steps, and in a confined space due to the redundancy caused by an array of actuators. We also show that active suction not only enables the use of softer tube foot actuators, but sensing the active suction line provides feedback to increase the efficiency of locomotion over obstacles.
|
|
11:20-11:30, Paper TuA-19.9 | |
Joint-Space CPG for Safe Foothold Planning and Body Pose Control During Locomotion and Climbing |
|
Sun, Ge | National University of Singapore |
Sartoretti, Guillaume Adrien | National University of Singapore (NUS) |
Keywords: Biologically-Inspired Robots, Legged Robots, Biomimetics
Abstract: From insects to larger mammals, legged animals can be seen easily traversing a wide variety of challenging environments, by carefully selecting, reaching, and exploiting high-quality contacts with the terrain. In contrast, existing robotic foothold planning methods remain computationally expensive, often relying on exhaustive search and/or (often offline) optimization methods, thus limiting their adoption for real-life robotic deployments. In this work, we propose a low-cost, bio-inspired foothold planning method for legged robots, which replicates the mechanism of the central nervous system of legged mammals. We develop a low-level joint-space CPG model along with a high-level vision-based controller that can inexpensively predict future foothold locations and locally optimize them based on a potential field based approach. Specifically, by reasoning about the quality of ground contacts and the robot's stability through the high-level vision-based controller, our CPG model smoothly and iteratively updates relevant locomotion parameters to both optimize foothold locations and body pose, directly in the joint space of the robot for easier implementation. We experimentally validate our control model on a modular hexapod on various locomotion tasks in obstacle-rich environments as well as on stair climbing. Our results show that our method enables stabler and steadier locomotion, yielding higher-quality feedback from onboard sensors by minimizing the effect of slippage and unexpected impacts.
|
|
TuA-20 |
Rm20 (Room 104) |
Distributed Robot Systems |
Regular session |
Chair: Roa, Maximo A. | DLR - German Aerospace Center |
Co-Chair: Prince Mathew, Joseph | Geroge Mason University |
|
10:00-10:10, Paper TuA-20.1 | |
Real-Time Distributed Multi-Robot Target Tracking Via Virtual Pheromones |
|
Prince Mathew, Joseph | Geroge Mason University |
Nowzari, Cameron | University of Pennsylvania |
Keywords: Distributed Robot Systems, Search and Rescue Robots, Autonomous Vehicle Navigation
Abstract: Actively searching for targets using a multi-agent system in an unknown environment poses a two-pronged problem, where on the one hand we need agents to cover as much of the environment as possible and on the other have a higher density of agents where there are potential targets to maximize detection performance. This paper proposes a fully distributed solution for an ad hoc network of agents to cooperatively search an unknown environment and actively track found targets. The solution combines a distributed pheromone-based coverage control strategy with a distributed target selection mechanism.
|
|
10:10-10:20, Paper TuA-20.2 | |
Conservative Filtering for Heterogeneous Decentralized Data Fusion in Dynamic Robotic Systems |
|
Dagan, Ofer | University of Colorado Boulder |
Ahmed, Nisar | University of Colorado Boulder |
Keywords: Distributed Robot Systems, Multi-Robot Systems, Sensor Fusion
Abstract: This paper presents a method for Bayesian multi-robot peer-to-peer data fusion where any pair of autonomous robots hold non-identical, but overlapping parts of a global joint probability distribution, representing real world inference tasks (e.g., mapping, tracking). It is shown that in dynamic stochastic systems, filtering, which corresponds to marginalization of past variables, results in direct and hidden dependencies between variables not mutually monitored by the robots, which might lead to an overconfident fused estimate. The paper makes both theoretical and practical contributions by providing (i) a rigorous analysis of the origin of the dependencies and (ii) a conservative filtering algorithm for heterogeneous data fusion in dynamic systems that can be integrated with existing fusion algorithms. This work uses factor graphs as both the analysis tool and the inference engine. Each robot in the network maintains a local factor graph and communicates only relevant parts of it (a sub-graph) to its neighboring robot. We discuss the applicability to various multi-robot robotic applications and demonstrate the performance using a multi-robot multi-target tracking simulation, showing that the proposed algorithm produces conservative estimates at each robot.
|
|
10:20-10:30, Paper TuA-20.3 | |
Decay-Based Error Correction in Collective Robotic Construction |
|
Chen, Jiahe | Cornell University |
Petersen, Kirstin Hagelskjaer | Cornell University |
Keywords: Distributed Robot Systems, Swarm Robotics, Assembly
Abstract: Multi-robot systems have been shown to build large-scale, blueprint-specific structures using distributed, environmentally-mediated coordination. Little attention, however, has been devoted to error propagation and mitigation. In this paper, we introduce a detailed simulation of TERMES, a prototypical construction system, in which robots have realistic error profiles. We use this simulator and 32 randomly generated 250-brick blueprints to show that action errors can have significant long-term effects. We study the spatio-temporal error distribution and introduce and characterize the efficacy of a simple decay-based error correction mechanism. Although inefficient, this type of error correction is promising because it can be performed by robots with the same limited sensory capabilities as those who place bricks. To limit the impact on the construction rate, we also examine decay mechanisms informed by spatial and temporal error distributions. The incorporation of decay in our building process increases the probability of successful completion by ~4, at the expense of ~1/4 decrease in construction rate.
|
|
10:30-10:40, Paper TuA-20.4 | |
Perceive, Represent, Generate: Translating Multimodal Information to Robotic Motion Trajectories |
|
Morgado Vital, Fábio Guilherme | INESC-ID |
Vasco, Miguel | INESC-ID |
Sardinha, Alberto | INESC-ID and IST |
Melo, Francisco S. | Instituto Superior Tecnico |
Keywords: Multi-Modal Perception for HRI, Deep Learning Methods, Human-Robot Collaboration
Abstract: We present Perceive-Represent-Generate (PRG), a novel three-stage framework that maps perceptual information of different modalities (e.g., visual or sound), corresponding to a series of instructions, to a sequence of movements to be executed by a robot. In the first stage, we perceive and preprocess the given inputs, isolating individual commands from the complete instruction provided by a human user. In the second stage we encode the individual commands into a multimodal latent space, employing a deep generative model. Finally, in the third stage we convert the latent samples into individual trajectories and combine them into a single dynamic movement primitive, allowing its execution by a robotic manipulator. We evaluate our pipeline in the context of a novel robotic handwriting task, where the robot receives as input a word through different perceptual modalities (e.g., image, sound), and generates the corresponding motion trajectory to write it, creating coherent and high-quality handwritten words.
|
|
10:40-10:50, Paper TuA-20.5 | |
SMA-NBO: A Sequential Multi-Agent Planning with Nominal Belief-State Optimization in Target Tracking |
|
Li, Tianqi | Texas A&M University |
Krakow, Lucas | Texas A&M University |
Gopalswamy, Swaminathan | Texas A&M University |
Keywords: Distributed Robot Systems, Cooperating Robots, Reactive and Sensor-Based Planning
Abstract: In target tracking with mobile multi-sensor systems, sensor deployment impacts the observation capabilities and the resulting state estimation quality. Based on a partially observable Markov decision process (POMDP) formulation comprised of the observable sensor dynamics, unobservable target states, and accompanying observation laws, we present a distributed information-driven solution approach to the multi-agent target tracking problem, namely, sequential multi-agent nominal belief-state optimization (SMA-NBO). SMA-NBO seeks to minimize the expected tracking error via receding horizon control including a heuristic expected cost-to-go (HECTG). SMA-NBO incorporates a computationally efficient approximation of the target belief-state over the horizon. The agent-by-agent decision-making is capable of leveraging on-board (edge) compute for selecting (sub-optimal) target-tracking maneuvers exhibiting non-myopic cooperative fleet behavior. The optimization problem explicitly incorporates semantic information defining target occlusions from a world model. To illustrate the efficacy of our approach, a random occlusion forest environment is simulated. SMA-NBO is compared to other baseline approaches. The simulation results show SMA-NBO 1) maintains tracking performance and reduces the computational cost by replacing the calculation of the expected target trajectory with a single sample trajectory based on maximum a posteriori estimation; 2) generates cooperative fleet decision by sequentially optimizing single-agent policy with efficient usage of other agents' policy of intent; 3) aptly incorporates the multiple weighted trace penalty (MWTP) HECTG, which improves tracking performance with a computationally efficient heuristic.
|
|
10:50-11:00, Paper TuA-20.6 | |
Influence of Variable Leg Elasticity on the Stability of Quadrupedal Gaits |
|
Del Fatti, Federico | Politecnico Di MIlano |
Sesselmann, Anna | DLR - German Aerospace Center |
Roa, Maximo A. | DLR - German Aerospace Center |
Keywords: Natural Machine Motion, Legged Robots, Passive Walking
Abstract: Several template models have been developed to facilitate the analysis of limit-cycles for quadrupedal locomotion. The parameters in the model are usually fixed; however, biology shows that animals change their leg stiffness according to the locomotion velocity, and this adaptability invariably affects the stability of the gait. This paper provides an analysis of the influence of this variable leg stiffness on the stability of different quadrupedal gaits. The analysis exploits a simplified quadrupedal model with compliant legs and shoulder joints represented as torsional springs. This model can reproduce the most common quadrupedal gaits observed in nature. The stability of such emerging gaits is then checked. Afterward, an optimization process is used to search for the system parameters that guarantee maximum gait stability. Our study shows that using the highest feasible leg swing frequency and adopting a leg stiffness that increases with the speed of locomotion noticeably improves the gait stability over a wide range of horizontal velocities while reducing the oscillations of the trunk. This insight can be applied in the design of novel elastic quadrupedal robots, where variable stiffness actuators could be employed to improve the overall locomotion behavior.
|
|
11:00-11:10, Paper TuA-20.7 | |
Learning to Act with Affordance-Aware Multimodal Neural SLAM |
|
Jia, Zhiwei | UC San Diego |
Lin, Kaixiang | Amazon |
Zhao, Yizhou | UCLA |
Gao, Qiaozi | Amazon |
Thattai, Govind | Amazon |
Sukhatme, Gaurav | University of Southern California |
Keywords: Multi-Modal Perception for HRI, SLAM, Integrated Planning and Learning
Abstract: Recent years have witnessed an emerging paradigm shift toward embodied artificial intelligence, in which an agent must learn to solve challenging tasks by interacting with its environment. There are several challenges in solving embodied multimodal tasks, including long-horizon planning, vision-and-language grounding, and efficient exploration. We focus on a critical bottleneck, namely the performance of planning and navigation. To tackle this challenge, we propose a Neural SLAM approach that, for the first time, utilizes several modalities for exploration, predicts an affordance-aware semantic map, and plans over it at the same time. This significantly improves exploration efficiency, leads to robust long- horizon planning, and enables effective vision-and-language grounding. With the proposed Affordance-aware Multimodal Neural SLAM (AMSLAM) approach, we obtain more than 40% improvement over prior published work on the ALFRED benchmark and set a new state-of-the-art generalization performance at a success rate of 23.48% on the test unseen scenes.
|
|
11:10-11:20, Paper TuA-20.8 | |
Learning to Assess Danger from Movies for Cooperative Escape Planning in Hazardous Environments |
|
Shree, Vikram | Cornell University |
Allen, Sarah | Cornell University |
Arruda Asfora, Beatriz | Cornell University |
Banfi, Jacopo | Massachusetts Institute of Technology |
Campbell, Mark | Cornell University |
Keywords: Multi-Modal Perception for HRI, Search and Rescue Robots
Abstract: There has been a plethora of work towards improving robot perception and navigation, yet their application in hazardous environments, like during a fire or an earthquake, is still at a nascent stage. We hypothesize two key challenges here: first, it is difficult to replicate such scenarios in the real world, which is necessary for training and testing purposes. Second, current systems are not fully able to take advantage of the rich multi-modal data available in such hazardous environments. To address the first challenge, we propose to harness the enormous amount of visual content available in the form of movies and TV shows, and develop a dataset that can represent hazardous environments encountered in the real world. The data is annotated with high-level danger ratings for realistic disaster images, and corresponding keywords are provided that summarize the content of the scene. In response to the second challenge, we propose a multi-modal danger estimation pipeline for collaborative human-robot escape scenarios. Our Bayesian framework improves danger estimation by fusing information from robot's camera sensor and language inputs from the human. Furthermore, we augment the estimation module with a risk-aware planner that helps in identifying safer paths out of the dangerous environment. Through extensive simulations, we exhibit the advantages of our multi-modal perception framework that gets translated into tangible benefits such as higher success rate in a collaborative human-robot mission.
|
|
11:20-11:30, Paper TuA-20.9 | |
Distributed Deployment with Multiple Moving Robots for Long Distance Complex Pipe Inspection |
|
Sato, Hiroto | Chuo University |
Uchiyama, Kosuke | Chuo University |
Ito, Fumio | Chuo University |
Sawahashi, Ryunosuke | Chuo University |
Nakamura, Taro | Chuo University |
Keywords: Distributed Robot Systems, Biomimetics, Flexible Robotics
Abstract: In this study, we propose a method to reduce the traveling load by designing a distributed robot arrangement that considers the shape of a pipeline for a cabled in-pipe mobile robot. The objective of this study is to establish a method to reduce the friction acting on the wires of a self-propelled in-pipe mobile robot and develop a robotic system that can inspect long-distance pipes. This study contributes the purpose of this research in that it proposes a model for estimating the tension acting on the wires of a mobile robot in a pipe and describes a method for reducing the traveling load using a distributed arrangement of robots. The friction acting on the wires in the bent pipe becomes a load for the robots moving in the pipe while towing the wires. Therefore, it is difficult for these robots to inspect complex, long-distance pipelines. Furthermore, we clarified the factors that cause friction forces in long-distance pipes and modeled and quantified the friction forces acting on the wires in the pipes so that each factor could be addressed. Using this model, we propose a distributed deployment method, in which multiple mobile robots are deployed. By driving two robots separated by a distance, we confirmed that two robots could move at 2.5 mm/s in a pipeline that could not be moved by a single robot. Moreover, we proposed a method for controlling the robot according to the load acting on it and confirmed that the robot equipped with this control could inspect a narrow pipe with an inner diameter of 108 mm, where a load of approximately 340 N acted on the back end of the robot, with an efficiency that was approximately 1.7 times higher. This result has the potential to enable the configuration and control of robots, which would be difficult to
|
|
TuA-OL1 |
Rm21 (on-line) |
Medical Robots and Systems 4 |
Regular session |
Chair: Au, K. W. Samuel | The Chinese University of Hong Kong |
Co-Chair: Wang, Jiaole | Harbin Institute of Technology, Shenzhen |
|
10:00-10:10, Paper TuA-OL1.1 | |
Time-Optimal Synchronous Terminal Trajectory Planning for Coupling Motions of Robotic Flexible Endoscope |
|
Wang, Xiangyu | Nankai University |
Yu, Ningbo | Nankai University |
Han, Jianda | Nankai University |
Fang, Yongchun | Institute of Robotics and Automatic Information System, College |
Keywords: Flexible Robotics, Surgical Robotics: Planning, Surgical Robotics: Steerable Catheters/Needles
Abstract: The robotic flexible endoscope is developed rapidly in the field of surgery robots due to its high flexibility and safety. However, some inherent features, e.g., high nonlinearity, material creep, complex dynamic hysteresis behaviors, and the unknown coupling effects between bending and twisting motions, can lead to the significant degradation on three-dimensional (3-D) positioning performance of the endoscope. Aiming at these challenges, this paper built a practical multi-motion hysteresis phenomenon model for the bending and twisting motions of the robotic flexible endoscope with consideration of the coupling effects. Then, the time-optimal synchronous terminal motion planner is first proposed for the 3-D motions of the robotic endoscope to decouple the coupling effects in an intuitive separate control scheme. Finally, a series of hardware experiments are conducted on a robotic flexible ureteroscope platform. The accuracy of the proposed model and the trajectory-planning-based decoupling strategy is comprehensively validated. Particularly, the experimental results with the proposed trajectory planner show the satisfactory performance of vibration suppression and over-shoot suppression.
|
|
10:10-10:20, Paper TuA-OL1.2 | |
Robotic Actuation and Control of a Catheter for Structural Intervention Cardiology |
|
Zhang, Xiu | Politecnico Di Milano |
Palumbo, Maria Chiara | Politecnico Di Milano |
Perico, Francesca | Politecnico Di Milano |
Magro, Mattia | Politecnico Di Milano |
Fortuna, Andrea | Politecnico Di Milano |
Magni, Tommaso | Politecnico Di Milano |
Votta, Emiliano | Politecnico Di Milano |
Segato, Alice | Politecnico Di Milano, Milano , Italy |
De Momi, Elena | Politecnico Di Milano |
Keywords: Surgical Robotics: Steerable Catheters/Needles, Medical Robots and Systems, Surgical Robotics: Planning
Abstract: Structural intervention cardiology (SIC) interventions are crucial procedures for correcting heart valves, wall, and muscle form defects. However, the possibility of embolization or perforation, as well as the lack of transparent vision and autonomous surgical equipment, make it difficult for the clinician. In this paper, we propose a robot-assisted tendon-driven catheter and machine learning-based path planner to overcome these challenges. Firstly, an analytical inverse kinematic model is constructed to convert the tip location in the Cartesian space to the tendons’ displacement. Then inverse reinforcement learning algorithm is employed to calculate the the optimal path to avoid possible collisions between the catheter tip and the atrial wall. Moreover, a closed-loop feedback controller is adopted to improve positioning accuracy in a direct distal position measurement manner. Simulation and experiments are designed and conducted to demonstrate the feasibility and performance of the proposed system.
|
|
10:20-10:30, Paper TuA-OL1.3 | |
A Kinematic Modeling and Control Scheme for Different Robotic Endoscopes: A Rudimentary Research Prototype |
|
Li, Weibing | Sun Yat-Sen University |
Ng, Wing Yin | The Chinese University of Hong Kong |
Zhang, Xue | CUHK |
Huang, Yisen | The Chinese University of Hong Kong |
Li, Yehui | The Chinese University of Hong Kong |
Song, Chengzhi | Chinese University of Hong Kong, |
Chiu, Philip, Wai-yan | Chinese University of Hong Kong |
Li, Zheng | The Chinese University of Hong Kong |
Keywords: Medical Robots and Systems, Flexible Robotics, Visual Tracking
Abstract: In image-guided robotic surgery, there exist different endoscopes coupled either with specialized surgical robots (SSRs) or general industrial robots (GIRs). In general, SSRs mechanically respect the remote-center-of-motion (RCM) constraints with directly and explicitly controllable degrees-of-freedom (DoFs), whereas GIRs meet the constraints at the algorithmic level in an implicit manner, with a loss of direct control of some RCM DoFs. Conventionally, different robotic endoscopes are treated as different monolithic systems. Then, kinematic models and control schemes are separately established to automate different robotic endoscopes. This means that a similar analysis must be followed for each system, which is tedious and time-consuming. This paper introduces a modular method to analyze the individual kinematics of the robotic holder, the endoscope shaft and the surgical endoscope with RCM constraints explicitly handled. For achieving automation, the integrated kinematics of a generic robotic endoscope is determined by combining the kinematics of the modular elements. Considering that cameras are intrinsically embedded sensors, a visual servo control scheme applicable to different robotic endoscopes is formulated to incorporate four explicit (virtual) DoFs characterizing the RCM constraints in the control law. Simulations and experiments performed using three different robotic endoscopes validate the effectiveness and practicality of the kinematic modeling and control scheme for automatic field-of-view adjustment.
|
|
10:30-10:40, Paper TuA-OL1.4 | |
Real-Time Intraoperative Surgical Guidance System in the Da Vinci Surgical Robot Based on Transrectal Ultrasound/photoacoustic Imaging with Photoacoustic Markers: An Ex Vivo Demonstration |
|
Song, Hyunwoo | The Johns Hopkins University |
Moradi, Hamid | University of British Columbia |
Jiang, Baichuan | Johns Hopkins University |
Xu, Keshuai | The Johns Hopkins University |
Wu, Yixuan | The Johns Hopkins University |
Taylor, Russell H. | The Johns Hopkins University |
Deguet, Anton | Johns Hopkins University |
Kang, Jin | The Johns Hopkins University |
Salcudean, Septimiu E. | University of British Columbia |
Boctor, Emad | The Johns Hopkins University |
Keywords: Medical Robots and Systems, Software-Hardware Integration for Robot Systems
Abstract: This paper introduces an integrated real-time intraoperative surgical guidance system, in which an endoscope camera of da Vinci surgical robot and a transrectal ultrasound (TRUS) transducer are co-registered using photoacoustic markers that are detected in both fluorescence (FL) and photoacoustic (PA) imaging. The co-registered system enables the TRUS transducer to track the laser spot illuminated by a pulsed-laser-diode attached to the surgical instrument, providing both FL and PA images of the surgical region-of-interest (ROI). As a result, the generated photoacoustic marker is visualized and localized in the da Vinci endoscopic FL images, and the corresponding tracking can be conducted by rotating the TRUS transducer to display the PA image of the marker. A quantitative evaluation revealed that the average registration and tracking errors were 0.84 mm and 1.16°, respectively. This study shows that the co-registered photoacoustic marker tracking can be effectively deployed intraoperatively using TRUS+PA imaging providing functional guidance of the surgical ROI.
|
|
10:40-10:50, Paper TuA-OL1.5 | |
Model-Free and Uncalibrated Visual-Feedback Control of Magnetically-Actuated Flexible Endoscopes |
|
Tan, Jiewen | Harbin Instittue of Technology (Shenzhen) |
Xue, Junnan | University |
Yang, Xing | Harbin Institute of Technolgoy, Shenzhen |
Yuan, Sishen | The Chinese University of Hong Kong |
Liu, Wei | Southern University of Science and Technology |
Ren, Hongliang | Chinese Univ Hong Kong (CUHK) & National Univ Singapore(NUS) |
Song, Shuang | Harbin Institute of Technology (Shenzhen) |
Wang, Jiaole | Harbin Institute of Technology, Shenzhen |
Keywords: Flexible Robotics, Surgical Robotics: Steerable Catheters/Needles, Visual Servoing
Abstract: Magnetically-actuated flexible endoscopes (MAFE) have been well used in minimally-invasive surgery because they can be steered by a magnetic field thus more flexible than traditional endoscopes. Model-free and uncalibrated visual feedback control makes it possible to manipulate MAFE with a magnetic field without external tracking systems. Because no extra sensor is required to obtain position and posture information, the size of MAFE can be made smaller. However, the traditional control method focuses on 2DoF control, which lacks control over the posture of the end of MAFE. In this way, the friction between MAFE and tissue may cause injury during the advancement of the endoscope. In this paper, we propose algorithms to enhance the pose control of MAFE to 4DoF and 5DoF based on model-free and uncalibrated visual-feedback control. Experiments in structured environments verify that the control algorithms can realize 4DoF manual navigation and 5DoF automatic navigation.
|
|
10:50-11:00, Paper TuA-OL1.6 | |
Design, Teleoperation Control and Experimental Validation of a Dexterous Robotic Flexible Endoscope for Laparoscopic Surgery |
|
Ma, Xin | Chinese Univerisity of HongKong |
Wang, Xuchen | The Chinese University of Hong Kong |
Cao, Rui | The Chinese University of Hong Kong |
Au, K. W. Samuel | The Chinese University of Hong Kong |
Keywords: Medical Robots and Systems, Surgical Robotics: Laparoscopy, Flexible Robotics
Abstract: Existing robotic endoscopes for laparoscopic surgery, predominantly rigid or limited in dexterity, occupy a large motion space1. The large occupied motion space necessitates large incisions and reduces the motion space for surgeons to simultaneously operate other surgical instruments. Meanwhile, surgeons only have limited view adjustment capability to avoid occlusion and they often have to lift/push some organs to observe occluded target lesions in some operations such as cholecystectomy. The situation gets worse when the operations are on obese patients. In this paper, we develop a novel dexterous robotic flexible endoscope (DRFE), which is comprised of a concentric cable-driven structure and a 2-DoF articulated joint attached to the end of DRFE, for laparoscopic surgery. The proposed design occupies much less motion space both inside and outside human body as compared to conventional robotic flexible endoscopes. When used in surgery, the part of the endoscope outside the body can remain still, which reduces the risk of expanding the incision and simplifies the structure of the remote center mechanism. Simulation and experimental studies are performed to validate the effectiveness of the proposed device in the improvement of vision occlusion and usability. Initial results reveal that the DRFE is highly dexterous and accurate in observing lesions with vision occlusion.
|
|
11:00-11:10, Paper TuA-OL1.7 | |
A Miniature Continuum Robot with Integrated Piezoelectric Beacon Transducers and Its Ultrasonic Shape Detection in Robot-Assisted Minimally Invasive Surgeries |
|
Yin, Zhanpeng | Nanjing University of Aeronautics and Astronautics |
Hong, Yan | Nanjing University of Aeronautics and Astronautics |
Sun, Xiaoyu | University of Bristol |
Shen, Zhiyuan | University of Bristol |
Zhang, Yingxuan | Nanjing University of Aeronautics and Astronautics |
Ju, Feng | Nanjing University of Aeronautics and Astronautics |
Drinkwater, Bruce | University of Bristol |
Keywords: Surgical Robotics: Steerable Catheters/Needles, Medical Robots and Systems, Flexible Robotics
Abstract: Minimally invasive surgeries (MIS) or natural orifice transluminal endoscopic surgeries (NOTES) such as the transurethral resection of bladder tumor (TURBT) require the surgical robot to be miniaturized to perform surgical procedures in confined spaces. However, the surgical robot’s tiny size poses problems in its fabrication and shape sensing. In this paper, a miniature continuum surgical robot is proposed with a unique laminated structure which can be fabricated through a 2D lamination process and converted into 3D through folding. This multi-material laminated structure also facilitates the integration of tiny piezoelectric transducers on the robot’s surface as beacons to generate ultrasonic waves for shape detection. A novel beacon total focusing method (b-TFM) algorithm is developed to process the received ultrasonic data and create a high-quality ultrasonic image from which the shape of the continuum robot can be extracted. The proposed robot and the ultrasonic shape detection method are validated through simulations and experiments. The error in the open-loop trajectory control is less than 4 mm without compensation, and the error in the ultrasonic shape detection is less than 1 mm. This confirms the possibility of improving the trajectory control accuracy by using the detected shape as a feedback for closed-loop control.
|
|
11:10-11:20, Paper TuA-OL1.8 | |
A Domain-Adapted Machine Learning Approach for Visual Evaluation and Interpretation of Robot-Assisted Surgery Skills |
|
Soleymani, Abed | University of Alberta |
Li, Xingyu | University of Alberta |
Tavakoli, Mahdi | University of Alberta |
Keywords: Medical Robots and Systems, Human-Robot Collaboration, Representation Learning
Abstract: In this study, we present an intuitive machine learning-based approach to evaluate and interpret surgical skills level of a participant working with robotic platforms. The proposed method is domain-adapted, i.e., jointly utilizes an end-to-end learning approach for smoothness detection and domain knowledge-based metrics such as fluidity and economy of motion for extracting skills-related features within a given trajectory. An advantage of our approach compared to similar stochastic or deep learning models is its intuitive and transparent manner for extraction and visualization of skills-related features within the data. We illustrate the performance of our proposed method on trials of the JIGSAWS data set as well as our own experimental data gathered from Phantom Premium 1.5A Haptic Device. This approach utilized tSNE technique and provides visualized low-dimensional representation for different trials that highlights nuanced information within the executive task and returns unusual or faulty trials as outliers far away from their normal skill or participant clusters. This information regarding the input trajectory can be used for evaluation and education applications such as learning curve analysis in surgical assessment and training programs.
|
|
11:20-11:30, Paper TuA-OL1.9 | |
Automating Surgical Peg Transfer: Calibration with Deep Learning Can Exceed Speed, Accuracy, and Consistency of Humans (I) |
|
Hwang, Minho | Daegu Gyeongbuk Instituute of Science and Technology (DGIST) |
Ichnowski, Jeffrey | UC Berkeley |
Thananjeyan, Brijen | UC Berkeley |
Seita, Daniel | Carnegie Mellon University |
Paradis, Samuel | University of California, Berkeley |
Fer, Danyal | University of California, San Francisco East Bay |
Low, Thomas | SRI International |
Goldberg, Ken | UC Berkeley |
Keywords: Calibration and Identification, Deep Learning in Grasping and Manipulation, Kinematics
Abstract: Peg transfer is a well-known surgical training task in the Fundamentals of Laparoscopic Surgery (FLS). While human sur-geons teleoperate robots such as the da Vinci to perform this task with high speed and accuracy, it is challenging to automate. This paper presents a novel system and control method using a da Vinci Research Kit (dVRK) surgical robot and a Zivid depth sensor, and a human subjects study comparing performance on three variants of the peg-transfer task: unilateral, bilateral without handovers, and bilateral with handovers. The system combines 3D printing, depth sensing, and deep learning for calibration with a new analytic inverse kinematics model and time-minimized motion controller. In a controlled study of 3384 peg transfer trials performed by the system, an expert surgical resident, and 9 volunteers, results suggest that the system achieves accuracy on par with the experienced surgical resident and is significantly faster and more consistent than the surgical resident and volunteers. The system also exhibits the highest consistency and lowest collision rate. To our knowledge, this is the first autonomous system to achieve “superhuman” performance on a standardized surgical task. All data is available at https://sites.google.com/view/surgicalpegtransfer.
|
|
TuA-OL2 |
Rm22 (on-line) |
Computer Vision for Automation 1 |
Regular session |
Chair: Shimonomura, Kazuhiro | Ritsumeikan University |
Co-Chair: Wang, Yu-Ping | Tsinghua University |
|
10:00-10:10, Paper TuA-OL2.1 | |
RGB-X Classification for Electronics Sorting |
|
Abhimanyu, Fnu | Carnegie Mellon University |
Choset, Howie | Carnegie Mellon University |
Travers, Matthew | Carnegie Mellon University |
Thillaivasan, Umesh | Apple Inc |
Zodage, Tejas | Birla Institute of Technology and Science |
Keywords: Computer Vision for Automation, Deep Learning for Visual Perception, Recognition
Abstract: Effectively disassembling and recovering materials from Waste Electrical and Electronic Equipment (WEEE) is a critical step in moving global supply chains from carbon-intensive, mined materials to recycled and renewable ones. Traditional recycling processes rely on shredding and sorting waste streams, but for WEEE, which is comprised of numerous dissimilar materials in a small volume, we explore targeted disassembly of numerous objects for improved material recovery. Since many WEEE objects can look very similar due to common features, but their material composition and internal component layout can vary, it is critical to have an accurate classifier for subsequent disassembly steps for accurate material separation and recovery. This work introduces RGB-X, a multi-modal image classification approach, that innovatively utilizes key features from external RGB images with those generated from X-ray images to very accurately classify electronic objects. More specifically, this work develops Iterative Class Activation Mapping (iCAM), a novel network architecture that explicitly focuses on the finer-details in the multi-modal feature maps that are needed for accurate electronic object classification. Electronic objects lack large and well annotated X-ray datasets for training due to expense and need of expert guidance. To overcome this constraint, we present a novel way of creating a synthetic dataset using domain randomization applied to the X-ray domain. The combined RGBX approach gives us an accuracy of 98.6% on 10 generations of modern smartphones, which is greater than their individual accuracies of 89.1% (RGB) and 97.9% (X-ray) independently. We provide experimental results3 to corroborate our results.
|
|
10:10-10:20, Paper TuA-OL2.2 | |
CPQNet: Contact Points Quality Network for Robotic Grasping |
|
Li, Zhihao | The Chinese University of Hong Kong (Shenzhen) |
Zeng, Pengfei | Shenzhen Institute of Artificial Intelligence and Robotics for S |
Su, Jionglong | Xi'an Jiaotong-Liverpool University |
Guo, Qingda | TOP-AIRS Hospital Intelligent Technique Joint Lab |
Ding, Ning | The Chinese University of Hong Kong, Shenzhen |
Zhang, Jiaming | The Chinese University of Hong Kong, Shenzhen |
Keywords: Computer Vision for Automation, Deep Learning in Grasping and Manipulation, Grasping
Abstract: In typical data-based grasping methods, a grasp based on parallel-jaw grippers is parameterized by the center of the gripper, the rotation angle, and the gripper opening width so as to predict the quality and pose of grasps at every pixel. In contrast, a grasp is represented using only two contact points for contact-points-based grasp representation, which allows for fusion with tactile sensors more naturally. In this work, we propose a method using contact-points-based grasp representation to get a robust grasp using only one contact points quality map generated by a neural network, which significantly reduces the complexity of the network with fewer parameters. We provide a synthetic dataset including depth image and contact points quality map generated by thousands of 3D models. We also provide the method for data generation, which can be used for contact-points-based multi-fingers grasp. Experiments show that contact points quality network can plan an available grasp in 0.15 seconds. The grasping success rate for unknown household objects is 94%. Our method is also available for deformable objects with a success rate of 95%. The dataset and reference code can be found on the project website: https://sites.google.com/view/cpqnet.
|
|
10:20-10:30, Paper TuA-OL2.3 | |
SESR: Self-Ensembling Sim-To-Real Instance Segmentation for Auto-Store Bin Picking |
|
Yang, Biqi | The Chinese University of Hong Kong |
Gao, Xiaojie | The Chinese University of Hong Kong |
Chen, Kai | The Chinese University of Hong Kong |
Cao, Rui | Hong Kong Centre for Logistics Robotics |
Feng, Yidan | Nanjing University of Aeronautics and Astronautics |
Li, Xianzhi | The Chinese University of Hong Kong |
Dou, Qi | The Chinese University of Hong Kong |
Fu, Chi-Wing | The Chinese University of Hong Kong |
Liu, Yunhui | Chinese University of Hong Kong |
Heng, Pheng Ann | The Chinese University of Hong Kong |
Keywords: Computer Vision for Automation, Object Detection, Segmentation and Categorization, Deep Learning Methods
Abstract: Instance segmentation is an important task for supporting robotic grasping in auto-store scenarios. Accurate segmentation usually relies on the quantity and quality of available annotated training data. However, it requires tremendous cost to obtain these labels. In this work, without requiring any human annotations on real data, our proposed self-ensembling sim-to-real network, namely SESR, is able to generate precise instance masks for a wide variety of supermarket goods. We design our SESR with a teacher model and a student model trained with a self-ensembling strategy. We adopt different levels of consistency to bridge the sim-to-real gap and boost the model generalization ability. Also, we compile an auto-store bin-picking dataset covering various goods. Extensive experiments on both unseen scenarios and unseen objects validate the effectiveness and superiority of our method over others, and the robot arm demonstrations further show that our segmentation results can support real-time auto-store bin picking.
|
|
10:30-10:40, Paper TuA-OL2.4 | |
Visual Odometry in HDR Environments by Using Spatially Varying Exposure Camera |
|
Ye, Keyang | National University of Defense Technology |
Gao, LiuZheng | National University of Defense Technology |
Guan, Banglei | National University of Defense Technology |
Keywords: Vision-Based Navigation, SLAM, Computer Vision for Automation
Abstract: The accuracy and robustness of visual odometry (VO) is significantly affected by the high dynamic range (HDR) environments, because traditional cameras have a limited dynamic range and inevitably miss information in both overexposed and underexposed areas. To overcome the above challenge, we use an spatially varying exposure (SVE) camera, which captures four images with different exposure levels simultaneously. Then, we propose a VO pipeline that leverages the advantages of the SVE camera. Specifically, we extract ORB features from four images in parallel firstly instead of fusing four images, then perform merging and filtering to provide more robust features. We demonstrate that the proposed system outperforms comparable state-of-the-art methods in terms of robustness and accuracy. The real-time performance of the proposed system is also guaranteed due to the elaborate design of the parallel algorithm.
|
|
10:40-10:50, Paper TuA-OL2.5 | |
Trifocal Tensor and Relative Pose Estimation from 8 Lines and Known Vertical Direction |
|
Guan, Banglei | National University of Defense Technology |
Vasseur, Pascal | Université De Picardie Jules Verne |
Demonceaux, Cédric | Université Bourgogne Franche-Comté |
Keywords: Vision-Based Navigation, SLAM, Computer Vision for Automation
Abstract: In this paper, we present a relative pose estimation algorithm based on lines knowing the vertical direction associated to each image. We demonstrate that a closed-form solution requiring only eight lines between three views is possible. As a linear solution, it is shown that our approach outperforms the standard trifocal estimation based on 13 triplets of lines and can be efficiently inserted into an hypothesize-and-test framework such as RANSAC. We also study our approach on different singular configurations of lines. The method is evaluated on both synthetic data and real-world sequences from KITTI and the Zürich Urban Micro Aerial Vehicle datasets. Our method is compared to 13 lines algorithm as well to points based methods such as 7-points, 5-points and 3-points.
|
|
10:50-11:00, Paper TuA-OL2.6 | |
Lidar with Velocity: Motion Distortion Correction of Point Clouds from Oscillating Scanning Lidars |
|
Yang, Wen | Southern University of Science and Technology |
Zheng, Gong | Southern University of Science and Technology |
Huang, Baifu | Southern University of Science and Technology |
Hong, Xiaoping | Southern University of Science and Technology |
Keywords: Computer Vision for Transportation, Visual Tracking, Intelligent Transportation Systems
Abstract: Lidar point cloud distortion from moving object is an important problem in autonomous driving, and recently becomes more demanding with the emerging of oscillating type lidars, which feature back-and-forth scanning patterns and complex distortions. Accurately correcting the point cloud distortion would not only describe the 3D moving objects more accurately, but also enable accurate estimation of moving objects' velocities with enhanced prediction and tracking capabilities. A lidar and camera fusion approach is proposed to correct the oscillating lidar distortions with full velocity estimation. Lidar measures the time-of-flight distance accurately in the radial direction but only with sparse angular information while camera as a complementary sensor could provide a dense angular resolution. In addition, the proposed framework utilizes a probabilistic Kalman-filter approach to combine the estimated velocities and track the moving objects with their real-time velocities and correct point clouds. The proposed framework is evaluated on real road data and consistently outperforms other methods. The complete framework is open-sourced (https://github.com/ISEE-Technology/lidar-with-vel ocity) to accelerate the adoption of the emerging lidars.
|
|
11:00-11:10, Paper TuA-OL2.7 | |
A Flexible and Robust Vision Trap for Automated Part Feeder Design |
|
Haugaard, Rasmus Laurvig | University of Southern Denmark |
Iversen, Thorbjørn Mosekjær | The Maersk Mc-Kinney Moller Institute, University of Southern De |
Buch, Anders Glent | University of Southern Denmark |
Kramberger, Aljaz | University of Southern Denmark |
Mathiesen, Simon | University of Southern Denmark |
Keywords: Computer Vision for Automation, Assembly
Abstract: Fast, robust, and flexible part feeding is essential for enabling automation of low volume, high variance assembly tasks. An actuated vision-based solution on a traditional vibratory feeder, referred to here as a vision trap, should in principle be able to meet these demands for a wide range of parts. However, in practice, the flexibility of such a trap is limited as an expert is needed to both identify manageable tasks and to configure the vision system. We propose a novel approach to vision trap design in which the identification of manageable tasks is automatic and the configuration of these tasks can be delegated to an automated feeder design system. We show that the trap's capabilities can be formalized in such a way that it integrates seamlessly into the ecosystem of automated feeder design. Our results on six canonical parts show great promise for autonomous configuration of feeder systems.
|
|
11:10-11:20, Paper TuA-OL2.8 | |
Leveraging Local Planar Motion Property for Robust Visual Matching and Localization |
|
Jiao, Yanmei | Zhejiang University |
Zhang, Qunkang | Zhejiang University |
Chen, Qunying | Hangzhou Jiazhi Technology Co., Ltd |
Fu, Bo | Zhejiang University, the State Key Laboratory of Industrial Cont |
Han, Fuzhang | Zhejiang University |
Wang, Yue | Zhejiang University |
Xiong, Rong | Zhejiang University |
Keywords: Localization, Service Robotics
Abstract: One primary difficulty preventing the visual localization for service robots is the robustness against changes, including environmental changes and perspective changes. In recent years, learning-based feature matching methods have been widely studied and effectively verified in practical applications. Learning-based feature matching effectively solves the problem of environmental changes, including illumination changes and man-made changes. However, there is still room for improvement dealing with large perspective changes. In this paper, we leverage local planar motion property to simplify the affine transform and propose an augmentation-based feature matching method that greatly enhances the robustness to perspective changes. The proposed feature matching approach maintains low matching costs as the augmentation is performed on the simplified affine matrix space. Combined with the motion property aided minimal solution for pose estimation, an end-to-end robust visual localization system is proposed which is shown to bring 67% improvement in localization performance under large perspective changes in publicly available OpenLORIS dataset, while increasing computational cost by only 20% by using batch processing techniques with a single GPU. In addition, a guide for map frame selection is presented to support robust localization with very sparse map frames in storage. Experiments on the classified dataset with environmental changes and perspective changes validate the effectiveness of the proposed system.
|
|
11:20-11:30, Paper TuA-OL2.9 | |
AFR: An Efficient Buffering Algorithm for Cloud Robotic Systems |
|
Wang, Yu-Ping | Tsinghua University |
Wang, Hao-Ning | Tsinghua University |
Zou, Zi-Xin | Tsinghua University |
Manocha, Dinesh | University of Maryland |
Keywords: Software, Middleware and Programming Environments, Networked Robots, Agent-Based Systems
Abstract: Communication between robots and the server is a major problem for cloud robotic systems. In this paper, we address the problem caused by data loss during such communications, and propose an efficient buffering algorithm, called AFR, to solve the problem. We model the problem into an optimization problem to maximize the received Quantity of Information (QoI). Our AFR algorithm is formally proved to achieve near-optimal QoI, which has a lower bound that is a constant multiple of the unrealizable optimal QoI. We implement our AFR algorithm in ROS without changing the interface or API for the applications. Our experiments on two cloud robot applications show that our AFR algorithm can efficiently and effectively reduce the impact of data loss. For the remote mapping application, the RMSE caused by data loss can be reduced by about 20%. For the remote tracking application, the probability of tracking failure caused by data loss can be reduced from about 40%-60% to under 10%. Meanwhile, our AFR algorithm introduces time overhead of under 10 microseconds.
|
|
TuB-1 |
Rm1 (Room A) |
Award Session VII |
Regular session |
Chair: Harada, Kensuke | Osaka University |
Co-Chair: Sugiura, Hisashi | Yanmar Co., Ltd |
|
12:30-12:45, Paper TuB-1.1 | |
Robot Learning of Mobile Manipulation with Reachability Behavior Priors (Finalist for IROS Best Paper Award on Mobile Manipulation Sponsored by OMRON Sinic X Corp.) |
|
Jauhri, Snehal | TU Darmstadt |
Peters, Jan | Technische Universität Darmstadt |
Chalvatzaki, Georgia | Technische Universität Darmastadt, Intelligent Robotic Systems |
Keywords: Mobile Manipulation, Reinforcement Learning, Transfer Learning
Abstract: Mobile Manipulation (MM) systems are ideal candidates for taking up the role of a personal assistant in unstructured real-world environments. Among other challenges, MM requires effective coordination of the robot's embodiments for executing tasks that require both mobility and manipulation. Reinforcement Learning (RL) holds the promise of endowing robots with adaptive behaviors, but most methods require prohibitively large amounts of data for learning a useful control policy. In this work, we study the integration of robotic reachability priors in actor-critic RL methods for accelerating the learning of MM for reaching and fetching tasks. Namely, we consider the problem of optimal base placement and the subsequent decision of whether to activate the arm for reaching a 6D target. For this, we devise a novel Hybrid RL method that handles discrete and continuous actions jointly, resorting to the Gumbel-Softmax reparameterization. Next, we train a reachability prior using data from the operational robot workspace, inspired by classical methods. Subsequently, we derive Boosted Hybrid RL (BHyRL), a novel algorithm for learning Q-functions by modeling them as a sum of residual approximators. Every time a new task needs to be learned, we can transfer our learned residuals and learn the component of the Q-function that is task-specific, hence, maintaining the task structure from prior behaviors. Moreover, we find that regularizing the target policy with a prior policy yields more expressive behaviors. We evaluate our method in simulation in reaching and fetching tasks of increasing difficulty, and we show the superior performance of BHyRL against baseline methods. Finally, we zero-transfer our learned 6D fetching policy with BHyRL to our MM robot TIAGo++.
|
|
12:45-13:00, Paper TuB-1.2 | |
A Hybrid Learning and Optimization Framework to Achieve Physically Interactive Tasks with Mobile Manipulators (Finalist for IROS Best Paper Award on Mobile Manipulation Sponsored by OMRON Sinic X Corp.) |
|
Zhao, Jianzhuang | Istituto Italiano Di Tecnologia |
Giammarino, Alberto | Istituto Italiano Di Tecnologia |
Lamon, Edoardo | Istituto Italiano Di Tecnologia |
Gandarias, Juan M. | Istituto Italiano Di Tecnologia |
De Momi, Elena | Politecnico Di Milano |
Ajoudani, Arash | Istituto Italiano Di Tecnologia |
Keywords: Compliance and Impedance Control, Mobile Manipulation, Imitation Learning
Abstract: This paper proposes a hybrid learning and optimization framework for mobile manipulators for complex and physically interactive tasks. The framework exploits the MOCA-MAN interface to obtain intuitive and simplified human demonstrations and Gaussian Mixture Model(GMM)/Gaussian Mixture Regression(GMR) to encode and generate the learned task requirements in terms of position, velocity, and force profiles. Next, using the desired trajectories and force profiles generated by GMM/GMR, the impedance parameters of a Cartesian impedance controller are optimized online through a Quadratic Program augmented with an energy tank to ensure the passivity of the controlled system. Two experiments are conducted to validate the framework, comparing our method with two approaches with constant stiffness (high and low). The results showed that the proposed method outperforms the other two cases in terms of trajectory tracking and generated interaction forces, even in the presence of disturbances such as unexpected end-effector collisions.
|
|
13:00-13:15, Paper TuB-1.3 | |
Contrastive 3D Shape Completion and Reconstruction for Agricultural Robots Using RGB-D Frames (Finalist for IROS Best Paper Award on Agri-Robotics Sponsored by YANMAR) |
|
Magistri, Federico | University of Bonn |
Marks, Elias Ariel | University of Bonn |
Nagulavancha, Sumanth | University of Bonn |
Vizzo, Ignacio | University of Bonn |
Läbe, Thomas | University of Bonn |
Behley, Jens | University of Bonn |
Halstead, Michael Allan | Bonn University |
McCool, Christopher Steven | University of Bonn |
Stachniss, Cyrill | University of Bonn |
Keywords: Robotics and Automation in Agriculture and Forestry, Deep Learning for Visual Perception, RGB-D Perception
Abstract: Monitoring plants and fruits is important in modern agriculture, with applications ranging from high-throughput phenotyping to autonomous harvesting. Obtaining highly accurate 3D measurements under real agricultural conditions is a challenging task. In this paper, we address the problem of estimating the 3D shape of fruits when only a partial view is available. We propose a pipeline that exploits high-resolution 3D data in the learning phase but only requires a single RGB-D frame to predict the 3D shape of a complete fruit during operation. To achieve this, we first learn a latent space of potential fruit appearances that we can decode into an SDF volume. With the pretrained, frozen decoder, we subsequently learn an encoder that can produce meaningful latent vectors from a single RGB-D frame. The experiments presented in this paper suggest that our approach can predict the 3D shape of whole fruits online, needing only 4ms for inference. We evaluate our approach in controlled environments and illustrate its deployment in greenhouses without modifications.
|
|
13:15-13:30, Paper TuB-1.4 | |
Instance Segmentation for Autonomous Log Grasping in Forestry Operations (Finalist for IROS Best Paper Award on Agri-Robotics Sponsored by YANMAR) |
|
Fortin, Jean-Michel | Université Laval |
Gamache, Olivier | Laval University |
Grondin, Vincent | Université Laval |
Pomerleau, Francois | Université Laval |
Giguère, Philippe | Université Laval |
Keywords: Robotics and Automation in Agriculture and Forestry, Data Sets for Robotic Vision, Object Detection, Segmentation and Categorization
Abstract: Wood logs picking is a challenging task to automate. Indeed, logs usually come in cluttered configurations, randomly orientated and overlapping. Recent work on log picking automation usually assume that the logs’ pose is known, with little consideration given to the actual perception problem. In this paper, we squarely address the latter, using a data-driven approach. First, we introduce a novel dataset, named TimberSeg 1.0, that is densely annotated, i.e., that includes both bounding boxes and pixel-level mask annotations for logs. This dataset comprises 220 images with 2500 individually segmented logs. Using our dataset, we then compare three neural network architectures on the task of individual logs detection and segmentation; two region-based methods and one attention-based method. Unsurprisingly, our results show that axis-aligned proposals, failing to take into account the directional nature of logs, underperform with 19.03 mAP. A rotation-aware proposal method significantly improve results to 31.83 mAP. More interestingly, a Transformer-based approach, without any inductive bias on rotations, outperformed the two others, achieving a mAP of 57.53 on our dataset. Our use case demonstrates the limitations of region-based approaches for cluttered, elongated objects. It also highlights the potential of attention-based methods on this specific task, as they work directly at the pixel-level. These encouraging results indicate that such a perception system could be used to assist the operators on the short-term, or to fully automate log picking operations in the future.
|
|
13:30-13:45, Paper TuB-1.5 | |
Explicitly Incorporating Spatial Information to Recurrent Networks for Agriculture (Finalist for IROS Best Paper Award on Agri-Robotics Sponsored by YANMAR) |
|
Smitt, Claus | University of Bonn |
Halstead, Michael Allan | Bonn University |
Ahmadi, Alireza | University of Bonn |
McCool, Christopher Steven | University of Bonn |
Keywords: Deep Learning for Visual Perception, Robotics and Automation in Agriculture and Forestry, Semantic Scene Understanding
Abstract: In agriculture, the majority of vision systems perform still image classification. Yet, recent work has highlighted the potential of spatial and temporal cues as a rich source of information to improve the classification performance. In this paper, we propose novel approaches to explicitly capture both spatial and temporal information to improve the classification of deep convolutional neural networks. We leverage available RGB-D images and robot odometry to perform inter-frame feature map spatial registration. This information is then fused within recurrent deep learnt models, to improve their accuracy and robustness. We demonstrate that this can considerably improve the classification performance with our best performing spatial-temporal model (ST-Atte) achieving absolute performance improvements for intersection-over-union (IoU[%]) of 4.7 for crop-weed segmentation and 2.6 for fruit (sweet pepper) segmentation. Furthermore, we show that these approaches are robust to variable framerates and odometry errors, which are frequently observed in real-world applications.
|
|
13:45-14:00, Paper TuB-1.6 | |
Multimodal Aerial-Tethered Robot for Tree Canopy Exploration (Finalist for IROS Best Paper Award on Agri-Robotics Sponsored by YANMAR) |
|
Kirchgeorg, Steffen | ETH Zürich |
Mintchev, Stefano | ETH Zurich |
Keywords: Aerial Systems: Mechanics and Control, Biologically-Inspired Robots, Robotics and Automation in Agriculture and Forestry
Abstract: Forest canopies are the biggest habitat for terrestrial life, yet our understanding of environmental processes and biodiversity inside the canopy continues to be limited due to labour and resource intensive data collection. Existing aerial and climbing robots also struggle to access these complex environments, while animals easily navigate them using multiple means of locomotion. Following this insight we present a robot with multimodal mobility obtained by combining aerial and tethered locomotion. After the robot is deployed at the top of the tree, it can descend with the tether and maneuver around leaves and branches with its thrusters. The tether increases robustness and safety and allows for resting as well as emergency retrieval of the system. The aerial locomotion grants the system the ability to move in a conical 3D space constrained by the tether. We modelled the static system and validated the impact of design parameters on it. A simple control architecture for teleoperation is discussed and its performance is analyzed. The proposed multimodal mobility is demonstrated in preliminary outdoor tests, which show how our robot can move within the canopy while continuously monitoring the environment.
|
|
TuB-2 |
Rm2 (Room B-1) |
Brain-Machine Interfaces and Natural Language Interaction |
Regular session |
Chair: Liarokapis, Minas | The University of Auckland |
Co-Chair: Meattini, Roberto | University of Bologna |
|
12:30-12:40, Paper TuB-2.1 | |
Lightmyography Based Decoding of Human Intention Using Temporal Multi-Channel Transformers |
|
de Godoy, Ricardo | The University of Auckland |
Dwivedi, Anany | University of Auckland |
Shahmohammadi, Mojtaba | University of Auckland |
Liarokapis, Minas | The University of Auckland |
Keywords: Brain-Machine Interfaces, Wearable Robotics
Abstract: For the development of muscle-machine interfaces (MuMIs), researchers have relied mainly on Electromyography (EMG) signals. However, these signals require complex hardware systems, as well as specialized signal processing and feature extraction methods. To overcome these issues, in our previous work, we proposed a novel MuMI for decoding human intention and motion, called Lightmyography (LMG). To improve the performance of this interface even further, in this work, we employ two novel deep learning techniques called Temporal Multi-Channel Transformer (TMC-T) and Temporal Multi-Channel Vision Transformer (TMC-ViT) for the classification of hand gestures based on the LMG data. The performance of these two Transformer-based methods is evaluated and compared with other well-known deep learning and classical machine learning methods. This work also addresses the influence of varying parameters defined during the training phase of decoding models, such as the size and shape of the input data packet. A series of data augmentation techniques were also employed to generate synthetic data and increase the dataset size so as to train deep learning models more efficiently.
|
|
12:40-12:50, Paper TuB-2.2 | |
Dynamic Network Model for Multi-Domain End-To-End Task-Oriented Dialogue System |
|
Zhao, Fengda | College of Information Science and Engineering, Yanshan Universi |
Qiu, Menglu | YanShan University |
Li, Xianshan | College of Information Science and Engineering, Yanshan Universi |
Guo, Dingding | YanShan University |
Keywords: Natural Dialog for HRI, Intention Recognition, Service Robotics
Abstract: Dialogue State Tracking (DST) is an important part in task-oriented dialog system, whose target is to infer the current dialog states and user intentions according to the dialog history information. To this end, we have achieved improvements to the existing work and proposed a dynamic network model suitable for multi-domain dialog, which can explicitly use domain information and better cope with zeroshot tasks. The model is composed of three modules: an encoder,a decoder and a slot classifier. The encoder module introduces a mixed-separate framework so that it can obtain the feature information of each domain on the premise of extracting the shared information between all domains. The experimental results show that the model achieves joint accuracy of 48.38% for the five domains of MultiWOZ, which is superior to existing models. Besides, by simulating the zero-shot scenario,the knowledge transferability of the model has also been well proven.Finally, in order to verify the effectiveness of the robot simulation system, this paper also uses the robot simulation technology to simulate the common tasks of helping users complete the task of taking items in the home environment service.
|
|
12:50-13:00, Paper TuB-2.3 | |
Hey Haru, Let’s Be Friends! Using the Tiers of Friendship to Build Rapport through Small Talk with the Tabletop Robot Haru |
|
Nichols, Eric | Honda Research Institute Japan |
Siskind, Sarah Rose | Hello SciCom |
Ivanchuk, Levko | University of Manitoba |
Pérez, Guillermo | 4i Intelligent Insights |
Waki, Kamino | Indiana University Bloomington |
Sabanovic, Selma | Indiana University Bloomington |
Gomez, Randy | Honda Research Institute Japan Co., Ltd |
Keywords: Robot Companions, Emotional Robotics, AI-Enabled Robotics
Abstract: Conversation can play an essential role in forging bonds between humans and social robots, however, participants need to feel like they are being listened to, remembered, and cared about in order to effectively build rapport. In this paper, we propose a novel strategy for conducting small talk with a social robot. Our approach is known as the Tiers of Friendship. It is centered around three core design elements: 1) Persuasive content and character is provided through topic modules created by professional creative writers to ensure engaging conversational content and a compelling personality for the social robot. 2) Conversational memory is achieved by allowing topic modules to specify required information that can learned through conversation or recalled from previous interactions and organizing topic modules into a hierarchy that enforces information requirements between topics. 3) Dynamicity in conversation is promoted through topic navigation that supports fluid transitions to topics of human interest and employs elements of random ordering to create fresh conversation experiences. In this paper, we show how the tiers of friendship can be used to generate conversation content for a social robot that encourages the development of rapport. We describe a working implementation of a small talk system for a social robot based on the tiers of friendship that combines off-the-shelf ASR and NLU components and custom robot behavior components implemented via behavior trees on ROS. Finally, in order to evaluate our approach’s effectiveness, we conduct an elicitation survey that evaluates conversations in terms of perceived engagement, personality traits, and rapport expectation and discuss the implications for social robotics.
|
|
13:00-13:10, Paper TuB-2.4 | |
Givenness Hierarchy Informed Optimal Document Planning for Situated Human-Robot Interaction |
|
Spevak, Kevin | Colorado School of Mines |
Han, Zhao | Colorado School of Mines |
Williams, Tom | Colorado School of Mines |
Dantam, Neil | Colorado School of Mines |
Keywords: Natural Dialog for HRI, AI-Based Methods, Task Planning
Abstract: Robots that use natural language in collaborative tasks must refer to objects in their environment. Recent work has shown the utility of the linguistic theory of the Givenness Hierarchy (GH) in generating appropriate referring forms. But before referring expression generation, collaborative robots must determine the content and structure of a sequence of utterances, a task known as document planning in the natural language generation community. This problem presents additional challenges for robots in situated contexts, where described objects change both physically and in the minds of their interlocutors. In this work, we consider how robots can “think ahead” about the objects they must refer to and how to refer to them, sequencing object references to form a coherent, easy to follow chain. Specifically, we leverage GH to enable robots to plan their utterances in a way that keeps objects at a high cognitive status, which enables use of concise, anaphoric referring forms. We encode these linguistic insights as a mixed integer program within a planning context, formulating constraints to concisely and efficiently capture GH-theoretic cognitive properties. We demonstrate that this GH-informed planner generates sequences of utterances with high inter-sentential coherence, which we argue should enable substantially more efficient and natural human-robot dialogue.
|
|
13:10-13:20, Paper TuB-2.5 | |
DoRO: Disambiguation of Referred Object for Embodied Agents |
|
Pramanick, Pradip | TCS Research & Innovation |
Sarkar, Chayan | TCS Research |
Paul, Sayan | TCS Research & Innovation |
Roychoudhury, Ruddra dev | TCS Research & Innovation |
Bhowmick, Brojeshwar | Tata Consultancy Services |
Keywords: Natural Dialog for HRI, Human Factors and Human-in-the-Loop, Human-Robot Collaboration
Abstract: Robotic task instructions often involve a referred object that the robot must locate (ground) within the environment. While task intent understanding is an essential part of natural language understanding, less effort is made to resolve ambiguity that may arise while grounding the task. Existing works use vision-based task grounding and ambiguity detection, suitable for a fixed view and a static robot. However, the problem magnifies for a mobile robot, where the ideal view is not known beforehand. Moreover, a single view may not be sufficient to locate all the object instances in the given area, which leads to inaccurate ambiguity detection. Human intervention is helpful only if the robot can convey the kind of ambiguity it is facing. In this article, we present DoRO (Disambiguation of Referred Object), a system that can help an embodied agent to disambiguate the referred object by raising a suitable query whenever required. Given an area where the intended object is, DoRO finds all the instances of the object by aggregating observations from multiple views while exploring & scanning the area. It then raises a suitable query using the information from the grounded object instances. Experiments conducted with the AI2Thor simulator show that DoRO not only detects the ambiguity more accurately but also raises verbose queries with more accurate information from the visual-language grounding.
|
|
13:20-13:30, Paper TuB-2.6 | |
Following Natural Language Instructions for Household Tasks with Landmark Guided Search and Reinforced Pose Adjustment |
|
Murray, Michael | University of Washington |
Cakmak, Maya | University of Washington |
Keywords: Natural Dialog for HRI, Multi-Modal Perception for HRI, Human-Centered Robotics
Abstract: We study the challenging problem of following natural language instructions on a mobile manipulator robot. This task is challenging because it requires the robot to integrate the semantics of the unconstrained natural language instructions with the robot's egocentric visual observations of the environment which are typically incomplete and noisy. To address these challenges, we propose a method that is able to use visible landmarks to more efficiently explore the environment in search of the objects described by the natural language instructions. Additionally, we propose using a pose adjustment policy during manipulation planning to help the robot recover from noisy visual observations. We show that this policy can be trained through experience with reinforcement learning as well as with human-in-the-loop feedback. We evaluate our approach on the popular ALFRED instruction following benchmark and show that these methods achieve state-of-the-art performance (35.41%) with a substantial (8.92% absolute) gap from prior work.
|
|
13:30-13:40, Paper TuB-2.7 | |
SEMG-Based Minimally Supervised Regression Using Soft-DTW Neural Networks for Robot Hand Grasping Control |
|
Meattini, Roberto | University of Bologna |
Bernardini, Alessandra | University of Bologna |
Palli, Gianluca | University of Bologna |
Melchiorri, Claudio | University of Bologna |
Keywords: Manipulation Planning, Multifingered Hands, Telerobotics and Teleoperation
Abstract: One of the major challenges in robotics consists in developing successful control strategies for robotic grasping devices. In this scenario, one of the most interesting approaches regards the exploitation of surface electromyography(sEMG.) In this work, we propose a novel sEMG-based minimally supervised regression approach capable of performing nonlinear fitting without the necessity for point-by-point training data labelling. The proposed method exploits a differentiable version of the Dynamic Time Warping (DTW) similarity - referred to as soft-DTW divergence - as loss function for a flexible neural network architecture. This is a different paradigm with respect to state-of-the-art approaches in which sEMG-based control of robot hands is mainly realized using supervised or unsupervised machine learning based regression. An experimental session was carried out involving 10 healthy subjects in an offline experiment for systematic and statistical evaluations, and an online experiment for the evaluation of the control of a robot hand. The reported results demonstrate that the proposed soft-DTW neural network can be trained by means of a labelling that does not require to be temporally aligned with the sEMG training dataset, while reporting performances comparable with a standard mean square error(MSE)-based neural network. Also, the subjects were able to successfully control a robot hand for grasping motions and tasks with error levels comparable to state-of-the-art regression approaches.
|
|
13:40-13:50, Paper TuB-2.8 | |
Hand Gesture Recognition Via Transient sEMG Using Transfer Learning of Dilated Efficient CapsNet: Towards Generalization for Neurorobotics |
|
Tyacke, Eion | New York University |
P J, Shreyas | International Institute of Information Technology, Bhubaneshwar |
Feng, Natalie | New York University |
Edlabadkar, Rama | Indian Institute of Technology Indore |
Zhou, Shucong | New York University |
Patel, Jay | New York University |
Hu, Qin | New York University |
Atashzar, S. Farokh | New York University (NYU), US |
Keywords: Brain-Machine Interfaces, Physically Assistive Devices, Neurorobotics
Abstract: There has been an accelerated surge in utilizing the deep neural network to decode central and peripheral activations of the human nervous system to boost the spatiotemporal resolution of neural interfaces used in human-centered robotic systems, such as prosthetics, and exoskeletons. Deep learning methods are proven to achieve high accuracy but are also challenged by their assumption of having access to massive training samples. {Objective:} In this letter, we propose Dilated Efficient CapsNet to improve the predictive performance when the available individual data is minimal and not enough to train an individualized network for controlling a personalized robotic system. {Method:} We proposed the concept of transfer learning for a new design of the dilated efficient capsular neural network to relax the need of having access to massive individual data and utilize the field knowledge which can be learned from a group of participants. In addition, instead of using complete sEMG signals, we only use the transient phase, reducing the volume of training samples to 20% of the original and maximizing the agility. {Results:} In experiments, we validate the performance with various amounts of injected personalized training data (from 25% to 100% of transient phase). The results support the use of the proposed transfer learning approach based on the dilated capsular neural network when the knowledge domain learned on a small number of subjects can be utilized to minimize the need for new data from new subjects. The model focuses only on the transient phase which is a challenging neural interfacing problem.
|
|
13:50-14:00, Paper TuB-2.9 | |
Deep Augmentation for Electrode Shift Compensation in Transient High-Density sEMG: Towards Application in Neurorobotics |
|
Sun, Tianyun | New York University |
Libby, Jacqueline | New York University |
JohnRoss, Rizzo | New York University |
Atashzar, S. Farokh | New York University (NYU), US |
Keywords: Brain-Machine Interfaces, Physically Assistive Devices, Neurorobotics
Abstract: Going beyond the traditional sparse multi-channel peripheral human-machine interface that has been used widely in neurorobotics, high-density surface electromyography (HD-sEMG) has shown significant potential for decoding upper-limb motor control. We have recently proposed heterogeneous temporal dilation of LSTM in a deep neural network architecture for a large number of gestures (>60), securing spatial resolution and fast convergence. However, several fundamental questions remain unanswered. One problem targeted explicitly in this paper is the issue of ``electrode shift,'' which can happen specifically for high-density systems and during doffing and donning the sensor grid. Another real-world problem is the question of transient versus plateau classification, which connects to the temporal resolution of neural interfaces and seamless control. In this paper, for the first time, we implement gesture prediction on the transient phase of HD-sEMG data while robustifying the human-machine interface decoder to electrode shift. For this, we propose the concept of deep data augmentation for transient HD-sEMG. We show that without using the proposed augmentation, a slight shift of 10mm may drop the decoder's performance to as low as 20%. Combining the proposed data augmentation with a 3D Convolutional Neural Network (CNN), we recovered the performance to 84.6% while securing a high spatiotemporal resolution, robustifying to the electrode shift, and getting closer to large-scale adoption by the end-users, enhancing resiliency.
|
|
TuB-3 |
Rm3 (Room B-2) |
Telerobotics and Teleoperation 1 |
Regular session |
Chair: Sakaino, Sho | University of Tsukuba |
Co-Chair: Zhu, Yaonan | Nagoya University |
|
12:30-12:40, Paper TuB-3.1 | |
Adaptive Wave Reconstruction through Regulated-BMFLC for Transparency-Enhanced Telerobotics Over Delayed Networks (I) |
|
Feizi, Navid | University of Western Ontario |
Patel, Rajnikant V. | The University of Western Ontario |
Kermani, Mehrdad R. | University of Western Ontario |
Atashzar, S. Farokh | New York University (NYU), US |
Keywords: Telerobotics and Teleoperation, Haptics and Haptic Interfaces, Physical Human-Robot Interaction
Abstract: Bilateral telerobotic systems have attracted a great deal of interest during the last two decades. The major challenges in this field are the transparency and stability of remote force rendering that are affected by network delays causing asynchrony between the actions and the corresponding reactions. In addition, the overactivation of stabilizers further degrades the fidelity of the rendered force field. In this paper, a real-time frequency-based delay compensation approach is proposed to maximize transparency while reducing the activation of the stabilization layer. The algorithm uses a Regulated Bound-limited Multiple Fourier Linear Combiner (R-BMFLC) to extract the dominant frequency of force waves. The estimated weights are used in conjunction with the relatively phase-lead harmonic kernels to reconstruct the signal and generate a compensated wave to reduce the effect of the delay. The reconstructed force will then pass through a modulated time-domain passivity controller to guarantee the stability of the system. We will show that the proposed technique will reduce the force tracking error by 40% and the activation of the stabilizer by 79%. It will be shown, for the first time, that through the utilization of online adaptive frequency-based prediction, the asynchrony between transmitted waves through delayed networks can be significantly mitigated while stability can be guaranteed with less activation of the stabilization layer.
|
|
12:40-12:50, Paper TuB-3.2 | |
Analysis of User Behavior and Workload During Simultaneous Tele-Operation of Multiple Mobile Manipulators |
|
Aoki, Tatsuya | Osaka University |
Nakamura, Tomoaki | The University of Electro-Communications |
Nagai, Takayuki | Osaka University |
Keywords: Telerobotics and Teleoperation, Product Design, Development and Prototyping, Multi-Robot Systems
Abstract: This paper discusses the tele-operation system for multiple mobile manipulators. If a single person could freely tele-operate multiple mobile manipulators simultaneously, it would be a great step toward the goal of ”avatar-symbiotic society” allowing people to live beyond the constraints of their bodies, space, and time. At present, however, such a teleoperation system has not been developed. Therefore, we built a prototype system to tele-operate two mobile manipulators and conducted a subject experiment on a pick-and-place task to investigate the tele-operator's task performance, workload and gaze behavior. By analyzing these results, we obtained guidelines to design tele-operation system for multiple robots.
|
|
12:50-13:00, Paper TuB-3.3 | |
Variable Impedance Control for Safety and Usability in Telemanipulation |
|
Schwarz, Stephan Andreas | Chemnitz University of Technology |
Thomas, Ulrike | Chemnitz University of Technology |
Keywords: Telerobotics and Teleoperation, Safety in HRI, Compliance and Impedance Control
Abstract: In recent years, haptic telemanipulation has been introduced to control robots remotely with an input device that generates force feedback. Compliant control strategies are needed to ensure safe interaction between humans and robots. Accurate and precise manipulation requires a stiff setup of the impedance parameters, while safety demands for low stiffness. This paper proposes an impedance-based control approach that combines stiff manipulation with a safety mechanism that adapts compliance when required. We introduce three system modes: operation, safety and recovery mode. If the external forces exceed a defined force threshold, the system switches to the compliant safety mode. A user input triggers the recovery process that increases the stiffness back to its nominal value. This paper suggests an energy tank, which limits the change of stiffness to ensure stability during recovering phase. We validate the functionality of this approach using a real telemanipulation setup and show that the suggested tank enables recovery even from large displacements.
|
|
13:00-13:10, Paper TuB-3.4 | |
Block-Based Novel Haptic Data Reduction for Time-Delayed Teleoperation |
|
Gui, Ming | Technical University of Munich |
Xu, Xiao | Technical University Munich |
Steinbach, Eckehard | Technical University of Munich |
Keywords: Telerobotics and Teleoperation, Haptics and Haptic Interfaces, Force and Tactile Sensing
Abstract: This work proposes a novel haptic data reduction scheme for time-delayed teleoperation by coding information as blocks. State-of-the-art (SOTA) haptic data reduction approaches are mainly sampled-based schemes. They encode haptic signals sample by sample in order to minimize the introduced coding delay. In contrast, our proposed block-based coding approach transmits a sample block as a single unit (haptic packet). Although it introduces additional algorithmic delays that are proportional to the block length, block coding has benefits since the packet rate is easy to control, the coding approach can be lossless, and the intra-block information can be employed to improve the force feedback quality. We further develop an energy adjustment approach that uses the information in a block to mitigate force oscillations caused by the Time Domain Passivity Approach. Simulation experiments and subjective tests demonstrate that our method reduces network load and significantly increases force feedback quality compared with the SOTA sample-based coding schemes, particularly for mid- to high-latency networks and low packet rates.
|
|
13:10-13:20, Paper TuB-3.5 | |
Skill-CPD: Real-Time Skill Refinement for Shared Autonomy in Manipulator Teleoperation |
|
Babaians, Edwin | Technical University of Munich |
Yang, Dong | Technical University of Munich |
Karimi, Mojtaba | Technical University of Munich |
Xu, Xiao | Technical University Munich |
Ayvasik, Serkut | Technical University of Munich |
Steinbach, Eckehard | Technical University of Munich |
Keywords: Telerobotics and Teleoperation, Manipulation Planning, Human-Centered Automation
Abstract: Advanced wireless communication networks provide lower latency and a higher transmission rate. Although this is an enabler for many new teleoperation applications, the risk of network instability or packet drop is still unavoidable. Real-time manipulator teleoperation requires data transmission with no discontinuity. Shared autonomy (SA) is a standard method to mitigate this issue. In this way, if the data from the remote side is unavailable, the controller can continue based on the previously observed models. However, due to the spatial gap between human and robot trajectories, indisputable fluctuations occur, which cause issues in teleoperation applications. This motivates us to propose a new skill refinement strategy to modify the previously trained skill and mitigate the sudden unwanted motions within the control takeover phase. To this end, our approach comprises applying the Hidden Semi-Markov Model (HSMM) and Linear Quadratic Tracker (LQT) in combination to learn and predict the user's intentions and then exploiting Coherent Point Drift (CPD) to refine the executable trajectory. We test our method both in simulation and in the real world for 2D English letter drawing and 3D robot-assisted feeding scenarios. Our experimental results using the Kinova Movo platform show that the proposed refinement approach generates a stable trajectory and mitigates the control switching inconsistency.
|
|
13:20-13:30, Paper TuB-3.6 | |
Haptic Teleoperation of High-Dimensional Robotic Systems Using a Feedback MPC Framework |
|
Cheng, Jin | ETH Zurich |
Abi-Farraj, Firas | ETH |
Farshidian, Farbod | ETH Zurich |
Hutter, Marco | ETH Zurich |
Keywords: Telerobotics and Teleoperation, Whole-Body Motion Planning and Control, Legged Robots
Abstract: Model Predictive Control (MPC) schemes have proven their efficiency in controlling high degree-of-freedom (DoF) complex robotic systems. However, they come at a high computational cost and an update rate of about tens of hertz. This relatively slow update rate hinders the possibility of stable haptic teleoperation of such systems since the slow feedback loops can cause instabilities and loss of transparency to the operator. This work presents a novel framework for transparent teleoperation of MPC-controlled complex robotic systems. In particular, we employ a feedback MPC approach and exploit its structure to account for the operator input at a fast rate which is independent of the update rate of the MPC loop itself. We demonstrate our framework on a mobile manipulator platform and show that it significantly improves haptic teleoperation's transparency and stability. We also highlight that the proposed feedback structure is constraint satisfactory and does not violate any constraints defined in the optimal control problem. To the best of our knowledge, this work is the first realization of the bilateral teleoperation of a legged manipulator using a whole-body MPC framework.
|
|
13:30-13:40, Paper TuB-3.7 | |
Manipulability-Aware Shared Locomanipulation Motion Generation for Teleoperation of Mobile Manipulators |
|
Torielli, Davide | Humanoids and Human Centered Mechatronics (HHCM), Istituto Itali |
Muratore, Luca | Istituto Italiano Di Tecnologia |
Tsagarakis, Nikos | Istituto Italiano Di Tecnologia |
Keywords: Telerobotics and Teleoperation, Motion Control, Redundant Robots
Abstract: The teleoperation of mobile manipulators may pose significant challenges, demanding complex interfaces and causing a substantial burden to the human operator due to the need to switch continuously from the manipulation of the arm to the control of the mobile platform. Hence, several works have considered to exploit shared control techniques to overcome this issue and, in general, to facilitate the task execution. This work proposes a manipulability-aware shared locomanipulation motion generation method to facilitate the execution of telemanipulation tasks with mobile manipulators. The method uses the manipulability level of the end-effector to control the generation of the mobile base and manipulator motions, facilitating their simultaneous control by the operator while executing telemanipulation tasks. Therefore, the operator can exclusively control the end-effector, while the underlying architecture generates the mobile platform commands depending on the end-effector manipulability level. The effectiveness of this approach is demonstrated with a number of experiments in which the CENTAURO robot, a hybrid leg-wheel platform with an anthropomorphic upper body, is teleoperated to execute a set of telemanipulation tasks.
|
|
13:40-13:50, Paper TuB-3.8 | |
Fast Reflexive Grasping with a Proprioceptive Teleoperation Platform |
|
SaLoutos, Andrew | Massachusetts Institute of Technology |
Stanger-Jones, Elijah | Massachusetts Institute of Technology |
Kim, Sangbae | Massachusetts Institute of Technology |
Keywords: Telerobotics and Teleoperation, Grasping, Human-Robot Collaboration
Abstract: We present a proprioceptive teleoperation system that uses a reflexive grasping algorithm to enhance the speed and robustness of pick-and-place tasks. The system consists of two manipulators that use quasi-direct-drive actuation to provide highly transparent force feedback. The end-effector has bimodal force sensors that measure 3-axis force information and 2-dimensional contact location. This information is used for anti-slip and re-grasping reflexes. When the user makes contact with the desired object, the re-grasping reflex aligns the gripper fingers with antipodal points on the object to maximize the grasp stability. The reflex takes only 150ms to correct for inaccurate grasps chosen by the user, so the user's motion is only minimally disturbed by the execution of the re-grasp. Once antipodal contact is established, the anti-slip reflex ensures that the gripper applies enough normal force to prevent the object from slipping out of the grasp. The combination of proprioceptive manipulators and reflexive grasping allows the user to complete teleoperated tasks with precision at high speed.
|
|
13:50-14:00, Paper TuB-3.9 | |
Design Interface Mapping for Efficient Free-Form Tele-Manipulation |
|
Unni Krishnan, Achyuthan | Worcester Polytechnic Institute |
Lin, Tsung-Chi | Worcester Polytechnic Institute |
Li, Zhi | Worcester Polytechnic Institute |
Keywords: Telerobotics and Teleoperation, Design and Human Factors, Human Factors and Human-in-the-Loop
Abstract: Motion tracking interfaces are intuitive for free-form teleoperation tasks. However, efficient manipulation control can be difficult with such interfaces because of issues like the interference of unintended motions and the limited precision of human motion control. The limitation in control efficiency reduces the operator's performance and increases their workload and frustration during robot teleoperation. To improve the efficiency, we proposed separating controlled degrees of freedom (DoFs) and adjusting the motion scaling ratio of a motion tracking interface. The motion tracking of handheld controllers from a Virtual Reality system was used for the interface. We separated the translation and rotational control into: 1) two controllers held in the dominant and non-dominant hands and 2) hand pose tracking and trackpad inputs of a controller. We scaled the control mapping ratio based on 1) the environmental constraints and 2) the teleoperator's control speed. We further conducted a user study to investigate the effectiveness of the proposed methods in increasing efficiency. Our results show that the separation of position and orientation control into two controllers and the environment-based scaling methods perform better than their alternatives.
|
|
TuB-4 |
Rm4 (Room C-1) |
Calibration and Identification |
Regular session |
Chair: Takemura, Kentaro | Tokai University |
Co-Chair: Aoyama, Tadayoshi | Nagoya University |
|
12:30-12:40, Paper TuB-4.1 | |
Simultaneous Calibration of Multiple Revolute Joints for Articulated Vision Systems Via SE(3) Kinematic Bundle Adjustment |
|
Zhou, Chengzhe | Shanghai Institute of Microsystem and Information Technology |
Sun, Qixuan | Shanghai Institute of Microsystem and Information Technology |
Wang, Kaifang | CAS |
Li, Jiamao | Shanghai Institute of Microsystem and Information Technology |
Zhang, Xiaolin | Shanghai Institute of Microsystem and Information Technology |
Keywords: Calibration and Identification, Computer Vision for Automation
Abstract: We propose a visual-based approach to calibrate kinematic structure of low degree-of-freedom (DoF) articulated systems. For industrial robots, the kinematic accuracy of end-effector (EE) is known with certitude. Standard hand-eye calibration (HEC) yields excellent eye-to-hand relations by explicitly estimating EE-mounted camera poses from the Perspective-n-Point (PnP) problem of a single calibration rig. However, these methods struggle when the ideal kinematic model is unknown or inaccurate, which are typical in customized serial chains such as articulated vision systems (AVS). Inspired by bundle adjustment (BA) in structure-from-motion (SfM) methods, we proposed an approach, dubbed KBA, to simultaneously localize multiple joint axes at the kinematic level instead of in the 3D space. To achieve this, we design a robust multi-checkerboard detector to enlarge the kinematic coverage of end-effector samples, which are usually limited by sensor's field of view (FoV). The overall optimization problem is formulated and solved using Lie group SE(3)---suitable for integration into modern simultaneous localization and mapping (SLAM) and SfM systems. In addition to simulated scenes, the experimental results with articulated monocular (AMV) and binocular vision (ABV) confirm that the proposed method is applicable to real serial kinematic chains and enables bio-mimicking 3D vision tasks such as stereo reconstruction in motion.
|
|
12:40-12:50, Paper TuB-4.2 | |
DXQ-Net: Differentiable LiDAR-Camera Extrinsic Calibration Using Quality-Aware Flow |
|
Jing, Xin | Zhejiang University |
Ding, Xiaqing | Zhejiang University |
Xiong, Rong | Zhejiang University |
Deng, Huanjun | Alibaba (Beijing) Software Services Co., Ltd |
Wang, Yue | Zhejiang University |
Keywords: Calibration and Identification, Deep Learning Methods
Abstract: Accurate LiDAR-camera extrinsic calibration is a precondition for many multi-sensor systems in mobile robots. Most calibration methods rely on laborious manual operations and calibration targets. While working online, the calibration methods should be able to extract information from the environment to construct the cross-modal data association. Convolutional neural networks (CNNs) have powerful feature extraction ability and have been used for calibration. However, most of the past methods solve the extrinsic as a regression task, without considering the geometric constraints involved. In this paper, we propose a novel end-to-end extrinsic calibration method named DXQ-Net, using a differentiable pose estimation module for generalization. We formulate a probabilistic model for LiDAR-camera calibration flow, yielding a prediction of uncertainty to measure the quality of LiDAR-camera data association. Testing experiments illustrate that our method achieves a competitive with other methods for the translation component and state-of-the-art performance for the rotation component. Generalization experiments illustrate that the generalization performance of our method is significantly better than other deep learning-based methods.
|
|
12:50-13:00, Paper TuB-4.3 | |
CSA-SVM Method for Internal Cavitation Defects Detection and Its Application of District Heating Pipes |
|
Chen, Yanran | Beijing University of Chemical Technology |
Ma, Shugen | Ritsumeikan University |
Li, Longchuan | Beijing University of Chemical Technology |
Li, Zhiqing | Beijing University of Chemical Technology |
Yang, Yulin | Beijing University of Chemical Technology |
Keywords: Calibration and Identification, In-Hand Manipulation, Marine Robotics
Abstract: The goal of this paper is to develop an ultrasonic detection device that can be mounted on an underwater snake vehicle (USV) for underwater district heating pipe (DHP) detection in the future. Ultrasonic detection technology (UDT) is the detection means used, and the cavitation defects in polyurethane (PUR) layer of DHPs are the object being detected. Due to the large thickness of PUR layer and the complex interface information of multi-layer structure, detecting defects of DHPs quantitatively is a difficult task. To address this issue, this paper proposes an approach that combines feature extraction and crow search algorithm (CSA) optimized support vector machine (SVM). Firstly, the main parameters and detection method of UDT are designed after investigation. Secondly, defective signals are pre-processed by signal processing to extract the features form three domains. Finally, four different classifiers are used to identify cavitation defects based on the feature-set. When compared among optimized random forest (RF), k-nearest neighbor (KNN), and ordinary SVM, the experimental results show that CSA-SVM had the highest accuracy in defect size prediction, and the validation-experiment verifies the practicability and feasibility of the CSA-SVM classifier. All experiments illustrate that the issue could be well solved by our method.
|
|
13:00-13:10, Paper TuB-4.4 | |
Industrial Robot Parameter Identification Using a Constrained Instrumental Variable Method |
|
Ardiani, Fabio | ONERA |
Janot, Alexandre | ONERA |
Benoussaad, Mourad | INP-ENI of Tarbes |
Keywords: Calibration and Identification, Industrial Robots
Abstract: Robot identification is a prolific topic that has a long history with results spanning recent decades. Recent years have witnessed a renew of interest in this problem due in part to a rapid increase in robotic hardware platforms capable of accurate model-based control. The most popular methods exploit the fact that the inverse dynamic model is linear to the dynamic parameters. Because we identify robots with closed-loop procedures, an Instrumental Variable approach called IDIM-IV (Inverse Dynamic Identification Model with Instrumental Variable estimation) that combines the direct and inverse dynamic models to prevent from correlation between errors has been successfully validated. However, IDIM-IV does not guarantee that the direct dynamic model will be well-posed during its iterations because of possible modeling errors. In this paper, we combine physical constraints and IDIM-IV to address this deficiency for IDIM-IV. This new constrained IV approach, called PC-IDIM-IV (Physically Consistent IDIM-IV), consists of two nested iterative algorithms: an outer one that is IDIM-IV and an inner one that accounts for the physical constraints solved by a Gauss-Newton algorithm. Experimental results and comparisons with other methods carried out with the TX40 robot show the feasibility of PC-IDIM-IV.
|
|
13:10-13:20, Paper TuB-4.5 | |
TEScalib: Targetless Extrinsic Self-Calibration of LiDAR and Stereo Camera for Automated Driving Vehicles with Uncertainty Analysis |
|
Hu, Haohao | Karlsruhe Institute of Technology |
Han, Fengze | KIT |
Bieder, Frank | Karlsruhe Institute of Technology |
Pauls, Jan-Hendrik | Karlsruhe Institute of Technology (KIT) |
Stiller, Christoph | Karlsruhe Institute of Technology |
Keywords: Calibration and Identification, Intelligent Transportation Systems, Computer Vision for Transportation
Abstract: In this paper, we present TEScalib, a novel extrinsic self-calibration approach of LiDAR and stereo camera using the geometric and photometric information of surrounding environments without any calibration targets for automated driving vehicles. Since LiDAR and stereo camera are widely used for sensor data fusion on automated driving vehicles, their extrinsic calibration is highly important. However, most of the LiDAR and stereo camera calibration approaches are mainly target-based and therefore time consuming. Even the newly developed targetless approaches in last years are either inaccurate or unsuitable for driving platforms. To address those problems, we introduce TEScalib. By applying a 3D mesh reconstruction-based point cloud registration, the geometric information is used to estimate the LiDAR to stereo camera extrinsic parameters accurately and robustly. To calibrate the stereo camera, a photometric error function is builded and the LiDAR depth is involved to transform key points from one camera to another. During driving, these two parts are processed iteratively. Besides that, we also propose an uncertainty analysis for reflecting the reliability of the estimated extrinsic parameters. Our TEScalib approach evaluated on the KITTI dataset achieves very promising results.
|
|
13:20-13:30, Paper TuB-4.6 | |
Extrinsic Calibration of a 2D Laser Rangefinder and a Depth-Camera Using an Orthogonal Trihedron |
|
Li, Zhengbin | Tianjin University, China |
Dong, Haiqing | AIIC of Midea Group (Shanhai) Co., Ltd |
Ding, Yabin | Tianjin University, China |
Liu, Dong | Midea Group (Shanghai) Co., Ltd |
Keywords: Calibration and Identification, Sensor Fusion, SLAM
Abstract: 2D laser rangefinders and depth-cameras are usually equipped on service robots. But there are rarely calibration methods of them. This paper proposes an extrinsic calibration method of a 2D laser rangefinder and a depth-camera using an orthogonal trihedron. The trihedron with orthogonal assumptions is taken as a reference frame to roughly estimate the relative pose between the sensors by solving a perspective-three-point (P3P) problem and basis-to-basis correspondence. Then, the estimated relative pose is refined via non-linear optimization based on line-to-plane constraints. Unlike other works which require enough motion, only one-shot observation is required, and it is insensitive to sensor ranging noise and the manufacturing errors of calibration targets. Verified by simulation and real experiments, the proposed method is simple, effective and accurate.
|
|
13:30-13:40, Paper TuB-4.7 | |
Continuous Calibration and Narrow Compensation Algorithm to Estimate a Joint Axis under the Various Conditions with Unit Sensor |
|
Choi, Jungsu | Yeungnam University |
Seo, Wonjeong | Yeungnam University |
Lee, Haseok | Yeungnam University |
Keywords: Calibration and Identification, Sensor Fusion, Wearable Robotics
Abstract: Wearable robots have been developed to aid or substitute the gait locomotion of humans. To assist gait locomotion based on the intention of a wearer, a gait pattern analysis is required with a wearable sensor by measuring body information, i.e., a joint angular velocity. However, measuring a precise joint angular velocity is difficult because the attachment position of a sensor has a curvature and an anatomical joint axis which is invisible. Therefore, a sensor calibration algorithm, which aligns a sensor axis into an anatomical joint axis, is required to provide an optimal assist for a wearer. Hence, in this paper, a new and simple sensor calibration algorithm is proposed with a unit sensor. Since a wearer shakes the body or collides with the ground when walking, the attachment position of a sensor may be changed. Thus, a continuous sensor compensation algorithm is also proposed. Additionally, the effectiveness of this new algorithm is demonstrated by gait locomotion experiments on various paths.
|
|
13:40-13:50, Paper TuB-4.8 | |
Visual-Inertial-Aided Online MAV System Identification |
|
Chen, Chuchu | University of Delaware |
Yang, Yulin | University of Delaware |
Geneva, Patrick | University of Delaware |
Lee, Woosik | University of Delaware |
Huang, Guoquan | University of Delaware |
Keywords: Calibration and Identification, Vision-Based Navigation, Aerial Systems: Applications
Abstract: System modeling and parameter identification of micro aerial vehicles (MAV) are crucial for robust autonomy, especially under highly dynamic motions. Visual-inertial-aided online parameter identification has recently seen research atten- tion due to the demanding of adaptation to platform configura- tion changes with minimal onboard sensor requirements. To this end, we design an online MAV system identification algorithm to tightly fuse visual, inertial and MAV aerodynamic informa- tion within a lightweight multi-state constraint Kalman filter (MSCKF) framework. In particular, while one could blindly fuse the MAV dynamic-induced relative motion constraints in EKF, we numerically show that due to the (quadrotor) MAV system modeling inaccuracy, they often become overconfident and negatively impact the state estimates. As such, we leverage the Schmidt-Kalman filter (SKF) for MAV system parameter identification to prevent corruption of state estimates. Through extensive simulations and real-world experiments, we validate the proposed SKF-based scheme and demonstrate its ability to perform robust system identification even in the presence of an inconsistent MAV dynamic model under different motions.
|
|
13:50-14:00, Paper TuB-4.9 | |
Efficient Extrinsic Calibration of Multi-Sensor 3D LiDAR Systems for Autonomous Vehicles Using Static Objects Information |
|
Ponton, Brahayam | MPI-IS |
Ferri, Magda | Ibeo Automotive Systems GmbH |
König, Lars | Ibeo Automotive Systems GmbH |
Bartels, Marcus | Hamburg University of Technology (TUHH) |
Keywords: Calibration and Identification, Sensor Fusion, Optimization and Optimal Control
Abstract: For an autonomous vehicle, the ability to sense its surroundings and to build an overall representation of the environment by fusing different sensor data streams is fundamental. To this end, the poses of all sensors need to be accurately determined. Traditional calibration methods are based on: 1) using targets specifically designed for calibration purposes in controlled environments, 2) optimizing a quality metric of the point clouds collected while traversing an unknown but static environment, or 3) optimizing the match among per-sensor incremental motion observations along a motion path fulfilling special requirements. In real scenarios, however, the online applicability of these methods can be limited, as they are typically highly dynamic, contain degenerate paths, and require fast computations. In this paper, we propose an approach that tackles some of these challenges by formulating the calibration problem as a joint but structured optimization problem of all sensor calibrations that takes as input a summary of the point cloud information consisting of ground points and pole detections. We demonstrate the efficiency and quality of the results of the proposed approach in a set of experiments with LiDAR simulation and real data from an urban trip.
|
|
TuB-5 |
Rm5 (Room C-2) |
Navigation Systems 4 |
Regular session |
Chair: Bezzo, Nicola | University of Virginia |
Co-Chair: Chinchali, Sandeep | The University of Texas at Austin |
|
12:30-12:40, Paper TuB-5.1 | |
Online Mapping and Motion Planning under Uncertainty for Safe Navigation in Unknown Environments (I) |
|
Pairet Artau, Èric | Technology Innovation Institute |
Hernández, Juan David | Cardiff University |
Carreras, Marc | Universitat De Girona |
Petillot, Yvan R. | Heriot-Watt University |
Lahijanian, Morteza | University of Colorado Boulder |
Keywords: Autonomous Vehicle Navigation, Motion and Path Planning, Motion Control
Abstract: Safe autonomous navigation is an essential and challenging problem for robots operating in highly unstructured or completely unknown environments. Under these conditions, not only robotic systems must deal with limited localisation information but also their manoeuverability is constrained by their dynamics and often suffer from uncertainty. In order to cope with these constraints, this article proposes an uncertainty-based framework for mapping and planning feasible motions online with probabilistic safety guarantees. The proposed approach deals with the motion, probabilistic safety, and online computation constraints by: (i) incrementally mapping the surroundings to build an uncertainty-aware representation of the environment, and (ii) iteratively (re)planning trajectories to goal that are kinodynamically feasible and probabilistically safe through a multi-layered sampling-based planner in the belief space. In-depth empirical analyses illustrate some important properties of this approach, namely: (a) the multilayered planning strategy enables rapid exploration of the high-dimensional belief space while preserving asymptotic optimality and completeness guarantees, and (b) the proposed routine for probabilistic collision checking results in tighter probability bounds in comparison to other uncertainty-aware planners in the literature. Furthermore, real-world in-water experimental evaluation on a nonholonomic torpedo-shaped autonomous underwater vehicle and simulated trials in an urban environment on an unmanned aerial vehicle demonstrate the efficacy of the method and its suitability for systems with limited on-board computational power.
|
|
12:40-12:50, Paper TuB-5.2 | |
Drift Reduced Navigation with Deep Explainable Features |
|
Mohd, Omama | IIIT-Hyderabad |
Venugopalaswamy Sriraman, Sundar Sripada | IIIT-Hyderabad |
Chinchali, Sandeep | The University of Texas at Austin |
Singh, Arun Kumar | University of Tartu |
Krishna, Madhava | IIIT Hyderabad |
Keywords: Autonomous Vehicle Navigation, Perception-Action Coupling, Visual Learning
Abstract: Modern autonomous vehicles (AVs) often rely on vision, LIDAR, and even radar-based simultaneous localization and mapping (SLAM) frameworks for precise localization and navigation. However, modern SLAM frameworks often lead to unacceptably high levels of drift (i.e., localization error) when AVs observe few visually distinct features or encounter occlusions due to dynamic obstacles. This paper argues that minimizing drift must be a key desiderata in AV motion planning, which requires an AV to take active control decisions to move towards feature-rich regions while also minimizing conventional control cost. To do so, we first introduce a novel data-driven perception module that observes LIDAR point clouds and estimates which features/regions an AV must navigate towards for drift minimization. Then, we introduce an interpretable model predictive controller (MPC) that moves an AV toward such feature-rich regions while avoiding visual occlusions and gracefully trading off drift and control cost. Our experiments on challenging, dynamic scenarios in the state-of-the-art CARLA simulator indicate our method reduces drift up to 76.76% compared to benchmark approaches.
|
|
12:50-13:00, Paper TuB-5.3 | |
A Robust and Fast Occlusion-Based Frontier Method for Autonomous Navigation in Unknown Cluttered Environments |
|
Mohammad, Nicholas | University of Virginia |
Bezzo, Nicola | University of Virginia |
Keywords: Autonomous Vehicle Navigation, Motion and Path Planning, Range Sensing
Abstract: Navigation through unknown, cluttered environments is a fundamental and challenging task for autonomous vehicles as they must deal with a myriad of obstacle configurations typically unknown a priori. Challenges arise because obstacles of unknown shapes and dimensions can create occlusions limiting sensor field of view and leading to uncertainty in motion planning. In this paper we propose to leverage such occlusions to quickly explore and cover unknown cluttered environments. Specifically, this work presents a novel occlusion-aware frontier-based approach that estimates gaps in point cloud data and shadows in the field of view to generate waypoints to navigate. Our scheme also proposes a breadcrumbing technique to save states of interest during exploration that can be exploited in future missions. For the latter aspect we focus primarily on the generation of the minimum number of breadcrumbs that will increase coverage and visibility of an explored environment. Extensive simulations and experiment results on an unmanned ground vehicle (UGV) are demonstrated to validate the proposed technique, showing improvements over traditional state of the art frontier-based exploration methods.
|
|
13:00-13:10, Paper TuB-5.4 | |
Bubble Planner: Planning High-Speed Smooth Quadrotor Trajectories Using Receding Corridors |
|
Ren, Yunfan | The University of Hong Kong |
Zhu, Fangcheng | The University of Hong Kong |
Liu, Wenyi | Harbin Institute of Technology, Shenzhen |
Wang, Zhepei | Zhejiang University |
Lin, Yi | Hong Kong University of Science and Technology |
Gao, Fei | Zhejiang University |
Zhang, Fu | University of Hong Kong |
Keywords: Autonomous Vehicle Navigation, Motion and Path Planning, Aerial Systems: Applications
Abstract: Quadrotors are agile platforms. With human experts, they can perform extremely high-speed flights in cluttered environments. However, fully autonomous flight at high speed remains a significant challenge. In this work, we propose a motion planning algorithm based on the corridor-constrained minimum control effort trajectory optimization (MINCO) framework. Specifically, we use a series of overlapping spheres to represent the free space of the environment and propose two novel designs that enable the algorithm to plan high-speed quadrotor trajectories in real-time. One is a sampling-based corridor generation method that generates spheres with large overlapped areas (hence overall corridor size) between two neighboring spheres. The second is a Receding Horizon Corridors (RHC) strategy, where part of the previously generated corridor is reused in each replan. Together, these two designs enlarge the corridor spaces in accordance with the quadrotor's current state and hence allow the quadrotor to maneuver at high speeds. We benchmark our algorithm against other state-of-the-art planning methods to show its superiority in simulation. Comprehensive ablation studies are also conducted to show the necessity of the two designs. The proposed method is finally evaluated on an autonomous LiDAR-navigated quadrotor UAV in woods environments, achieving flight speeds over 13.7 m/s without any prior map of the environment or external localization facility.
|
|
13:10-13:20, Paper TuB-5.5 | |
Temporal Context for Robust Maritime Obstacle Detection |
|
Zust, Lojze | University of Ljubljana |
Kristan, Matej | University of Ljubljana |
Keywords: Autonomous Vehicle Navigation, Marine Robotics, Semantic Scene Understanding
Abstract: Robust maritime obstacle detection is essential for fully autonomous unmanned surface vehicles (USVs). The currently widely adopted segmentation-based obstacle detection methods are prone to misclassification of object reflections and sun glitter as obstacles, producing many false positive detections, effectively rendering the methods impractical for USV navigation. However, water-turbulence-induced temporal appearance changes on object reflections are very distinctive from the appearance dynamics of true objects. We harness this property to design WaSR-T, a novel maritime obstacle detection network, that extracts the temporal context from a sequence of recent frames to reduce ambiguity. By learning the local temporal characteristics of object reflection on the water surface, WaSR-T substantially improves obstacle detection accuracy in the presence of reflections and glitter. Compared with existing single-frame methods, WaSR-T reduces the number of false positive detections by 41% overall and by over 53% within the danger zone of the boat, while preserving a high recall, and achieving new state-of-the-art performance on the challenging MODS maritime obstacle detection benchmark. The code, pretrained models and extended datasets are available at: https://github.com/lojzezust/WaSR-T
|
|
13:20-13:30, Paper TuB-5.6 | |
Information-Aware Guidance for Magnetic Anomaly Based Navigation |
|
Ramos, J Humberto | University of Florida |
Shin, Jaejeong | Cornell University |
Volle, Kyle | National Research Council Postdoctoral Program |
Buzaud, Paul | University of Florida |
Brink, Kevin | Air Force Research Lab |
Ganesh, Prashant | University of Florida |
Keywords: Autonomous Vehicle Navigation, Task and Motion Planning, Optimization and Optimal Control
Abstract: In the absence of an absolute positioning system, such as GPS, autonomous vehicles are subject to accumulation of position error which can interfere with reliable performance. Improved navigational accuracy without GPS enables vehicles to achieve a higher degree of autonomy and reliability, both in terms of decision making and safety. This paper details the use of two path planning algorithms for autonomous agents when using magnetic field anomalies to localize themselves within a map. Both techniques use the information content in the environment in distinct ways and is aimed at reducing the localization uncertainty. The first method is based on a nonlinear observability metric of the vehicle model, while the second is an information-theory based technique which minimizes the expected entropy of the system. These conditions are used to design guidance laws that minimize the localization uncertainty and is verified both in simulation and hardware experiments.
|
|
13:30-13:40, Paper TuB-5.7 | |
Unified Automatic Control of Vehicular Systems with Reinforcement Learning (I) |
|
Yan, Zhongxia | Massachusetts Institute of Technology |
Kreidieh, Aboudy | UC Berkeley |
Vinitsky, Eugene | UC Berkeley |
Bayen, Alexandre | UC Berkeley |
Wu, Cathy | MIT |
Keywords: Path Planning for Multiple Mobile Robots or Agents, Reinforcement Learning, Automation Technologies for Smart Cities
Abstract: Emerging vehicular systems with increasing proportions of automated components present opportunities for optimal control to mitigate congestion and increase efficiency. There has been recent interest in applying deep reinforcement learning (DRL) to these nonlinear dynamical systems for the automatic design of effective control strategies. Despite conceptual advantages of DRL being model-free, studies typically nonetheless rely on training setups that are painstakingly specialized to specific vehicular systems. This is a key challenge to efficient analysis of diverse vehicular and mobility systems. To this end, this article contributes a streamlined methodology for vehicular microsimulation and discovers high performance control strategies with minimal manual design. A variable-agent, multi-task approach is presented for optimization of vehicular Partially Observed Markov Decision Processes. The methodology is experimentally validated on mixed autonomy traffic systems, where fractions of vehicles are automated; empirical improvement, typically 15-60% over a human driving baseline, is observed in all configurations of six diverse open or closed traffic systems. The study reveals numerous emergent behaviors resembling wave mitigation, traffic signaling, and ramp metering. Finally, the emergent behaviors are analyzed to produce interpretable control strategies.
|
|
13:40-13:50, Paper TuB-5.8 | |
SphereMap: Dynamic Multi-Layer Graph Structure for Rapid Safety-Aware UAV Planning |
|
Musil, Tomáš | Czech Technical University in Prague, Faculty of Electrical Engi |
Petrlik, Matej | Czech Technical University in Prague, Faculty of Electrical Engi |
Saska, Martin | Czech Technical University in Prague |
Keywords: Autonomous Vehicle Navigation, Planning under Uncertainty, Mapping
Abstract: A flexible topological representation consisting of a two-layer graph structure built on-board an Unmanned Aerial Vehicle (UAV) by continuously filling the free space of an occupancy map with intersecting spheres is proposed in this letter. Most state-of-the-art planning methods find the shortest paths while keeping the UAV at a pre-defined distance from obstacles. Planning over the proposed structure reaches this pre-defined distance only when necessary, maintaining a safer distance otherwise, while also being orders of magnitude faster than other state-of-the-art methods. Furthermore, we demonstrate how this graph representation can be converted into a lightweight shareable topological-volumetric map of the environment, which enables decentralized multi-robot cooperation. The proposed approach was successfully validated in several kilometers of real subterranean environments, such as caves, devastated industrial buildings, and in the harsh and complex setting of the final event of the DARPA SubT Challenge, which aims to mimic the conditions of real search and rescue missions as closely as possible, and where our approach achieved the 2nd place in the virtual track.
|
|
13:50-14:00, Paper TuB-5.9 | |
AIB-MDP: Continuous Probabilistic Motion Planning for Automated Vehicles by Leveraging Action Independent Belief Spaces |
|
Naumann, Maximilian | Bosch Center for Artificial Intelligence |
Stiller, Christoph | Karlsruhe Institute of Technology |
Keywords: Autonomous Vehicle Navigation, Motion and Path Planning
Abstract: While automated research vehicles are already populating the roads, their commercial availability at scale is still to come. Presumably, one of the key challenges is to derive behaviors that are safe and comfortable but at the same time not overcautious, despite considerable uncertainties. These uncertainties stem from imperfect perception, occlusions and limited sensor range, but also from the unknown future behavior of other traffic participants. A holistic uncertainty treatment, for example in a general POMDP formulation, often induces a strong limitation on the action space due to the need for real-time capability. Further, related approaches often do not account for the need for verifiable safety, including traffic rule compliance. The proposed approach is targeted towards scenarios with clear precedence. It is based on an MDP with an action-independent belief (AIB-MDP): We assume that the future belief over the trajectories of other traffic participants is independent of the ego vehicle’s behavior. Thus, the future belief can be predicted and simplified in an upstream module, independent of motion planning. This modularization facilitates subsequent ego motion planning in a continuous action space despite the thorough uncertainty consideration.
|
|
TuB-6 |
Rm6 (Room D) |
SLAM 5 |
Regular session |
Chair: Oishi, Takeshi | The University of Tokyo |
Co-Chair: Karimi, Mojtaba | Technical University of Munich |
|
12:30-12:40, Paper TuB-6.1 | |
InCOpt: Incremental Constrained Optimization Using the Bayes Tree |
|
Qadri, Mohamad | Carnegie Mellon University |
Sodhi, Paloma | Carnegie Mellon University |
Mangelson, Joshua | Brigham Young University |
Dellaert, Frank | Georgia Institute of Technology |
Kaess, Michael | Carnegie Mellon University |
Keywords: SLAM, Localization, Mapping
Abstract: In this work, we investigate the problem of incrementally solving constrained non-linear optimization problems formulated as factor graphs. Prior incremental solvers were either restricted to the unconstrained case or required periodic batch relinearizations of the objective and constraints which are expensive and detract from the online nature of the algorithm. We present InCOpt, an Augmented Lagrangian-based incremental constrained optimizer that views matrix operations as message passing over the Bayes tree. We first show how the linear system, resulting from linearizing the constrained objective, can be represented as a Bayes tree. We then propose an algorithm that views forward and back substitutions, which naturally arise from solving the Lagrangian, as upward and downward passes on the tree. Using this formulation, InCOpt can exploit properties such as fluid/online relinearization leading to increased accuracy without a sacrifice in runtime. We evaluate our solver on different applications (navigation and manipulation) and provide an extensive evaluation against existing constrained and unconstrained solvers.
|
|
12:40-12:50, Paper TuB-6.2 | |
S3LAM: Structured Scene SLAM |
|
Gonzalez, Mathieu | IRT B<>com |
Marchand, Eric | Univ Rennes, Inria, CNRS, IRISA |
Kacete, Amine | IRT B<>com |
Royan, Jerome | IRT B-Com |
Keywords: SLAM
Abstract: We propose a new SLAM system that uses the semantic segmentation of objects and structures in the scene. Semantic information is relevant as it contains high level information which may make SLAM more accurate and robust. Our contribution is twofold: i) A new SLAM system based on ORB-SLAM2 that creates a semantic map made of clusters of points corresponding to objects instances and structures in the scene. ii) A modification of the classical Bundle Adjustment formulation to constrain each cluster using geometrical priors, which improves both camera localization and reconstruction and enables a better understanding of the scene. We evaluate our approach on sequences from several public datasets and show that it improves camera pose estimation with respect to state of the art.
|
|
12:50-13:00, Paper TuB-6.3 | |
Fast Structural Representation and Structure-Aware Loop Closing for Visual SLAM |
|
Xie, Shuxiang | The University of Tokyo |
Ishikawa, Ryoichi | The University of Tokyo |
Sakurada, Ken | National Institute of Advanced Industrial Science and Technology |
Onishi, Masaki | National Inst. of AIST |
Oishi, Takeshi | The University of Tokyo |
Keywords: SLAM, Localization, Range Sensing
Abstract: Perceptual Aliasing is one of the main problems in simultaneous localization and mapping (SLAM). Wrong associations between different places may lead to failure of the whole map. Research on structure information is rarely investigated among existing solutions to this problem. In cases of visual SLAM without sensors, such as LiDAR or Inertial Measurement Unit (IMU), structure information can rarely be obtained due to the sparsity of 3D points, which also makes structure analysis complex. This study provides a spherical harmonics (SH) based fast structural representation (SH-FS) in visual SLAM using sparse point clouds, which extracts the structure information from sparse points into single vector. SH-FS was applied in conventional feature-based loop closing process. Furthermore, a structure-aware loop closing method in visual SLAM was proposed to improve the robustness of SLAM systems. Moreover, our methods show a favorable performance in extensive experiments on different large-scale real world datasets.
|
|
13:00-13:10, Paper TuB-6.4 | |
Efficient 2D Graph SLAM for Sparse Sensing |
|
Zhou, Hanzhi | University of Virginia |
Hu, Zichao | University of Texas at Austin |
Liu, Sihang | University of Virginia |
Khan, Samira | University of Virginia |
Keywords: SLAM
Abstract: Simultaneous localization and mapping (SLAM) plays a vital role in mapping unknown spaces and aiding autonomous navigation. Virtually all state-of-the-art solutions today for 2D SLAM are designed for dense and accurate sensors such as laser range-finders (LiDARs). However, these sensors are not suitable for resource-limited nano robots, which become increasingly capable and ubiquitous nowadays, and these robots tend to mount economical and low-power sensors that can only provide sparse and noisy measurements. This introduces a challenging problem called SLAM with sparse sensing. This work addresses the problem by adopting the form of the state-of-the-art graph-based SLAM pipeline with a novel frontend and an improvement for loop closing in the backend, both of which are designed to work with sparse and uncertain range data. Experiments show that the maps constructed by our algorithm have superior quality compared to prior works on sparse sensing. Furthermore, our method is capable of running in real-time on a modern PC with an average processing time of 1/100th the input interval time.
|
|
13:10-13:20, Paper TuB-6.5 | |
Spectral Measurement Sparsification for Pose-Graph SLAM |
|
Doherty, Kevin | Massachusetts Institute of Technology |
Rosen, David | Northeastern University |
Leonard, John | MIT |
Keywords: SLAM, Localization, Mapping
Abstract: Simultaneous localization and mapping (SLAM) is a critical capability in autonomous navigation, but in order to scale SLAM to the setting of "lifelong" SLAM, particularly under memory or computation constraints, a robot must be able to determine what information should be retained and what can safely be forgotten. In graph-based SLAM, the number of edges (measurements) in a pose graph determines both the memory requirements of storing a robot's observations and the computational expense of algorithms deployed for performing state estimation using those observations; both of which can grow unbounded during long-term navigation. To address this, we propose a spectral approach for pose graph sparsification which maximizes the algebraic connectivity of the sparsified measurement graphs, a key quantity which has been shown to control the estimation error of pose graph SLAM solutions. Our algorithm, MAC (for "maximizing algebraic connectivity"), which is based on convex relaxation, is simple and computationally inexpensive, and admits formal post hoc performance guarantees on the quality of the solutions it provides. In experiments on benchmark pose-graph SLAM datasets, we show that our approach quickly produces high-quality sparsification results which retain the connectivity of the graph and, in turn, the quality of corresponding SLAM solutions, as compared to a baseline approach which does not consider graph connectivity.
|
|
13:20-13:30, Paper TuB-6.6 | |
BOEM-SLAM: A Block Online EM Algorithm for the Visual-Inertial SLAM Backend |
|
Pogue, Alexandra | UCLA |
Chang, Tsang-Kai | University of California, Los Angeles |
Mehta, Ankur | UCLA |
Keywords: SLAM, Optimization and Optimal Control, Localization
Abstract: In this paper we present BOEM-SLAM, a backend for visual-inertial SLAM systems capable of creating a globally consistent trajectory and map without retaining the entire history of data. By leveraging the hidden Markov model structure, BOEM-SLAM can summarize historical data into sufficient statistics and then discard it. As a data-efficient algorithm, BOEM-SLAM addresses the growing computational costs and storage requirements of the SLAM backend. To demonstrate the performance of our algorithm we compare BOEM-SLAM to other fundamental approaches on both synthetic data and the EuRoC dataset. For evaluation on the EuRoC dataset, we use the open source okvis frontend and apply the Lie group state space representation and visual outlier removal. Overall, BOEM-SLAM shows a considerably lower computation time with comparable estimation performance. For example, the processing time of BOEM-SLAM is 50 times smaller than the optimization-based method using simulated data and 5 times smaller than the optimization-based method in the EuRoC dataset experiments.
|
|
13:30-13:40, Paper TuB-6.7 | |
RO-LOAM: 3D Reference Object-Based Trajectory and Map Optimization in LiDAR Odometry and Mapping |
|
Oelsch, Martin | Technical University of Munich |
Karimi, Mojtaba | Technical University of Munich |
Steinbach, Eckehard | Technical University of Munich |
Keywords: SLAM, Localization, Mapping
Abstract: We propose an extension to the LiDAR Odometry and Mapping framework (LOAM) that enables reference object-based trajectory and map optimization. Our approach assumes that the location and geometry of a large reference object are known, e.g., as a CAD model from Building Information Modeling (BIM) or a previously captured dense point cloud model. We do not expect the reference object to be present in every LiDAR scan. Our approach uses the poses of the LOAM algorithm as an initial guess to refine them with scan-to-model alignment. To evaluate if the alignment was accurate, an EKF-based motion prior filtering step is employed. Subsequently, the past trajectory is optimized by adding the model-aligned pose as a pose graph constraint and the map of the LOAM algorithm is corrected to improve future localization and mapping. We evaluate our approach with data captured in a visual airplane inspection scenario inside an aircraft hangar. A 3D LiDAR sensor is mounted via a gimbal on an Unmanned Aerial Vehicle (UAV) and is continuously actuated. We compare the localization accuracy of the LOAM and R-LOAM algorithms when enabling or disabling our proposed reference object-based trajectory and map optimization extension. For three recorded datasets, enabling the proposed extension yields a reduction in Absolute Pose Error compared to conventional LOAM and R-LOAM, while being able to run online. This reduces drift and improves map quality.
|
|
13:40-13:50, Paper TuB-6.8 | |
TwistSLAM: Constrained SLAM in Dynamic Environment |
|
Gonzalez, Mathieu | IRT B<>com |
Marchand, Eric | Univ Rennes, Inria, CNRS, IRISA |
Kacete, Amine | IRT B<>com |
Royan, Jerome | IRT B-Com |
Keywords: SLAM
Abstract: Classical visual simultaneous localization and mapping (SLAM) algorithms usually assume the environment to be rigid. This assumption limits the applicability of those algorithms as they are unable to accurately estimate the camera poses and world structure in real life scenes containing moving objects (e.g. cars, bikes, pedestrians, etc.). To tackle this issue, we propose TwistSLAM: a semantic, dynamic and stereo SLAM system that can track dynamic objects in the environment. Our algorithm creates clusters of points according to their semantic class. Thanks to the definition of inter-cluster constraints modeled by mechanical joints (function of the semantic class), a novel constrained bundle adjustment is then able to jointly estimate both poses and velocities of moving objects along with the classical world structure and camera trajectory. We evaluate our approach on several sequences from the public KITTI dataset and demonstrate quantitatively that it improves camera and object tracking compared to state-of-the-art approaches.
|
|
13:50-14:00, Paper TuB-6.9 | |
Fast Sparse LiDAR Odometry Using Self-Supervised Feature Selection on Intensity Images |
|
Guadagnino, Tiziano | Sapienza University of Rome |
Chen, Xieyuanli | N/A |
Sodano, Matteo | Photogrammetry and Robotics Lab, University of Bonn |
Behley, Jens | University of Bonn |
Grisetti, Giorgio | Sapienza University of Rome |
Stachniss, Cyrill | University of Bonn |
Keywords: SLAM, Vision-Based Navigation
Abstract: Ego-motion estimation is a fundamental building block of any autonomous system that needs to navigate in an environment. In large-scale outdoor scenes, 3D LiDARs are often used for this task, as they provide a large number of range measurements at high precision. In this paper, we propose a novel approach that exploits the intensity channel of 3D LiDAR scans to compute an accurate odometry estimate at a high frequency. In contrast to existing methods that operate on full point clouds, our approach extracts a sparse set of salient points from intensity images using data-driven feature extraction architectures originally designed for RGB images. These salient points are then used to compute the relative pose between successive scans. Furthermore, we propose a novel self-supervised procedure to fine-tune the feature extraction network online during navigation, which exploits the estimated relative motion but does not require ground truth data. The the experimental evaluation suggests that the proposed approach provides a solid ego-motion estimation at a much higher frequency than the sensor frame rate while improving its estimation accuracy online.
|
|
TuB-7 |
Rm7 (Room E) |
Medical Robots and Systems 5 |
Regular session |
Chair: Navab, Nassir | TU Munich |
Co-Chair: Haraguchi, Daisuke | Tokyo National College of Technology |
|
12:30-12:40, Paper TuB-7.1 | |
Design and Evaluation of a Robotic Forceps with Flexible Wrist Joint Made of PEEK Plastic |
|
Zhou, Dongbo | Tokyo Institute of Technology |
Haraguchi, Daisuke | Tokyo National College of Technology |
Keywords: Medical Robots and Systems, Tendon/Wire Mechanism, Force and Tactile Sensing
Abstract: This paper presents the design and evaluation of a robotic forceps in which the wrist joint is made of polyetheretherketone (PEEK) plastic. Due to the structural improvement of enclosing a flexible backbone of PTFE tube inside the wrist joint, the developed forceps simultaneously fulfills the requirements of small diameter, bending dexterity, sufficient axial stiffness, and even the control rigidity. The proposed robotic forceps employs pneumatic drive with wire actuation mechanism. For accurate motion control and external force estimation, we developed a novel inverse dynamics model considering the coupled dynamic effect between the wrist joint and the gripper motions. The position control accuracy of the wrist joint bending angle is at the level of 1° and is not affected by the simultaneous open–close motion of the gripper. The translation and grasping forces are beyond 4.5 N, enabling powerful tasks in laparoscopic surgery. Furthermore, the robotic forceps is capable of multi-DOF external force estimation. The estimation accuracy of the translation force is about 0.2 N, and the estimation accuracy of grasping force remains within 0.2 N regardless of the bending angle of the wrist joint. The performance evaluation results demonstrate that the developed forceps is eligible to be used in robotic-assisted laparoscopic surgery.
|
|
12:40-12:50, Paper TuB-7.2 | |
External and Internal Sensor Fusion Based Localization Strategy for 6-DOF Pose Estimation of a Magnetic Capsule Robot |
|
Li, Keyu | The Chinese University of Hong Kong |
Xu, Yangxin | Yuanhua Robotics, Perception & AI Technologies Ltd |
Zhao, Ziqi | Southern University of Science and Technology |
Meng, Max Q.-H. | The Chinese University of Hong Kong |
Keywords: Medical Robots and Systems
Abstract: This paper introduces a novel localization approach for active capsule endoscopy that, for the first time, combines external magnetic field sensing and internal inertial sensing to realize 6-DOF pose estimation of a magnetic capsule robot. It utilizes an inertial measurement unit embedded in the capsule with an external magnetic sensor array to estimate the 6-DOF pose of the capsule, which does not require complicated structures of the capsule and the actuator or the implementation of specific motions of the magnets, and can achieve accurate and real-time localization of the capsule in a large workspace. We formulate the localization model and analyze the singularities of the method, and present the design approach to determine the configuration of the localization system for efficient and accurate localization in a 0.5m * 0.5m * 0.2m workspace. Simulation and real-world experiments are conducted to validate the effectiveness of the proposed localization strategy. Our results show that the proposed method can achieve a localization accuracy of 5.35 mm and 1.46° in position and orientation in the real-time tracking task at an update rate of 60 Hz. The presented method can be integrated with any magnetic actuation method to achieve closed-loop control of a magnetic capsule robot in the human body.
|
|
12:50-13:00, Paper TuB-7.3 | |
Robotic Auscultation Over Clothes for Eliminating Gender Bias |
|
Tsumura, Ryosuke | National Institute of Advanced Industrial Science and Technology |
Umezawa, Akihiro | Waseda University |
Morishima, Yuko | University of Tsukuba |
Iwata, Hiroyasu | Waseda University |
Koseki, Yoshihiko | AIST |
Nitta, Naotaka | National Institute of Advanced Industrial Science and Technology |
Yoshinaka, Kiyoshi | National Institute of Advanced Industrial Science and Technology |
Keywords: Medical Robots and Systems
Abstract: During auscultation, patients in difficult age often feel embarrassed and uncomfortable when exposing their chests to doctors of different gender and being touched physically by doctors. We assume that an auscultation with robot technology can address the aforementioned gender-related issue. Toward eliminating gender bias during auscultation exam, this paper proposes a robotic platform which enables to perform the automated auscultation over clothes. Our developed system is comprised of two folds: a depth image-based estimation system of the listening positions over clothes with RGB-D camera and a contact force adjustment system for minimizing the acoustic attenuation due to the clothes with a passive-actuated end- effector. Our preliminary results demonstrated the robotic platform enables to estimate the listening locations to hear the sounds of four cardiac valves over the clothes by combining the estimated skeletal structure with statistical anatomical data and acquire the maximized acoustic quality over the clothes by adjusting the contact force. The developed robotic platform has the potential to address the gender-related issues in auscultation.
|
|
13:00-13:10, Paper TuB-7.4 | |
ANN-Based Optimization of Human Gait Data Obtained from a Robot-Mounted 3D Camera: A Multiple Sclerosis Case Study |
|
Guffanti, Diego | Universidad Politécnica De Madrid |
Brunete, Alberto | Universidad Politécnica De Madrid |
Hernando, Miguel | Universidad Politécnica De Madrid |
Gambao, Ernesto | Universidad Politecnica De Madrid |
Álvarez, David | Universidad Politécnica De Madrid |
Keywords: Medical Robots and Systems, Rehabilitation Robotics, Bioinspired Robot Learning
Abstract: Assessment of gait consistency requires testing over a long walking distance. Robot-mounted 3D cameras represent a cost-effective, markerless technology of human gait analysis that can be applied for this purpose. However, the use of robotic platforms for gait analysis is limited by the low accuracy of 3D cameras. The aim of this study is to improve the accuracy of kinematic and spatio-temporal estimations obtained from a robot-mounted 3D camera by applying a supervised learning process, and then to verify the effectiveness of the proposed method for clinical use. Artificial neural networks have been trained using the reference data provided by a Vicon system, which lead to improved estimations. Then, gait characteristics in Multiple Sclerosis patients has been measured. Significant differences respect to Healthy Controls have been found mainly for hip flexion, pelvis tilt, pelvis rotation and stride length. The improvement of robot-mounted 3D camera estimations and their application to the analysis of gait impairment in a natural environment show the flexibility and adaptability that this setup can provide.
|
|
13:10-13:20, Paper TuB-7.5 | |
Magnetic Microrobot Control Using an Adaptive Fuzzy Sliding-Mode Method |
|
Mousavi, Alireza | University of Essex |
Khaksar, Hesam | The School of Computer Science and Electronic Engineering, Unive |
Ahmed, Awais | DGIST |
Choi, Hongsoo | Daegu Gyeongbuk Institute of Science and Technology (DGIST) |
Kafash hoshiar, Ali | University of Essex |
Keywords: Medical Robots and Systems, Micro/Nano Robots, Motion Control
Abstract: The magnetic medical microrobots are influenced by diverse factors such as the medium, the geometry of the microrobot, and the imaging procedure. It is worth noting that the size limitations make it difficult or even impossible to obtain reliable physical properties of the system. In this research, to achieve a precise microrobot control using minimum knowledge about the system, an Adaptive Fuzzy Sliding-Mode Control (AFSMC) scheme is designed for the motion control problem of the magnetically actuated microrobots in presence of input saturation constraint. The AFSMC input consists of a fuzzy system designed to approximate an unknown nonlinear dynamical system and a robust term considered for mismatch compensation. According to the designed adaptation laws, the asymptotic stability is proved based on the Lyapunov theorem and Barbalat’s lemma. In order to evaluate the effectiveness of the proposed method, a comparative simulation study is conducted.
|
|
13:20-13:30, Paper TuB-7.6 | |
An Easy-To-Deploy Combined Nasal/Throat Swab Robot with Sampling Dexterity and Resistance to External Interference |
|
Chen, Wei | The Chinese University of Hong Kong |
Chen, Zhi | Hefei University |
Lu, Yiang | The Chinese University of Hong Kong |
Cao, Hanwen | Chinese University of Hong Kong |
Zhou, Jianshu | The Chinese University of Hong Kong |
Tong, Michael CF | The Chinese University of Hong Kong |
Liu, Yunhui | Chinese University of Hong Kong |
Keywords: Medical Robots and Systems, Soft Robot Applications
Abstract: Robots have been used extensively in the battle against the COVID-19 pandemic since its outbreak. One prominent direction is the use of robots for swab sampling, which not only solves the shortage of medical staffs, but also prevents them from being infected during the face-to-face sampling. However, massive deployment of sampling robots is still not achieved due to their high cost, safety concern, deployment complexity, et al. In this paper, we propose a flexible, safe and easy-to-deploy swab robot in a compact bench-top system. The sampling flexibility is enabled by a soft-rigid hybrid continuum mechanism, where the bio-mimetic rigid interior and soft exterior design guarantee the robot with both flexibility and safety. Besides, the integration of 3-D fiber Bragg grating (FBG) based shape sensor and multi-axis force sensor provides enhanced closed-loop control performance. A dedicated constrained compliance control (CCC) algorithm was developed to tackle those unexpected interactions during sampling, which ensures the validity and safety of the robot sampling under disturbance. The robot can perform nasal/throat swab sampling tasks as dexterous as human manual operation. Various experiments are carried out to validate our system and prove its feasibility, flexibility, high safety, and efficiency for both nasal/throat swab sampling tasks. The proposed easy-to-deploy system is promising to be massive duplicated for robotic swab sampling. end{abstract}
|
|
13:30-13:40, Paper TuB-7.7 | |
Towards Autonomous Atlas-Based Ultrasound Acquisitions in Presence of Articulated Motion |
|
Jiang, Zhongliang | Technische Universitat Munchen |
Gao, Yuan | Shanghai Jiao Tong University |
Xie, Le | Shanghai Jiao Tong University |
Navab, Nassir | TU Munich |
Keywords: Medical Robots and Systems, Sensor-based Control, Sensor Fusion
Abstract: Robotic ultrasound (US) imaging aims at overcoming some of the limitations of free-hand US examinations, e.g. difficulty in guaranteeing intra- and inter-operator repeatability. However, due to anatomical and physiological variations between patients and relative movement of anatomical substructures, it is challenging to robustly generate optimal trajectories to examine the anatomies of interest, in particular, when they comprise articulated joints. To address this challenge, this paper proposes a vision-based approach allowing autonomous robotic US limb scanning. To this end, an atlas MRI template of a human arm with annotated vascular structures is used to generate trajectories and register and project them onto patients' skin surfaces for robotic ultrasound acquisition. To effectively segment and accurately reconstruct the targeted 3D vessel, we make use of spatial continuity in consecutive US frames by incorporating channel attention modules into a U-Net-type neural network. The automatic trajectory generation method is evaluated on five volunteers with various articulated joint angles. In all cases, the system could successfully acquire the planned vascular structure on volunteers' limbs. For one volunteer the MRI scan was also available allowing the evaluation of the average radius of the scanned artery estimated from the proposed robotic ultrasound acquisition, resulting in a radius estimation (1.2pm0.05 mm) compared to the MRI ground truth (1.2pm0.04 mm).
|
|
13:40-13:50, Paper TuB-7.8 | |
Optimization of Surgical Robotic Instrument Mounting in a Macro-Micro Manipulator Setup for Improving Task Execution (I) |
|
Cursi, Francesco | Imperial College London |
Bai, Weibang | Imperial College London |
Yeatman, Eric | Imperial College London |
Kormushev, Petar | Imperial College London |
Keywords: Optimization and Optimal Control, Surgical Robotics: Laparoscopy, Mechanism Design
Abstract: In minimally invasive robotic surgery, the surgical instrument is usually inserted inside the patient’s body through a small incision, which acts as a remote center of motion (RCM). Serial-link manipulators can be used as macro robots on which microsurgical robotic instruments are mounted to increase the number of degrees of freedom of the system and ensure safe task and RCM motion execution. However, the surgical instrument needs to be placed in an appropriate configuration when completing the motion tasks. The contribution of this article is to present a novel framework that preoperatively identifies the best base configuration, in terms of Roll, Pitch, and Yaw angles, of the microsurgical instrument with respect to the macro serial-link manipulator’s end effector in order to achieve the maximum accuracy and dexterity in performing specified tasks. The framework relies on hierarchical quadratic programming for the control, genetic algorithm for the optimization, and on a resilience to error strategy to make sure deviations from the optimum do not affect the system’s performance. Simulation results show that the mounting configuration of the surgical instrument significantly impacts the performance of the whole macro–micro manipulator in executing the desired motion tasks, and both the simulation and experimental results demonstrate that the proposed optimization method improves the overall performance.
|
|
TuB-8 |
Rm8 (Room F) |
Robotics and Automation in Agriculture and Construction 1 |
Regular session |
Chair: Tazaki, Yuichi | Kobe University |
Co-Chair: Zhang, Liangjun | Baidu |
|
12:30-12:40, Paper TuB-8.1 | |
Excavation of Fragmented Rocks with Multi-Modal Model-Based Reinforcement Learning |
|
Zhu, Yifan | University of Illinois at Urbana-Champaign |
Wang, Liyang | Baidu USA |
Zhang, Liangjun | Baidu |
Keywords: Robotics and Automation in Construction, Manipulation Planning, Force and Tactile Sensing
Abstract: This paper presents a multi-modal model-based reinforcement learning (MBRL) approach to the excavation of fragmented rocks, which are very challenging to model due to their highly variable sizes and geometries, and visual occlusions. A multi-modal recurrent neural network (RNN) learns the dynamics of bucket-terrain interaction from a small physical dataset, with a discrete set of motion primitives encoded with domain knowledge as the action space. Then a model predictive controller (MPC) tracks a global reference path using multi-modal feedback. We show that our RNN-based dynamics function achieves lower prediction errors compared to a feed-forward neural network baseline, and the MPC is able to significantly outperform manually designed strategies on such a challenging task.
|
|
12:40-12:50, Paper TuB-8.2 | |
Model Learning and Predictive Control for Autonomous Obstacle Reduction Via Bulldozing |
|
Wagner, W. Jacob | University of Illinois at Urbana-Champaign and Construction Engi |
Driggs-Campbell, Katherine | University of Illinois at Urbana-Champaign |
Soylemezoglu, Ahmet | US Army Corps of Engineers |
Keywords: Robotics and Automation in Construction, Model Learning for Control, Mining Robotics
Abstract: We investigate how employing model learning methods in concert with model predictive control (MPC) can be used to automate obstacle reduction to mitigate risks to Combat Engineers operating construction equipment in an active battlefield. We focus on the task of earthen berm removal using a bladed vehicle. We introduce a novel data-driven formulation for earthmoving dynamics that enables prediction of the vehicle and detailed terrain state over a one second horizon. In a simulation environment, we first record demonstrations from a human operator and then train two different earthmoving models to produce predictions of the high-dimensional state using under six minutes of data. Optimization over the learned model is performed to select an action sequence, constrained to a 2D space of template action trajectories. Simple recovery controllers are implemented to improve controller performance when the model predictions degrade. This system yields near human-level performance on a berm removal task, indicating that model learning and predictive control is a promising data-efficient approach to autonomous earthmoving.
|
|
12:50-13:00, Paper TuB-8.3 | |
Design and Motion Planning for a Reconfigurable Robotic Base |
|
Pankert, Johannes | ETH Zuerich |
Valsecchi, Giorgio | Robotic System Lab, ETH |
Baret, Davide | ETH Zürich |
Zehnder, Jon Noa | ETH Zürich |
Pietrasik, Lukasz Leszek | ETH Zurich |
Bjelonic, Marko | ETH Zurich |
Hutter, Marco | ETH Zurich |
Keywords: Robotics and Automation in Construction, Motion Control, Mechanism Design
Abstract: A robotic platform for mobile manipulation needs to satisfy two contradicting requirements for many real-world applications: A compact base is required to navigate through cluttered indoor environments, while the support needs to be large enough to prevent tumbling or tip over, especially during fast manipulation operations with heavy payloads or forceful interaction with the environment. This paper proposes a novel robot design that fulfills both requirements through a versatile footprint. It can reconfigure its footprint to a narrow configuration when navigating through tight spaces and to a wide stance when manipulating heavy objects. Furthermore, its triangular configuration allows for high-precision tasks on uneven ground by preventing support switches. A model predictive control strategy is presented that unifies planning and control for simultaneous navigation, reconfiguration, and manipulation. It converts task-space goals into whole-body motion plans for the new robot. The proposed design has been tested extensively with a hardware prototype. The footprint reconfiguration allows to almost completely remove manipulation induced vibrations. The control strategy proves effective in both lab experiment and during a real-world construction task.
|
|
13:00-13:10, Paper TuB-8.4 | |
External Load Estimation of Hydraulically Driven Construction Machinery from Cylinder Pressures and Link Accelerations |
|
Shimamura, Naotake | Kobe University |
Katayama, Raita | Kobe University |
Nagano, Hikaru | Kobe University |
Tazaki, Yuichi | Kobe University |
Yokokohji, Yasuyoshi | Kobe University |
Keywords: Robotics and Automation in Construction, Force and Tactile Sensing, Telerobotics and Teleoperation
Abstract: Remotely controlled hydraulically driven robots are expected to play an important role in extreme environments such as disaster sites, and force feedback is effective for improving the fidelity of the remote environment and the work efficiency. However, it is not reasonable to attach a force sensor directly to the end-point of a hydraulically driven robot. In a previous study, the authors showed that the impact forces, which are important information to improve the fidelity of the remote environment and work efficiency, can be estimated by using the information of acceleration of each link in addition to the cylinder pressures. In this paper, we investigated how many accelerometers and where those accelerometers should be attached on each link by using an index called GDOP (Geometric Dilution Of Precision) to improve the accuracy of impact force estimation. The experimental results show that although the estimation accuracy could not be improved significantly by rearranging the accelerometers, the effect of reducing the noise of the estimated load due to the sensor noise was confirmed.
|
|
13:10-13:20, Paper TuB-8.5 | |
Connected Reconfiguration of Polyominoes Amid Obstacles Using RRT* |
|
Garcia Gonzalez, Javier | University of Houston |
Yannuzzi, Michael | University of Houston |
Kramer, Peter | TU Braunschweig |
Rieck, Christian | Technische Universität Braunschweig |
Becker, Aaron | University of Houston |
Keywords: Robotics and Automation in Construction, Computational Geometry, Task and Motion Planning
Abstract: This paper investigates using a sampling-based approach, the RRT*, to reconfigure a 2D set of connected tiles in complex environments, where multiple obstacles might be present. Since the target application is automated building of discrete, cellular structures using mobile robots, there are constraints that determine what tiles can be picked up and where they can be dropped off during reconfiguration. We compare our approach to two algorithms as global and local planners, and show that we are able to find more efficient build sequences using a reasonable amount of samples, in environments with varying degrees of obstacle space. See overview video at https://youtu.be/Fp0MUag8po4.
|
|
13:20-13:30, Paper TuB-8.6 | |
Autonomous Mobile 3D Printing of Large-Scale Trajectories |
|
Sustarevas, Julius | University College London |
Kanoulas, Dimitrios | University College London |
Julier, Simon | University College London |
Keywords: Robotics and Automation in Construction, Motion and Path Planning, Additive Manufacturing
Abstract: Mobile 3D Printing (M3DP), using printing-in-motion, is a powerful paradigm for automated construction. A mobile robot, equipped with its own power, materials and an arm-mounted extruder, simultaneously navigates and creates its environment. Such systems can be highly scalable, parallelizable and flexible. However, planning and controlling the motion of the arm and base at the same time is challenging and most deployments either avoid robot-base motion entirely or use human prescribed robot-base paths. In a previous paper, we developed a high-level planning algorithm to automate M3DP given a print task. The generated robot-base paths avoid collisions and maintain task reachability. In this paper, we extend this work to robot control. We develop and compare three different ways to integrate the long-duration planned path with a short horizon Model Predictive Controller. Experiments are carried out via a new M3DP system — Armstone. We evaluate and demonstrate our algorithm in a 250 m long multi- layer print which is about 5 times longer than any previous physical printing-in-motion system.
|
|
13:30-13:40, Paper TuB-8.7 | |
Soil-Adaptive Excavation Using Reinforcement Learning |
|
Egli, Pascal Arturo | RSL, ETHZ |
Hutter, Marco | ETH Zurich |
Kerscher, Simon | Eth Zurich |
Jud, Dominic | ETH Zurich |
Keywords: Robotics and Automation in Construction, Reinforcement Learning, Hydraulic/Pneumatic Actuators
Abstract: In this letter, we present an excavation controller for a full-sized hydraulic excavator that can adapt online to different soil characteristics. Soil properties are hard to predict and can vary even within one scoop, which requires a controller that can adapt online to the encountered soil conditions. The objective is to fill the bucket with excavation material while respecting machine limitations to prevent stalling or lifting of the machine. To this end, we train a control policy in simulation using RL. The soil interactions are modeled based on the FEE with heavily randomized soil parameters to expose the agent to a wide range of different conditions. The agent learns to output joint velocity commands, which can be directly applied to the standard proportional valves of the real machine. We test the controller on a 12-ton excavator in different types of soils. The experiments demonstrate that the controller can adapt online to changing conditions without the explicit knowledge of the soil parameters, solely from proprioceptive observations, which are easily measurable.
|
|
13:40-13:50, Paper TuB-8.8 | |
Loading an Autonomous Large-Scale Dump Truck: Path Planning Based on Motion Data from Human-Operated Construction Vehicles |
|
Akegawa, Tetsu | Tohoku University |
Ohno, Kazunori | Tohoku University |
Kojima, Shotaro | Tohoku University |
Miyamoto, Naoto | Tohoku Univ |
Suzuki, Taro | Chiba Institute of Technology |
Komatsu, Tomohiro | KOWATECH Co |
Suzuki, Takahiro | Tohoku University |
Shibata Yukinori, Shibata | Sato Komuten Co |
Asano, Kimitaka | Sanyo-Technics Co |
Tadokoro, Satoshi | Tohoku University |
Keywords: Robotics and Automation in Construction, Motion and Path Planning, Field Robots
Abstract: A large-scale dump truck that automatically transports earth and sand in cooperation with a human-operated backhoe is of interest to the construction industry. A human-operated dump truck generally drives slightly past the desired loading position and then backs up to it for loading the sediment. The turning and loading positions are subjectively decided according to the working posture of the backhoe and the surrounding environment, and the safety margin of cooperative works. Backhoe operators want to perform the same maneuvers for human-operated/automated dump trucks. The movements of the autonomous vehicle should be similar to those of a human-operated one. However, it is difficult to derive a human-like path that does more than minimize costs. This study proposes a path-planning method that generates a path including a turning back, according to the changing backhoe position and surrounding conditions. We modeled the positional relationship during loading between a backhoe and dump truck, determining the loading and turning positions and related parameters from operational data collected in trials with human-operated construction vehicles. The proposed method allowed the autonomous dump truck path to resemble a human-like one. The authors have retrofitted an existing large-scale six-wheeled dump truck for automatic operation. Automatic loading in cooperation with a human-operated backhoe was realized all 17 times using the retrofitted dump. The average stopping accuracy was 0.57 m and 9.7 deg.
|
|
13:50-14:00, Paper TuB-8.9 | |
Towards Autonomous Visual Navigation in Arable Fields |
|
Ahmadi, Alireza | University of Bonn |
Halstead, Michael Allan | Bonn University |
McCool, Christopher Steven | University of Bonn |
Keywords: Robotics and Automation in Agriculture and Forestry, Vision-Based Navigation, Agricultural Automation
Abstract: Autonomous navigation of a robot in agricultural fields is essential for every task from crop monitoring to weed management and fertilizer application. Many current approaches rely on accurate GPS, however, such technology is expensive and can be impacted by lack of coverage. As such, autonomous navigation through sensors that can interpret their environment (such as cameras) is important to achieve the goal of autonomy in agriculture. In this paper, we introduce a purely vision-based navigation scheme that is able to reliably guide the robot through row-crop fields using computer vision and signal processing techniques without manual intervention. Independent of any global localization or mapping, this ap- proach is able to accurately follow the crop-rows and switch between the rows, only using onboard cameras. The proposed navigation scheme can be deployed in a wide range of fields with different canopy shapes in various growth stages, creating a crop agnostic navigation approach. This was completed under various illumination conditions using simulated and real fields where we achieve an average navigation accuracy of 3.82cm with minimal human intervention (hyper-parameter tuning) on BonnBot-I. Keywords — Robotics and Automation in Agriculture and Forestry; Agricultural Automation; Vision-Based Navigation.
|
|
TuB-9 |
Rm9 (Room G) |
Recognition |
Regular session |
Chair: Marshall, Joshua A. | Queen's University |
Co-Chair: Wu, Zhenyu | Nanyang Technological University |
|
12:30-12:40, Paper TuB-9.1 | |
STEADY: Simultaneous State Estimation and Dynamics Learning from Indirect Observations |
|
Wei, Jiayi | University of Texas at Austin |
Holtz, Jarrett | University of Texas at Austin |
Dillig, Isil | UT Austin |
Biswas, Joydeep | University of Texas at Austin |
Keywords: Dynamics, Probabilistic Inference, Machine Learning for Robot Control
Abstract: Accurate kinodynamic models play a crucial role in many robotics applications such as off-road navigation and high-speed driving. Many state-of-the-art approaches for learning stochastic kinodynamic models, however, require precise measurements of robot states as labeled input/output examples, which can be hard to obtain in outdoor settings due to limited sensor capabilities and the absence of ground truth. In this work, we propose a new technique for learning neural stochastic kinodynamic models from noisy and indirect observations by performing simultaneous state estimation and dynamics learning. The proposed technique iteratively improves the kinodynamic model in an expectation-maximization loop, where the E Step samples posterior state trajectories using particle filtering, and the M Step updates the dynamics to be more consistent with the sampled trajectories via stochastic gradient ascent. We evaluate our approach on both simulation and real-world benchmarks and compare it with several baseline techniques. Our approach not only achieves significantly higher accuracy but is also more robust to observation noise, thereby showing promise for boosting the performance of many other robotics applications.
|
|
12:40-12:50, Paper TuB-9.2 | |
Smart Explorer: Recognizing Objects in Dense Clutter Via Interactive Exploration |
|
Wu, Zhenyu | Beijing University of Posts and Telecommunications |
Wang, Ziwei | Tsinghua University |
Wei, Zibu | Tsinghua University |
Wei, Yi | Tsinghua University |
Yan, Haibin | Beijing University of Posts and Telecommunications |
Keywords: Recognition, RGB-D Perception
Abstract: Recognizing objects in dense clutter accurately makes significant contribution to a wide variety of robotic manipulation tasks including grasping, packing, rearranging and many others. However, the visual recognition model usually misses objects because of the significant occlusion among instances and causes incorrect prediction due to the visual ambiguity with the high object crowdedness. In this paper, we propose an interactive exploration framework called Smart Explorer for recognizing all objects in dense clutters. Our Smart Explorer physically interacts with the clutter to maximize the recognition performance while minimize the number of motions, where the false positives and negatives can be alleviated effectively with the optimal accuracy-efficiency trade-offs. Specifically, we first collect the multi-view RGB-D images of the clutter and reconstruct the corresponding point cloud. By aggregating the instance segmentation of RGB images across views, we acquire the instance-wise point cloud partition of the clutter through which the existed classes and the number of objects for each class are predicted. The pushing actions for effective physical interaction are generated to sizably reduce the recognition uncertainty that consists of the instance segmentation entropy and multi-view object disagreement. Therefore, the optimal accuracy-efficiency trade-off of object recognition in dense clutter is achieved via iterative instance prediction and physical interaction. Extensive experiments demonstrate that our Smart Explorer boosts acquires promising recognition accuracy with only a few actions, which also outperforms the random pushing by a large margin.
|
|
12:50-13:00, Paper TuB-9.3 | |
LSDNet: A Lightweight Self-Attentional Distillation Network for Visual Place Recognition |
|
Peng, Guohao | Nanyang Technological University |
Huang, Yifeng | Nanyang Technological University |
Li, Heshan | Nanyang Technological University |
Wu, Zhenyu | Nanyang Technological University |
Wang, Danwei | Nanyang Technological University |
Keywords: Recognition, Representation Learning, Deep Learning for Visual Perception
Abstract: Abstract— Visual Place Recognition (VPR) has become an indispensable capacity for mobile robots to operate in large-scale environments. Existing methods in this field mostly focus on exploring high-performance encoding strategies, while few attempts are devoted to lightweight models that balance performance and computational cost. In this work, we propose a Lightweight Self-attentional Distillation Network (LSDNet) aiming to obtain advantages of both performance and efficiency. (1) From a performance perspective, an attentional encoding strategy is proposed to integrate crucial information in the scene. It extends the NetVLAD architecture with a self-attention module to facilitate non-local information interaction between local features. Through further visual word vector rescaling, the final image representation can benefit from both non-local spatial integration and cluster-wise weighting. (2) From an efficiency perspective, LSDNet is built upon a lightweight backbone. To maintain comparable performance to large backbone models, a dual distillation strategy is proposed. It prompts LSDNet to learn both encoding patterns in the hidden space and feature distributions in the encoding space from the teacher model. Through distillation-augmented training, LSDNet is able to rival the teacher model and outperform SOTA global representations with the same lightweight backbone.
|
|
13:00-13:10, Paper TuB-9.4 | |
STUN: Self-Teaching Uncertainty Estimation for Place Recognition |
|
Cai, Kaiwen | University of Liverpool |
Lu, Chris Xiaoxuan | University of Edinburgh |
Huang, Xiaowei | University of Liverpool |
Keywords: Recognition, Localization, Probabilistic Inference
Abstract: Place recognition is key to Simultaneous Localization and Mapping (SLAM) and spatial perception. However, a place recognition in the wild often suffers from erroneous predictions due to image variations, e.g., viewpoints and street appearance. Integrating uncertainty estimation into the life cycle of place recognition is a promising method to mitigate the impact of variations on place recognition performance. However, existing uncertainty estimation approaches in this vein are either computationally inefficient (e.g., Monte Carlo dropout) or at the cost of dropped accuracy. This paper proposes STUN, a self-teaching framework that learns to simultaneously predict the place and estimate the prediction uncertainty given an input image. To this end, we first train a teacher net using a standard metric learning pipeline to produce embedding priors. Then, supervised by the pretrained teacher net, a student net with an additional variance branch is trained to finetune the embedding priors and estimate the uncertainty sample by sample. When it comes to the online inference phase, we only use the student net to generate a place prediction in conjunction with the uncertainty. When compared with place recognition systems that are ignorant to the uncertainty, our framework features the uncertainty estimation for free without sacrificing any prediction accuracy or incurring extra computation loads. Our experimental results on the large-scale Pittsburgh30k dataset demonstrate that STUN outperforms the state-of-the-art methods in both recognition accuracy and the quality of uncertainty estimation.
|
|
13:10-13:20, Paper TuB-9.5 | |
Self-Supervised Reinforcement Learning for Active Object Detection |
|
Fang, Fen | I2R |
Liang, Wenyu | Institute for Infocomm Research, A*STAR |
Wu, Yan | A*STAR Institute for Infocomm Research |
Xu, Qianli | Institute for Infocomm Research |
Lim, Joo Hwee | I2R A*STAR |
Keywords: Recognition, Deep Learning for Visual Perception, Integrated Planning and Learning
Abstract: Active object detection (AOD) offers significant advantage in expanding the perceptual capacity of a robotics system. AOD is formulated as a sequential action decision process to determine optimal viewpoints to identify objects of interest in a visual scene. While reinforcement learning (RL) has been successfully used to solve many AOD problems, conventional RL methods suffer from (i) sample inefficiency, and (ii) unstable out- come due to inter-dependencies of action type (direction of view change) and action range (step size of view change). To address these issues, we propose a novel self-supervised RL method, which employs self-supervised representations of viewpoints to initialize the policy network, and a self-supervised loss on action range to enhance the network parameter optimization. The output and target pairs of self-supervised learning loss are automatically generated from the policy network online prediction and a range shrinkage algorithm (RSA), respectively. The proposed method is evaluated and benchmarked on two public datasets (T-LESS and AVD) using on-policy and off-policy RL algorithms. The results show that our method enhances detection accuracy and achieves faster convergence on both datasets. By evaluating on a more complex environment with a larger state space (where viewpoints are more densely sampled), our method achieves more robust and stable performance. Our experiment on real robot application scenario to disambiguate similar objects in a cluttered scene has also demonstrated the effectiveness of the proposed method.
|
|
13:20-13:30, Paper TuB-9.6 | |
ReINView: Re-Interpreting Views for Multi-View 3D Object Recognition |
|
Xu, Ruchang | Beijing University of Technology |
Ma, Wei | Beijing University of Technology |
Mi, Qing | Beijing University of Technology |
Zha, Hongbin | Peking University |
Keywords: Recognition, Computer Vision for Manufacturing, Computer Vision for Automation
Abstract: Multi-view-based 3D object recognition is important in robot-environment interaction. However, recent methods simply extract features from each view via convolutional neural networks (CNNs) and then fuse these features together to make predictions. These methods ignore the inherent ambiguities of each view caused due to 3D-2D projection. To address this problem, we propose a novel deep framework for multi-view-based 3D object recognition. Instead of fusing the multi-view features directly, we design a re-interpretation module (ReINView) to eliminate the ambiguities at each view. To achieve this, ReINView re-interprets view features patch by patch by using their context from nearby views, considering that local patches are generally co-visible at nearby viewpoints. Since contour shapes are essential for 3D object recognition as well, ReINView further performs view-level re-interpretation, in which we use all the views as context sources since the target contours to be re-interpreted are globally observable. The re-interpreted multi-view features can better reflect the 3D global and local structures of the object. Experiments on both ModelNet40 and ModelNet10 show that the proposed model outperforms state-of-the-art methods in 3D object recognition.
|
|
13:30-13:40, Paper TuB-9.7 | |
Dual-Camera High Magnification Surveillance System with Non-Delay Gaze Control and Always-In-Focus Function in Indoor Scenes |
|
Zhang, Tianyi | Chiba University |
Hu, Shaopeng | Hiroshima University |
Shimasaki, Kohei | Hiroshima University |
Ishii, Idaku | Hiroshima University |
Namiki, Akio | Chiba University |
Keywords: Environment Monitoring and Management, Surveillance Robotic Systems
Abstract: This study proposes a dual-camera system for indoor high magnification surveillance which is capable of achieving always-in-focus and non-delay gaze control based on high-speed vision. The users are enabled to move the mouse freely on the wide-view screen while observing its in-focal zoom-in monitoring video in real-time. The proposed system consists of a wide-angle camera for wide-view and a Galvano mirror-enabled ultra-fast pan-tilt-zoom (PTZ) camera for zoom-in view. To achieve always-in-focus, a high-speed focus scanning system is proposed that is comprised of a high-speed camera, a parfocal zoom lens, and a gear mechanism. Through continuously reciprocating rotational motion of the focusing ring driven by the servo motor, the high-speed camera captures sets of images with varying focal distances. Moreover, we proposed a most-in-focus (MIF) frame extraction algorithm to select the sharpest images as output. The experimental results are obtained to confirm the effectiveness of our system.
|
|
13:40-13:50, Paper TuB-9.8 | |
Point Label Aware Superpixels for Multi-Species Segmentation of Underwater Imagery |
|
Raine, Scarlett | Queensland University of Technology |
Marchant, Ross | Queensland University of Technology |
Kusy, Brano | CSIRO |
Maire, Frederic | Queensland University of Technology |
Fischer, Tobias | Queensland University of Technology |
Keywords: Environment Monitoring and Management, Semantic Scene Understanding
Abstract: Monitoring coral reefs using underwater vehicles increases the range of marine surveys and availability of historical ecological data by collecting significant quantities of images. Analysis of this imagery can be automated using a model trained to perform semantic segmentation, however it is too costly and time-consuming to densely label images for training supervised models. In this letter, we leverage photo-quadrat imagery labeled by ecologists with sparse point labels. We propose a point label aware method for propagating labels within superpixel regions to obtain augmented ground truth for training a semantic segmentation model. Our point label aware superpixel method utilizes the sparse point labels, and clusters pixels using learned features to accurately generate single-species segments in cluttered, complex coral images. Our method outperforms prior methods on the UCSD Mosaics dataset by 3.62% for pixel accuracy and 8.35% for mean IoU for the label propagation task, while reducing computation time reported by previous approaches by 76%. We train a DeepLabv3+ architecture and outperform state-of-the-art for semantic segmentation by 2.91% for pixel accuracy and 9.65% for mean IoU on the UCSD Mosaics dataset and by 4.19% for pixel accuracy and 14.32% for mean IoU on the Eilat dataset.
|
|
13:50-14:00, Paper TuB-9.9 | |
Mapping of Spatiotemporal Scalar Fields by Mobile Robots Using Gaussian Process Regression |
|
Sears, Thomas M. C. | Queen's University |
Marshall, Joshua A. | Queen's University |
Keywords: Environment Monitoring and Management, Marine Robotics, Field Robots
Abstract: Spatiotemporal maps are data-driven estimates of time changing phenomena. For environmental science, rather than collect data from an array of static sensors, a mobile sensor platform could reduce setup time and cost, maintain flexibility to be deployed to any area of interest, and provide active feedback during observations. While promising, mapping is challenging with mobile sensors because vehicle constraints limit not only where, but also when observations can be made. By assuming spatial and temporal correlations in the data through kernel functions, this paper uses Gaussian process regression (GPR) to generate a maximum likelihood estimate of the phenomenon while also tracking the estimate uncertainty. Spatiotemporal mapping by GPR is simulated for a single fixed-path mobile robot observing a latent spatiotemporal scalar field. The learned spatiotemporal map captures the structure of the latent scalar field with the largest uncertainties in areas the robot never visited.
|
|
TuB-10 |
Rm10 (Room H) |
Safety in HRI |
Regular session |
Chair: Chalvatzaki, Georgia | Technische Universität Darmastadt, Intelligent Robotic Systems for Assistance Group |
Co-Chair: Alagi, Hosam | Karlsruhe Institute of Technology |
|
12:30-12:40, Paper TuB-10.1 | |
PSM: A Predictive Safety Model for Body Motion Based on the Spring-Damper Pendulum |
|
Tafrishi, Seyed Amir | Tohoku University |
Ravankar, Ankit A. | Faculty of Mechanical and Aerospace Engineering, Tohoku Universi |
Hirata, Yasuhisa | Tohoku University |
Keywords: Safety in HRI, Human-Centered Robotics, Probability and Statistical Methods
Abstract: Quantifying the safety of the human body orientation is an important issue in human-robot interaction. Knowing the changing physical constraints on human motion can improve inspection of safe human motions and bring essential information about stability and normality of human body orientations with real-time risk assessment. Also, this information can be used in cooperative robots and monitoring systems to evaluate and interact in the environment more freely. Furthermore, the workspace area can be more deterministic with the known physical characteristics of safety. Based on this motivation, we propose a novel predictive safety model (PSM) that relies on the information of an inertial measurement unit on the human chest. The PSM encompasses a 3-Dofs spring-damper pendulum model that predicts human motion based on a safe motion dataset. The estimated safe orientation of humans is obtained by integrating a safety dataset and an elastic spring-damper model in a way that the proposed approach can realize complex motions at different safety levels. We did experiments in a real-world scenario to verify our novel proposed model. This novel approach can be used in different guidance/assistive robots and health monitoring systems to support and evaluate the human condition, particularly elders.
|
|
12:40-12:50, Paper TuB-10.2 | |
Physical Adversarial Attack on a Robotic Arm |
|
Jia, Yifan | Singapore University of Technology and Design |
Poskitt, Christopher M. | Singapore Management University |
Sun, Jun | Singapore Management University |
Chattopadhyay, Sudipta | Singapore University of Technology and Design |
Keywords: Safety in HRI, Localization, AI-Based Methods
Abstract: Collaborative Robots (cobots) are regarded as highly safety-critical cyber-physical systems (CPSs) owing to their close physical interactions with humans. In settings such as smart factories, they are frequently augmented with AI. For example, in order to move materials, cobots utilize object detectors based on deep learning models. Deep learning, however, has been demonstrated as vulnerable to adversarial attacks: a minor change (noise) to benign input can fool the underlying neural networks and lead to a different result. While existing works have explored such attacks in the context of picture/object classification, less attention has been given to attacking neural networks used for identifying object locations and demonstrating that this can actually lead to a physical attack in a real CPS. In this paper, we propose a method to generate adversarial patches for the object detectors of CPSs, to miscalibrate them and cause potentially dangerous physical effects. In particular, we evaluate our method on an industrial robotic arm for card gripping, demonstrating that it can be misled into clipping the operator’s hand instead of the card. To our knowledge, this is the first work to attack object locations and lead to an incident on human users by an actual system.
|
|
12:50-13:00, Paper TuB-10.3 | |
Regularized Deep Signed Distance Fields for Reactive Motion Generation |
|
Liu, Puze | Technische Universität Darmstadt |
Zhang, Kuo | TU-Darmstadt |
Tateo, Davide | Technische Universität Darmstadt |
Jauhri, Snehal | TU Darmstadt |
Peters, Jan | Technische Universität Darmstadt |
Chalvatzaki, Georgia | Technische Universität Darmastadt, Intelligent Robotic Systems |
Keywords: Safety in HRI, Machine Learning for Robot Control, Deep Learning Methods
Abstract: Autonomous robots should operate in real-world dynamic environments and collaborate with humans in tight spaces. A key component for allowing robots to leave structured lab and manufacturing settings is their ability to evaluate online and real-time collisions with the world around them. Distance-based constraints are fundamental for enabling robots to plan their actions and act safely, protecting both humans and their hardware. However, different applications require different distance resolutions, leading to various heuristic approaches for measuring distance fields w.r.t. obstacles, which are computationally expensive and hinder their application in dynamic obstacle avoidance use-cases. We propose Regularized Deep Signed Distance Fields (ReDSDF), a single neural implicit function that can compute smooth distance fields at any scale, with fine-grained resolution over high-dimensional manifolds and articulated bodies like humans, thanks to our effective data generation and a simple inductive bias during training. We demonstrate the effectiveness of our approach in representative simulated tasks for whole-body control (WBC) and safe Human-Robot Interaction (HRI) in shared workspaces. Finally, we provide proof of concept of a real-world application in a HRI handover task with a mobile manipulator robot.
|
|
13:00-13:10, Paper TuB-10.4 | |
Safe and Ergonomic Human-Drone Interaction in Warehouses |
|
Proia, Silvia | Politecnico Di Bari |
Cavone, Graziana | University of Roma Tre |
Camposeo, Antonio | Politecnico Di Bari |
Ceglie, Fabio | Politecnico Di Bari |
Carli, Raffaele | Politecnico Di Bari |
Dotoli, Mariagrazia | Politecnico Di Bari |
Keywords: Safety in HRI, Motion and Path Planning, Aerial Systems: Applications
Abstract: This paper presents an application of human-drone interaction (HDI) for inventory management in a warehouse 4.0 that aims at improving the operators’ safety and well-being together with increasing efficiency and reducing production costs. In our work, the speed and separation monitoring (SSM) methodology is applied for the first time to HDI, in analogy to the human-robot interaction (HRI) ISO safety requirements as well as the rapid upper limb assessment (RULA), for evaluating the operator’s ergonomic posture during the interaction with the drone. With the aim of validating the proposed approach in a realistic scenario, a quadrotor is controlled to perform a pick and place task along a desired trajectory, from the picking bay to the palletizing area where the operator is located, avoiding collisions with the warehouse shelves by implementing the artificial potential field technique (APF) for planning and the linear quadratic regulator (LQR) and iterative LQR (iLQR) algorithms for tracking. The obtained results of the HDI architecture simulations are presented and discussed in detail proving the effectiveness of the proposed method for a safe and ergonomic HDI.
|
|
13:10-13:20, Paper TuB-10.5 | |
Robot Contact Reflexes: Adaptive Maneuvers in the Contact Reflex Space |
|
Vorndamme, Jonathan | Chair of Robotics and Systems Intelligence, Technical University |
Figueredo, Luis Felipe Cruz | Technical University of Munich (TUM) |
Haddadin, Sami | Technical University of Munich |
Keywords: Safety in HRI, Performance Evaluation and Benchmarking, Reactive and Sensor-Based Planning
Abstract: In order to transform a robot into an intelligent machine it needs to be enabled to react to unforeseen events (most importantly collisions) during task execution and have a plan on how to continue the task afterwards. This requires a flexible operational framework that allows to define adaptive reactions and interactions with the motion generation and task planning stage. Within this work we first reason about the choices the robot has for reactions to unforeseen events such as collisions with respect to safety of humans in the workspace, the robot itself and the environment as well as the successful task execution. We further present a flexible reflex engine together with a concept of integration into the motion generation and control work flow. The reflex engine and it’s reflex maneuvers are a combination of state machines and decision trees that take into account the state of the robot and the world. It is capable of choosing safe reactions and can differentiate between different levels of contact severity and according reaction sets. Several reflex maneuvers are evaluated towards safety performance criteria in real robot experiments using an ISO/TS 15066 conform measurement device. Some of the tested reflexes are furthermore integrated into an implementation of the proposed approach for a simple real world example task where the robot needs to pickup a container and dispose it’s content into a bin.
|
|
13:20-13:30, Paper TuB-10.6 | |
Suppressing Delay-Induced Oscillations in Physical Human-Robot Interaction with an Upper-Limb Exoskeleton Using Rate-Limiting |
|
Sun, Jianwei | University of California Los Angeles |
Ferguson, Peter | University of California Los Angeles |
Rosen, Jacob | University of California, Los Angeles |
Keywords: Safety in HRI, Physical Human-Robot Interaction, Prosthetics and Exoskeletons
Abstract: In physical human-robot interaction (pHRI) enabled by admittance control, delay-induced oscillations arising from both the neuromuscular time-delays of the human and electromechanical delays of the robot can cause unsafe instability in the system. This study presents and evaluates rate-limiting as a means to overcome such instability, and provides a new perspective on how rate-limiting can benefit pHRI. Specifically, a rate-limited and time-delayed human-in-the-loop (HITL) model is analyzed to show not only how the rate-limiter can transform an unstable equilibrium (due to time-delay) into a stable limit-cycle, but also how a desired upper-bound on the range of persistent oscillations can be achieved by appropriately setting the rate-limiter threshold. In addition, a study involving 10 subjects and the EXO-UL8 upper-limb exoskeleton, and consisting of 16 trials - 4 rate-limiter thresholds by 4 time-delays - is performed to: (1) validate the relationships between time-delays, rate-limits, and position bounds on persistent oscillations, and (2) demonstrate the effectiveness of rate-limiting for recovery from delay-induced oscillations without interfering with regular operation. Agreement of experimental results with the theoretical developments supports the feasibility of incorporating rate-limiting in admittance-controlled pHRI systems as a safety mechanism.
|
|
13:30-13:40, Paper TuB-10.7 | |
Safety Compliant Control for Robotic Manipulator with Task and Input Constraints |
|
Murtaza, Muhammad Ali | Georgia Institute of Technology |
Aguilera, Sergio | Georgia Institute of Technology |
Waqas, Muhammad | University of Southern California |
Hutchinson, Seth | Georgia Institute of Technology |
Keywords: Safety in HRI, Robot Safety, Dexterous Manipulation
Abstract: Increasingly, robots are working in close proximity with humans, and in dynamically changing environments. We present a new control architecture that guarantees safety under such conditions, while simultaneously ensuring that task goals are satisfied. Our approach combines Control Barrier Functions (to provide safety guarantees) with Rapidly Exponentially Stabilizing Control Lyapunov functions (to satisfy task constraints and ensure exponential convergence). We formulate the problem completely in the operational space, using super ellipsoids to define safe sets which are dynamically adapted to restrict the robot's operational space in response to moving obstacles (including humans). Control barrier functions, which can be implemented in real-time using quadratic programming, ensure the forward invariance of these safe sets, while minimally perturbing prescribed control trajectories. We demonstrate our approach on a seven degree-of-freedom Kuka LBR iiwa robotic manipulator, both in a physics engine-based simulation and with real hardware experiments.
|
|
13:40-13:50, Paper TuB-10.8 | |
Safe and Efficient Exploration of Human Models During Human-Robot Interaction |
|
Pandya, Ravi | Carnegie Mellon University |
Liu, Changliu | Carnegie Mellon University |
Keywords: Safety in HRI, Robust/Adaptive Control, Modeling and Simulating Humans
Abstract: Many collaborative human-robot tasks require the robot to stay safe and work efficiently around humans. Since the robot can only stay safe with respect to its own model of the human, we want the robot to learn a good model of the human in order to act both safely and efficiently. This paper studies methods that enable a robot to safely explore the space of a human-robot system to improve the robot's model of the human, which will consequently allow the robot to access a larger state space and better work with the human. In particular, we introduce active exploration under the framework of energy-function based safe control, investigate the effect of different active exploration strategies, and finally analyze the effect of safe active exploration on both analytical and neural network human models.
|
|
13:50-14:00, Paper TuB-10.9 | |
Evaluation of On-Robot Capacitive Proximity Sensors with Collision Experiments for Human-Robot Collaboration |
|
Alagi, Hosam | Karlsruhe Institute of Technology |
Ergun, Serkan | University of Klagenfurt |
Ding, Yitao | Chemnitz University of Technology |
Huck, Tom Philip | Karlsruhe Institute of Technology |
Thomas, Ulrike | Chemnitz University of Technology |
Zangl, Hubert | Alpen-Adria-Universitaet Klagenfurt |
Hein, Björn | University of Applied Sciences Karlsruhe |
Keywords: Safety in HRI, Robot Safety, Performance Evaluation and Benchmarking
Abstract: A robot must comply with very restrictive safety standards in close human-robot collaboration applications. These standards limit the robot’s performance because of speed reductions to avoid potentially large forces exerted on humans during collisions. On-robot capacitive proximity sensors (CPS) can serve as a solution to allow higher speeds and thus better productivity. They allow early reactive measures before contacts occur to reduce the forces during collisions. An open question on designing the systems is the selection of an adequate activation distance to trigger safety measures for a specific robot while considering latency and detection robustness. Furthermore, the systems’ actual effectiveness of impact attenuation and performance gain has not been evaluated before. In this work, we define and conduct a unified test procedure based on collision experiments to determine these parameters and investigate the performance gain. Two capacitive proximity sensor systems are evaluated on this test strategy on two robots. A significant performance increase can be achieved, since a small detection distance doubles robot operation speed while maintaining the same contact force as without Capacitive Proximity Sensor (CPS). This work can serve as a reference guide for designing, configuring and implementing future on-robot CPS.
|
|
TuB-11 |
Rm11 (Room I) |
Humanoid and Bipedal Locomotion |
Regular session |
Chair: Kajita, Shuuji | Chubu University |
Co-Chair: Park, Jaeheung | Seoul National University |
|
12:30-12:40, Paper TuB-11.1 | |
Terrain-Adaptive, ALIP-Based Bipedal Locomotion Controller Via Model Predictive Control and Virtual Constraints |
|
Gibson, Grant | University of Michigan |
Dosunmu-Ogunbi, Oluwami | University of Michigan |
Gong, Yukai | University of Michigan |
Grizzle, J.W | University of Michigan |
Keywords: Humanoid and Bipedal Locomotion, Legged Robots
Abstract: This paper presents a gait controller for bipedal robots to achieve highly agile walking over various terrains given local slope and friction cone information. Without these considerations, untimely impacts can cause a robot to trip and inadequate tangential reaction forces at the stance foot can cause slippages. We address these challenges by combining, in a novel manner, a model based on an Angular Momentum Linear Inverted Pendulum (ALIP) and a Model Predictive Control (MPC) foot placement planner that is executed by the method of virtual constraints. The process starts with abstracting from the full dynamics of a Cassie 3D bipedal robot, an exact low-dimensional representation of its center of mass dynamics, parameterized by angular momentum. Under a piecewise planar terrain assumption and the elimination of terms for the angular momentum about the robot's center of mass, the centroidal dynamics about the contact point become linear and have dimension four. Importantly, we include the intra-step dynamics at uniformly-spaced intervals in the MPC formulation so that realistic workspace constraints on the robot's evolution can be imposed from step-to-step. The output of the low-dimensional MPC controller is directly implemented on a high-dimensional Cassie robot through the method of virtual constraints. In experiments, we validate the performance of our control strategy for the robot on a variety of surfaces with varied inclinations and textures.
|
|
12:40-12:50, Paper TuB-11.2 | |
Robust Contact State Estimation in Humanoid Walking Gaits |
|
Piperakis, Stylianos | Foundation for Research and Technology - Hellas (FORTH) |
Maravgakis, Michael | Institute of Computer Science, Foundation for Research and Techn |
Kanoulas, Dimitrios | University College London |
Trahanias, Panos | Foundation for Research and Technology – Hellas (FORTH) |
Keywords: Humanoid and Bipedal Locomotion, Legged Robots
Abstract: In this article, we propose a deep learning framework that provides a unified approach to the problem of leg contact detection in humanoid robot walking gaits. Our formulation accomplishes to accurately and robustly estimate the contact state probability for each leg (i.e., stable or slip/no contact). The proposed framework employs solely proprioceptive sensing and although it relies on simulated ground-truth contact data for the classification process, we demonstrate that it generalizes across varying friction surfaces and different legged robotic platforms and, at the same time, is readily transferred from simulation to practice. The framework is quantitatively and qualitatively assessed in simulation via the use of ground-truth contact data and is contrasted against state-of-the-art methods with an ATLAS, a NAO, and a TALOS humanoid robot. Furthermore, its efficacy is demonstrated in base estimation with a real TALOS humanoid. To reinforce further research endeavors, our implementation is offered as an open-source ROS/Python package, coined Legged Contact Detection (LCD).
|
|
12:50-13:00, Paper TuB-11.3 | |
Uniform Global Exponential Stabilizing Passivity-Based Tracking Controller Applied to Planar Biped Robots |
|
Arpenti, Pierluigi | Università Degli Studi Di Napoli "Federico II" |
Donaire, Alejandro | Queensland University of Technology |
Ruggiero, Fabio | Università Di Napoli Federico II |
Lippiello, Vincenzo | University of Naples FEDERICO II |
Keywords: Humanoid and Bipedal Locomotion, Underactuated Robots, Legged Robots
Abstract: This paper presents a novel control approach, based on the interconnection and damping-assignment passivity-based control (IDA-PBC), to achieve stable and periodic walking for underactuated planar biped robots with one degree of underactuation. The system's physical structure is preserved by assigning a target port-Hamiltonian dynamics to the closed-loop system, which also ensures passivity. The control design ensures that the tracking error to the desired periodic gait converges exponentially to zero, and the convergence rate can be adjusted via gain tuning. Besides, through the hybrid zero dynamics, the stability of the full-order system can be retrieved from the stability of the orbit created in a lower-dimensional manifold. The proposed approach is the first example of a tracking controller based on the IDA-PBC applied to underactuated biped robots. Numerical simulations on a five-link planar biped robot with unactuated ankles validate the approach and show the performance of the closed-loop system.
|
|
13:00-13:10, Paper TuB-11.4 | |
Learning Dynamic Bipedal Walking across Stepping Stones |
|
Duan, Helei | Oregon State University |
Malik, Ashish | Oregon State University |
Gadde, Mohitvishnu S. | Oregon State University |
Dao, Jeremy | Oregon State University |
Fern, Alan | Oregon State University |
Hurst, Jonathan | Oregon State University |
Keywords: Humanoid and Bipedal Locomotion, Legged Robots, Reinforcement Learning
Abstract: In this work, we propose a learning approach for 3D dynamic bipedal walking when footsteps are constrained to stepping stones. While recent work has shown progress on this problem, real-world demonstrations have been limited to relatively simple open-loop, perception-free scenarios. Our main contribution is a more advanced learning approach that enables real-world demonstrations, using the Cassie robot, of closed-loop dynamic walking over moderately difficult stepping-stone patterns. Our approach first uses reinforcement learning (RL) in simulation to train a controller that maps footstep commands onto joint actions without any reference motion information. We then learn a model of that controller's capabilities, which enables prediction of feasible footsteps given the robot's current dynamic state. The resulting controller and model are then integrated with a real-time overhead camera system for detecting stepping stone locations. For evaluation, we develop a benchmark set of stepping stone patterns, which are used to test performance in both simulation and the real world. Overall, we demonstrate that sim-to-real learning is extremely promising for enabling dynamic locomotion over stepping stones. We also identify challenges remaining that motivate important future research directions.
|
|
13:10-13:20, Paper TuB-11.5 | |
Humanoid Balance Control Using Centroidal Angular Momentum Based on Hierarchical Quadratic Programming |
|
Kim, Myeong-Ju | Seoul National University |
Lim, Daegyu | Seoul National University |
Park, Gyeongjae | Seoul National University |
Park, Jaeheung | Seoul National University |
Keywords: Humanoid and Bipedal Locomotion, Optimization and Optimal Control
Abstract: Maintaining balance to external pushes is one of the most important features for a humanoid to walk in a real environment. In particular, methods for counteracting to pushes using the centroidal angular momentum (CAM) control have been actively developed. In this paper, a CAM control scheme based on hierarchical quadratic programming (HQP) is proposed. The scheme of the CAM control consists of CAM tracking control and initial pose return control, which is hierarchically operated based on HQP to ensure the priority of CAM tracking performance. The proposed method is implemented in a capture point (CP) feedback control framework. Through simulations and experiments, the proposed method demonstrated more stable balance control performance than the previous method when the humanoid is walking in the presence of external perturbation.
|
|
13:20-13:30, Paper TuB-11.6 | |
Resolved Motion Control for 3D Underactuated Bipedal Walking Using Linear Inverted Pendulum Dynamics and Neural Adaptation |
|
Paredes, Victor | The Ohio State University |
Hereid, Ayonga | Ohio State University |
Keywords: Humanoid and Bipedal Locomotion, Legged Robots, Neural and Fuzzy Control
Abstract: We present a framework to generate periodic trajectory references for a 3D under-actuated bipedal robot, using a linear inverted pendulum (LIP) based controller with adaptive neural regulation. We use the LIP template model to estimate the robot's center of mass (CoM) position and velocity and formulate a discrete controller that outputs the next footstep location to achieve a desired walking velocity. This controller is equipped on the frontal plane with a Neural-Network-based adaptive term that reduces the model mismatch between the template and physical robot that particularly affects the lateral motion. Then, the foot placement location computed for the LIP model is used to generate task space trajectories (CoM and swing foot trajectories) for the actual robot to realize stable walking. We use a real-time QP-based inverse kinematics algorithm that produces joint references from the task space trajectories in real-time, which makes the formulation independent of the knowledge of the robot dynamics. Finally, we implemented and evaluated the proposed approach in simulation and hardware experiments with a Digit robot obtaining stable periodic locomotion for both cases.
|
|
13:30-13:40, Paper TuB-11.7 | |
Preemptive Foot Compliance to Lower Impact During Biped Robot Walking Over Unknown Terrain |
|
Guadarrama-Olvera, Julio Rogelio | Technical University of Munich |
Kajita, Shuuji | Chubu University |
Cheng, Gordon | Technical University of Munich |
Keywords: Humanoid and Bipedal Locomotion, Sensor-based Control, Compliance and Impedance Control
Abstract: In this work, we present a novel method for ankle/foot compliance for biped humanoid robots walking over uneven terrain. Based on distributed plantar proximity sensing, we developed the Preemptive Foot Compliance (PFC) control that generates a Preemptive Ground Reaction Wrench that modifies the foot orientation to maximize the largest contact area to dampen the impact force produced at landing. PFC can be easily included in any walking controller for flat ground and become capable of walking on uneven terrains. The PFC was tested on two full-size humanoid robots running two different walking controllers, originally designed only for flat-ground walking. Both robots increase their capability to walk over uneven terrain and the walking impacts in flat grown were reduced by approximately 80%.
|
|
13:40-13:50, Paper TuB-11.8 | |
Improved Biped Walking Performance Around the Kinematic Singularities of Biomimetic Four-Bar Knees |
|
Smyrli, Aikaterini | National Technical University of Athens |
Papadopoulos, Evangelos | National Technical University of Athens |
Keywords: Passive Walking, Humanoid and Bipedal Locomotion, Biomimetics
Abstract: This paper studies the effects of replacing pin-joint knees in passive dynamic bipedal walkers with biomimetic four-bar knees. The kinetic model of the four-bar knees is presented in detail, and an analytical model of the passive walking dynamics is derived. The resulting four-bar kneed biped is compared with a pin-joint kneed walker, for their passive walking performance. The geometry of the four-bar knees used in the study is based on human anatomical data. It is found that the biomimetic four-bar knee configuration works to the advantage of the biped, especially around the extended-knee singular position. The four-bar knees are found to overperform the pin-joint ones, resulting in significant reduction of peak impact loads and energetic expenditure.
|
|
13:50-14:00, Paper TuB-11.9 | |
Comparison of EKF-Based Floating Base Estimators for Humanoid Robots with Flat Feet |
|
Ramadoss, Prashanth | Italian Institute of Technology, Genova |
Romualdi, Giulio | Fondazione Istituto Italiano Di Tecnologia |
Dafarra, Stefano | Istituto Italiano Di Tecnologia |
Traversaro, Silvio | Istituto Italiano Di Tecnologia |
Pucci, Daniele | Italian Institute of Technology |
Keywords: Humanoid and Bipedal Locomotion, Humanoid Robot Systems, Legged Robots
Abstract: Extended Kalman filtering is a common approach to achieve floating base estimation of a humanoid robot. These filters rely on measurements from an Inertial Measurement Unit (IMU) and relative forward kinematics for estimating the base position-and-orientation and its linear velocity along with the augmented states of feet position-and-orientation. We refer to such filters as flat-foot filters. However, the availability of only partial measurements often poses the question of consistency in the filter design. In this paper, we perform an experimental comparison of state-of-the-art flat-foot filters based on the representation choice of state, observation, matrix Lie group error and system dynamics evaluated for filter consistency and trajectory errors. The comparison is performed over simulated and real-world experiments conducted on the iCub humanoid platform. It is observed that filters on Lie groups that exploit properties of invariant filtering tend to perform better as consistent estimators while discrete-time filters in general provide higher accuracy along observable directions.
|
|
TuB-12 |
Rm12 (Room J) |
Computer Vision for Automation 2 |
Regular session |
Chair: Civera, Javier | Universidad De Zaragoza |
Co-Chair: Itsumi, Hayato | NEC |
|
12:30-12:40, Paper TuB-12.1 | |
ParaPose: Parameter and Domain Randomization Optimization for Pose Estimation Using Synthetic Data |
|
Hagelskjær, Frederik | University of Southern Denmark |
Buch, Anders Glent | University of Southern Denmark |
Keywords: Computer Vision for Automation, Object Detection, Segmentation and Categorization
Abstract: Pose estimation is the task of determining the 6D position of an object in a scene. Pose estimation aid the abilities and flexibility of robotic set-ups. However, the system must be configured towards the use case to perform adequately. This configuration is time-consuming and limits the usability of pose estimation and, thereby, robotic systems. Deep learning is a method to overcome this configuration procedure by learning parameters directly from the dataset. However, obtaining this training data can also be very time-consuming. The use of synthetic training data avoids this data collection problem, but a configuration of the training procedure is necessary to overcome the domain gap problem. Additionally, the pose estimation parameters also need to be configured. This configuration is jokingly known as grad student descent as parameters are manually adjusted until satisfactory results are obtained. This paper presents a method for automatic configuration using only synthetic data. This is accomplished by learning the domain randomization during network training, and then using the domain randomization to optimize the pose estimation parameters. The developed approach shows state-of-the-art performance of 82.0 % recall on the challenging OCCLUSION dataset, outperforming all previous methods with a large margin. These results prove the validity of automatic set-up of pose estimation using purely synthetic data.
|
|
12:40-12:50, Paper TuB-12.2 | |
Jacobian Computation for Cumulative B-Splines on SE(3) and Application to Continuous-Time Object Tracking |
|
Tirado, Javier | University of Zaragoza |
Civera, Javier | Universidad De Zaragoza |
Keywords: Computer Vision for Automation, Visual Tracking, Kinematics
Abstract: In this paper we propose a method that estimates the SE(3) continuous trajectories (orientation and translation) of the dynamic rigid objects present in a scene, from multiple RGB-D views. Specifically, we fit the object trajectories to cumulative B-Splines curves, which allow us to interpolate, at any intermediate time stamp, not only their poses but also their linear and angular velocities and accelerations. Additionally, we derive in this work the analytical SE(3) Jacobians needed by the optimization, being applicable to any other approach that uses this type of curves. To the best of our knowledge this is the first work that proposes 6-DoF continuous-time object tracking, which we endorse with significant computational cost reduction thanks to our analytical derivations. We evaluate our proposal in synthetic data and in a public benchmark, showing competitive results in localization and significant improvements in velocity estimation in comparison to discrete-time approaches.
|
|
12:50-13:00, Paper TuB-12.3 | |
Intensity Image-Based LiDAR Fiducial Marker System |
|
Shan, Jinjun | York University |
Liu, Yibo | York University |
Schofield, Hunter | York University |
Keywords: Computer Vision for Automation, Range Sensing, Object Detection, Segmentation and Categorization
Abstract: The fiducial marker system for LiDAR is crucial for the robotic application but it is still rare to date. In this paper, an Intensity Image-based LiDAR Fiducial Marker (IILFM) system is developed. This system only requires an unstructured point cloud with intensity as the input and it has no restriction on marker placement and shape. A marker detection method that locates the predefined 3D fiducials in the point cloud through the intensity image is introduced. Then, an approach that utilizes the detected 3D fiducials to estimate the LiDAR 6-DOF pose that describes the transmission from the world coordinate system to the LiDAR coordinate system is developed. Moreover, all these processes run in real-time (approx 40 Hz on Livox Mid-40 and approx 143 Hz on VLP-16). Qualitative and quantitative experiments are conducted to demonstrate that the proposed system has similar convenience and accuracy as the conventional visual fiducial marker system. The codes and results are available at: https://github.com/York-SDCNLab/IILFM.
|
|
13:00-13:10, Paper TuB-12.4 | |
Multi-Modal Non-Isotropic Light Source Modelling for Reflectance Estimation in Hyperspectral Imaging |
|
Mehami, Jasprabhjit | University of Technology Sydney |
Falque, Raphael | University of Technology Sydney |
Vidal-Calleja, Teresa A. | University of Technology Sydney |
Alempijevic, Alen | University of Technology Sydney |
Keywords: RGB-D Perception, Sensor Fusion, Calibration and Identification
Abstract: Estimating reflectance is key when working with hyperspectral cameras. The modelling of light sources can aid reflectance estimation, however, it is commonly overlooked. The key contribution of this paper is a physics-based, data-driven model formed by a Gaussian Process (GP) with a unique mean function capable of modelling a light source with an asymmetric radiant intensity distribution (RID) and a configurable attenuation function. This is referred to as the light-source-mean model. Moreover, we argue that by utilising multi-modal sensing information, we can achieve improved reflectance estimation using the proposed light source model with shape information obtained by depth cameras. An existing reflectance estimation method, that solves the dichromatic reflectance model (DRM) via quadratic programming optimisation, is augmented with terms that allow input of shape information. Experiments in simulation show that the light-source-mean GP model had less error when compared to a parametric model. The improved reflectance estimation outperforms existing methods in simulation by reducing the error by 96.8% on average when compared to existing works. We further validate the improved reflectance estimation method through a multi-modal classification application.
|
|
13:10-13:20, Paper TuB-12.5 | |
LiSnowNet: Real-Time Snow Removal for LiDAR Point Clouds |
|
Yu, Ming-Yuan | University of Michigan |
Vasudevan, Ram | University of Michigan |
Johnson-Roberson, Matthew | University of Michigan |
Keywords: Deep Learning for Visual Perception, Computer Vision for Automation, Deep Learning Methods
Abstract: Light Detection And Rangings (LiDARs) have been widely adopted to modern self-driving vehicles, providing 3D information of the scene and surrounding objects. However, adverser weather conditions still pose significant challenges to LiDARs since point clouds captured during snowfall can easily be corrupted. The resulting noisy point clouds degrade downstream tasks such as mapping. Existing works in de-noising point clouds corrupted by snow are based on nearest-neighbor search, and thus do not scale well with modern LiDARs which usually capture 100k or more points at 10Hz. In this paper, we introduce an unsupervised de-noising algorithm, LiSnowNet, running 52x faster than the state-of-the-art methods while achieving superior performance in de-noising. Unlike previous methods, the proposed algorithm is based on a deep convolutional neural network and easily be deployed with hardware accelerators such as GPUs. In addition, we demonstrate how to use the proposed method for mapping even with corrupted point clouds.
|
|
13:20-13:30, Paper TuB-12.6 | |
Application of Ghost-DeblurGAN to Fiducial Marker Detection |
|
Liu, Yibo | York University |
Haridevan, Amaldev | York University |
Schofield, Hunter | York University |
Shan, Jinjun | York University |
Keywords: Computer Vision for Automation, Object Detection, Segmentation and Categorization, Autonomous Vehicle Navigation
Abstract: Feature extraction or localization based on the fiducial marker could fail due to motion blur in real-world robotic applications. To solve this problem, a lightweight generative adversarial network, named Ghost-DeblurGAN, for real-time motion deblurring is developed in this paper. Furthermore, on account that there is no existing deblurring benchmark for such a task, a new large-scale dataset, YorkTag, is proposed that provides pairs of sharp/blurred images containing fiducial markers. With the proposed model trained and tested on YorkTag, it is demonstrated that when applied along with fiducial marker systems to motion-blurred images, Ghost-DeblurGAN improves the marker detection significantly. The datasets and codes used in this paper are available at: https://github.com/York-SDCNLab/Ghost-DeblurGAN.
|
|
13:30-13:40, Paper TuB-12.7 | |
Learning Important Regions Via Attention for Video Streaming on Cloud Robotics |
|
Itsumi, Hayato | NEC |
Beye, Florian | NEC |
Vitthal, Charvi | NEC Corporation |
Nihei, Koichi | NEC Corporation |
Keywords: Computer Vision for Automation, Mobile Manipulation, Networked Robots
Abstract: Cloud robotics, i.e., controlling robots from the cloud, make it possible to perform more complex processes, make robots smaller, and coordinate multi-robots by sharing information between robots and utilizing abundant computing resources. In cloud robotics, robots need to transmit videos to the cloud in real time to recognize their surroundings. Lowering the video quality reduces the bitrate in low bandwidth environments; however, this may lead to control errors and mis-recognition due to lack of detailed image features. Even with 5G, bandwidth fluctuates widely, especially in moving robots, making it difficult to upload high quality video consistently. To reduce bitrate while preserving Quality of Control (QoC), we propose a method of learning the important regions for a pre-trained autonomous agent using self-attention, and transmitting the video to the agent by controlling the image quality of each region on the basis of the estimated importance. The evaluation results demonstrate that our approach can maintain QoC while reducing the bitrate to 26% by setting important regions to high quality and the rest to low quality.
|
|
13:40-13:50, Paper TuB-12.8 | |
Unsupervised Simultaneous Learning for Camera Re-Localization and Depth Estimation from Video |
|
Taguchi, Shun | Toyota Central R&D Labs., Inc |
Hirose, Noriaki | TOYOTA Central R&D Labs., INC |
Keywords: Computer Vision for Automation, Localization, Mapping
Abstract: We present an unsupervised simultaneous learning framework for the task of monocular camera re-localization and depth estimation from unlabeled video sequences. Monocular camera re-localization refers to the task of estimating the absolute camera pose from an instance image in a known environment, which has been intensively studied for alternative localization in GPS-denied environments. In recent works, camera re-localization methods are trained via supervised learning from pairs of camera images and camera poses. In contrast to previous works, we propose a completely unsupervised learning framework for camera re-localization and depth estimation, requiring only monocular video sequences for training. In our framework, we train two networks that estimate the scene coordinates using directions and the depth map from each image which are then combined to estimate the camera pose. The networks can be trained through the minimization of loss functions based on our loop closed view synthesis. In experiments with the 7-scenes dataset, the proposed method outperformed the re-localization of the state-of-the-art visual SLAM, ORB-SLAM3. Our method also outperforms state-of-the-art monocular depth estimation in a trained environment.
|
|
13:50-14:00, Paper TuB-12.9 | |
HyperPocket: Generative Point Cloud Completion |
|
Spurek, Przemysław | Faculty of Mathematics and Computer Science |
Kasymov, Artur | Jagiellonian University |
Mazur, Marcin | Jagiellonian University in Krakow |
Diana, Janik | Jagiellonian University |
Tadeja, Slawomir Konrad | Jagiellonian University |
Lukasz, Struski | Jagiellonian University |
Jacek, Tabor | Jagiellonian University |
Trzcinski, Tomasz | Warsaw University of Technology |
Keywords: Computer Vision for Automation, Recognition, Deep Learning Methods
Abstract: Scanning real-life scenes with modern registration devices typically give incomplete point cloud representations, mostly due to the limitations of the scanning process and 3D occlusions. Therefore, completing such partial representations remains a fundamental challenge of many computer vision applications. Most of the existing approaches aim to solve this problem by learning to reconstruct individual 3D objects in a synthetic setup of an uncluttered environment, which is far from a real-life scenario. In this work, we reformulate the problem of point cloud completion into an objects hallucination task. Thus, we introduce a novel autoencoder-based architecture called HyperPocket that disentangles latent representations and, as a result, enables the generation of multiple variants of the completed 3D point clouds. Furthermore, we split point cloud processing into two disjoint data streams and leverage a hypernetwork paradigm to fill the spaces, dubbed pockets, that are left by the missing object parts. As a result, the generated point clouds are smooth, plausible, and geometrically consistent with the scene. Moreover, our method offers competitive performances to the other state-of-the-art models, enabling a plethora of novel applications.
|
|
TuB-13 |
Rm13 (Room K) |
Marine Robotics 1 |
Regular session |
Chair: Tanner, Herbert G. | University of Delaware |
Co-Chair: Pang, Yusong | Delft University of Technology |
|
12:30-12:40, Paper TuB-13.1 | |
Flexible Collision-Free Platooning Method for Unmanned Surface Vehicle with Experimental Validations |
|
Du, Bin | Shanghai Jiao Tong University |
Lin, Bin | Shanghai Jiao Tong University |
Xie, Wei | Shanghai Jiao Tong University |
Zhang, Weidong | Shanghai JiaoTong University |
Negenborn, R.R. | Delft University of Technology |
Pang, Yusong | Delft University of Technology |
Keywords: Marine Robotics, Autonomous Vehicle Navigation, Collision Avoidance
Abstract: This paper addresses the flexible formation problem for unmanned surface vehicles in the presence of obstacles. Building upon the leader-follower formation scheme, a hybrid line-of-sight based flexible platooning method is proposed for follower vehicle to keep tracking the leader ship. A fusion artificial potential field collision avoidance approach is tailored to generate optimal collision-free trajectories for the vehicle to track. To steer the vehicle towards and stay within the neighborhood of the generated collision-free trajectory, a nonlinear model predictive controller is designed. Experimental results are presented to validate the efficiency of proposed method, showing that the the unmanned surface vehicle is able to track the leader ship without colliding with the surrounded static obstacles in the considered experiments.is able to track the leader ship without colliding with the surrounded static obstacles in the considered experiments.
|
|
12:40-12:50, Paper TuB-13.2 | |
Development and Field Testing of an Optimal Path Following ASV Controller for Marine Surveys |
|
Baxevani, Kleio | University of Delaware |
Otto, Grant | University of Delaware |
Tanner, Herbert G. | University of Delaware |
Trembanis, Arthur | University of Delaware |
Keywords: Marine Robotics, Field Robots, Motion and Path Planning
Abstract: Marine autonomous vehicles deployed to conduct marine geophysical surveys are becoming an increasingly used asset in the commercial, academic, and defense industries. However, the ability to collect high-quality data from applicable sensors is directly related to the robustness of vehicle motion caused by environmental disturbances. In this paper we designed and integrated a new path following controller on an autonomous surface vehicle (ASV) that minimizes the linear and angular accelerations on the sensor’s local frame. Simulation and experimental results verify reduction of vehicle motion, improvement in path following, and improvement in preliminary sonar data quality compared to that of the existing proportional-yaw path following controller.
|
|
12:50-13:00, Paper TuB-13.3 | |
Inertial-Measurement-Based Catenary Shape Estimation of Underwater Cables for Tethered Robots |
|
Drupt, Juliette | Université De Toulon |
Dune, Claire | Université De Toulon |
Comport, Andrew Ian | CNRS-I3S/UNS |
Seillier, Sabine | COSMER - Université De Toulon |
Hugel, Vincent | University of Toulon |
Keywords: Marine Robotics
Abstract: This paper deals with the estimation of the shape of a catenary for a negatively buoyant cable, connecting a pair of underwater robots in a robot chain. The new estimation method proposed here is based on the calculation of local tangents thanks to the data issued from inertial measurement units (IMUs), which are attached to the cable near its ends. This method is compared with a vision-based estimation method that was developed previously. Experiments are conducted, in the air and in a pool, using a motion capture system for ground truth. The results obtained show that the new method significantly improves the estimation of the catenary height. Actually, the identification of the cable shape is not affected by the limits of the camera’s field of view and by the image projection, resulting in increased accuracy and range, without singularities.
|
|
13:00-13:10, Paper TuB-13.4 | |
Motion Attribute-Based Clustering and Collision Avoidance of Multiple In-Water Obstacles by Autonomous Surface Vehicle |
|
Jeong, Mingi | Dartmouth College |
Quattrini Li, Alberto | Dartmouth College |
Keywords: Marine Robotics, Collision Avoidance, Autonomous Vehicle Navigation
Abstract: Navigation and obstacle avoidance in aquatic environments for autonomous surface vehicles (ASVs) in high-traffic maritime scenarios is still an open challenge, as the Convention on the International Regulations for Preventing Collisions at Sea (COLREGs) is not defined for multi-encounter situations. Current state-of-the-art methods resolve single-to-single encounters with sequential actions and assume that other obstacles follow COLREGs. Our work proposes a novel real-time non-myopic obstacle avoidance method, allowing an ASV that has only partial knowledge of the surroundings within the sensor radius to navigate in high-traffic maritime scenarios. Specifically, we achieve a holistic view of the feasible ASV action space able to avoid deadlock scenarios, by proposing (1) a clustering method based on motion attributes of other obstacles, (2) a geometric framework for identifying the feasible action space, and (3) a multi-objective optimization to determine the best action. Theoretical analysis and extensive realistic experiments in simulation considering real-world traffic scenarios demonstrate that our proposed real-time obstacle avoidance method is able to achieve safer trajectories than other state-of-the-art methods and that is robust to uncertainty present in the current information available to the ASV.
|
|
13:10-13:20, Paper TuB-13.5 | |
Training Dynamic Motion Primitives Using Deep Reinforcement Learning to Control a Robotic Tadpole |
|
Hameed, Imran | The Hong Kong Polytechnic University |
Chao, Xu | The Hong Kong Polytechnic University |
Navarro-Alarcon, David | The Hong Kong Polytechnic University |
Jing, Xingjian | City University of Hong Kong |
Keywords: Marine Robotics, Bioinspired Robot Learning, Reinforcement Learning
Abstract: Developing a good control strategy for biomimetic robots is challenging. Robust control methods require an accurate model of the robot. Nowadays, model-free methods are being extensively explored for the control and navigation of terrestrial robots. In this paper, we consider a novel deep reinforcement learning-based model-free swimming control for our bio-inspired robotic tadpole. To realize this, we utilize dynamic motion primitives, which can represent a large range of motion behaviors, and combine them with a decoupled reinforcement learning framework. The proposed architecture optimizes the motion primitives first to develop a travelling wave undulation pattern in the tail and then to navigate the robot along different predefined paths. Through this framework, effective swimming gait emerges, and the robot is able to navigate well on the surface of water. This framework combines the optimization potential of deep reinforcement learning with stability and generalization properties of dynamic motion primitives. We train and test our method on a simulated model of the robot to demonstrate the effectiveness of the method and also conduct experimental testing on the real robot to verify the results.
|
|
13:20-13:30, Paper TuB-13.6 | |
Hydrodynamic Parameters Estimation Using Varying Forces and Numerical Integration Fitting Method |
|
Gartner, Nicolas | Université De Toulon |
Richier, Mathieu | University of Toulon |
Dune, Claire | Université De Toulon |
Hugel, Vincent | University of Toulon |
Keywords: Marine Robotics, Dynamics, Performance Evaluation and Benchmarking
Abstract: This paper presents a new experimental method for the estimation of hydrodynamic parameters of underwater vehicles, namely the added mass and drag coefficients. The principle is to accurately record the movement of the vehicle in response to a known time-varying actuating force, and find an optimal solution for the parameters that allows simulating a velocity trajectory that fits the recorded trajectory. The optimal solution is obtained using a Numerical Integration Fitting method, which, employing an error minimization algorithm, successively generates velocity trajectories through numerical integration until the parameters converge. This method is applied to determine the hydrodynamic parameters of a torpedo-shaped Autonomous Underwater Vehicle (AUV). The hydrodynamic parameters obtained are compared and found to be close to their theoretical values. In addition, the use of a time-varying force is shown to provide better and more reliable results than the use of a constant or zero external force.
|
|
13:30-13:40, Paper TuB-13.7 | |
An Underwater Target Perception Framework for Underwater Operation Scene |
|
Gao, Jue | Gao Jue |
Zhu, Chi | Maebashi Institute of Technology |
Keywords: Object Detection, Segmentation and Categorization, Surveillance Robotic Systems, Cognitive Modeling
Abstract: This paper proposes an underwater target perception framework to comprehensively explore target information in underwater scenes, to improve the work efficiency and safety of underwater operations. This framework adopts a layered processing mechanism including water column imaging, constant false alarm rate detection (CFAR) detection, and local feature analysis, to accurately distinguish between false targets, static targets, and dynamic targets in the underwater scene, and obtain the motion trajectory of dynamic targets. The experiment is designed to simulate the underwater operation scene, and the results prove the effectiveness of the proposed framework.
|
|
TuB-14 |
Rm14 (Room 501) |
Soft Robot Materials and Design 2 |
Regular session |
Chair: Rossiter, Jonathan | University of Bristol |
Co-Chair: Kwon, Dong-Soo | KAIST |
|
12:30-12:40, Paper TuB-14.1 | |
Self-Morphing Soft Parallel-And-Coplanar Electroadhesive Grippers Based on Laser-Scribed Graphene Oxide Electrodes |
|
Guo, Jianglong | Harbin Institute of Technology (Shenzhen) |
Kuhnel, Djen | Ecole Polytechnique Fédérale De Lausanne |
Qi, Qiukai | University of Bristol |
Xiang, Chaoqun | Bristol Robotics Laboratory |
Ho, Van | Japan Advanced Institute of Science and Technology |
Faul, Charl | University of Bristol |
Rossiter, Jonathan | University of Bristol |
Keywords: Grippers and Other End-Effectors, Soft Sensors and Actuators, Soft Robot Materials and Design
Abstract: Electroadhesion is a versatile and controllable adhesion mechanism that has been used extensively in robotics. Soft electroadhesion embodies electrostatic adhesion in soft materials and is required for shape-adaptive and safe grasping of curved objects and delicate materials. In this work, we present a soft electroadhesive fabrication method based on laser scribing graphene oxide on a silicone film, which is cost-effective, facile and green. The method can be used to generate complex electroadhesive patterns without molds or stencils. We then present a 2D finite element model to demonstrate the shape-changing behavior and electric field distributions of a dual-mode parallel dielectric elastomer actuation and coplanar electroadhesion structure. The soft electroadhesive fabrication method based on laser-scribed graphene oxide electrodes and its experimental characterization results, together with its shape-morphing simulation model are expected to enable the wider adoption of soft electroadhesion in future robotics.
|
|
12:40-12:50, Paper TuB-14.2 | |
Embeddable Coiled Soft Sensor-Based Joint Angle Sensing for Flexible Surgical Manipulator |
|
Yi, Yesung | Korea Advanced Institute of Science and Technology |
Youn, Jung-Hwan | Korea Advanced Institute of Science and Technology (KAIST) |
Kyung, Ki-Uk | Korea Advanced Institute of Science & Technology (KAIST) |
Kwon, Dong-Soo | KAIST |
Keywords: Medical Robots and Systems, Soft Sensors and Actuators, Flexible Robotics
Abstract: Tendon-driven flexible endoscopic surgical robots have been developed to access narrow curved paths without incision. Robot shape information is essential for precise control and to prevent unwanted tissue damage. In this paper, we propose a joint angle sensing method using coiled soft sensors to estimate the shape of the hyperredundant manipulator, which is commonly used in flexible endoscopic surgical robots. The soft sensors can be fabricated with small size and are highly stretchable, such that by being pre-stretched, they can be integrated between individual joints, maintain a center hollow, and sense both compression and extension. The pre-stretch length is experimentally selected by using the sensor linearity to maximize the potential sensitivity. We validated the proposed design using a two-degree of freedom (DOF) single joint manipulator by implementing two sensors; sensors at all joints could sense joint angle independently and simultaneously with a root-mean-square error (RMSE) less than 2.53°. Based on the proposed method, a two-DOF configuration of the hyperredundant manipulator that can be used in real applications was achieved, following a constant curvature model in real time with values RMSE of 2.30° and 2.63°, for pitch and yaw joint angle respectively.
|
|
12:50-13:00, Paper TuB-14.3 | |
A Soft Fluidic Sensor-Actuator for Active Sensing of Force and Displacement in Biomedical Applications |
|
Jones, Joanna | University of Sheffield |
Damian, Dana | University of Sheffield |
Keywords: Soft Sensors and Actuators, Force and Tactile Sensing, Hydraulic/Pneumatic Actuators
Abstract: Achieving compact and biocompatible actuators with sensing capabilities is a key challenge for the safety critical and highly patient-specific biomedical field. In this study, a compact and versatile soft fluidic sensor-actuator capable of measuring both force and displacement in static and dynamic conditions is presented. Pressure and resistance are shown to be interchangeable in predicting load and sensor-actuator height, and showed good repeatability and distinction between the loaded and constrained conditions tested. Furthermore the sensor-actuator is demonstrated in a probe application and showed comparable findings to a tensile test machine when tested on three objects of varying stiffness. Overall, this sensor-actuator has the potential to be a key building block for biomedical robots that require large expansion, as well as continuous monitoring of both displacement and force.
|
|
13:00-13:10, Paper TuB-14.4 | |
GSG: A Granary-Shaped Soft Gripper with Mechanical Sensing Via Snap-Through Structure |
|
Dong, Huixu | Zhejiang University |
Chen, Chao-Yu | National University of Singapore |
Qiu, Chen | Nanyang Technological University |
Yeow, Chen-Hua | National University of Singapore |
Yu, Haoyong | National University of Singapore |
Keywords: Grasping, Soft Sensors and Actuators, Soft Robot Materials and Design
Abstract: Soft robotic grippers have attracted considerable attention in terms of the advantages of the high compliance and robustness to variance in object geometry; however, they are still limited by the corresponding sensing capabilities. We propose a novel soft gripper that looks like a 慻ranary?in the geometrical shape with a snap-through bistable mechanism fabricated by an ordered mold technology, which consists of a palm chamber, shell, cap and three fingers. It can achieve 憇ensing?mechanically and perform pinching, enveloping and caging grasps for objects with various profiles. In particular, the snap-through bistable mechanism in the proposed gripper allows us to reduce the complexity of the mechanism, control, sensing designs. The grasping behavior is activated once the gripper抯 deformation or perceived pressure arrives at a certain value. First, after the theoretical model for snap-through behavior is constructed, the modularized design of the gripper is described in detail. Then, the ordered molding method is employed to fabricate the proposed gripper. Finally, the finite element (FE) simulations are conducted to verify the built theoretical model. Further, a series of grasping experiments are carried out to assess the performance of the proposed gripper on grasping and sensing. The experimental results illustrate that the proposed gripper can manipulate a variety of soft and rigid objects and remain stable even though it undergoes external disturbances. (YouTube video: https://youtu.be/74h9A-qlv28)
|
|
13:10-13:20, Paper TuB-14.5 | |
FireFly: An Insect-Scale Aerial Robot Powered by Electroluminescent Soft Artificial Muscles |
|
Kim, Suhan | Massachusetts Institute of Technology (MIT) |
Hsiao, Yi-Hsuan | Massachusetts Institute of Technology |
Chen, Yu Fan | Massa | |