| |
Last updated on October 25, 2019. This conference program is tentative and subject to change
Technical Program for Tuesday November 5, 2019
|
TuAT1 Regular session, L1-R1 |
Add to My Program |
Calibration and Identification |
|
|
Chair: Zhang, Xuebo | Nankai University, |
Co-Chair: Stoyanov, Danail | University College London |
|
11:00-11:15, Paper TuAT1.1 | Add to My Program |
A Novel Robust Approach for Correspondence-Free Extrinsic Calibration |
Hu, Xiao | Technical University of Denmark |
Olesen, Daniel | Technical University of Denmark |
Knudsen, Per | Technical University of Denmark |
Keywords: Calibration and Identification
Abstract: Extrinsic calibration is a necessary step when using heterogeneous sensors for robotics applications. Most existing methods work under the assumption that the prior data correspondence is known. Considering data loss and false measurements, the correspondence may not be accessible in practice. To solve this problem without knowing the correspondence, several probabilistic methods have been proposed. However, an implicit restriction on input data limits their application. Therefore, in this paper, we propose a more stable correspondence-free method with two improvements that can relax the restrictions on inputs and improve the calibration accuracy. The first improvement finds consistent sets from raw inputs using screw invariants, which significantly improve the robustness in case of outliers and data loss. A new optimization method on matrix Lie group is proposed as the second improvement, which demonstrates better accuracy. The experimental results on both numerical and real data show the superiority and robustness of the proposed method.
|
|
11:15-11:30, Paper TuAT1.2 | Add to My Program |
Automatic Multi-Sensor Extrinsic Calibration for Mobile Robots |
Zuñiga-Noël, David | University of Malaga |
Ruiz-Sarmiento, J.R. | University of Malaga |
Gomez-Ojeda, Ruben | University of Málaga |
González-Jiménez, Javier | University of Málaga |
Keywords: Calibration and Identification
Abstract: In order to fuse measurements from multiple sensors mounted on a mobile robot, it is needed to express them in a common reference system through their relative spatial transformations. In this paper, we present a method to estimate the full 6DoF extrinsic calibration parameters of multiple heterogeneous sensors (Lidars, Depth and RGB cameras) suitable for automatic execution on a mobile robot. Our method computes the 2D calibration parameters (x, y, yaw) through a motion-based approach, while for the remaining 3 parameters (z, pitch, roll) it requires the observation of the ground plane for a short period of time. What set this proposal apart from others is that: i) all calibration parameters are initialized in closed form, and ii) the scale ambiguity inherent to motion estimation from a monocular camera is explicitly handled, enabling the combination of these sensors and metric ones (Lidars, stereo rigs, etc.) within the same optimization framework. We provide a formal definition of the problem, as well as of the contributed method, for which a C++ implementation has been made publicly available. The suitability of the method has been assessed in simulation and with real data from indoor and outdoor scenarios. Finally, improvements over state-of-the-art motion-based calibration proposals are shown through experimental evaluation.
|
|
11:30-11:45, Paper TuAT1.3 | Add to My Program |
Automatic Calibration of Multiple 3D LiDARs in Urban Environments |
Jiao, Jianhao | The Hong Kong University of Science and Technology |
Yu, Yang | Hong Kong University of Science and Technology |
Liao, Qing hai | Hong Kong University of Science and Technology |
Ye, Haoyang | The Hong Kong University of Science and Technology |
Fan, Rui | The Hong Kong University of Science and Technology |
Liu, Ming | Hong Kong University of Science and Technology |
Keywords: Calibration and Identification, Computer Vision for Transportation, Intelligent Transportation Systems
Abstract: Multiple LiDARs have progressively emerged on autonomous vehicles for rendering a rich view and dense measurements. However, the lack of precise calibration negatively affects their potential applications. In this paper, we propose a novel system that enables automatic multi-LiDAR calibration without any calibration target, prior environment information, and initial values of the extrinsic parameters. Our approach starts with a hand-eye calibration for automatic initialization by aligning the motions of each sensor. The initial results are then refined with an appearance-based method by minimizing a cost function constructed by point-plane distance. Experimental results on simulated and real-world data sets demonstrate the reliability and accuracy of our calibration approach. The proposed approach can calibrate a multi-LiDAR system with the rotation and translation errors less than 0.04rad and 0.1m respectively for a mobile platform.
|
|
11:45-12:00, Paper TuAT1.4 | Add to My Program |
Hand-Eye Calibration with a Remote Centre of Motion |
Pachtrachai, Krittin | University College London |
Vasconcelos, Francisco | University College London |
Dwyer, George | University College London |
Hailes, Stephen | University College London , Dept. of Computer Science , Gower S |
Stoyanov, Danail | University College London |
Keywords: Calibration and Identification, Formal Methods in Robotics and Automation, Computer Vision for Medical Robotics
Abstract: In the eye-in-hand robot configuration, hand-eye calibration plays a vital role in completing the link between the robot and camera coordinate systems. Calibration algorithms are mature and provide accurate transformation estimations for an effective camera-robot link but rely on a sufficiently wide range of calibration data to avoid errors and degenerate configurations. This can be difficult in the context of keyhole surgical robots because they are mechanically constrained to move around a remote centre of motion (RCM) which is located at the trocar port. The trocar limits the range of feasible calibration poses that can be obtained and results in ill-conditioned hand-eye constraints. In this paper, we propose a new approach to deal with this problem by incorporating the RCM constraints into the hand-eye formulation. We show that this not only avoids ill-conditioned constraints but is also more accurate than classic hand-eye calibration with a free 6DoF motion, due to solving simpler equations that take advantage of the reduced DoF. We validate our method using simulation to test numerical stability and a physical implementation on an RCM constrained KUKA LBR iiwa 14 R820 equipped with a NanEye stereo camera.
|
|
12:00-12:15, Paper TuAT1.5 | Add to My Program |
Model Free Calibration of Wheeled Robots Using Gaussian Process |
Nutalapati, Mohan Krishna | Indian Institute of Technology Kanpur |
Arora, Lavish | Indian Institute of Technology Kanpur |
Bose, Anway | Indian Institute of Technology, Kanpur |
Rajawat, Ketan | IIT Kanpur |
Hegde, Rajesh M | Indian Institute of Technology Kanpur |
Keywords: Calibration and Identification, Kinematics, Wheeled Robots
Abstract: Robotic calibration allows for the fusion of data from multiple sensors such as odometers, cameras, etc., by providing appropriate relationships between the corresponding reference frames. For wheeled robots equipped with camera/lidar along with wheel encoders, calibration entails learning the motion model of the sensor or the robot in terms of the data from the encoders and generally carried out before performing tasks such as simultaneous localization and mapping (SLAM). This work puts forward a novel Gaussian Process-based non-parametric approach for calibrating wheeled robots with arbitrary or unknown drive configurations. The procedure is more general as it learns the entire sensor/robot motion model in terms of odometry measurements. Different from existing non-parametric approaches, our method relies on measurements from the onboard sensors and hence does not require the ground truth information from external motion capture systems. Alternatively, we propose a computationally efficient approach that relies on the linear approximation of the sensor motion model. Finally, we perform experiments to calibrate robots with un-modelled effects to demonstrate the accuracy, usefulness, and flexibility of the proposed approach.
|
|
12:15-12:30, Paper TuAT1.6 | Add to My Program |
A Robust Extrinsic Calibration Framework for Vehicles with Unscaled Sensors |
Walters, Celyn | University of Surrey |
Mendez Maldonado, Oscar Alejandro | University of Surrey |
Hadfield, Simon | University of Surrey |
Bowden, Richard | University of Surrey |
Keywords: Calibration and Identification, Sensor Fusion, Autonomous Vehicle Navigation
Abstract: Accurate extrinsic sensor calibration is essential for both autonomous vehicles and robots. Traditionally this is an involved process requiring calibration targets, known fiducial markers and is generally performed in a lab. Moreover, even a small change in the sensor layout requires recalibration. With the anticipated arrival of consumer autonomous vehicles, there is demand for a system which can do this automatically, after deployment and without specialist human expertise. To solve these limitations, we propose a flexible framework which can estimate extrinsic parameters without an explicit calibration stage, even for sensors with unknown scale. Our first contribution builds upon standard hand-eye calibration by jointly recovering scale. Our second contribution is that our system is made robust to imperfect and degenerate sensor data, by collecting independent sets of poses and automatically selecting those which are most ideal. We show that our approach's robustness is essential for the target scenario. Unlike previous approaches, ours runs in real time and constantly estimates the extrinsic transform. For both an ideal experimental setup and a real use case, comparison against these approaches shows that we outperform the state-of-the-art. Furthermore, we demonstrate that the recovered scale may be applied to the full trajectory, circumventing the need for scale estimation via sensor fusion.
|
|
TuAT2 Regular session, L1-R2 |
Add to My Program |
Deep Learning for Aerial Systems |
|
|
Chair: Saha, Indranil | IIT Kanpur |
Co-Chair: Hoenig, Wolfgang | California Institute of Technology |
|
11:00-11:15, Paper TuAT2.1 | Add to My Program |
DeepControl: Energy-Efficient Control of a Quadrotor Using a Deep Neural Network |
Varshney, Pratyush | Indian Institute of Technology, Kanpur |
Nagar, Gajendra | Indian Institute of Technology, Kanpur |
Saha, Indranil | IIT Kanpur |
Keywords: Deep Learning in Robotics and Automation, Aerial Systems: Mechanics and Control, Optimization and Optimal Control
Abstract: Synthesis of a feedback controller for nonlinear dynamical systems like a quadrotor requires to deal with the trade-off between performance and online computation requirement of the controller. Model predictive controllers (MPC) provide excellent control performance, but at the cost of high online computation. In this paper, we present our experience in approximating the behavior of an MPC for a quadrotor with a feed-forward neural network. To facilitate the collection of training data, we create a faithful model of the quadrotor and use Gazebo simulator to collect sufficient training data. The deep neural network (DNN) controller learned from the training data has been tested on various trajectories to compare its performance with a model-predictive controller. Our experimental results show that our DNN controller can provide almost similar trajectory tracking performance at a lower control computation cost, which helps in increasing the flight time of the quadrotor. Moreover, the hardware requirement for our DNN controller is significantly less than that for the MPC controller. Thus, the use of DNN based controller also helps in reducing the overall price of a quadrotor.
|
|
11:15-11:30, Paper TuAT2.2 | Add to My Program |
Informed Region Selection for Efficient UAV-Based Object Detectors: Altitude-Aware Vehicle Detection with CyCAR Dataset |
Kouris, Alexandros | Imperial College London |
Kyrkou, Christos | University of Cyprus |
Bouganis, Christos-Savvas | Imperial College London |
Keywords: Deep Learning in Robotics and Automation, Object Detection, Segmentation and Categorization, Aerial Systems: Perception and Autonomy
Abstract: Deep Learning-based object detectors enhance the capabilities of remote sensing platforms, such as Unmanned Aerial Vehicles (UAVs), in a wide spectrum of machine vision applications. However, the integration of deep learning introduces heavy computational requirements, preventing the deployment of such algorithms in scenarios that impose low-latency constraints during inference, in order to make mission-critical decisions in real-time. In this paper, we address the challenge of efficient deployment of region-based object detectors in aerial imagery, by introducing an informed methodology for extracting candidate detection regions (proposals). Our approach considers information from the UAV on-board sensors, such as flying altitude and light-weight computer vision filters, along with prior domain knowledge to intelligently decrease the number of region proposals by eliminating false-positives at an early stage of the computation, reducing significantly the computational workload while sustaining the detection accuracy. We apply and evaluate the proposed approach on the task of vehicle detection. Our experiments demonstrate that state-of-the-art detection models can achieve up to 2.6x faster inference by employing our altitude-aware data-driven methodology. Alongside, we introduce and provide to the community a novel vehicle-annotated and altitude-stamped dataset of real UAV imagery, captured at numerous flying heights under a wide span of traffic scenarios.
|
|
11:30-11:45, Paper TuAT2.3 | Add to My Program |
Sim-To-(Multi)-Real: Transfer of Low-Level Robust Control Policies to Multiple Quadrotors |
Molchanov, Artem | University of Southern California |
Chen, Tao | University of Southern California |
Hoenig, Wolfgang | California Institute of Technology |
Preiss, James | USC |
Ayanian, Nora | University of Southern California |
Sukhatme, Gaurav | University of Southern California |
Keywords: Deep Learning in Robotics and Automation, Aerial Systems: Mechanics and Control, Learning and Adaptive Systems
Abstract: Quadrotor stabilizing controllers often require careful, model-specific tuning for safe operation. We use reinforcement learning to train policies in simulation that transfer remarkably well to multiple different physical quadrotors. Our policies are low-level, i.e., we map the rotorcrafts' state directly to the motor outputs. The trained control policies are very robust to external disturbances and can withstand harsh initial conditions such as throws. We show how different training methodologies (change of the cost function, modeling of noise, use of domain randomization) might affect flight performance. To the best of our knowledge, this is the first work that demonstrates that a simple neural network can learn a robust stabilizing low-level quadrotor controller (without the use of a stabilizing PD controller) that is shown to generalize to multiple quadrotors.
|
|
11:45-12:00, Paper TuAT2.4 | Add to My Program |
Low Level Control of a Quadrotor with Deep Model-Based Reinforcement Learning |
Lambert, Nathan | University of California, Berkeley |
Drew, Daniel S. | UC Berkeley |
Yaconelli, Joseph | University of Oregon |
Levine, Sergey | UC Berkeley |
Calandra, Roberto | Facebook |
Pister, Kristofer S. J. | University of California, Berkeley |
Keywords: Deep Learning in Robotics and Automation, Aerial Systems: Mechanics and Control, Model Learning for Control
Abstract: Designing effective low-level robot controllers often entail platform-specific implementations that require manual heuristic parameter tuning, significant system knowledge, or long design times. With the rising number of robotic and mecha- tronic systems deployed across areas ranging from industrial automation to intelligent toys, the need for a general approach to generating low-level controllers is increasing. To address the challenge of rapidly generating low-level controllers, we argue for using model-based reinforcement learning (MBRL) trained on relatively small amounts of automatically generated (i.e., without system simulation) data. In this paper, we explore the capabilities of MBRL on a Crazyflie centimeter-scale quadrotor with rapid dynamics to predict and control at <50Hz. To our knowledge, this is the first use of MBRL for controlled hover of a quadrotor using only on-board sensors, direct motor input signals, and no initial dynamics knowledge. Our controller leverages rapid simulation of a neural network forward dynamics model on a GPU-enabled base station, which then transmits the best current action to the quadrotor firmware via radio. In our experiments, the quadrotor achieved hovering capability of up to 6 seconds with 3 minutes of experimental training data.
|
|
12:00-12:15, Paper TuAT2.5 | Add to My Program |
A Convolutional Neural Network Feature Detection Approach to Autonomous Quadrotor Indoor Navigation |
Garcia, Adriano | Binghamton University |
Mittal, Sandeep S | Binghamton University |
Kiewra, Edward | SUNY Binghamton |
Ghose, Kanad | State University of New York, Binghamton |
Keywords: Deep Learning in Robotics and Automation, Visual-Based Navigation, Aerial Systems: Perception and Autonomy
Abstract: Object detection, extended to recognize and localize indoor structural features, is used to enable a quadrotor drone to autonomously navigate through indoor environments. The video stream from a monocular front-facing camera on-board a quadrotor drone is fed to an off-board system that runs a Convolutional Neural Network (CNN) object detection algorithm to identify specific features such as dead-ends, doors, and intersections in hallways. Using pixel-scale dimensions of the bounding boxes around the recognized objects, the distance to intersections, dead-ends and doorways can be estimated accurately using a Support Vector Regression (SVR) model to generate flight control commands for consistent real-time autonomous navigation at flight speeds approaching 2 m/s.
|
|
12:15-12:30, Paper TuAT2.6 | Add to My Program |
Long Range Neural Navigation Policies for the Real World |
Wahid, Ayzaan | Google |
Toshev, Alexander | Google |
Fiser, Marek | Google |
Lee, Tsang-Wei Edward | Google |
Keywords: Deep Learning in Robotics and Automation
Abstract: Learned Neural Network based policies have shown promising results for robot navigation. However, most of these approaches fall short of being used on a real robot due to the extensive simulated training they require. These simulations lack the visuals and dynamics of the real world, which makes it infeasible to deploy on a real robot. We present a novel Neural Net based policy, NavNet, which allows for easy deployment on a real robot. It consists of two sub policies – a high level policy which can understand real images and perform long range planning expressed in high level commands; a low level policy that can translate the long range plan into low level commands on a specific platform in a safe and robust manner. For every new deployment, the high level policy is trained on an easily obtainable scan of the environment modeling its visuals and layout. We detail the design of such an environment and how one can use it for training a final navigation policy. Further, we demonstrate a learned low-level policy. We deploy the model in a large office building and test it extensively, achieving 0.80 success rate over long navigation runs and outperforming SLAM-based models in the same settings.
|
|
TuAT3 Regular session, L1-R3 |
Add to My Program |
Learning and Adaptive Systems I |
|
|
Chair: Calinon, Sylvain | Idiap Research Institute |
Co-Chair: Caldwell, Darwin G. | Istituto Italiano Di Tecnologia |
|
11:00-11:15, Paper TuAT3.1 | Add to My Program |
Uncertainty-Aware Imitation Learning Using Kernelized Movement Primitives |
Silvério, João | Idiap Research Institute |
Huang, Yanlong | Istituto Italiano Di Tecnologia |
Abu-Dakka, Fares | Aalto University |
Rozo, Leonel | Bosch Center for Artificial Intelligence |
Caldwell, Darwin G. | Istituto Italiano Di Tecnologia |
Keywords: Learning from Demonstration, Learning and Adaptive Systems
Abstract: During the past few years, probabilistic approaches to imitation learning have earned a relevant place in the literature. One of their most prominent features, in addition to extracting a mean trajectory from task demonstrations, is that they provide a variance estimation. The intuitive meaning of this variance, however, changes across different techniques, indicating either variability or uncertainty. In this paper we leverage kernelized movement primitives (KMP) to provide a new perspective on imitation learning by predicting variability, correlations and uncertainty about robot actions. This rich set of information is used in combination with optimal controller fusion to learn actions from data, with two main advantages: i) robots become safe when uncertain about their actions and ii) they are able to leverage partial demonstrations, given as elementary sub-tasks, to optimally perform a higher level, more complex task. We showcase our approach in a painting task, where a human user and a KUKA robot collaborate to paint a wooden board. The task is divided into two sub-tasks and we show that using our approach the robot becomes compliant (hence safe) outside the training regions and executes the two sub-tasks with optimal gains.
|
|
11:15-11:30, Paper TuAT3.2 | Add to My Program |
Bayesian Gaussian Mixture Model for Robotic Policy Imitation |
Pignat, Emmanuel | Idiap Research Institute, Martigny, Switzerland |
Calinon, Sylvain | Idiap Research Institute |
Keywords: Learning from Demonstration, Learning and Adaptive Systems
Abstract: A common approach to learn robotic skills is to imitate a policy demonstrated by a supervisor. One of the existing problems is that, due to the compounding of small errors and perturbations, the robot may leave the states where demonstrations were given. If no strategy is employed to provide a guarantee on how the robot will behave when facing unknown states, catastrophic outcomes can happen. An appealing approach is to use Bayesian methods, which offer a quantification of the action uncertainty given the state. Bayesian methods are usually more computationally demanding and require more complex design choices than their non-Bayesian alternatives, which limits their application. In this work, we present a Bayesian method that is both simple to set up, computationally efficient and that can adapt to a wide range of problems. These advantages make this method very convenient for imitation of robotic manipulation tasks in the continuous domain. We exploit the provided uncertainty to fuse the imitation policy with other policies. The approach is validated on a Panda robot with three tasks using different control input/state pairs.
|
|
11:30-11:45, Paper TuAT3.3 | Add to My Program |
High-Dimensional Motion Segmentation by Variational Autoencoder and Gaussian Processes |
Nagano, Masatoshi | University of Electro-Communications |
Nakamura, Tomoaki | The University of Electro-Communications |
Nagai, Takayuki | Osaka University |
Mochihashi, Daichi | Institute of Statistical Mathematics |
Kobayashi, Ichiro | Ochanomizu University |
Takano, Wataru | Osaka University |
Keywords: Probability and Statistical Methods, Learning and Adaptive Systems, Deep Learning in Robotics and Automation
Abstract: Humans perceive continuous high-dimensional information by dividing it into significant segments such as words and units of motion. We believe that such unsupervised segmentation is also important for robots to learn topics such as language and motion. To this end, we previously proposed a hierarchical Dirichlet process--Gaussian process--hidden semi-Markov model (HDP-GP-HSMM). However, an important drawback to this model is that it cannot divide high-dimensional time-series data. Further, low-dimensional features must be extracted in advance. Segmentation largely depends on the design of features, and it is difficult to design effective features, especially in the case of high-dimensional data. To overcome this problem, this paper proposes a hierarchical Dirichlet process--variational autoencoder--Gaussian process--hidden semi-Markov model (HVGH). The parameters of the proposed HVGH are estimated through a mutual learning loop of the variational autoencoder and our previously proposed HDP-GP-HSMM. Hence, HVGH can extract features from high-dimensional time-series data, while simultaneously dividing it into segments in an unsupervised manner. In an experiment, we used various motion-capture data to show that our proposed model estimates the correct number of classes and more accurate segments than baseline methods. Moreover, we show that the proposed method can learn latent space suitable for segmentation.
|
|
11:45-12:00, Paper TuAT3.4 | Add to My Program |
Learning Barrier Functions for Constrained Motion Planning with Dynamical Systems |
Saveriano, Matteo | University of Innsbruck |
Lee, Dongheui | Technical University of Munich |
Keywords: Learning from Demonstration, Learning and Adaptive Systems, Motion and Path Planning
Abstract: Stable dynamical systems are a flexible tool to plan robotic motions in real-time. In the robotic literature, dynamical system motions are typically planned without considering possible limitations in the robot’s workspace. This work presents a novel approach to learn workspace constraints from human demonstrations and to generate motion trajectories for the robot that lie in the constrained workspace. Training data are incrementally clustered into different linear subspaces and used to fit a low dimensional representation of each subspace. By considering the learned constraint subspaces as zeroing barrier functions, we are able to design a control input that keeps the system trajectory within the learned bounds. This control input is effectively combined with the original system dynamics preserving eventual asymptotic properties of the unconstrained system. Simulations and experiments on a real robot show the effectiveness of the proposed approach.
|
|
12:00-12:15, Paper TuAT3.5 | Add to My Program |
Active Learning of Reward Dynamics from Hierarchical Queries |
Basu, Chandrayee | UC Merced |
Bıyık, Erdem | Stanford University |
He, Zhixun | UC Merced |
Singhal, Mukesh | UC Merced |
Sadigh, Dorsa | Stanford University |
Keywords: Learning and Adaptive Systems, Optimization and Optimal Control, Probability and Statistical Methods
Abstract: Enabling robots to act according to human preferences across diverse environments is a crucial task, extensively studied by both roboticists and machine learning researchers. To achieve it, human preferences are often encoded by a reward function which the robot optimizes for. This reward function is generally static in the sense that it does not vary with time or the interactions. Unfortunately, such static reward functions do not always adequately capture human preferences, especially, in non-stationary environments: Human preferences change in response to the emergent behaviors of the other agents in the environment. In this work, we propose learning reward dynamics that can adapt in non-stationary environments with several interacting agents. We define reward dynamics as a tuple of reward functions, one for each mode of interaction, and mode-utility functions governing transitions between the modes. Reward dynamics thereby encodes not only different human preferences but also how the preferences change. Our contribution is in the way we adapt preference-based learning into a hierarchical approach that aims at learning not only reward functions but also how they evolve based on interactions. We derive a probabilistic observation model of how people will respond to the hierarchical queries. Our algorithm leverages this model to actively select hierarchical queries that will maximize the volume removed from a continuous hypothesis space of reward dynamics. We empirically demonstrate reward dynamics can match human preferences accurately.
|
|
12:15-12:30, Paper TuAT3.6 | Add to My Program |
Neural-Learning Trajectory Tracking Control of Flexible-Joint Robot Manipulators with Unknown Dynamics |
Chen, Shuyang | Rensselaer Polytechnic Institute |
Wen, John | Rensselaer Polytechnic Institute |
Keywords: Deep Learning in Robotics and Automation, Neural and Fuzzy Control
Abstract: Fast and precise motion control is important for industrial robots in manufacturing applications. However, some collaborative robots sacrifice precision for safety, particular for high motion speed. The performance degradation is caused by the inability of the joint servo controller to address the uncertain nonlinear dynamics of the robot arm, e.g., due to joint flexibility. We consider two approaches to improve the trajectory tracking performance through feedforward compensation. The first approach uses iterative learning control, with the gradient-based iterative update generated from the robot forward dynamics model. The second approach uses dynamic inversion to directly compensate for the robot forward dynamics. If the forward dynamics is strictly proper or is non-minimum-phase (e.g., due to time delays), its stable inverse would be non-causal. Both approaches require dynamical models. This paper presents results of using recurrent neural networks (RNNs) to approximate these dynamical models – forward dynamics in the first case, inverse dynamics (possibly non-causal) in the second case. We use the bi-directional RNN to capture the noncausality. The RNNs are trained based on a collection of commanded trajectories and the actual robot responses. We use a Baxter robot to evaluate the two approaches. The Baxter robot exhibits significant joint flexibility due to the series-elastic joint actuators. Both approaches achieve sizable improvement over the uncompensated robot motion, for both random joint trajectories and Cartesian motion. The inverse dynamics method is particularly attractive as it may be used to more accurately track a user input as in teleoperation.
|
|
TuAT4 Regular session, L1-R4 |
Add to My Program |
Award Session I |
|
|
Chair: Liu, Yunhui | Chinese University of Hong Kong |
Co-Chair: Amato, Nancy | University of Illinois |
|
11:00-11:15, Paper TuAT4.1 | Add to My Program |
Planning Reactive Manipulation in Dynamic Environments |
Schmitt, Philipp Sebastian | Siemens Corporate Technology |
Wirnshofer, Florian | Siemens AG |
Wurm, Kai M. | Siemens AG Corporate Technology |
v. Wichert, Georg | Siemens AG |
Burgard, Wolfram | University of Freiburg |
Keywords: Manipulation Planning
Abstract: When robots perform manipulation tasks, they need to determine their own movement, as well as how to make and break contact with objects in their environment. Reasoning about the motions of robots and objects simultaneously leads to a constrained planning problem in a high-dimensional state-space. Additionally, when environments change dynamically motions must be computed in real-time. To this end, we propose a feedback planner for manipulation. We model manipulation as constrained motion and use this model to automatically derive a set of constraint-based controllers. These controllers are used in a switching-control scheme, where the active controller is chosen by a reinforcement learning agent. Our approach is capable of addressing tasks with second-order dynamics, closed kinematic chains, and time-variant environments. We validated our approach in simulation and on a real, dual-arm robot. Extensive simulation of three distinct robots and tasks show a significant increase in robustness compared to a previous approach.
|
|
11:15-11:30, Paper TuAT4.2 | Add to My Program |
Bounded-Error LQR-Trees |
Ames, Barrett | Duke University |
Konidaris, George | Brown University |
Keywords: Optimization and Optimal Control, AI-Based Methods
Abstract: We present a feedback motion planning algorithm, Bounded-Error LQR-Trees, that leverages reinforcement learning theory to find a policy with a bounded amount of error. The algorithm composes locally valid linear-quadratic regulators (LQR) into a nonlinear controller, similar to how LQR-Trees constructs its policy, but minimizes the cost of the constructed policy by minimizing the Bellman Residual, which is estimated in the overlapping regions of LQR controllers. We prove a sample-based upper bound on the true Bellman Residual, and demonstrate a five-fold reduction in cost over previous methods on a simple underactuated nonlinear system.
|
|
11:30-11:45, Paper TuAT4.3 | Add to My Program |
Interaction-Aware Decision Making with Adaptive Strategies under Merging Scenarios |
Hu, Yeping | University of California, Berkeley |
Nakhaei, Alireza | Honda Research Institute USA |
Tomizuka, Masayoshi | University of California |
Fujimura, Kikuo | Honda Research Institute |
Keywords: Intelligent Transportation Systems, Autonomous Agents, Deep Learning in Robotics and Automation
Abstract: In order to drive safely and efficiently under merging scenarios, autonomous vehicles should be aware of their surroundings and make decisions by interacting with other road participants. Moreover, different strategies should be made when the autonomous vehicle is interacting with drivers having different level of cooperativeness. Whether the vehicle is on the merge-lane or main-lane will also influence the driving maneuvers since drivers will behave differently when they have the right-of-way than otherwise. Many traditional methods have been proposed to solve decision making problems under merging scenarios. However, these works either are incapable of modeling complicated interactions or require implementing hand-designed rules which cannot properly handle the uncertainties in real-world scenarios. In this paper, we proposed an interaction-aware decision making with adaptive strategies (IDAS) approach that can let the autonomous vehicle negotiate the road with other drivers by leveraging their cooperativeness under merging scenarios. A single policy is learned under the multi-agent reinforcement learning (MARL) setting via the curriculum learning strategy, which enables the agent to automatically infer other drivers' various behaviors and make decisions strategically. A masking mechanism is also proposed to prevent the agent from exploring states that violate common sense of human judgment and increase the learning efficiency. An exemplar merging scenario was used to implement and examine the proposed method.
|
|
11:45-12:00, Paper TuAT4.4 | Add to My Program |
Bee+: A 95-Mg Four-Winged Insect-Scale Flying Robot Driven by Twinned Unimorph Actuators |
Yang, Xiufeng | University of Southern California |
Chen, Ying | University of Southern California |
Chang, Longlong | University of Southern California |
Calderon, Ariel, A | University of Southern California |
Perez-Arancibia, Nestor O | University of Southern California (USC) |
Keywords: Micro/Nano Robots, Biologically-Inspired Robots, Aerial Systems: Mechanics and Control
Abstract: We introduce Bee +, a 95-mg four-winged microrobot with improved controllability and open-loop-response characteristics with respect to those exhibited by state-of-the-art two-winged microrobots with the same size and similar weight (i.e., the 75-mg Harvard RoboBee). The key innovation that made possible the development of Bee + is the introduction of an extremely light (28-mg) pair of twinned unimorph actuators, which enabled the design of a new microrobotic mechanism that flaps four wings independently. A first main advantage of the proposed design, compared to those of two-winged flyers, is that by increasing the number of actuators from two to four, the number of direct control inputs increases from three to four when simple sinusoidal excitations are employed. A second advantage of Bee + is that its four-wing configuration and flapping mode naturally damp the rotational disturbances that commonly affect the yaw degree of freedom of two-winged microrobots. In addition, the proposed design greatly reduces the complexity of the associated fabrication process compared to those of other microrobots, as the unimorph actuators are fairly easy to build. Lastly, we hypothesize that given the relatively low wing-loading affecting their flapping mechanisms, the life expectancy of Bee +s must be considerably higher than those of the two-winged counterparts. The functionality and basic capabilities of the robot are demonstrated through a set of simple control experiments.
|
|
TuAT5 Regular session, L1-R5 |
Add to My Program |
Robot Safety |
|
|
Chair: Wen, Li | Beihang University |
Co-Chair: Tang, Chaoquan | China University of Mining and Technology |
|
11:00-11:15, Paper TuAT5.1 | Add to My Program |
Lambda-Field: A Continuous Counterpart of the Bayesian Occupancy Grid for Risk Assessment |
Laconte, Johann | Institut Pascal |
Debain, Christophe | Irstea |
Chapuis, Roland | Institut Pascal |
Pomerleau, Francois | Laval University |
Aufrere, Romuald | Clermont Auvergne University |
Keywords: Robot Safety, Autonomous Vehicle Navigation, Intelligent Transportation Systems
Abstract: In a context of autonomous robots, one of the most important tasks is to ensure the safety of the robot and its surrounding. The risk of navigation is usually said to be the probability of collision. This notion of risk is not well defined in the literature, especially when dealing with occupancy grids. The Bayesian occupancy grid is the most used method to deal with complex environments. However, this is not fitted to compute the risk along a path by its discrete nature. In this article, we present a new way to store the occupancy of the environment that allows the computation of risk along a given path. We then define the risk as the force of collision that would occur for a given obstacle. Using this framework, we are able to generate navigation paths ensuring the safety of the robot.
|
|
11:15-11:30, Paper TuAT5.2 | Add to My Program |
Online Active Safety for Robotic Manipulators |
Singletary, Andrew | California Institute of Technology |
Nilsson, Petter | California Institute of Technology |
Gurriet, Thomas | California Institute of Technology |
Ames, Aaron | California Institute of Technology |
Keywords: Robot Safety, Industrial Robots, Optimization and Optimal Control
Abstract: Future manufacturing environments will see an increased need for cooperation between humans and machines. In this paper we propose a method that allows industrial manipulators to safely operate around humans. This approach guarantees that the manipulator will never collide with human operators while performing its normal tasks. This is done in an near-optimal way by considering how forward reachable sets of human operators grow with time, and by continuously updating these reachable sets based on current position estimates of the operators near the robot. An implicit active set invariance filter is then used to constrain the system---in a minimally invasive way---to stay in the complement of that forward reachable set. We demonstrate this approach in simulation on an industrial robotic arm: the ABB IRB 6640.
|
|
11:30-11:45, Paper TuAT5.3 | Add to My Program |
Black Block Recorder: Immutable Black Box Logging for Robots Via Blockchain |
White, Ruffin | University of California San Diego |
Caiazza, Gianluca | Ca Foscari University of Venice |
Cortesi, Agostino | Università Ca' Foscari Venezia |
Cho, Young Im | Gachon University |
Christensen, Henrik Iskov | UC San Diego |
Keywords: Robot Safety, Networked Robots, Software, Middleware and Programming Environments
Abstract: Event data recording is crucial in robotics research, providing prolonged insights into a robot's situational understanding, progression of behavioral state, and resulting outcomes. Such recordings are invaluable when debugging complex robotic applications or profiling experiments ex post facto. As robotic developments mature into production, both the roles and requirements of event logging will broaden, to include serving as evidence for auditors and regulators investigating accidents or fraud. Given the growing number of high profile public incidents involving self-driving automotives resulting in fatality and regulatory policy making, it is paramount that the integrity, authenticity and non-repudiation of such event logs are maintained to ensure accountability. Being mobile cyber-physical systems, robots present new threats and vulnerabilities beyond traditional IT: unsupervised physical system access or postmortem collusion between robot and OEM could result in the truncation or alteration of prior records. In this work, we address immutablization of log records via integrity proofs and distributed ledgers with special considerations for mobile and public service robot deployments.
|
|
11:45-12:00, Paper TuAT5.4 | Add to My Program |
DISC: A Large-Scale Virtual Dataset for Simulating Disaster Scenarios |
Jeon, Hae-Gon | Carnegie Mellon University |
Im, Sunghoon | KAIST |
Lee, Byeong-Uk | KAIST |
Choi, Dong-Geol | KAIST |
Hebert, Martial | CMU |
Kweon, In So | KAIST |
Keywords: Robot Safety, Performance Evaluation and Benchmarking, Computer Vision for Other Robotic Applications
Abstract: In this paper, we present the first large-scale synthetic dataset for visual perception in disaster scenarios, and analyze state-of-the-art methods for multiple computer vision tasks with reference baselines. We simulated before and after disaster scenarios such as fire and building collapse for fifteen different locations in realistic virtual worlds. The dataset consists of more than 300K high-resolution stereo image pairs, all annotated with ground-truth data for semantic segmentation, depth, optical flow, surface normal estimation and camera pose estimation. To create realistic disaster scenes, we manually augmented the effects with 3D models using physical-based graphics tools. We use our dataset to train state-of-the-art methods and evaluate how well these methods can recognize the disaster situations and produce reliable results on virtual scenes as well as real-world images. The results obtained from each task are then used as inputs to the proposed visual odometry network for generating 3D maps of buildings on fire. Finally, we discuss challenges for future research.
|
|
12:00-12:15, Paper TuAT5.5 | Add to My Program |
The Role of Robot Payload in the Safety Map Framework |
Hamad, Mazin | Technical University of Munich (TUM) |
Mansfeld, Nico | German Aerospace Center (DLR) |
Abdolshah, Saeed | Nagoya University |
Haddadin, Sami | Technical University of Munich |
Keywords: Robot Safety, Physical Human-Robot Interaction, Gripper and Other End-Effectors
Abstract: In the practical application of robots, usually certain tools are attached to the robot to manipulate objects and carry out various tasks. Besides adding gravitational load to the robot that results in large joint torques, such payloads may influence the collision safety characteristics through changing surface curvature properties, effective mass and robot speed along a given direction of motion. In this paper, we evaluate the contribution of a known, unactuated payload that is attached to the robot end effector to the robot inertial parameters and maximum task velocity. The proposed mass update approach relies on the analysis of the kinetic energy matrices. The velocity maximization is tackled by formulating static optimization problems with different constraints on angular motion of the end-effector. Finally, we present simulation results with a PUMA 560 robot that has an exemplary payload attached to its end-effector, which enables us to check the validity and accuracy of the developed methods. We further show the application for updating the effective mass and obtaining velocity optimization solutions in the context of Safety Maps and also via different geometrical visualizations.
|
|
12:15-12:30, Paper TuAT5.6 | Add to My Program |
Concept and Validation of a Large-Scale Human-Machine Safety System Based on Real-Time UWB Indoor Localization |
Wang, Wei | Cooperate Research of Robert Bosch |
Zeng, Zhuoqi | Bosch (China) Investment Ltd |
Ding, Wan | Bosch (China) Investment Ltd |
Yu, Huajun | Bosch (China) Investment Ltd |
Rose, Hannes | Bosch (China) Investment Ltd |
Keywords: Robot Safety, Product Design, Development and Prototyping, Performance Evaluation and Benchmarking
Abstract: In production line, the conventional industrial robots and automatic machines require machinery safety protection to guarantee the safety of human operators. A scalable and easy-to-configure safety system concept called “Real-time Safety Virtual Positioning” (RSVP) is proposed, which could act as potentially key enabler for agile production systems by eliminating fixed safety installation and thus increasing productivity and flexibility. The RSVP provides easy access to robots of automatic assembly lines in plants, e.g., automotive OEM, and supports virtualization and transparent to fully automatic and semi-automatic assembly line by knowing the position of persons and relevant objects (e.g., tools, finished and/or semi-finished goods, and materials). The focus of the paper will discuss the functional safety certification realization (concept approved by TÜV (Technical Inspection Association)) and validation details of indoor localization-based safety system developments. The detailed strategy of the functional safety requirements, danger diagnosis and reaction approach, communication among safety controller, robot and machine, and safe failure reaction are listed and analyzed. The physical hardware framework and software architecture of the safety system are built and developed with the 1oo2-architecture according to the Performance Levels (PL d) of ISO 13849 and the Safety Integrity Levels (SIL 2) of IEC 61508. The implicated algorithms and data process of the UWB-based (Ultra-Wide Band) indoor localization system are introduced. The safety system concept is validated and verified in an ABS (Anti-lock Braking System) production line with human-robot co-existence environment.
|
|
TuAT6 Regular session, L1-R6 |
Add to My Program |
Aerial Robotics I |
|
|
Chair: He, Yuqing | Shenyang Institute of Automation, Chinese Academy OfSciences |
Co-Chair: Ollero, Anibal | University of Seville |
|
11:00-11:15, Paper TuAT6.1 | Add to My Program |
Small-Scale Compliant Dual Arm with Tail for Winged Aerial Robots (I) |
Suarez, Alejandro | University of Seville |
Pérez García, Manuel | University of Seville |
Heredia, Guillermo | University of Seville |
Ollero, Anibal | University of Seville |
Keywords: Aerial Systems: Mechanics and Control, Aerial Systems: Applications
Abstract: Winged aerial robots represent an evolution of aerial manipulation robots, replacing the multirotor vehicles by fixed or flapping wing platforms. The development of this morphology is motivated in terms of efficiency, endurance and safety in some inspection operations where multirotor platforms may not be suitable. This paper presents a first prototype of compliant dual arm as preliminary step towards the realization of a winged aerial robot capable of perching and manipulating with the wings folded. The dual arm provides 6 DOF (degrees of freedom) for end effector positioning in a human-like kinematic configuration, with a reach of 25 cm (half-scale w.r.t. the human arm), and 0.2 kg weight. The prototype is built with micro metal gear motors, measuring the joint angles and the deflection with small potentiometers. The paper covers the design, electronics, modeling and control of the arms. Experimental results in test-bench validate the developed prototype and its functionalities, including joint position and torque control, bimanual grasping, the dynamic equilibrium with the tail, and the generation of 3D maps with laser sensors attached at the arms.
|
|
11:15-11:30, Paper TuAT6.2 | Add to My Program |
Design and Implementation of a Contact Aerial Manipulator System for Glass-Wall Inspection Tasks (I) |
Meng, Xiangdong | Shenyang Institute of Automation, Chinese Academy Sciences |
He, Yuqing | Shenyang Institute of Automation, Chinese Academy of Sciences |
Han, Jianda | Shenyang Institute of Automation, Chinese AcademyofSciences |
Keywords: Aerial Systems: Applications, Aerial Systems: Mechanics and Control, Contact Modelling
Abstract: Glass curtain walls have been widely used in modern architecture. This makes it urgent to inspect and clean these glasses at regular intervals. Up to now, most of these work is performed by workers, which is expensive and inefficient. Therefore, a novel robot—the contact aerial manipulator system—is developed. The new designed system presents priorities in the aspects of high flexibility and easy operation. In this paper, the system mechanical structure is first introduced. Subsequently, the hybrid force/motion control framework is utilized to realize the precise and steady motion on the two-dimensional plane and maintain a certain sustained contact force, simultaneously. Finally, two flight experiments (including continuous square-wave trajectory tracking and aerial drawing task) are performed and the results indicate that the developed contact aerial manipulator works and presents good performance.
|
|
11:30-11:45, Paper TuAT6.3 | Add to My Program |
Achievement of Online Agile Manipulation Task for Aerial Transformable Multilink Robot |
Shi, Fan | The University of Tokyo |
Zhao, Moju | The University of Tokyo |
Anzai, Tomoki | The University of Tokyo |
Ito, Keita | The University of Tokyo |
Chen, Xiangyu | The University of Tokyo |
Okada, Kei | The University of Tokyo |
Inaba, Masayuki | The University of Tokyo |
Keywords: Aerial Systems: Applications, Aerial Systems: Mechanics and Control
Abstract: Transformable aerial robots are favorable in aerial manipulation tasks for their flexible ability to change configuration during the flight. By assuming robot keeping in the mild motion, the previous researches sacrifice aerial agility to simplify the complex non-linear system into a single rigid body with a linear controller. In this paper, we present a framework towards agile swing motion for the transformable multi-links aerial robot. We introduce a computational-efficient non-linear model predictive controller and joints motion primitive framework to achieve agile transforming motions and validate with a novel robot named HYRURS-X. Finally, we implement our framework under a table tennis task to validate the online and agile performance.
|
|
11:45-12:00, Paper TuAT6.4 | Add to My Program |
Towards a Robust Aerial Cinematography Platform: Localizing and Tracking Moving Targets in Unstructured Environments |
Bonatti, Rogerio | Carnegie Mellon University |
Ho, Cherie | Carnegie Mellon University |
Wang, Wenshan | Shanghai Jiao Tong University, Research Institute of Robotics |
Choudhury, Sanjiban | University of Washington |
Scherer, Sebastian | Carnegie Mellon University |
Keywords: Aerial Systems: Applications, Aerial Systems: Perception and Autonomy
Abstract: The use of drones for aerial cinematography has revolutionized several applications and industries that require live and dynamic camera viewpoints such as entertainment, sports, and security. However, safely controlling a drone while filming a moving target usually requires multiple expert human operators; hence the need for an autonomous cinematographer. Current approaches have severe real-life limitations such as requiring fully scripted scenes, high-precision motion-capture systems or GPS tags to localize targets, and prior maps of the environment to avoid obstacles and plan for occlusion. In this work, we overcome such limitations and propose a complete system for aerial cinematography that combines: (1) a vision-based algorithm for target localization; (2) a real-time incremental 3D signed-distance map algorithm for occlusion and safety computation; and (3) a real-time camera motion planner that optimizes smoothness, collisions, occlusions and artistic guidelines. We evaluate robustness and real-time performance in series of field experiments and simulations by tracking dynamic targets moving through unknown, unstructured environments. Finally, we verify that despite removing previous limitations, our system achieves state-of-the-art performance.
|
|
12:00-12:15, Paper TuAT6.5 | Add to My Program |
Aerial Animal Biometrics: Individual Friesian Cattle Recovery and Visual Identification Via an Autonomous UAV with Onboard Deep Inference |
Andrew, William | University of Bristol |
Greatwood, Colin | University of Bristol |
Burghardt, Tilo | University of Bristol |
Keywords: Aerial Systems: Applications, Agricultural Automation, Robotics in Agriculture and Forestry
Abstract: This paper describes a computationally-enhanced M100 UAV platform with an onboard deep learning inference system for integrated computer vision and navigation. The system is able to autonomously find and visually identify by coat pattern individual Holstein Friesian cattle in freely moving herds. We propose an approach that utilises three deep convolutional neural network architectures running live onboard the aircraft: (1) a YOLOv2-based species detector, (2) a dual-stream deep network delivering exploratory agency, and (3) an InceptionV3-based biometric long-term recurrent convolutional network for individual animal identification. We evaluate the performance of each of the components offline, and also online via real-world field tests comprising 147 minutes of autonomous low altitude flight in a farm environment over a dispersed herd of 17 heifer dairy cows. We report error free identification performance on this online experiment. The presented proof-of-concept system is the first of its kind. It represents a practical step towards autonomous biometric identification of individual animals from the air in open pasture environments for tag-less AI support in farming and ecology.
|
|
12:15-12:30, Paper TuAT6.6 | Add to My Program |
Geometric and Physical Constraints for Drone-Based Head Plane Crowd Density Estimation |
Liu, Weizhe | EPFL |
Lis, Krzysztof | EPFL |
Salzmann, Mathieu | EPFL CVLab |
Fua, Pascal | EPFL |
Keywords: Aerial Systems: Applications, Computer Vision for Other Robotic Applications, Recognition
Abstract: State-of-the-art methods for counting people in crowded scenes rely on deep networks to estimate crowd density in the image plane. While useful for this purpose, this image- plane density has no immediate physical meaning because it is subject to perspective distortion. This is a concern in sequences acquired by drones because the viewpoint changes often. This distortion is usually handled implicitly by either learning scale- invariant features or estimating density in patches of different sizes, neither of which accounts for the fact that scale changes must be consistent over the whole scene. In this paper, we explicitly model the scale changes and reason in terms of people per square-meter. We show that feeding the perspective model to the network allows us to enforce global scale consistency and that this model can be obtained on the fly from the drone sensors. In addition, it also enables us to enforce physically-inspired temporal consistency constraints that do not have to be learned. This yields an algorithm that outperforms state-of-the-art methods in inferring crowd density from a moving drone camera especially when perspective effects are strong.
|
|
TuAT7 Regular session, L1-R7 |
Add to My Program |
Computer Vision and Applications I |
|
|
Chair: Rameau, Francois | KAIST, RCV Lab |
Co-Chair: Liu, Peilin | Shanghai Jiao Tong Universit |
|
11:00-11:15, Paper TuAT7.1 | Add to My Program |
Infrastructure-Free NLoS Obstacle Detection for Autonomous Cars |
Naser, Felix Maximilian | Massachusetts Institute of Technology (MIT) |
Gilitschenski, Igor | Massachusetts Institute of Technology |
Amini, Alexander | Massachusetts Institute of Technology |
Christina, Liao | MIT |
Rosman, Guy | Massachusetts Institute of Technology |
Karaman, Sertac | Massachusetts Institute of Technology |
Rus, Daniela | MIT |
Keywords: Computer Vision for Other Robotic Applications, Object detection, segmentation, categorization, Autonomous Vehicle Navigation
Abstract: Current perception systems mostly require direct line of sight to anticipate and ultimately prevent potential collisions at intersections with other road users. We present a fully integrated autonomous system capable of detecting shadows or weak illumination changes on the ground caused by a dynamic obstacle in NLoS scenarios. This additional virtual sensor ``ShadowCam'' extends the signal range utilized so far by computer-vision ADASs. We show that (1) our algorithm maintains the mean classification accuracy of around 70% even when it doesn't rely on infrastructure -- such as AprilTags -- as an image registration method. We validate (2) in real-world experiments that our autonomous car driving in night time conditions detects a hidden approaching car earlier with our virtual sensor than with the front facing 2-D LiDAR.
|
|
11:15-11:30, Paper TuAT7.2 | Add to My Program |
Action Recognition Based on 3D Skeleton and RGB Frame Fusion |
Liu, Guiyu | Shanghai JitaoTong University |
Qian, Jiuchao | Shanghai Jiao Tong University |
Wen, Fei | Shanghai Jiao Tong University |
Zhu, Xiaoguang | Shanghai Jiao Tong University |
Ying, Rendong | Shanghai Jiao Tong University |
Liu, Peilin | Shanghai Jiao Tong Universit |
Keywords: Computer Vision for Other Robotic Applications, RGB-D Perception, Cognitive Human-Robot Interaction
Abstract: Action recognition has wide applications in assisted living, health monitoring, surveillance, and human-computer interaction. In traditional action recognition methods, RGB video-based ones are effective but computationally inefficient, while skeleton-based ones are computationally efficient but do not make use of low-level detail information. This work considers action recognition based on a multimodal fusion between the 3D skeleton and the RGB image. We design a neural network that uses a 3D skeleton sequence and a single middle frame from an RGB video as input. Specifically, our method picks up one frame in a video and extracts spatial features from it using two attention modules, a self-attention module, and a skeleton-attention module. Further, temporal features are extracted from the skeleton sequence via a BI-LSTM subnetwork. Finally, the spatial features and the temporal features are combined via a feature fusion network for action classification. A distinct feature of our method is that it uses only a single RGB frame rather than an RGB video. Accordingly, it has a light-weighted architecture and is more efficient than RGB video-based methods. Comparative evaluation on two public datasets, NTU-RGBD and SYSU, demonstrates that our method can achieve competitive performance compared with state-of-the-art methods.
|
|
11:30-11:45, Paper TuAT7.3 | Add to My Program |
Estimating Metric Scale Visual Odometry from Videos Using 3D Convolutional Networks |
Koumis, Alexander | University of Southern California |
Preiss, James | USC |
Sukhatme, Gaurav | University of Southern California |
Keywords: Computer Vision for Other Robotic Applications, Visual Learning, Aerial Systems: Perception and Autonomy
Abstract: We present an end-to-end deep learning approach for performing metric scale-sensitive regression tasks such visual odometry with a single camera and no additional sensors. We propose a novel 3D convolutional architecture, 3DC-VO, that can leverage temporal relationships over a short moving window of images to estimate linear and angular velocities. The network makes local predictions on stacks of images that can be integrated to form a full trajectory. We apply 3DC-VO to the KITTI visual odometry benchmark and the task of estimating a pilot's control inputs from a first-person video of a quadrotor flight. Our method exhibits increased accuracy relative to comparable learning-based algorithms trained on monocular images. We also show promising results for quadrotor control input prediction when trained on a new dataset collected with a UAV simulator.
|
|
11:45-12:00, Paper TuAT7.4 | Add to My Program |
Unsupervised Traffic Accident Detection in First-Person Videos |
Yao, Yu | University of Michigan |
Xu, Mingze | Indiana University |
Wang, Yuchen | Indiana University |
Crandall, David | Indiana University |
Atkins, Ella | University of Michigan |
Keywords: Computer Vision for Transportation, Computer Vision for Automation
Abstract: Recognizing abnormal events such as traffic violations and accidents in natural driving scenes is essential for successful autonomous driving and advanced driver assistance systems. However, most work on video anomaly detection suffers from two crucial drawbacks. First, they assume cameras are fixed and videos have static backgrounds, which is reasonable for surveillance applications but not for vehicle-mounted cameras. Second, they pose the problem as one-class classification, relying on arduously hand-labeled training datasets that limit recognition to anomaly categories that have been explicitly trained. This paper proposes an unsupervised approach for traffic accident detection in first-person (dashboard-mounted camera) videos. Our major novelty is to detect anomalies by predicting the future locations of traffic participants and then monitoring the prediction accuracy and consistency metrics with three different strategies. We evaluate our approach using a new dataset of diverse traffic accidents, AnAn Accident Detection (A3D), as well as another publicly-available dataset. Experimental results show that our approach outperforms the state-of-the-art. Code and the dataset developed in this work are available at: https://github.com/MoonBlvd/tad-IROS2019
|
|
12:00-12:15, Paper TuAT7.5 | Add to My Program |
Vehicular Multi-Camera Sensor System for Automated Visual Inspection of Electric Power Distribution Equipment |
Park, Jinsun | KAIST |
Shin, Ukcheol | KAIST(Korea Advanced Institute of Science and Technology) |
Shim, Gyu Min | KAIST |
Joo, Kyungdon | Korea Advanced Institute of Science and Technology (KAIST) |
Rameau, Francois | KAIST, RCV Lab |
Kim, Junhyuck | KEPCO |
Choi, Dong-Geol | KAIST |
Kweon, In So | KAIST |
Keywords: Computer Vision for Other Robotic Applications, Computer Vision for Automation, Deep Learning in Robotics and Automation
Abstract: In this paper, we present a multi-camera sensor system along with its control algorithm for automated visual inspection from a moving vehicle. To accomplish this task, we propose a unique hardware configuration consisting of a frontal stereo vision system, six lateral cameras motorized to tilt, and a GPS/IMU sensor mounted on the roof of a car. From the frontal stereo system, we detect electric poles and estimate their corresponding 3D positions. Based on this 3D estimation, the tilt angles of the motorized lateral cameras are controlled in real-time to capture high resolution images of the equipment - typically installed a few meters above the road surface. In addition, inertial odometry information from the GPS/IMU module is utilized for pose estimation, object localization, and re-identification among cameras. Experimental results demonstrate the efficiency and robustness of our system for automated electric equipment maintenance, which can reduce human effort significantly.
|
|
12:15-12:30, Paper TuAT7.6 | Add to My Program |
Representation Learning Via Parallel Subset Reconstruction for 3D Point Cloud Generation |
Matsuzaki, Kohei | KDDI Research, Inc |
Tasaka, Kazuyuki | KDDI Research Inc |
Keywords: Computer Vision for Other Robotic Applications, Deep Learning in Robotics and Automation, Computational Geometry
Abstract: Three-dimensional (3D) point cloud processing has attracted a great deal of attention in computer vision, robotics, and the machine learning community because of significant progress in deep neural networks on 3D data. Another trend in the community is learning of generative models based on generative adversarial networks. In this paper, we propose a framework for 3D point cloud generation based on a combination of auto-encoders and generative adversarial networks. The framework first trains auto-encoders to learn latent representations, and then trains generative adversarial networks in the learned latent space. We focus on improving the training method for auto-encoders in order to generate 3D point clouds with higher fidelity and coverage. We add parallel sub-decoders that reconstruct subsets of the input point cloud. In order to construct these subsets, we introduce a point sampling algorithm that imposes a method to sample spatially localized point sets. These local subsets are utilized to measure local reconstruction losses. We train auto-encoders to learn an effective latent representation for both global and local shape reconstruction based on the multi-task learning strategy. Furthermore, we add global and local adversarial losses to generate more plausible point clouds. Quantitative and qualitative evaluations demonstrate that the proposed method outperforms state-of-the-art method on the task of 3D point cloud generation.
|
|
TuAT8 Cutting Edge Forum, LG-R8 |
Add to My Program |
Autonomous Driving: Contributions from Intelligent Robotics, AI and ITS |
|
|
Chair: Ang Jr, Marcelo H | National University of Singapore |
Co-Chair: Martinet, Philippe | INRIA |
|
11:00-11:15, Paper TuAT8.1 | Add to My Program |
Autonomous Driving: Contributions from Intelligent Robotics, AI and ITS |
Martinet, Philippe | INRIA |
Laugier, Christian | INRIA |
Stiller, Christoph | Karlsruhe Institute of Technology |
Sotelo Vázquez, Miguel Ángel | University of Alcalá |
Ang Jr, Marcelo H | National University of Singapore |
|
|
TuAT9 Regular session, LG-R9 |
Add to My Program |
Social Human-Robot Interaction I |
|
|
Chair: Guimarães Macharet, Douglas | Universidade Federal De Minas Gerais |
Co-Chair: Lim, Yoonseob | Korea Institute of Science and Technology |
|
11:00-11:15, Paper TuAT9.1 | Add to My Program |
Identifying Opportunities for Relationship-Focused Robotic Interventions in Strained Hierarchical Relationships |
Pettinati, Michael | Georgia Institute of Technology |
Arkin, Ronald | Georgia Tech |
Keywords: Social Human-Robot Interaction, Agent-Based Systems
Abstract: When disagreements arise in hierarchical relationships, relationship members sometimes prefer conflict management strategies that avoid or quickly end the overt conflict even if the relationship is left in a state of dissatisfaction. Our lab has proposed that a peripheral robotic agent may be able to support these types of relationships during conflict. In this paper, we present the results of an IRB- approved human-robot interaction study that examines how the members of a hierarchical relationship involved in conflict respond to the presence of an unengaged robot. This study serves as a baseline for additional studies. The unengaged robot appears to have a minimal influence on the interaction. The observed conflicts followed the patterns typically described in mediation literature. Our lab previously proposed a computational model to identify weakness and alienation in these relationships. We discuss a partial implementation of this model, and its ability to recognize problems in certain relationships within the data collected. Based on our observations, and the performance of the model’s partial implementation, we suggest considerations that need to be made for an intervening robotic agent.
|
|
11:15-11:30, Paper TuAT9.2 | Add to My Program |
Fast Adaptation with Meta-Reinforcement Learning for Trust Modelling in Human-Robot Interaction |
Gao, Yuan | Uppsala University |
Sibirtseva, Elena | KTH Royal Institute of Technology |
Castellano, Ginevra | Uppsala University |
Kragic, Danica | KTH |
Keywords: Social Human-Robot Interaction, Cognitive Human-Robot Interaction, Learning and Adaptive Systems
Abstract: In socially assistive robotics, an important research area is the development of adaptation techniques and their effect on human-robot interaction. We present a meta-learning based policy gradient method for addressing the problem of adaptation in human-robot interaction and also investigate its role as a mechanism for trust modelling. By building an escape room scenario in mixed reality with a robot, we test our hypothesis that bi-directional trust can be influenced by different adaptation algorithms. We found that our proposed model increased the perceived trustworthiness of the robot and influenced the dynamics of gaining human's trust. Additionally, participants evaluated that the robot perceived them as more trustworthy during the interactions with the meta-learning based adaptation compared to the previously studied statistical adaptation model.
|
|
11:30-11:45, Paper TuAT9.3 | Add to My Program |
Are You with Me? Determining the Association of Individuals and the Collective Social Space |
D. G. Silva, Alan | UFMG |
Guimarães Macharet, Douglas | Universidade Federal De Minas Gerais |
Keywords: Social Human-Robot Interaction, Cognitive Human-Robot Interaction, Semantic Scene Understanding
Abstract: The increasing use of autonomous mobile robots in different parts of society, and not restricted only to industrial environments, makes it important to propose techniques that will allow them to behave in the most socially acceptable way as possible. In most real-world scenarios, individuals in the environment are interacting with each other and are arranged into groups. Therefore, it is paramount the proposition of techniques to efficiently and correctly identify and represent such groups. This information can be useful in different tasks such as approaching and initiating an interaction, escorting, and the navigation itself. In this work, we propose a novel graph-based approach to evaluate the possible association of individuals in the environment based on their position and body orientation. Next, based on this association, we propose a representation of the combined social space of individuals in the same group. The methodology was evaluated using synthetic and real-world datasets, showing that it achieves results comparable to or better than the state-of-the-art.
|
|
11:45-12:00, Paper TuAT9.4 | Add to My Program |
Are You Hearing or Listening? the Effect of Task Performance in Verbal Behavior with Smart Speaker |
Park, Chaewon | Korea Institute of Science and Technology |
Choi, Jongsuk | Korea Inst. of Sci. and Tech |
Sung, Jee Eun | Ewha Womans University |
Lim, Yoonseob | Korea Institute of Science and Technology |
Keywords: Social Human-Robot Interaction, Cognitive Human-Robot Interaction, Service Robots
Abstract: Human has an ability to adjust utterance depending on the state of interlocutor. In this study, we explore the verbal behaviors of human through interaction with two smart speakers that have different level of task competence. We analyzed (1) linguistic behaviors appeared in user’s utterance, (2) length of the uttered speech, and (3) required pragmatics skills to understand the user’s intent. As a result, there were no significant difference in linguistic behaviors and length of the speech while user interacts with speakers with different task competence. In addition, various pragmatics elements were equally utilized and especially, implied intentions were frequently observed in user’s short utterance even under simple interaction scenarios.
|
|
12:00-12:15, Paper TuAT9.5 | Add to My Program |
The Robot Show Must Go On: Effective Responses to Robot Failures |
Fallatah, Abrar | Oregon State University |
Urann, Jeremy | Oregon State University |
Knight, Heather | Oregon State University |
Keywords: Social Human-Robot Interaction, Entertainment Robotics, Cognitive Human-Robot Interaction
Abstract: This paper consists of a failure analysis of two robot performance productions. Both included three-week rehearsal periods, and culminated in live performances including both robots and humans. To develop these productions, a theater artist collaborated with a robotics lab to develop, (1) a narrative dance performed live on stage, and, (2) an improvisational performance in a public space. While the interdisciplinary team did not set out to explore robot failures, during the eighteen rehearsals and two live performances, failures played an ever-present role. This paper presents the technical and choreographic failures encountered, and details strategies for addressing, planning for, and rehearsing responses to robot failures on stage. In addition to scaffolding future robot theater performances, we discuss how these strategies apply to other customer- and audience-facing robots, including sponsor demos. The on-stage exploration of robot chairs and human performers also suggests that humans can conceptualize minimal robots as both characters and props, moving fluidly from one to the other. We hope these insights ensure that future audiences will *want* the robot shows to go on, as well as expand ideas about the types of robots that can be cast in future human-robot productions.
|
|
12:15-12:30, Paper TuAT9.6 | Add to My Program |
Right of Way, Assertiveness and Social Recognition in Human-Robot Doorway Interaction |
Thomas, Jack | Simon Fraser University |
Vaughan, Richard | Simon Fraser University |
Keywords: Social Human-Robot Interaction
Abstract: We expand on previous work for negotiating human-robot navigation contention around doorways to produce a more socially-compliant autonomous robot behaviour. Our goal is to improve the integration of robots navigating in human environments by eliciting human recognition of the robot's right of way. This is achieved by incorporating feedback from a user study of our previous system to create a more communicative, reciprocal, and assertive behaviour. Our contribution includes both the updated behaviour and a new user study that evaluates and compares the system to its predecessor. Results show that participants are more likely to respect the robot's right of way given the new robot behaviour, but their responses also highlight the challenges of socially integrating robots into human spaces.
|
|
TuAT10 Regular session, LG-R10 |
Add to My Program |
SLAM I |
|
|
Chair: Roumeliotis, Stergios | University of Minnesota |
Co-Chair: Colosi, Mirco | Sapienza, University of Rome |
|
11:00-11:15, Paper TuAT10.1 | Add to My Program |
Active SLAM Using Connectivity Graphs As Priors |
Soragna, Alberto | Sapienza University of Rome |
Baldini, Marco | KUKA Roboter GmbH |
Joho, Dominik | University of Freiburg |
Kuemmerle, Rainer | KUKA Roboter GmbH |
Grisetti, Giorgio | Sapienza University of Rome |
Keywords: SLAM, Autonomous Vehicle Navigation
Abstract: Mobile robots can be considered completely autonomous if they embed active algorithms for Simultaneous Localization And Mapping (SLAM). This means that the robot is able to autonomously, or actively, explore and create a reliable map of the environment, while simultaneously estimating its pose. In this paper, we propose a novel framework to robustly solve the active SLAM problem, in scenarios in which some prior information about the environment is available in the form of a topo-metric graph. This information is typically available or can be easily developed in industrial environments, but it is usually affected by uncertainties. In particular, the distinguishing features of our approach are: the inclusion of prior information for solving the active SLAM problem; the exploitation of this information to pursue active loop closure; the on-line correction of the inconsistencies in the provided data. We present some experiments, that are performed in different simulated environments: the results suggest that our method improves on state-of-the-art approaches, as it is able to deal with a wide variety of possibly large uncertainties.
|
|
11:15-11:30, Paper TuAT10.2 | Add to My Program |
Map-Aware SLAM with Sparse Map Features |
Burger, Patrick | Bundeswehr University Munich |
Naujoks, Benjamin | UniBw Munich |
Wuensche, Hans J | UniBw Munich |
Keywords: SLAM, Localization, Autonomous Vehicle Navigation
Abstract: Localization is a key capability for autonomous vehicles. High-Definition maps are a popular method to represent the environment and to enable precise localization. However, the creation is very demanding and it is not always guaranteed to receive accurate map information especially for unstructured areas. In this paper, we introduce a novel probabilistic localization and mapping framework that brings together the advantages of sparse feature maps, multi-target tracking for landmark detection, probabilistic global vehicle localization and a graph-based formulation to achieve a consistent map. The front-end of our Simultaneous Localization and Mapping (SLAM) framework is based on Monte Carlo Localization (MCL). Our novel measurement model integrates a virtual topological Path-Map with sparse map features to obtain global localization. The graph-based back-end optimizes online the vehicle trajectory and the landmarks’ configuration to create a globally aligned map. Furthermore, our method allows weaker requirements in terms of accuracy of the sparse feature map as we represent the degree of uncertainty by means of probabilistic distribution. Additionally, the sparse map representation needs substantially less memory than other approaches, which is an advantage for autonomous vehicles. The framework has been tested and evaluated in real experiments for several autonomous runs. The results demonstrate the robustness of our system.
|
|
11:30-11:45, Paper TuAT10.3 | Add to My Program |
RISE-SLAM: A Resource-Aware Inverse Schmidt Estimator for SLAM |
Ke, Tong | University of Minnesota |
Wu, Kejian | University of Minnesota |
Roumeliotis, Stergios | University of Minnesota |
Keywords: SLAM, Localization, Mapping
Abstract: In this paper, we present the RISE-SLAM algorithm for performing visual-inertial simultaneous localization and mapping (SLAM), while improving estimation consistency. Specifically, in order to achieve real-time operation, existing approaches often assume previously-estimated states to be perfectly known, which leads to inconsistent estimates. Instead, based on the idea of the Schmidt-Kalman filter, which has processing cost linear in the size of the state vector but quadratic memory requirements, we derive a new consistent approximate method in the information domain, which has linear memory requirements and adjustable (constant to linear) processing cost. In particular, this method, the resource-aware inverse Schmidt estimator (RISE), allows trading estimation accuracy for computational efficiency. Furthermore, and in order to better address the requirements of a SLAM system during an exploration vs. a relocalization phase, we employ different configurations of RISE (in terms of the number and order of states updated) to maximize accuracy while preserving efficiency. Lastly, we evaluate the proposed RISE-SLAM algorithm on publicly-available datasets and demonstrate its superiority, both in terms of accuracy and efficiency, as compared to alternative visual-inertial SLAM systems.
|
|
11:45-12:00, Paper TuAT10.4 | Add to My Program |
Better Lost in Transition Than Lost in Space: SLAM State Machine |
Colosi, Mirco | Sapienza, University of Rome |
Haug, Sebastian | Robert Bosch GmbH |
Biber, Peter | Robert Bosch GmbH |
Arras, Kai Oliver | Bosch Research |
Grisetti, Giorgio | Sapienza University of Rome |
Keywords: SLAM, Localization, Mapping
Abstract: A Simultaneous Localization and Mapping (SLAM) system is a complex program consisting of several interconnected components with different functionalities such as optimization, tracking or loop detection. Whereas the literature addresses in detail how enhancing the algorithmic aspects of the individual components improves SLAM performance, the modal aspects, such as when to localize, relocalize or close a loop, are usually left aside. In this paper, we address the modal aspects of a SLAM system and show that the design of the modal controller has a strong impact on SLAM performance in particular in terms of robustness against unforeseen events such as sensor failures, perceptual aliasing or kidnapping. We preset a novel taxonomy for the components of a modern SLAM system, investigate their interplay and propose a highly modular architecture of a generic SLAM system using the Unified Modeling Language (UML) state machine formalism. The result, called SLAM state machine, is compared to the modal controller of several state-of-the-art SLAM systems and evaluated in two experiments. We demonstrate that our state machine handles unforeseen events much more robustly than the state-of-the-art systems.
|
|
12:00-12:15, Paper TuAT10.5 | Add to My Program |
Stereo Visual Inertial LiDAR Simultaneous Localization and Mapping |
Shao, Weizhao | Carnegie Mellon University |
Vijayarangan, Srinivasan | Carnegie Mellon University |
Li, Cong | Carnegie Mellon University |
Kantor, George | Carnegie Mellon University |
Keywords: SLAM, Localization, Mapping
Abstract: Simultaneous Localization and Mapping (SLAM) is a fundamental task to mobile and aerial robotics. LiDAR based systems have proven to be superior compared to vision based systems due to its accuracy and robustness. In spite of its superiority, pure LiDAR based systems fail in certain degenerate cases like traveling through a tunnel. We propose Stereo Visual Inertial LiDAR (VIL) SLAM that performs better on these degenerate cases and has comparable performance on all other cases. VIL-SLAM accomplishes this by incorporating tightly-coupled stereo visual inertial odometry (VIO) with LiDAR mapping and LiDAR enhanced visual loop closure. The system generates loop-closure corrected 6-DOF LiDAR poses in real-time and 1cm voxel dense maps near real-time. VIL-SLAM demonstrates improved accuracy and robustness compared to state-of-the-art LiDAR methods.
|
|
12:15-12:30, Paper TuAT10.6 | Add to My Program |
Fast and Incremental Loop Closure Detection Using Proximity Graphs |
An, Shan | JD.COM |
Che, Guangfu | Jd.com |
Zhou, Fangru | Jd.com |
Liu, Xianglong | Beihang University |
Ma, Xin | Shandong University |
Chen, Yu | JD.com |
Keywords: SLAM, Localization, Recognition
Abstract: Visual loop closure detection, which can be considered as an image retrieval task, is an important problem in SLAM (Simultaneous Localization and Mapping) systems. The frequently used bag-of-words (BoW) models can achieve high precision and moderate recall. However, the requirement for lower time costs and fewer memory costs for mobile robot applications is not well satisfied. In this paper, we propose a novel loop closure detection framework titled "FILD" (Fast and Incremental Loop closure Detection), which focuses on an on-line and incremental graph vocabulary construction for fast loop closure detection. The global and local features of frames are extracted using the Convolutional Neural Networks (CNN) and SURF on the GPU, which guarantee extremely fast extraction speeds. The graph vocabulary construction is based on one type of proximity graph, named Hierarchical Navigable Small World (HNSW) graphs, which is modified to adapt to this specific application. In addition, this process is coupled with a novel strategy for real-time geometrical verification, which only keeps binary hash codes and significantly saves on memory usage. Extensive experiments on several publicly available datasets show that the proposed approach can achieve fairly good recall at 100% precision compared to other state-of-the-art methods. The source code can be downloaded at https://github.com/AnshanTJU/FILD for further studies.
|
|
TuAT11 Regular session, LG-R11 |
Add to My Program |
Medical Robot: Design |
|
|
Chair: Yang, Guang-Zhong | Imperial College London |
Co-Chair: Gruijthuijsen, Caspar | KU Leuven |
|
11:00-11:15, Paper TuAT11.1 | Add to My Program |
An Open-Source 7-Axis, Robotic Platform to Enable Dexterous Procedures within CT Scanners |
Schreiber, Dimitri A. | University of California |
Shak, Daniel B. | University of California San Diego |
Norbash, Alexander M. | University of California San Diego |
Yip, Michael C. | University of California, San Diego |
Keywords: Medical Robots and Systems, Telerobotics and Teleoperation, Mechanism Design
Abstract: This paper describes the design, manufacture, and performance of a highly dexterous, low-profile, 7 Degree-of-Freedom (DOF) robotic arm for CT-guided percutaneous needle biopsy. Direct CT guidance allows physicians to localize tumours quickly; however, needle insertion is still performed by hand. This system is mounted to a fully active gantry superior to the patient's head and teleoperated by a radiologist. Unlike other similar robots, this robot's fully serial-link approach uses a unique combination of belt and cable drives for high-transparency and minimal-backlash, allowing for an expansive working area and numerous approach angles to targets all while maintaining a small in-bore cross-section less than 16cm^2. Simulations verified the system's expansive collision free work-space and ability to hit targets across the entire chest, as required for lung cancer biopsy. Targeting error is on average <1mm on a teleoperated accuracy task, illustrating the system's sufficient accuracy to perform biopsy procedures. The system is designed for lung biopsies due to the large working volume that is required for reaching peripheral lung lesions, though, with its large working volume and small in-bore cross-sectional area, the robotic system is effectively a general-purpose CT-compatible manipulation device for percutaneous procedures. Finally, with the considerable development time undertaken in designing a precise and flexible-use system and with the desire to reduce the burden of other researchers in developing algorithms for image-guided surgery, this system provides open-access, and to the best of our knowledge, is the first open-hardware image-guided biopsy robot of its kind.
|
|
11:15-11:30, Paper TuAT11.2 | Add to My Program |
A Handheld Master Controller for Robot-Assisted Microsurgery |
Zhang, Dandan | Imperial College London |
Guo, Yao | Imperial College London |
Chen, Junhong | Imperial College London |
Liu, Jindong | Imperial College London |
Yang, Guang-Zhong | Imperial College London |
Keywords: Medical Robots and Systems, Telerobotics and Teleoperation
Abstract: Accurate master-slave control is important for Robot-Assisted Microsurgery (RAMS). This paper presents a handheld master controller for the operation and training of RAMS. A 9-axis Inertial Measure Unit (IMU) and a micro camera are utilized to form the sensing system for the handheld controller. A new hybrid marker pattern is designed to achieve reliable visual tracking, which integrated QR codes, Aruco markers, and chessboard vertices. Real-time multi-sensor fusion is implemented to further improve the tracking accuracy. The proposed handheld controller has been verified on an in-house microsurgical robot to assess its usability and robustness. User studies were conducted based on a trajectory following task, which indicated that the proposed handheld controller had comparable performance with the Phantom Omni, demonstrating its potential applications in microsurgical robot control and training.
|
|
11:30-11:45, Paper TuAT11.3 | Add to My Program |
A Hybrid Active/Passive Wrist Approach for Increasing Virtual Fixture Stiffness in Comanipulated Robotic Minimally Invasive Surgery |
Gruijthuijsen, Caspar | KU Leuven |
Borghesan, Gianni | KU Leuven |
Reynaerts, Dominiek | Division Production Engineering, Machine Design andAutomation, K |
Vander Poorten, Emmanuel B | KU Leuven |
Keywords: Surgical Robotics: Laparoscopy, Haptics and Haptic Interfaces, Optimization and Optimal Control
Abstract: In Minimally Invasive Surgery (MIS), the incision point acts as a fulcrum about which the surgical instrument pivots. Most robotic MIS systems foresee procedures to carefully align the robot with the incision in the patient. Such procedures disrupt the normal workflow. This hampers the clinical translation of such assistive technology. In contrast, Backdrivable, Wristed (BW) robots, especially when operated as comanipulation devices, have a minimal impact on the traditional surgical workflow in MIS, as they are readily equipped with algorithms that automate fulcrum point estimation. To improve safety, accuracy and/or task completion time, virtual fixtures can be implemented at the distal instrument tip, using an assistive BW system. This letter formalizes the problems related to such distal virtual fixtures. It then develops a hybrid active/passive wrist control strategy to improve the performance of virtual fixtures with BW systems while minimizing stability issues. Aside from an analytical example, experimental validation is added to show that the new algorithm increases the achievable stiffness of distal virtual fixtures.
|
|
11:45-12:00, Paper TuAT11.4 | Add to My Program |
Three-Degrees-Of-Freedom Passive Gravity Compensation Mechanism Applicable to Robotic Arm with Remote Center of Motion for Minimally Invasive Surgery |
Kim, Chang Kyun | Korea Advanced Institute of Science and Technology (KAIST) |
Chung, Deok Gyoon | KAIST |
Hwang, Minho | Korea Advanced Institute of Science and Technology (KAIST) |
Cheon, Byungsik Cheon | Kohyoung Company |
Kim, Hansoul | Korea Advanced Institute of Science and Technology |
Kim, Joonhwan | Korea Advanced Institute of Science and Technology(KAIST) |
Kwon, Dong-Soo | KAIST |
Keywords: Surgical Robotics: Laparoscopy, Mechanism Design, Robot Safety
Abstract: For safety enhancement reasons, passive gravity compensation is widely applied in robotic systems used in minimally invasive surgery (MIS). MIS robotic systems have a remote center of motion (RCM) in which a surgical instrument conducts a fulcrum motion around a point of invasion. RCM mechanisms include three-degrees-of-freedom (3-DoF): roll, pitch, and translation. Existing studies to date have focused on multi-degrees-of-freedom (MDoF) gravity compensation mechanisms by installing springs and wires in a robot. However, a gravity compensation mechanism with 3-DoF that simultaneously uses all three directional movements (roll, pitch, and translation) has not yet been researched. Here, we propose a novel gravity compensation mechanism applicable to a 3-DoF MIS robotic arm with an RCM mechanism. When a translational motion is exerted, the proposed gravity compensator can adjust the roll-pitch-directional compensating torque by utilizing a reduction gear box and wire cable. To verify the 3-DoF gravity compensation, a gravity-compensated robotic arm for MIS and customized torque sensors were manufactured and calibrated. Results showed the proposed static balancing mechanism can compensate for the gravitational torque with respect to roll, pitch, and translation. The total torque error along the roll and pitch axis was less than 0.38 Nm. In particular, the torque variation due to the translational motion was less than 0.13 Nm.
|
|
12:00-12:15, Paper TuAT11.5 | Add to My Program |
Design and Verification of a Portable Master Manipulator Based on an Effective Workspace Analysis Framework |
Zhang, Dandan | Imperial College London |
Liu, Jindong | Imperial College London |
Zhang, Lin | Imperial College London |
Yang, Guang-Zhong | Imperial College London |
Keywords: Surgical Robotics: Laparoscopy, Medical Robots and Systems
Abstract: Master manipulators represent a key component of Robot-Assisted Minimally Invasive Surgery (RAMIS). In this paper, an Analytic Hierarchy Process (AHP) method is used to construct an effective workspace analysis framework, which can assist the configuration selection and design evaluation of a portable master manipulator for surgical robot control and training. The proposed framework is designed based on three criteria: 1) compactness, 2) workspace quality, and 3) mapping efficiency. A hardware prototype, called the Hamlyn Compact Robotic Master (Hamlyn CRM), is constructed following the proposed framework. Experimental verification of the platform is conducted on the da Vinci Research Kit (dVRK) with which a da Vinci robot is controlled as a slave. The proposed CRM is compared with Phantom Omni, a commercial portable master device, with results demonstrating the relative merits of the new platform in terms of task completion time, average control speed and number of clutching.
|
|
12:15-12:30, Paper TuAT11.6 | Add to My Program |
Macro-Micro Multi-Arm Robot for Single-Port Access Surgery |
Vandebroek, Tom | KU Leuven |
Ourak, Mouloud | University of Leuven |
Gruijthuijsen, Caspar | KU Leuven |
Javaux, Allan | KU Leuven |
Legrand, Julie | KULeuven |
Vercauteren, Tom | King's College London |
Ourselin, Sebastien | University College London |
Deprest, Jan | University Hospital Leuven |
Vander Poorten, Emmanuel B | KU Leuven |
Keywords: Surgical Robotics: Laparoscopy, Product Design, Development and Prototyping, Multi-Robot Systems
Abstract: Minimally invasive surgery is now a well established field in surgery but continuous efforts are made to reduce invasiveness even further. This paper proposes a novel concept of small-diameter multi-arm instrument for Single-Port Access Surgery. The concept introduces a novel combination of backbone and actuation principles in a macro-micro fashion to achieve an excellent decoupling of the triangulation platform (macro) and of the end-effectors (micro). Concentric tube robots are used for the triangulation platform, while compliant fluidic-actuated bending segments are used as end-effectors. The fluidic actuation is advantageous as it minimally interferes with the triangulation platform. The triangulation platform on the other hand provides a stable base for the end-effectors such that large distal actuation bandwidth can be achieved. A specific embodiment for Spina Bifida repair is developed and proposed. The surgical and technical requirements as well as the mechanical design are presented in details. A first prototype is built and characterization experiments are conducted to evaluate its performance.
|
|
TuAT12 Regular session, LG-R12 |
Add to My Program |
Human Detection and Tracking |
|
|
Chair: Rocco, Paolo | Politecnico Di Milano |
Co-Chair: Bera, Aniket | University of Maryland |
|
11:00-11:15, Paper TuAT12.1 | Add to My Program |
Real-Time Monitoring of Human Task Advancement |
Maderna, Riccardo | Politecnico Di Milano |
Lanfredini, Paolo | Politecnico Di Milano |
Zanchettin, Andrea Maria | Politecnico Di Milano |
Rocco, Paolo | Politecnico Di Milano |
Keywords: Human Detection and Tracking, Human Factors and Human-in-the-Loop, Task Planning
Abstract: In collaborative robotics applications, the human behaviour is a major source of uncertainty. Predicting the evolution of the current human activity might be beneficial to the effectiveness of task planning, as it enables a higher level of coordination of robot and human activities. This paper addresses the problem of monitoring the advancement of human tasks in real-time giving an estimate of their expected duration. The proposed method relies on dynamic time warping to align the current activity with a reference template. No training phase is required, as the prototypical execution is learnt online from previous instances of the same activity. The applicability and performance of the method within an industrial context have been verified on a realistic assembly task.
|
|
11:15-11:30, Paper TuAT12.2 | Add to My Program |
Deep Orientation: Fast and Robust Upper Body Orientation Estimation for Mobile Robotic Applications |
Lewandowski, Benjamin | Ilmenau University of Technology |
Seichter, Daniel | Ilmenau University of Technology |
Wengefeld, Tim | Ilmenau University of Technology |
Pfennig, Lennard | University |
Drumm, Helge | Technische Universität Ilmenau |
Gross, Horst-Michael | Ilmenau University of Technology |
Keywords: Human Detection and Tracking, RGB-D Perception, Deep Learning in Robotics and Automation
Abstract: An essential feature for navigating socially with a mobile robot is the upper body orientation of persons in its vicinity. For example, in a supermarket orientation indicates whether a person is looking at goods on the shelves or where a person is likely to go. However, given limited computing and battery capabilities, it is not possible to rely on high- performance graphics cards to run large, computationally expensive deep neural networks for orientation estimation in real time. Nevertheless, deep learning performs quite well for regression problems. Therefore, we tackle the problem of upper body orientation estimation with small yet efficient deep neural networks on a mobile robot in this paper. We employ a fast person detection approach as preprocessing that outputs fixed size person images before the actual estimation of the orientation is done. The combination with lightweight networks allows us to estimate a continuous angle in real time, even using a CPU only. We experimentally evaluate the performance of our system on a new, self-recorded data set consisting of more than 100,000 RGB-D samples from 37 persons, which is made publicly available. We also do an extensive comparison of different network architectures and output encodings for their applicability in estimating orientations. Furthermore, we show that depth images are more suitable for the task of orientation estimation than RGB images or the combination of both.
|
|
11:30-11:45, Paper TuAT12.3 | Add to My Program |
Wearable Activity Recognition for Robust Human-Robot Teaming in Safety-Critical Environments Via Hybrid Neural Networks |
Frank, Andrea | University of California, San Diego |
Kubota, Alyssa | University of California San Diego |
Riek, Laurel D. | University of California San Diego |
Keywords: Human Detection and Tracking, Robot Safety, Human-Centered Robotics
Abstract: In this work, we present a novel non-visual HAR system that achieves state-of-the-art performance on realistic SCE tasks via a single wearable sensor. We leverage surface electromyography and inertial data from a low-profile wearable sensor to attain performant robot perception while remaining unobtrusive and user-friendly. By capturing both convolutional and temporal features with a hybrid CNN-LSTM classifier, our system is able to robustly and effectively classify complex, full-body human activities with only this single sensor. We perform a rigorous analysis of our method on two datasets representative of SCE tasks, and compare performance with several prominent HAR algorithms. Results show our system substantially outperforms rival algorithms in identifying complex human tasks from minimal sensing hardware, achieving F1-scores up to 84% over 31 strenuous activity classes. To our knowledge, we are the first to robustly identify complex full-body tasks using a single, unobtrusive sensor feasible for real-world use in SCEs. Using our approach, robots will be able to more reliably understand human activity, enabling them to safely navigate sensitive, crowded spaces.
|
|
11:45-12:00, Paper TuAT12.4 | Add to My Program |
Normal Distribution Mixture Matching Based Model Free Object Tracking Using 2D LIDAR |
Choi, Baehoon | Yonsei University |
Jo, HyungGi | Yonsei University |
Kim, Euntai | Yonsei University |
Keywords: Human Detection and Tracking, Sensor Fusion, Intelligent Transportation Systems
Abstract: In this paper, a novel normal distribution mixture matching based model free object tracking algorithm using 2D LIDAR is proposed. Each target object is modeled as a normal distribution mixture that captures the distribution of the points scanned from the surface of the object. This novel representation enables normal distribution transform (NDT) to accurately estimate the motion of objects, even if the shape of the points differs depending on where it is observed. Our evaluation of the proposed algorithm shows good performance in practical applications. In addition, we provides an alternative way of segmentation and data association using occupancy grid map to avoid a problem that defines a distance metric between the mixture and the point cloud. As a result, the proposed algorithm works in real time in our experiments.
|
|
12:00-12:15, Paper TuAT12.5 | Add to My Program |
Privacy-Preserving Robot Vision with Anonymized Faces by Extreme Low Resolution |
Kim, Myeung Un | Ulsan National Institute of Science and Technology (UNIST) |
Lee, Harim | Ulsan National Institute of Science and Technology |
Yang, Hyun Jong | Ulsan National Institute of Science and Technology |
Ryoo, Michael S. | Indiana University Bloomington |
Keywords: Human Detection and Tracking, Surveillance Systems, Computer Vision for Automation
Abstract: As smart cameras are becoming ubiquitous in mobile robot systems, there is an increasing concern in camera devices invading people's privacy by recording unwanted images. We want to fundamentally protect privacy by blurring unwanted blocks in images, such as faces, yet ensure that the robots can understand the video for their perception. In this paper, we propose a novel mobile robot framework with a deep learning-based privacy-preserving camera system. The proposed camera system detects privacy-sensitive blocks, i.e., human face, from extreme low resolution (LR) images, and then dynamically enhances the resolution of only privacy-insensitive blocks, e.g., backgrounds. Keeping all the face blocks to be extreme LR of 15x15 pixels, we can guarantee that human faces are never at high resolution (HR) in any of processing or memory, thus yielding strong privacy protection even from cracking or backdoors. Our camera system produces an image on a real-time basis, the human faces of which are in extreme LR while the backgrounds are in HR. We experimentally confirm that our proposed face detection camera system outperforms the state-of-the-art small face detection algorithm, while the robot performs ORB-SLAM2 well even with videos of extreme LR faces. Therefore, with the proposed system, we do not too much sacrifice robot perception performance to protect privacy.
|
|
12:15-12:30, Paper TuAT12.6 | Add to My Program |
DensePeds: Pedestrian Tracking in Dense Crowds Using FRVO and Sparse Features |
Chandra, Rohan | University of Maryland |
Bhattacharya, Uttaran | UMD College Park |
Bera, Aniket | Dr |
Manocha, Dinesh | University of Maryland |
Keywords: Human Detection and Tracking
Abstract: We present a pedestrian tracking algorithm, DensePeds, that tracks individuals in highly dense crowds > 2 pedestrians per square meter. Our approach is designed for videos captured from front-facing or elevated cameras. We present a new motion model called Front RVO for predicting pedestrian movements in dense situations using collision avoidance constraints and combine it with state-of-the-art Mask R-CNN to compute sparse feature vectors that reduce the loss of pedestrian tracks (false negatives). We evaluate DensePeds on the standard MOT benchmarks as well as a new dense crowd dataset. In practice, our approach is 4.5X faster than prior tracking algorithms on the MOT benchmark and we are state-of-the-art in dense crowd videos by over 2.6% on the absolute scale on average.
|
|
TuAT13 Regular session, LG-R13 |
Add to My Program |
Humanoid and Bipedal Locomotion I |
|
|
Chair: Takanishi, Atsuo | Waseda University |
Co-Chair: Tsagarakis, Nikos | Istituto Italiano Di Tecnologia |
|
11:00-11:15, Paper TuAT13.1 | Add to My Program |
TIP Model: A Combination of Unstable Subsystems for Lateral Balance in Walking |
Firouzi, Vahid | University of Tehran |
Ahmad Sharbafi, Maziar | Technical University of Darmstadt |
Seyfarth, Andre | TU Darmstadt |
Keywords: Humanoid and Bipedal Locomotion, Biologically-Inspired Robots, Legged Robots
Abstract: Balancing or postural stability is one of important locomotor subfunctions in bipedal gaits. The inverted pendulum and virtual pivot point (VPP) are common modeling approaches to analyze balance control in human and robot walking. In this paper, we employ the VPP concept to investigate posture control in the frontal plane. The outcomes demonstrate that unlike posture control in the sagittal plane, the VPP in the frontal plane is placed below center of mass. This finding explains a novel hybrid strategy for lateral stability in human walking. The here proposed model shows that switching between unstable inverted virtual pendulums generates stable posture control in the frontal plane. This outcome is consistent within a group of seven human subjects walking at normal and slow speeds.
|
|
11:15-11:30, Paper TuAT13.2 | Add to My Program |
Avoiding Obstacles During Push Recovery Using Real-Time Vision Feedback |
Jeong, Hyobin | KAIST |
Kim, Joon-Ha | Korea Advanced Institute of Science and Technology(KAIST) |
Sim, Okkee | KAIST |
Oh, Jun Ho | Korea Advanced Inst. of Sci. and Tech |
Keywords: Humanoid and Bipedal Locomotion, Computer Vision for Other Robotic Applications, Robotics in Hazardous Fields
Abstract: This paper introduces an obstacle-avoiding algorithm for bipedal robots, especially in push recovery situations. Typically, There are many algorithms that plan footstep to avoid obstacles based on vision recognition data. However, if the robot is pushed, the planned footprint will change, and thus, there is no guarantee that it will avoid obstacles. Although modified stepping positions can be limited, the robot’s stability is not assured. Our proposed algorithm focuses on avoiding obstacles through vision recognition in push recovery situations and generating compensation actions for instability by restricting modified footsteps. We fuse vision feedback with our previous push recovery algorithm, which optimizes the ankle, hip, and stepping strategies. We build simple grid data using vision recognition and apply it to the inequality constraint of the stepping position. We validate the effectiveness of our algorithm using the bipedal platform GAZELLE with the Kinect V2 RGBD sensor.
|
|
11:30-11:45, Paper TuAT13.3 | Add to My Program |
Effect of Planning Period on MPC-Based Navigation for a Biped Robot in a Crowd |
Ciocca, Matteo | INRIA |
Wieber, Pierre-Brice | INRIA Rhône-Alpes |
Fraichard, Thierry | INRIA |
Keywords: Humanoid and Bipedal Locomotion, Humanoid Robots, Collision Avoidance
Abstract: We control a biped robot moving in a crowd with a Model Predictive Control (MPC) scheme that generates stable walking motions, with automatic footstep placement. Most walking strategies propose to re-plan the walking motion to adapt to changing environments only once at every footstep. This is because a footstep is planted on the ground, it usually stays there at a constant position until the next footstep is initiated, what naturally constrains the capacity for the robot to react and adapt its motion in between footsteps. The objective of this paper is to measure if re-planning the walking motion more often than once at every footstep can lead to an improvement in collision avoidance when navigating in a crowd. Our result is that re-planning twice (or more) during each footstep leads to a significant reduction of the number of collisions when walking in a crowd, but depends on the density of the crowd.
|
|
11:45-12:00, Paper TuAT13.4 | Add to My Program |
Unified Balance Control for Biped Robots Including Modification of Footsteps with Angular Momentum and Falling Detection Based on Capturability |
Kojio, Yuta | The University of Tokyo |
Ishiguro, Yasuhiro | The University of Tokyo |
Nguyen, Kim-Ngoc-Khanh | The University of Tokyo |
Sugai, Fumihito | The University of Tokyo |
Kakiuchi, Yohei | The University of Tokyo |
Okada, Kei | The University of Tokyo |
Inaba, Masayuki | The University of Tokyo |
Keywords: Humanoid and Bipedal Locomotion, Humanoid Robots, Legged Robots
Abstract: In this paper, we propose walking balance control based on Caputurability. The proposed method consists of five strategies: (i) moving Zero Moment Point (ZMP) in the support polygon (ii) landing position modification (iii) landing timing modification (iv) angular momentum control (v) falling detection and fall control. Walking pattern generation calculates the ZMP so that the Capture Point (CP) reaches the position of the supporting foot at the end of the double support phase. Owing to the asymmetry of the reachable landing region, landing timing modification is different in the sagittal and lateral planes, and the step time is extended in the lateral plane depending on the direction of disturbances. The torque around the center of gravity to avoid falling is realized through whole-body inverse kinematics with constraints on the angular momentum. In addition, we propose falling detection considering the reachable landing region. We verified the effectiveness of the proposed method through experiments in which the biped robot was disturbed by pushing during tether-free walking. The robot could prevent breakdown by detecting possible falling and performed knee bending motions to suppress damage.
|
|
12:00-12:15, Paper TuAT13.5 | Add to My Program |
Estimating the Center of Mass and the Angular Momentum Derivative for Legged Locomotion — a Recursive Approach |
Bailly, François | LAAS-CNRS |
Carpentier, Justin | INRIA |
Benallegue, Mehdi | AIST Japan |
Watier, Bruno | LAAS, CNRS, Université Toulouse 3 |
Soueres, Philippe | LAAS-CNRS |
Keywords: Humanoid and Bipedal Locomotion, Humanoid Robots
Abstract: Estimating the center of mass position and the angular momentum derivative of legged systems is essential for both controlling legged robots and analyzing human motion. In this paper, a novel recursive approach to concurrently and accurately estimate these two quantities together is introduced. The proposed method employs kinetic and kinematic measurements from classic sensors available in robotics and biomechanics, to effectively exploits the accuracy of each measurement in the spectral domain. The soundness of the proposed approach is first validated on a simulated humanoid robot, where ground truth data is available, against an Extend Kalman Filter. The results demonstrate that the proposed method reduces the estimation error on the center of mass position with regard to kinematic estimation alone, whereas at the same time, it provides an accurate estimation of the derivative of angular momentum. Finally, the effectiveness of the proposed method is illustrated on real measurements, obtained from walking experiments with the HRP-2 humanoid robot.
|
|
12:15-12:30, Paper TuAT13.6 | Add to My Program |
Online Relative Footstep Optimization for Legged Robots Dynamic Walking Using Discrete-Time Model Predictive Control |
Xin, Songyan | Istituto Italiano Di Tecnologia (IIT) |
Orsolino, Romeo | Istituto Italiano Di Tecnologia |
Tsagarakis, Nikos | Istituto Italiano Di Tecnologia |
Keywords: Humanoid and Bipedal Locomotion, Legged Robots, Motion and Path Planning
Abstract: We present a unified control framework that generates dynamic walking motions for biped and quadruped robots with online relative footstep optimization. The footstep optimization is formulated as a discrete-time Model Predictive Control problem which determines future footstep locations. The framework has a hierarchical structure consisting of three layers: footstep planner, trajectory generator and whole-body controller. The footstep planner plans next footstep position based on Linear Inverted Pendulum (LIP) model. Relative footstep optimization is proposed to enable automatic footstep planning without the use of any predefined footstep sequences. The trajectory generator will generate CoM and feet trajectory given the next footstep placement. In order to generalize to quadruped robots, ``virtual leg'' concept has been used to coordinate leg pair movement. The whole-body inverse dynamic controller calculates joint torques to track given Cartesian reference trajectories. To include under-actuation into consideration, contact vertices formulation of ground reaction forces (GRFs) has been adopted. Generalized whole-body controller can handle biped robot with line feet as well as quadruped robots with point feet walking with dynamic gaits. Several simulations have been performed to demonstrate the robustness and generality of the proposed framework.
|
|
TuAT14 Regular session, LG-R14 |
Add to My Program |
Space Robotics |
|
|
Chair: Secchi, Cristian | Univ. of Modena & Reggio Emilia |
Co-Chair: Higa, Shoya | Jet Propulsion Laboratory |
|
11:00-11:15, Paper TuAT14.1 | Add to My Program |
Time-Delay Compensation Using Energy Tank for Satellite Dynamics Robotic Simulators |
De Stefano, Marco | German Aerospace Center (DLR) |
Vezzadini, Luca | Univ. of Modena & Reggio Emilia |
Secchi, Cristian | Univ. of Modena & Reggio Emilia |
Keywords: Space Robotics and Automation, Industrial Robots, Motion Control
Abstract: In this work we present a novel approach which compensates the destabilising effects of the time delay intrinsic in the control loop of an admittance-controlled robot employed for satellite dynamics simulation. The method is based on an energy storing element, the tank, which is exploited by the controller to preserve the passivity of the system and to avoid instability. Furthermore, we compare the performance of the proposed method with existing energy-based approaches, namely time-domain-passivity and wave variable transformation. The performance comparison and robustness of the methods are analysed in a Montecarlo simulation and validated experimentally.
|
|
11:15-11:30, Paper TuAT14.2 | Add to My Program |
A Sweeping and Grinding Methods Combined Hybrid Sampler for Asteroid Exploration |
Dong, Chengcheng | Southeast University |
Zhang, Jun | Southeast University |
Jiang, Chaojun | Southeast University |
Huang, Fanzhang | Southeast University |
Lu, Xi | Shanghai Institute of Satellite Engineering |
Huang, Fan | Shanghai Institute of Satellite Engineering |
Song, Aiguo | Southeast University |
Keywords: Space Robotics and Automation, Mining Robotics
Abstract: Successful sampling on the surface of asteroids is difficult because of their weightless environment and unknown material mechanical property. This work presents an asteroid sampler based on sweeping and grinding methods to improve the success rate of sampling. The sampler uses two brushes rotating clockwise and counterclockwise to collect sample particles on the surface of asteroids. When encountering the hard rock or sample particles with large cohesion, the sampler adopts a drill bit to grind them to loose samples suitable for collecting by the brushes. The interaction between the brushes and the regolith is modeled and the sweeping mechanism is designed. A simple grinding mechanism is also designed. Numerical simulation and prototype experiments, at different parameters including blades number, rotational speed, and feeding speed of the brushes, mechanical property of the sample, and gravity, were conducted for validating the proposed methods. The 280g sampler prototype with 8 blades of brushes could collect about 19g regolith simulant in 25s in earth environment. The drill bits could work together with the brushes to improve the sampling efficiency through DEM simulation in case of large cohesion among sample particles. The sampler will be a good choice for installing into an asteroid rover in exploration.
|
|
11:30-11:45, Paper TuAT14.3 | Add to My Program |
Non-Myopic Planetary Exploration Combining in Situ and Remote Measurements |
Kodgule, Suhit | Carnegie Mellon University |
Candela, Alberto | Carnegie Mellon University |
Wettergreen, David | Carnegie Mellon University |
Keywords: Space Robotics and Automation, Motion and Path Planning
Abstract: Remote sensing can provide crucial information for planetary rovers. However, they must validate these orbital observations with in situ measurements. Typically, this involves validating hyperspectral data using a spectrometer on-board the field robot. In order to achieve this, the robot must visit sampling locations that jointly improve a model of the environment while satisfying sampling constraints. However, current planners follow sub-optimal greedy strategies that are not scalable to larger regions. We demonstrate how the problem can be effectively defined in an MDP framework and propose a planning algorithm based on Monte Carlo Tree Search, which is devoid of the common drawbacks of existing planners and also provides superior performance. We evaluate our approach using hyperspectral imagery of a well-studied geologic site in Cuprite, Nevada.
|
|
11:45-12:00, Paper TuAT14.4 | Add to My Program |
Vision-Based Estimation of Driving Energy for Planetary Rovers Using Deep Learning and Terramechanics |
Higa, Shoya | Jet Propulsion Laboratory |
Iwashita, Yumi | NASA / Caltech Jet Propulsion Laboratory |
Otsu, Kyohei | California Institute of Technology |
Ono, Masahiro | California Institute of Technology |
Lamarre, Olivier | University of Toronto, Space and Terrestrial Autonomous Robotic |
Didier, Annie | NASA JPL |
Hoffmann, Mark | JPL |
Keywords: Space Robotics and Automation, Wheeled Robots, Deep Learning in Robotics and Automation
Abstract: This paper presents a prediction algorithm of driving energy for future Mars rover missions. The majority of future Mars rovers would be solar-powered, which would require energy-optimal driving to maximize the range with limited energy. The essential and arguably the most challenging technology for realizing energy-optimal driving is the capability to predict the driving energy, which is needed to construct an energy-aware cost function for path planning. In this paper, we propose vision-based algorithms to remotely predict the driving energy consumption using machine learning. Specifically, we develop and compare two machine-learning models in this paper, namely VeeGer-EnergyNet and Veeger-TerramechanicsNet, respectively. The former is trained directly using recorded power, while the latter estimates terrain parameters from the images using a simplified-terramechanics model, and calculate the power based on the model. The two approaches are fully automated self-supervised learning algorithms. To combine RGB and depth images efficiently with high accuracy, we propose a new network architecture called Two-PNASNet-5, which is based on PNASNet-5. We collected a new dataset to verify the effectiveness of the proposed approaches. Comparison of the two approaches showed that Veeger-TerramechanicsNet had better performance than VeeGer-EnergyNet.
|
|
12:00-12:15, Paper TuAT14.5 | Add to My Program |
Improved Planetary Rover Inertial Navigation and Wheel Odometry Performance through Periodic Use of Zero-Type Constraints |
Kilic, Cagri | West Virginia University |
Gross, Jason | West Virginia University |
Ohi, Nicholas | West Virginia University |
Watson, Ryan | West Virginia University |
Strader, Jared | West Virginia University |
Swiger, Thomas | West Virginia University |
Harper, Scott | West Virginia University |
Gu, Yu | West Virginia University |
Keywords: Space Robotics and Automation, Wheeled Robots, Nonholonomic Mechanisms and Systems
Abstract: We present an approach to enhance wheeled planetary rover dead-reckoning localization performance by leveraging the use of zero-type constraint equations in the navigation filter. Without external aiding, inertial navigation solutions inherently exhibit cubic error growth. Furthermore, for planetary rovers that are traversing diverse types of terrain, wheel odometry is often unreliable for use in localization, due to wheel slippage. For current Mars rovers, computer vision-based approaches are generally used whenever there is a high possibility of positioning error; however, these strategies require additional computational power, energy resources, adequate features in the environment, and significantly slow down the rover traverse speed. To this end, we propose a navigation approach that compensates for the high likelihood of odometry errors by providing a reliable navigation solution that leverages non-holonomic vehicle constraints as well as state-aware pseudo-measurements (e.g., zero velocity and zero angular rate) updates during periodic stops. By using this, computationally expensive visual-based corrections could be performed less often. Experimental tests that compare against GPS-based localization are used to demonstrate the accuracy of the proposed approach. The source code, post-processing scripts, and example datasets associated with the paper are published in a public repository.
|
|
12:15-12:30, Paper TuAT14.6 | Add to My Program |
Modeling and Force Control of a Terramechanical Wheel-Soil Contact for a Robotic Manipulator Used in the Planetary Rover Design Process |
Wachter, Jan | Karlsruhe Institute of Technology |
Mikut, Ralf | Karlsruhe Institute of Technology |
Buse, Fabian | Institute of System Dynamic and Control - German Aerospace Cente |
Keywords: Force Control, Contact Modelling, Space Robotics and Automation
Abstract: The German Aerospace Center (DLR) has developed the Terramechanics Robotics Locomotion Lab (TROLL) to provide a feasible testing facility for developing planetary exploration rovers, as well as validating terramechanical models. A robotic manipulator is used to provide the required degrees of freedom to the mounted wheel or subsystem, making it necessary to actively control the interaction force of the wheel-soil contact. This paper is concerned with the development of a feasible force control strategy for the testbench wheel-soil contact during single-wheel experiments. For this purpose, a terramechanical model has been developed to accurately map the dynamic processes relevant for the force control design, which is later used in a testbench simulation framework to predict and evaluate the performance of control strategies. The Adaptive Admittance Control (AAC) scheme developed is adapting the gain based on the current control deviation, the rotational velocity of the wheel and an estimated soil stiffness during the experiment. The AAC is evaluated using a benchmark single-wheel experiment and shows superior performance compared to standard admittance control.
|
|
TuAT15 Regular session, LG-R15 |
Add to My Program |
Motion and Path Planning I |
|
|
Chair: Kim, H. Jin | Seoul National University |
Co-Chair: Zhang, Shiwu | University of Science and Technology of China |
|
11:00-11:15, Paper TuAT15.1 | Add to My Program |
Predictive Inverse Kinematics: Optimizing Future Trajectory through Implicit Time Integration and Future Jacobian Estimation |
Ayusawa, Ko | AIST |
Suleiman, Wael | University of Sherbrooke |
Yoshida, Eiichi | National Inst. of AIST |
Keywords: Kinematics, Motion and Path Planning, Optimization and Optimal Control
Abstract: This paper presents an inverse kinematics (IK) method which can control future velocities and accelerations for multi-body systems. The proposed IK method is formulated as a quadratic programing (QP) that optimizes future joint trajectories. The features of the proposed IK are: (1) the evaluation of accelerations at future time instances, (2) the trajectory representation that can implicitly integrate the time integral formula into QP, (3) the computation of future Jacobian matrices based on the comprehensive theory of differential kinematics proposed in our previous work. Those features enable a stable and fast IK computation while evaluating the future accelerations. We also conducted thorough numerical studies to show the efficiency of the proposed method.
|
|
11:15-11:30, Paper TuAT15.2 | Add to My Program |
Computing 3D From-Region Visibility Using Visibility Integrity |
Zhi, Jixuan | George Mason University |
Hao, Yue | George Mason University |
Vo, Christopher | George Mason University |
Morales, Marco | Instituto Tecnológico Autónomo De México |
Lien, Jyh-Ming | George Mason University |
Keywords: Motion and Path Planning, Computational Geometry, Simulation and Animation
Abstract: Visibility integrity (VI) is a measurement of similarity between the visibilities of regions. It can be used to approximate the visibility of coherently moving targets, called group visibility. It has been shown that computing visibility integrity using agglomerative clustering takes O(n^4*logn) for n samples. Here, we present a method that speeds up the computation of visibility integrity and reduces the time complexity from O(n^4*logn) to O(n^2). Based on the idea of visibility integrity, we show for the first time that the visibility-integrity roadmap (VIR), a data structure that partitions a space into zones, can be calculated efficiently in 3D. More specifically, we offer two different offline approaches, a naive one and a kernel-based one, to compute a VIR. In addition, we demonstrate how to apply a VIR to solve group visibility and group following problems in 3D. We propose a planning algorithm for the camera to maintain visibility of group targets by querying the VIR. We evaluate our approach in different 3D simulation environments and compare it to other planning methods.
|
|
11:30-11:45, Paper TuAT15.3 | Add to My Program |
Edge-Preserving Camera Trajectories for Improved Optical Character Recognition on Static Scenes with Text |
Katoch, Rohan | Georgia Institute of Technology |
Ueda, Jun | Georgia Institute of Technology |
Keywords: Motion and Path Planning, Recognition, Computer Vision for Other Robotic Applications
Abstract: Camera systems in fast motion suffer from the effects of motion blur, which degrades image quality and can have a large impact on the performance of visual tasks. The degradation in image quality can be mitigated through the use of image reconstruction. Blur effects and the resulting reconstruction performance are highly dependent on the point-spread function resulting from camera motion. This work focuses on the motion planning problem for a camera system with boundary conditions on time and position, with the objective of improving the performance of optical character recognition. Tuned edge-preserving trajectories are shown to result in higher recognition accuracy when compared to inverse error and linear trajectories. Simulation and experimental results provide quantitative measures to verify edge-preservation and greater recognition performance.
|
|
11:45-12:00, Paper TuAT15.4 | Add to My Program |
Virtual Region Based Multi-Robot Path Planning in an Unknown Occluded Environment |
Roy, Dibyendu | Tata Consultancy Servives Limited |
Chowdhury, Arijit | TCS Research and Innovation |
Maitra, Madhubanti | JADAVPUR UNIVERSITY |
Bhattacharya, Samar | Jadavpur University |
Keywords: Path Planning for Multiple Mobile Robots or Agents, Swarms, Autonomous Agents
Abstract: This paper introduces an arbitrary-region based shape-control methodology for a swarm-robotics framework to conquer the traditional obstacle-avoidance problem. In this control approach, during the movement of a team of robots through an unknown occluded environment, a virtual boundary for the robotic crowd that could pass through even a very narrow passage, as a swarm, has been created based on the agents’/robots’ sensing information. The primary challenge is to force each and every agent to fit into the arbitrary boundary, thus created, so that they can form a swarm and eventually progress towards the goal while avoiding obstacles. Consequently, the process maximizes the inter-agent cohesion. Furthermore, the virtual region is subdivided into so-called cells designated by a non-overlapping circular geometry. The utmost endeavor is to make an agent place itself at the center of a circular cell and a mechanism has been devised, where each agent is duly attracted and eventually is resident at the center of each cell completely blanketing the entire virtual region. To place each agent inside the stipulated contour, a novel circle-packing algorithm has been proposed such that the contour is fitted with a number of identical circles. Finally, the simulation results along with hardware implementation demonstrate the effectiveness of the proposed control technique.
|
|
12:00-12:15, Paper TuAT15.5 | Add to My Program |
Fast Trajectory Planning for Multiple Quadrotors Using Relative Safe Flight Corridor |
Park, Jungwon | Seoul National University |
Kim, H. Jin | Seoul National University |
Keywords: Path Planning for Multiple Mobile Robots or Agents, Swarms, Collision Avoidance
Abstract: This paper presents a new trajectory planning method for multiple quadrotors in obstacle-dense environments. We suggest a relative safe flight corridor (RSFC) to model safe region between a pair of agents, and it is used to generate linear constraints for inter-collision avoidance by utilizing the convex hull property of relative Bernstein polynomial. Our approach employs a graph-based multi-agent pathfinding algorithm to generate an initial trajectory, which is used to construct a safe flight corridor (SFC) and RSFC. We express the trajectory as a piecewise Bernstein polynomial and formulate the trajectory planning problem into one quadratic programming problem using linear constraints from SFC and RSFC. The proposed method can compute collision-free trajectory for 16 agents within a second and for 64 agents less than a minute, and it is validated both through simulation and indoor flight test.
|
|
12:15-12:30, Paper TuAT15.6 | Add to My Program |
Predictive Inverse Kinematics for Redundant Manipulators with Task Scaling and Kinematic Constraints (I) |
Faroni, Marco | University of Brescia |
Beschi, Manuel | National Research Council of Italy |
Pedrocchi, Nicola | National Research Council of Italy |
Visioli, Antonio | University of Brescia |
Keywords: Motion and Path Planning, Kinematics, Optimization and Optimal Control
Abstract: The paper presents a fast online predictive method to solve the task-priority differential inverse kinematics of redundant manipulator under kinematic constraints. It implements a task-scaling technique to preserve the desired geometrical task, when the trajectory is infeasible for the robot capabilities. Simulation results demonstrate the effectiveness of the methodology.
|
|
TuAT16 Regular session, LG-R16 |
Add to My Program |
Grasping I |
|
|
Chair: Kroeger, Torsten | Karlsruher Institut Für Technologie (KIT) |
Co-Chair: Watanabe, Tetsuyou | Kanazawa University |
|
11:00-11:15, Paper TuAT16.1 | Add to My Program |
Robot Learning of Shifting Objects for Grasping in Cluttered Environments |
Berscheid, Lars | Karlsruhe Institute of Technology |
Meißner, Pascal | Karlsruhe Institute of Technology |
Kroeger, Torsten | Karlsruher Institut Für Technologie (KIT) |
Keywords: Grasping, AI-Based Methods, Perception for Grasping and Manipulation
Abstract: Robotic grasping in cluttered environments is often infeasible due to obstacles preventing possible grasps. Then, pre-grasping manipulation like shifting or pushing an object becomes necessary. We developed a robotic system that can learn, in addition to grasping, to shift objects in such a way that their grasp probability increases. Our research contribution is threefold: First, we present an algorithm for learning the optimal pose of manipulation primitives like clamping or shifting. Second, we learn non-prehensible actions that explicitly increase the grasping probability. Making one skill (shifting) directly dependent on another (grasping) removes the need of sparse rewards, leading to more data-efficient learning. Third, we apply a real-world solution to the industrial task of bin picking, resulting in the ability to empty bins completely. The system is trained in a self-supervised manner with around 25000 grasp and 2500 shift actions. Our robot is able to grasp and file objects with 274 picks per hour. Furthermore, we demonstrate the system's ability to generalize to novel objects.
|
|
11:15-11:30, Paper TuAT16.2 | Add to My Program |
Deep Reinforcement Learning for Robotic Pushing and Picking in Cluttered Environment |
Deng, Yuhong | Tsinghua Univerisity |
Guo, Xiaofeng | Tsinghua Univerisity |
Wei, Yixuan | Tsinghua Univerisity |
Lu, Kai | Tsinghua Univerisity |
Fang, Bin | Tsinghua University |
Guo, Di | Tsinghua University |
Liu, Huaping | Tsinghua University |
Sun, Fuchun | Tsinghua Univerisity |
Keywords: Grasping, AI-Based Methods
Abstract: In this paper, a novel robotic grasping system is established to automatically pick up objects in cluttered scenes. A composite robotic hand composed of a suction cup and a gripper is designed for grasping the object stably. The suction cup is used for lifting the object from the clutter first and the gripper for grasping the object accordingly. We utilize the affordance map to provide pixel-wise lifting point candidates for the suction cup. To obtain a good affordance map, the active exploration mechanism is introduced to the system. An effective metric is designed to calculate the reward for the current affordance map, and a deep Q-Network (DQN) is employed to guide the robotic hand to actively explore the environment until the generated affordance map is suitable for grasping. Experimental results have demonstrated that the proposed robotic grasping system is able to greatly increase the success rate of the robotic grasping in cluttered scenes.
|
|
11:30-11:45, Paper TuAT16.3 | Add to My Program |
Vision-Based Automatic Control of a 5-Fingered Simulated Assistive Robotic Manipulator for Activities of Daily Living |
Wang, Chen | Imperial College London |
Freer, Daniel | Imperial College London |
Liu, Jindong | Imperial College London |
Yang, Guang-Zhong | Imperial College London |
Keywords: Grasping, Computer Vision for Automation, Multifingered Hands
Abstract: Assistive robotic manipulators (ARMs) play an important role for people with upper-limb disabilities and the elderly to assist them in fulfilling Activities of Daily Living (ADLs). However, as the objects to be handled in ADLs differ in size, shape and manipulation constraints, many two or three fingered end-effectors of ARMs have difficulty robustly interacting with these objects. In this paper, we proposed vision-based control of a 5-fingered manipulator (Schunk SVH), allowing it to automatically change its shape based on object classification using computer vision combined with deep learning. The control method is tested in a simulated environment and achieves a more robust grasp with the properly shaped five-fingered hand than with a comparable three-fingered gripper (Barrett Hand) using the same control sequence. In addition, the final optimal grasp pose (x, y, and theta) is learned through a deep regressor in the penultimate stage of the grasp. This method correctly identifies the optimal grasp pose in 78.35% of the test cases when considering all three parameters.
|
|
11:45-12:00, Paper TuAT16.4 | Add to My Program |
Recalling Candidates of Grasping Method from an Object Image Using Neural Network |
Sanada, Makoto | Ritsumeikan University |
Matsuo, Tadashi | Ritsumeikan University |
Shimada, Nobutaka | Ritsumeikan University |
Shirai, Yoshiaki | Ritsumeikan University |
Keywords: Grasping, Computer Vision for Automation, Visual Learning
Abstract: Robots are required to support people’s work. In order to alleviate the burden on people, it is desirable that robot can automatically generate and execute complicated motions according to simple directions from people. However, there are multiple grasping methods for one object. In order to select a motion suitable for the direction, it is important to estimate candidates of grasping method for the object. In this research, we purpose to recall candidates of grasping position and hand shape from an object image. In learning, a network that outputs a plurality of grasping method candidates for one object image to each channel of a multi-channel image is used. At this time, a plurality of grasping methods are not learned at same time, learned one by one. The similar grasping method for the similar object shape is automatically clustered to each output channel in the learning process, and a grasping method having a characteristic difference is presented as a candidate. We show the usefulness of this method using experimental examples.
|
|
12:00-12:15, Paper TuAT16.5 | Add to My Program |
Domain-Independent Unsupervised Detection of Grasp Regions to Grasp Novel Objects |
Pharswan, Siddhartha Vibhu | Indian Institute of Technology Kanpur |
Vohra, Mohit | Indian Institute of Technology, Kanpur |
Kumar, Ashish | Indian Institute of Technology, Kanpur |
Behera, Laxmidhar | IIT Kanpur |
Keywords: Grasping, Computer Vision for Other Robotic Applications, Factory Automation
Abstract: One of the main challenges in the vision-based grasping is the selection of feasible grasp regions while interacting with novel objects. Recent approaches exploit the power of convolutional neural network (CNN) to achieve accurate grasping at the cost of high computational power and time. In this paper, we present a novel unsupervised learning-based algorithm for the selection of feasible grasp regions. Unsupervised learning infers the pattern in dataset without any external labels. We applied k-means clustering at every sampling stage on image plane to identify the grasp regions, followed by axes assignment method. We define a novel concept of Grasp Decide Index (GDI) to select the best grasp pose in the image plane. We have conducted several experiments in clutter or isolated environment on standard objects of Amazon Robotics Challenge 2017 and Amazon Picking Challenge 2016. We compared the results with prior learning-based approaches to validate the robustness and adaptive nature of our algorithm for a variety of novel objects in different domains.
|
|
12:15-12:30, Paper TuAT16.6 | Add to My Program |
Near-Contact Grasping Strategies from Awkward Poses: When Simply Closing Your Fingers Is Not Enough |
Ong, Yi Herng | Oregon State University |
Morrow, John | Oregon State University |
Qiu, Yu | Oregon State University |
Gupta, Kartik | Oregon State University |
Balasubramanian, Ravi | Oregon State University |
Grimm, Cindy | Oregon State University |
Keywords: Grasping, Human Factors and Human-in-the-Loop
Abstract: Grasping a simple object from the side is easy --- unless the object is almost as big as the hand or space constraints require positioning the robot hand awkwardly with respect to the object. We show that humans --- when faced with this challenge --- adopt coordinated finger movements which enable them to successfully grasp objects even from these awkward poses. We also show that it is relatively straight forward to implement these strategies autonomously. Our human-studies approach asks participants to perform grasping task by either ``puppetteering'' a robotic manipulator that is identical~(geometrically and kinematically) to a popular underactuated robotic manipulator~(the Barrett hand), or using sliders to control the original Barrett hand. Unlike previous studies, this enables us to directly capture and compare human manipulation strategies with robotic ones. Our observation is that, while humans employ underactuation, how they use it is fundamentally different (and more effective) than that found in existing hardware.
|
|
TuAT17 Regular session, LG-R17 |
Add to My Program |
Micro/Nano Robots I |
|
|
Chair: Feng, Lin | Beihang University |
Co-Chair: Zhang, Xuping | Aarhus University |
|
11:00-11:15, Paper TuAT17.1 | Add to My Program |
Automated Macro-Micro Manipulation for Robotic Microinjection with Computer Vision |
Zhang, Huipeng | Beijing University of Technology |
Su, Liying | Beijing University of Technology |
Wei, Hongmiao | Beijing University of Technology |
Yu, Yueqing | Beijing University of Technology |
Zhang, Xuping | Aarhus University |
Keywords: Biological Cell Manipulation, Automation at Micro-Nano Scales, Computer Vision for Automation
Abstract: Extensive research efforts have been made toward automating the microinjection of biological cells by leveraging micro-robotic technologies. However, to best knowledge of the authors, there is no report on the automation of the time- consuming process: moving the injection tools (a micropipette, a grippers, etc.) and cells into the field of view (FOV) of micro- scope from the macro FOV(outside the microscopic FOV). This paper presents a novel macro-micro conversion strategy, and a grid detection and positioning algorithm to automate the time-consuming step of moving the injection tools and cells to the microscopic FOV. The proposed solution can free the techni -cian from the laborious hand-eye coordination operations for moving the injection tools and cells to the target position within the microscopic FOV. Furthermore, this paper proposes an auto-focusing algorithm to automate the operation step: moving down the gripper from the air outside the culture media and then precisely clamping a cell in the liquid environment for injection. In the proposed solution, the active window-based auto-focusing algorithm is developed to solve the challenging problem: the image information is lost due to the “viscous effect” taking place when the gripper jaw touches the water surface. The proposed solutions are tested and validated by the microinjection experiments of zebrafish embryos using the in-house develop micro-robotic system. The technologies and strategies proposed in this paper significantly improve the automation level of the cell microinjections, and can be easily extended to any other micromanipulation of biological cells.
|
|
11:15-11:30, Paper TuAT17.2 | Add to My Program |
A Robotic Surgery Approach to Mitochondrial Transfer Amongst Single Cells |
Shakoor, Adnan | City University of Hong Kong |
Xie, Mingyang | Nanjing University of Aeronautics & Astronautics |
Pan, Fei | City University of Hong Kong |
Gao, Wendi | CityU of Hong Kong |
Sun, Jiayu | City University of Hong Kong |
Sun, Dong | City University of Hong Kong |
Keywords: Biological Cell Manipulation, Automation at Micro-Nano Scales, Surgical Robotics: Steerable Catheters/Needles
Abstract: Introducing alterations in the mtDNA sequence is challenging but needed for potential therapies and basic studies. Direct microinjection of mitochondria into small cells has been considered inefficient and impractical. To address this issue, we present a highly efficient and precise robotic approach for automatically transferring mitochondria from one single cell to another. A microfluidic cell positioning device is used to pattern two different types of cells in one dimensional array, and an image processing algorithm is applied to identify the location of the mitochondria and cell. A visual feedback control mechanism is developed to enhance the mitochondrial extraction efficiency. A robust adaptive sliding control algorithm is developed to precisely control an X–Y stage to accomplish the extraction of mitochondria from A type cell followed by injection of the mitochondria into B type cell automatically. The system can transfer mitochondria from one cell to another with an average duration of 15 s/mitochondria. Experiments of mitochondrial transfer from THP1 and NB4 cells to THP1 cells and fibroblasts are conducted to show the effectiveness of the developed approach
|
|
11:30-11:45, Paper TuAT17.3 | Add to My Program |
A Magnetically Transduced Whisker for Angular Displacement and Moment Sensing |
Kim, Suhan | Carnegie Mellon University |
Velez, Camilo | Carnegie Mellon University |
Patel, Dinesh | Carnegie Mellon University |
Bergbreiter, Sarah | Carnegie Mellon University |
Keywords: Biomimetics, Force and Tactile Sensing, Biologically-Inspired Robots
Abstract: This work presents the design, modeling, and fabrication of a whisker-like sensor capable of measuring the whisker's angular displacement as well as the applied moments at the base of the whisker. The sensor takes advantage of readily accessible and low-cost 3D magnetic sensors to transduce whisker deflections, and a planar serpentine spring structure at the whisker base is used to provide a mechanical suspension for the whisker to rotate. The sensor prototype was characterized, calibrated, and compared with analytical models of the spring system and the magnetic field. The prototype showed a moment sensing range of 1.1 N-mm when deflected up to 19.7 degrees. The sensitivity of the sensor was 0.38 degree/LSB for the angular displacement sensing, and 0.021 N-mm/LSB for the moment sensing. A fully integrated system is demonstrated to display real-time information from the whisker on a graphical interface.
|
|
11:45-12:00, Paper TuAT17.4 | Add to My Program |
Active Whisker Placement and Exploration for Rapid Object Recognition |
Pearson, Martin | Bristol Robotics Laboratory |
Salman, Mohammed | Bristol University , Bristol Robotics Laboratory |
Keywords: Biomimetics, Neurorobotics, Biologically-Inspired Robots
Abstract: Identifying objects using sparse tactile sensor arrays requires movement across the surface and the integration of sensory information to support hypotheses. In this study we demonstrate a surface placement control strategy to orient and position a biomimetic tactile whisker array relative to the object surface. This reduces the variance in tactile view, or representation, of each region of the object surface thus relaxing the prior pose dependence for object recall. In addition we evaluate two regional search strategies in an attempt to further improve the robustness and speed of object classification.
|
|
12:00-12:15, Paper TuAT17.5 | Add to My Program |
On-Chip Three-Dimension Cell Rotation Using Whirling Flows Generated by Oscillating Asymmetrical Microstructures |
Song, Bin | BEIHANG UNIVERSITY |
Feng, Yanmin | Beihang University |
Zhou, Qiang | Beihang University |
Feng, Lin | Beihang University |
Keywords: Biological Cell Manipulation, Micro/Nano Robots
Abstract: The capability to precisely rotate cells and other micrometer-sized biological samples is invaluable in biomedical, bioengineering, and biophysical applications. We propose a novel on-chip three-dimension (3D) cell rotation method based on whirling flows created by oscillating asymmetrical microstructures. In an acoustic field excited by the vibration of a piezoelectric transducer, two different modes of microvortices were generated around the specially designed microstructures, which were utilized to precisely achieve the in-plane and out-of-plane rotational manipulation. Experiments of the different sizes of microparticles and oocytes are conducted to demonstrate the claimed functions. We also investigated the effect of various parameters on the acoustically induced flow such as the frequency, the driving voltage and the distance from the microstructure tip to the oocyte center, indicating the rotational rate can be effectively tuned on demand for single-cell studies. Rotation by using this method demonstrated overall out-of-plane and in-plane rotational orientation control in a quite simple way. And comparing with the conventional works, our acoustofluidic cell rotation approach is simple-to-fabricate and easy-to-operate, permitting rotation irrespective of the magnetic, optical, or electrical properties of the specimen under investigation.
|
|
12:15-12:30, Paper TuAT17.6 | Add to My Program |
Adaptive Dynamic Control for Magnetically Actuated Medical Robots |
Barducci, Lavinia | University of Leeds |
Pittiglio, Giovanni | University of Leeds |
Norton, Joseph | University of Leeds |
Obstein, Keith | Vanderbilt University |
Valdastri, Pietro | University of Leeds |
Keywords: Robust/Adaptive Control of Robotic Systems, Medical Robots and Systems, Dynamics
Abstract: In the present work we discuss a novel dynamic control approach for magnetically actuated robots, by proposing an adaptive control technique, robust towards parametric uncertainties and unknown bounded disturbances. The former generally arise due to partial knowledge of the robots’ dynamic parameters, such as inertial factors, the latter are the outcome of unpredictable interaction with unstructured environments. In order to show the application of the proposed approach, we consider controlling the Magnetic Flexible Endoscope (MFE) which is composed of a soft-tethered Internal Permanent Magnet (IPM), actuated with a single External Permanent Magnet (EPM). We provide with experimental analysis to show the possibility of levitating the MFE - one of the most difficult tasks with this platform - in case of partial knowledge of the IPM’s dynamics and no knowledge of the tether’s behaviour. Experiments in an acrylic tube show a reduction of contact of the 32% compared to non-levitating techniques and 1.75 times faster task completion with respect to levitating techniques. More realistic experiments, performed in a colon phantom, show that levitating the capsule achieves faster and smoother exploration and that the minimum time for completing the task is attained by the proposed approach.
|
|
TuAT18 Regular session, LG-R18 |
Add to My Program |
Localization I |
|
|
Chair: Tan, Xiaobo | Michigan State University |
Co-Chair: Akai, Naoki | Nagoya University |
|
11:00-11:15, Paper TuAT18.1 | Add to My Program |
Low-Cost Sonar Navigation System |
Guadagnino, Tiziano | Sapienza University of Rome |
Della Corte, Bartolomeo | Sapienza University of Rome |
Grisetti, Giorgio | Sapienza University of Rome |
Keywords: Localization, Calibration and Identification, Mapping
Abstract: In this paper, we present a sonar-based navigation system, designed to deploy a fleet of autonomous mobile platforms at a reasonable cost. In educational and hobbyist contexts, a large number of robots is required. By means of classical navigation approaches, every robot should be provided with accurate vision or range sensors. This limits the maximum number of robots in the fleet, due to the unaffordable cost of these sensors. In contrast to that, our system requires a single platform equipped with a higher quality sensor, used to perform calibration and mapping tasks. The rest of the fleet, able to localize and navigate, is equipped solely with low-cost sonars, providing a notable reduction in the overall cost. We achieve this task by presenting a novel calibration procedure to estimate the sonars extrinsic, and by adapting a classical Monte Carlo localization algorithm to the sonar model, focusing on efficiency. We release an open source implementation of the system to the community.
|
|
11:15-11:30, Paper TuAT18.2 | Add to My Program |
Misalignment Recognition Using Markov Random Fields with Fully Connected Latent Variables for Detecting Localization Failures |
Akai, Naoki | Nagoya University |
Morales Saiki, Luis Yoichi | Nagoya University |
Hirayama, Takatsugu | Nagoya University |
Murase, Hiroshi | Nagoya University |
Keywords: Localization, Failure Detection and Recovery, Recognition
Abstract: Recognizing misalignment between sensor measurements and objects that exist on a map due to inaccuracies in localization estimation is challenging. This can be attributed to the fact that the sensor measurements are individually modelled for solving the localization problem, resulting in entire relations of the measurements being ignored. This paper proposes a misalignment recognition method using Markov random fields with fully connected latent variables for the detection of localization failures. The proposed method estimates the classes of each sensor measurement that are aligned, misaligned, and obtained from unknown obstacles. The full connection allows us to consider the entire relation of the measurements. A misalignment can be exactly recognized even when partial sensor measurements overlap with mapped objects. Based on the class estimation results, we are able to distinguish whether the localization has failed or not. The proposed method was compared with six alternative methods, including a convolutional neural network, using datasets composed of success and failure localization samples. Experimental results show that classification accuracy of the localization samples using the proposed method exceeds 95 % and outperforms the other examined methods.
|
|
11:30-11:45, Paper TuAT18.3 | Add to My Program |
Outlier-Robust State Estimation for Humanoid Robots |
Piperakis, Stylianos | Foundation for Research and Technology – Hellas (FORTH) |
Kanoulas, Dimitrios | Istituto Italiano Di Tecnologia |
Tsagarakis, Nikos | Istituto Italiano Di Tecnologia |
Trahanias, Panos | Foundation for Research and Technology – Hellas (FORTH) |
Keywords: Localization, Humanoid and Bipedal Locomotion, Sensor Fusion
Abstract: Contemporary humanoids are equipped with visual and LiDAR sensors that are effectively utilized for Visual Odometry (VO) and LiDAR Odometry (LO). Unfortunately, such measurements commonly suffer from outliers in a dynamic environment, since frequently it is assumed that only the robot is in motion and the world is static. To this end, robust state estimation schemes are mandatory in order for humanoids to symbiotically co-exist with humans in their daily dynamic environments. In this article, the robust Gaussian Error-State Kalman Filter for humanoid robot locomotion is presented. The introduced method automatically detects and rejects outliers without relying on any prior knowledge on measurement distributions or finely tuned thresholds. Subsequently, the proposed method is quantitatively and qualitatively assessed in realistic conditions with the full-size humanoid robot WALK-MAN v2.0 and the mini-size humanoid robot NAO to demonstrate its accuracy and robustness when outlier VO/LO measurements are present. Finally, in order to reinforce further research endeavours, our implementation is released as an open-source ROS/C++ package.
|
|
11:45-12:00, Paper TuAT18.4 | Add to My Program |
Robust Outdoor Self-Localization in Changing Environments |
Muhammad, Haris | Frankfurt University of Applied Sciences |
Franzius, Mathias | Honda Research Institute (HRI) |
Bauer Wersing, Ute | Frankfurt University of Applied Sciences |
Keywords: Localization, Mapping, Service Robots
Abstract: In outdoor scenarios changing conditions (e.g., seasonal, weather and lighting effects) have a substantial impact on the appearance of a scene, which often prevents successful visual localization. The application of an unsupervised Slow Feature Analysis (SFA) on the images captured by an autonomous robot enables self-localization from a single image. However, changes occurring during the training phase or over a more extended period can affect the learned representations. To address the problem, we propose to join long-term recordings from an outdoor environment based on their position correspondences. The established hierarchical model trained on raw images performs well, but as an extension, we extract Fourier components of the views and use that for learning of spatial representations, which reduces the computation time and makes it adequate to run on an ARM embedded system. We present the experimental results from a simulated environment and real-world outdoor recordings collected over a full year, which has effects like different day time, weather, seasons and dynamic objects. Results show an increasing invariance w.r.t. changing conditions over time, thus an outdoor robot can improve its localization performance during operation.
|
|
12:00-12:15, Paper TuAT18.5 | Add to My Program |
Randomized Sensor Selection for Nonlinear Systems with Application to Target Localization |
Bopardikar, Shaunak D. | Michigan State University |
Ennasr, Osama | Michigan State University |
Tan, Xiaobo | Michigan State University |
Keywords: Localization, Probability and Statistical Methods, Sensor Networks
Abstract: Given a nonlinear dynamical system, this paper considers the problem of selecting a subset of the total set of sen- sors that has provable guarantees on standard metrics related to the nonlinear observability Gramian. The key contribution is a simple randomized algorithm that samples the sensors uni- formly without replacement, and yields probabilistic guarantees on the minimum eigenvalue or the inverse of the condition number of the nonlinear observability Gramian relative to that of the complete set of sensors. Numerical studies reveal that the utility of the theoretical results lies in the regime of large total number of sensors wherein the combinatorial nature of the problem presents a significant computational challenge. The results are demonstrated numerically on a problem of moving target localization using an Extended Kalman Filter (EKF) in two scenarios: one using range sensors and another with time-difference-of-arrival (TDOA) measurements. A graceful degradation of performance with a decreased number of sensors is observed when compared to the use of all of the sensors for localization. It is also observed that for certain metrics, the proposed approach provides an improvement over a heuristic that selects the sensors in a greedy manner based on the contribution of an additional sensor toward the observability Gramian metric.
|
|
12:15-12:30, Paper TuAT18.6 | Add to My Program |
On the Bayes Filter for Shared Autonomy |
Luft, Lukas | Freiburg University |
Boniardi, Federico | University of Freiburg |
Schaefer, Alexander | Freiburg University |
Büscher, Daniel | Albert-Ludwigs-Universität Freiburg |
Burgard, Wolfram | University of Freiburg |
Keywords: Localization, Probability and Statistical Methods
Abstract: The Bayes filter is the basis for many state-of-the-art on-line robot localization algorithms. In the literature, its derivation typically requires the robot controls to be chosen independently of all other variables. However, this assumption is not valid for a robotic system which is to act purposefully. The contribution of this paper is twofold: We prove that the Bayes filter is also exact for an autonomous system which chooses the controls not randomly but depending on any subset of all observable variables. We further show how to augment the filter if a human agent chooses controls on the basis of parts of the state space that are not directly accessible to the robot. In this case, modeling the agent's purpose improves the pose estimate as the control provides additional information about the hidden state space. A careful derivation of the Bayes filter then leads to an additional pseudo measurement update step. Real-world experiments with our teleoperated mobile robot and evaluations on the KITTI dataset show that the localization accuracy significantly improves if we augment a particle filter with this pseudo measurement update. We further present an analytical example for an augmented Kalman filter, which leads to a more accurate estimate than the standard Kalman filter.
|
|
TuAT19 Regular session, LG-R19 |
Add to My Program |
AI-Based Methods for Robotics |
|
|
Chair: Tan, Jindong | University of Tennessee, Knoxville |
Co-Chair: Kyrki, Ville | Aalto University |
|
11:00-11:15, Paper TuAT19.1 | Add to My Program |
A Model-Based Human Activity Recognition for Human-Robot Collaboration |
Lee, Sang Uk | Massachusetts Institute of Technology |
Hofmann, Andreas | MIT |
Williams, Brian | MIT |
Keywords: AI-Based Methods, Recognition, Failure Detection and Recovery
Abstract: Human activity recognition is a crucial ingredient in safe and efficient human–robot collaboration. In this paper, we present a new model-based human activity recognition system called logical activity recognition system (LCARS). LCARS requires much less training data compared to learning-based works. Compared to other model-based works, LCARS requires minimal domain-specific modeling effort from users. The minimal modeling is for two reasons: i) we provide a systematic and intuitive way to encode domain knowledge for LCARS and ii) LCARS automatically constructs a probabilistic estimation model from the domain knowledge. Requiring minimal training data and modeling effort allows LCARS to be easily applicable to various scenarios. We verify this through simulations and experiments.
|
|
11:15-11:30, Paper TuAT19.2 | Add to My Program |
Augmenting Knowledge through Statistical, Goal-Oriented Human-Robot Dialog |
Amiri, Saeid | SUNY Binghamton |
Bajracharya, Sujay | Cleveland State University |
Goktolga, Cihangir | SUNY Binghamton |
Thomason, Jesse | University of Washington |
Zhang, Shiqi | SUNY Binghamton |
Keywords: Probability and Statistical Methods, Service Robots
Abstract: Some robots can interact with humans using natural language, and identify service requests through human-robot dialog. However, few robots are able to improve their language capabilities from this experience. In this paper, we develop a dialog agent for robots that is able to interpret user commands using a semantic parser, while asking clarification questions using a probabilistic dialog manager. This dialog agent is able to augment its knowledge base and improve its language capabilities by learning from dialog experiences, e.g., adding new entities and learning new ways of referring to existing entities. We have extensively evaluated our dialog system in simulation as well as with human participants through MTurk and real-robot platforms. We demonstrate that our dialog agent performs better in efficiency and accuracy in comparison to baseline learning agents. Demo video can be found at url{https://youtu.be/DFB3jbHBqYE}
|
|
11:30-11:45, Paper TuAT19.3 | Add to My Program |
Inverse Dynamics Modeling of Robotic Manipulator with Hierarchical Recurrent Network |
Sun, Pengfei | Capital Normal University |
Shao, Zhenzhou | Capital Normal University |
Qu, Ying | The University of Tennessee, Knoxville |
Guan, Yong | Capital Normal University |
Tan, Jindong | University of Tennessee, Knoxville |
Keywords: AI-Based Methods, Dynamics, Industrial Robots
Abstract: Inverse dynamics modeling is a critical problem for the computed-torque control of robotic manipulator. This paper presents a novel recurrent network based on the modified Simple Recurrent Unit (SRU) with hierarchical memory (SRU-HM), which is achieved by the nested SRU structure. In this way, it enables the capability to retain the long-term information in the distant past, compared with the conventional stacked structure. The hidden state of SRU is able to provide more complete information relevant to current prediction. Experimental results demonstrate that the proposed method can improve the accuracy of dynamics model greatly, and outperforms the state-of-the-art methods.
|
|
11:45-12:00, Paper TuAT19.4 | Add to My Program |
Bayesian Optimization for Policy Search in High-Dimensional Systems Via Automatic Domain Selection |
Froehlich, Lukas | Robert Bosch GmbH |
Klenske, Edgar | Bosch Research |
Daniel, Christian | Bosch |
Zeilinger, Melanie N. | ETH Zurich |
Keywords: AI-Based Methods, Learning and Adaptive Systems
Abstract: Bayesian Optimization (BO) is an effective method for optimizing expensive-to-evaluate black-box functions with a wide range of applications for example in robotics, system design and parameter optimization. However, scaling BO to problems with large input dimensions (> 10) remains an open challenge. In this paper, we propose to leverage results from optimal control to scale BO to higher dimensional control tasks and to reduce the need for manually selecting the optimization domain. The contributions of this paper are twofold: 1) We show how we can make use of a learned dynamics model in combination with a model-based controller to simplify the BO problem by focusing onto the most relevant regions of the optimization domain. 2) Based on (1) we present a method to find an embedding in parameter space that reduces the effective dimensionality of the optimization problem. To evaluate the effectiveness of the proposed approach, we present an experimental evaluation on real hardware, as well as simulated tasks including a 48-dimensional policy for a quadcopter.
|
|
12:00-12:15, Paper TuAT19.5 | Add to My Program |
Long-Term Prediction of Motion Trajectories Using Path Homology Clusters |
Carvalho, Joao Frederico | KTH Royal Technical Institute |
Vejdemo-Johansson, Mikael | CUNY College of Staten Island |
Pokorny, Florian T. | KTH Royal Institute of Technology |
Kragic, Danica | KTH |
Keywords: AI-Based Methods, Learning from Demonstration, Human Detection and Tracking
Abstract: In order for robots to share their workspace with people, they need to reason about human motion efficiently. In this work we leverage large datasets of paths in order to infer local models that are able to perform long-term predictions of human motion. Further, since our method is based on simple dynamics, it is conceptually simple to understand and allows one to interpret the predictions produced, as well as to extract a cost function that can be used for planning. The main difference between our method and similar systems, is that we employ a map of the space and translate the motion of groups of paths into vector fields on that map. We test our method on synthetic data and show its performance on the Edinburgh forum pedestrian long-term tracking dataset [1] where we were able to outperform a Gaussian Mixture Model tasked with extracting dynamics from the paths.
|
|
12:15-12:30, Paper TuAT19.6 | Add to My Program |
Feedback-Based Fabric Strip Folding |
Petrik, Vladimir | Aalto University |
Kyrki, Ville | Aalto University |
Keywords: AI-Based Methods, Manipulation Planning, Motion Control
Abstract: Accurate manipulation of a deformable body such as a piece of fabric is difficult because of its many degrees of freedom and unobservable properties affecting its dynamics. To alleviate these challenges, we propose the application of feedback-based control to robotic fabric strip folding. The feedback is computed from the low dimensional state extracted from a camera image. We trained the controller using reinforcement learning in simulation which was calibrated to cover the real fabric strip behaviors. The proposed feedback-based folding was experimentally compared to two state-of-the-art folding methods and our method outperformed both of them in terms of accuracy.
|
|
TuAT20 Regular session, LG-R20 |
Add to My Program |
Biologically-Inspired Robots I |
|
|
Chair: Yun, Dongwon | Daegu Gyeongbuk Institute of Science and Technology (DGIST) |
Co-Chair: Asano, Yuki | The University of Tokyo |
|
11:00-11:15, Paper TuAT20.1 | Add to My Program |
Guinea Fowl Jumping Robot with Balance Control Mechanism: Modeling, Simulation, and Experiment Results |
Kim, Myeong-Jin | Daegu Gyeonbuk Institute of Science and Technology |
Yun, Dongwon | Daegu Gyeongbuk Institute of Science and Technology (DGIST) |
Keywords: Biologically-Inspired Robots, Biomimetics, Dynamics
Abstract: Recently, diverse research has actively been conducted to control the posture of jumping robots using an inertial tail mechanism. However, the inertial tail mechanism has a high probability of collision with obstacles. In this study, a momentum wheel mechanism is proposed to achieve the same attitude control performance while reducing the volume occupied by the inertial tail mechanism. To verify the performance of the momentum wheel mechanism, we proposed a jumping robot with a momentum wheel mechanism and performed a dynamic analysis, simulation, and experiments on a jumping robot with a momentum wheel mechanism. In addition, it has been demonstrated that the momentum wheel mechanism can contribute to control of the body angle of the jumping robot. As a result, the momentum wheel mechanism can enhance the stability of the jumping robot more than the tail mechanism, and the momentum wheel mechanism contributes to the attitude control of the body angle, which allows the jumping robot to perform continuous jumping.
|
|
11:15-11:30, Paper TuAT20.2 | Add to My Program |
Carpie: A Soft, Mechanically-Reconfigurable Worm Robot |
Ahmadian, Pouya | University of Toronto (UofT) |
Natividad, Rainier | National University of Singapore |
Yeow, Chen-Hua | National University of Singapore |
Keywords: Biologically-Inspired Robots, Biomimetics, Soft Robot Materials and Design
Abstract: Soft robots are a burgeoning archetype in robotics due to their ability to perform intricate movements easily and seamlessly. They serve as an ideal concept for realistically emulating animal movements. However, the majority of soft robots today are unable to vary their motions due to the coupled interaction of the nature of their kinematics and structural composition. We therefore created Carpie, a robotic caterpillar adapted from a modular, pneumatic actuator. Carpie is designed to perform a variety of caterpillar movements by tuning its mechanical structure through physical reconfiguration. The robot has a completely soft body that enables a variety of movements and contortions into different shapes, making Carpie is perhaps the perfect example to showcase a soft robot’s aptness for animal mimicry. We analyzed the robot’s control aspect. Its capabilities were measured by performing gait velocity studies on three different configurations. Each configuration was executed by applying various modules on Carpie’s structure without fully refabricating the entire assembly. We also analyzed the accompanying change in the robot’s maximum height during gait. Finally, we demonstrate how physical reconfiguration was able to radically alter Carpie’s movement, allowing it to perform a turning motion.
|
|
11:30-11:45, Paper TuAT20.3 | Add to My Program |
A Spring-Aided Two-Dimensional Electromechanical Spine Architecture for Bio-Inspired Robots |
Ku, Bonhyun | University of Illinois at Urbana-Champaign |
Wang, Sunyu | University of Illinois at Urbana-Champaign |
Banerjee, Arijit | University of Illinois at Urbana-Champaign |
Keywords: Biologically-Inspired Robots, Biomimetics
Abstract: This paper presents a distributed six-link two-dimensional electromechanical actuator that emulates a biological spine. The gearless actuator is made by stacking modules of E-shaped cores along with integrated springs. A coil excitation induces magnetic flux in the core, which produces electromechanical force in the air gap between two adjacent cores. The proposed actuator is driven by dc-dc converters with current control. Electromechanical force analysis for the proposed spine architecture with different air-gap distances and spring analysis that improve the actuator's performance are discussed. Experimental results show different biological motions with a prototype design. The prototype can produce a maximum torque of 1.4 N-m.
|
|
11:45-12:00, Paper TuAT20.4 | Add to My Program |
Effect of Arm Swinging and Trunk Twisting on Bipedal Locomotion |
Onishi, Ryo | Osaka Institute of Technology |
Kitamura, Ryoma | Osaka Institute of Technology |
Takuma, Takashi | Osaka Institute of Technology |
Kase, Wataru | Osaka Institute of Technology |
Keywords: Biologically-Inspired Robots, Humanoid and Bipedal Locomotion, Compliant Joint/Mechanism
Abstract: It is expected that human locomotion, in which the trunk twists and arms swing, provides stability as well as compensation for the angular momentum because the trunk and arms are placed over legs and then their position and the acceleration influence the bipedal locomotion. This paper presents a bipedal robot equipped with a trunk, embedding a viscoelastic joint and two arms away from the midline of the trunk. Physical experiments show that a passive oscillation of the viscoelastic trunk joint around the yaw axis vertical to the ground and a swing of heavy arms provided an oscillation of zero moment point (ZMP) in the lateral direction and a small oscillation in the anteroposterior direction. This is despite the fact that the robot does not have a roll joint and the arm swings back and forth. Numerical analysis supports the results of the physical experiments through a simple model and by expressing ZMP trajectory, showing that the arm swinging facilitates anti-phase locomotion in which right arm and left leg swings ahead and vice versa.
|
|
12:00-12:15, Paper TuAT20.5 | Add to My Program |
Proto-Object Based Saliency for Event-Driven Cameras |
Iacono, Massimiliano | Istituto Italiano Di Tecnologia |
D'Angelo, Giulia | Istituto Italiano Di Tecnologia |
Glover, Arren | Istituto Italiano Di Tecnologia |
Tikhanoff, Vadim | Italian Institute of Technology |
Niebur, Ernst | John Hopkins University |
Bartolozzi, Chiara | Istituto Italiano Di Tecnologia |
Keywords: Biologically-Inspired Robots, Humanoid Robots, Computer Vision for Other Robotic Applications
Abstract: Autonomous robots can rely on attention mechanisms to explore complex scenes and select salient stimuli relevant for behaviour. Stimulus selection should be fast to efficiently allocate available (and limited) computational resources to process in detail a subset of the otherwise overwhelmingly large sensory input. The amount of processing required is a product of the amount of data sampled by a robot's sensors; while a standard RGB camera produces a fixed amount of data for every pixel of the sensor, an textit{event-camera} produces data only for where there is a contrast change in the field of view, and does so with a lower latency. In this paper, we describe the implementation of a state-of-the-art bottom-up attention model, based on structuring the visual scene in terms of proto-objects. As an event-camera encodes different visual information compared to frame-based cameras, the original algorithm must be adapted and modified. We find that the event-camera's inherent detection of edges removes the need for some early stages of processing in the model. We describe the modifications, compare the event-driven algorithm to the original, and validate the potential for use on the iCub humanoid robot.
|
|
12:15-12:30, Paper TuAT20.6 | Add to My Program |
Task-Specific Self-Body Controller Acquisition by Musculoskeletal Humanoids: Application to Pedal Control in Autonomous Driving |
Kawaharazuka, Kento | The University of Tokyo |
Tsuzuki, Kei | University of Tokyo |
Makino, Shogo | The University of Tokyo |
Onitsuka, Moritaka | The University of Tokyo |
Shinjo, Koki | The University of Tokyo |
Asano, Yuki | The University of Tokyo |
Okada, Kei | The University of Tokyo |
Kawasaki, Koji | The University of Tokyo |
Inaba, Masayuki | The University of Tokyo |
Keywords: Biologically-Inspired Robots, Learning and Adaptive Systems, Humanoid Robots
Abstract: The musculoskeletal humanoid has many benefits that human beings have, but the modeling of its complex flexible body is difficult. Although we have developed an online acquisition method of the nonlinear relationship between joints and muscles, we could not completely match the actual robot and its self-body image. When realizing a certain task, the direct relationship between the control input and task state needs to be learned. So, we construct a neural network representing the time-series relationship between the control input and task state, and realize the intended task state by applying the network to a real-time control. In this research, we conduct accelerator pedal control experiments as one application, and verify the effectiveness of this study.
|
|
TuPS1 Poster session, L1-R0, Zone I |
Add to My Program |
Late Breaking Result Poster Session 1 |
|
|
|
12:30-13:30, Paper TuPS1.1 | Add to My Program |
Position Control of Wire-Suspended Hand for Long-Reach Aerial Manipulation |
Miyazaki, Ryo | Ritsumeikan University |
Paul, Hannibal | Ritsumeikan University |
Shimonomura, Kazuhiro | Ritsumeikan University |
Keywords: Aerial Systems: Applications, Aerial Systems: Mechanics and Control, Field Robots
Abstract: We propose an aerial robot system with a long-reach manipulator to allow the manipulated target to be at the desired distance away from the airframe. The system consists of a hexrotor with winch mechanism and wire-suspended hand. The winch mechanism is used to adjust the distance of the manipulator from the unmanned aerial vehicle (UAV). The wire-suspended hand is constructed with 4 ducted fans, IMU sensor, dual-gripper and vision system. The hand is autonomously able to stabilize its swing motion by controlling the ducted fans and position itself to the desired target using the vision system. Through experiment, we verified that for a 3m suspended hand, it is possible to control its position within 0.6m in X axis and 0.24m in Y axis of the hand.
|
|
12:30-13:30, Paper TuPS1.2 | Add to My Program |
Hong Hu - an Efficient and Versatile Tail-Sitter VTOL UAV Platform: Design, Implementation and Control |
Xu, Wei | University of Hong Kong |
Gu, Haowei | Hong Kong University of Science and Technology |
Zhang, Fu | University of Hong Kong |
Keywords: Aerial Systems: Mechanics and Control
Abstract: This work introduces the design, implementation, and control of an efficient and versatile tail-sitter VTOL UAV platform - Hong Hu. A multidisciplinary design and optimization framework is used to optimize the aircraft’s aerodynamic configuration and propulsion system. Then the carbon fiber structured aircraft is manufactured based on the optimal design. The resulting Hong Hu UAV platform could be used for various outdoor applications. In this abstract, we demonstrate its application in large-scale surveying and mapping
|
|
12:30-13:30, Paper TuPS1.3 | Add to My Program |
Adapting Weed Growth Predictions for Mechanical Weeding Agbots |
McAllister, Wyatt | University of Illinois at Urbana-Champaign |
Whitman, Joshua | University of Illinois |
Davis, Adam | USDA-ARS in Urbana, IL |
Chowdhary, Girish | University of Illinois at Urbana Champaign |
Keywords: Agricultural Automation, Multi-Robot Systems, Planning, Scheduling and Coordination
Abstract: Consider a team of mechanical-weeding robots managing herbicide-resistant weeds on any row-crop farm in the US. This team of robots needs to predict the weed growth across the whole farm in order to make intelligent decisions on robot coordination cite{mcallister2018weed}. However, none of the robots can observe the whole field, and even together, they can only observe a very limited part of the field at any given time. Furthermore, the robots do not have high bandwidth access to a High-Performance Computing (HPC) facility, nor would such a facility be able to provide highly accurate predictions of weed growth without real-time access to weed growth data being gathered through the weeding process, as weed distribution depends highly on the field and the day. The number of emerged seedlings at a given location in the field is governed by the latent seed bank density within the soil at that location, which vary randomly over different fields. For this reason, it is necessary to adapt any model trained on general weed growth data to the specific field. We enable our team of robots to efficiently learn a predictive model of the specific field by adapting a general model which was trained using data from many fields. The predictive model enables this team of robots to pursue more intelligent coordination strategies, leveraging our earlier work cite{mcallister2018weed}, rather than resorting to grid search over the whole field or using pre-determined heuristics to plan its actions. Our work addresses the key challenge facing multi-agent field-robotic systems operating in spatiotemporally changing domains: managing the trade-off between learning a robust generalized model, and learning a model which is specialized in order to most accurately explain the current observations of the agents. Our work is based on the Evolving-Gaussian Process model, and is trained on ecologically realistic simulations of weed growth cite{mcallister2018weed}. The E-GP model has the valuable benefit of being able to predict spatiotemporal dynamics throughout a continuous domain using a finite linear dynamical system embedded in a feature space. We advance this modeling technique by showing the model may be decoupled into two components, the first based on aggregated past observations, and the second adapted in real-time to capture the unique properties of the current environment.
|
|
12:30-13:30, Paper TuPS1.4 | Add to My Program |
HouseExpo: A Large-Scale 2D Indoor Layout Dataset for Learning-Based Algorithms |
Li, Tingguang | The Chinese University of Hong Kong |
Ho, Danny | The Chinese University of Hong Kong |
Li, Chenming | The Chinese University of Hong Kong |
Zhu, Delong | The Chinese University of Hong Kong |
Wang, Chaoqun | The Chinese University of HongKong |
Meng, Max Q.-H. | The Chinese University of Hong Kong |
Keywords: AI-Based Methods, Deep Learning in Robotics and Automation
Abstract: With the significant achievements in the AI field, the investigation of learning-based methods in robotics area has received more and more attention in recent years. However, for the learning-based methods, the issue of data requirement must be addressed first since the size and diversity of training data are crucial for the performance. However, it is still challenging for the existing datasets of 2D environments to meet such demand since their size as well as the variability are limited. On the other hand, the processing time to build the map through Simultaneous Localization And Mapping (SLAM) is time-costing, which is a bottleneck in training the neuron networks which routinely involves millions of trial-and-error episodes. These issues motivate the authors to develop a large-scale dataset, HouseExpo, and a fast simulation platform, Pseudo-SLAM, to improve the training efficiency.
|
|
12:30-13:30, Paper TuPS1.5 | Add to My Program |
Urban Street Trajectory Prediction with Multi-Class LSTM Networks |
Li, Xin | Meituan-Dianping Group |
Zhu, Yanliang | Meituan-Dianping |
Qian, Deheng | MeiTuan |
Ren, Dongchun | Meituan-Dianping |
Keywords: AI-Based Methods, Deep Learning in Robotics and Automation, Probability and Statistical Methods
Abstract: Predicting the trajectory of traffic-agents is an essential module for autonomous driving system. It is a great challenge on urban streets due to various lighting conditions and traffic densities. In this paper, we propose multi-class LSTM encoder-decoder networks which involve multi-branches. Each branch has an encoder-decoder structure and predict one class traffic-agent. For each agent, we predict n trajectories by different noises and select better trajectory by manual rule.
|
|
12:30-13:30, Paper TuPS1.6 | Add to My Program |
Depth-Image-Based Textureless-Object Picking by DCNN and Visual Servoing |
Jiang, Ping | Toshiba Corporation |
Ishihara, Yoshiyuki | Toshiba Corporation |
Sugiyama, Nobukatsu | Toshiba Corporation |
Oaki, Junji | Toshiba Corporation |
Tokura, Seiji | Corporation |
Sugahara, Atsushi | Toshiba Corporation |
Ogawa, Akihito | TOSHIBA CORPORATION |
Keywords: AI-Based Methods, Grasping, Big Data in Robotics and Automation
Abstract: A robot picking system incorporating a deepconvolutional-neural-network (DCNN) and depth-image-based visual servoing (VS) is presented for textureless-object picking. The DCNN predicts the grasp point belonging to the highest graspable surface among a pile of boxes. Then, based on the predicted grasp point, the visual servoing system tracks the graspable surface and controls the manipulator to pick the target box. The system has been validated by an experiment for consecutively picking 10 piled up boxes. As a result, the DCNN successfully predicted the grasp points, and the success rate of the picking task was 87.3%.
|
|
12:30-13:30, Paper TuPS1.7 | Add to My Program |
Explainable One-Shot Meta-Learning to Imitate Motion Segments of Unseen Human-Robot Interactions |
Tian, Nan | University of California, Berkeley |
Tanwani, Ajay Kumar | UC Berkeley |
Sojoudi, Somayeh | UC Berkeley |
Keywords: AI-Based Methods, Learning from Demonstration, Social Human-Robot Interaction
Abstract: Hybrid Cloud-Edge robotic systems are scalable, adaptable, and intelligent. However, network latency between the Cloud and the Edge presents a problem for responsive human-robot interactions (HRI). Generative models shared between the Cloud and the Edge for motion segmentation and synthesis can help mitigate network latency, as we showed previously when a human tele-operated a dynamic robot to draw handwritten letters in two phases: (i) direct tele-operation with latency; (ii) automatic letter drawing so the robot completes the letter before the tele-operator. These generative models are GMM-HSMM models trained with supervision on labels of alphabetical letters. Similar motion segments across different letters can represent unlabeled structural concepts. Discovering and interpreting these concepts can help us train more expressive generative models for motion segmentation. It can also help us train a new model faster from previously unseen samples based on an ensemble of models. In this work, we use one-shot meta-learning to observe an unseen demonstration to: (i) discover and explain motion segments within the unseen demonstrations automatically; (ii) reconstruct a new generative model from ensemble of models for human-robot interactive controls.
|
|
12:30-13:30, Paper TuPS1.8 | Add to My Program |
Data-Based Modeling of Contact State in Robotic Assembly |
Na, Minwoo | Korea University |
Song, Jae-Bok | Korea University |
Keywords: Assembly, Contact Modelling
Abstract: Since robotic assembly is involved in contact between assembly parts, the robot or parts may be damaged when the assembly fails. To prevent this, the contact force generated at contact should be monitored. In this study, the force and torque data generated during the battery assembly using a 6 DOF industrial robot is statistically modeled in the time domain using the Gaussian process.
|
|
12:30-13:30, Paper TuPS1.9 | Add to My Program |
Complexity Conditioned Goals for Reinforcement Learning Agents |
Gupta, Kashish | University of British Columbia |
Najjaran, Homayoun | University of British Columbia |
Keywords: Autonomous Agents, AI-Based Methods
Abstract: The proposed work aims at the introduction and utilization of a complexity conditioned goals technique to enable a Reinforcement Learning Agent to train efficiently. RL agents (Sparse or Densely rewarded, Multi-Goal), in general, pursue randomly sampled goals irrespective of their training status or "ability". A premature agent (early phases of learning) seeking a complex goal extends training time and challenges as the episode rollout would most certainly fail and not contribute to learning. The proposed method bases the sampled goals on an exponentially growing complexity factor that sets "easier" goals for the agent improving the likelihood of a successful rollout. The experiments executed on a Bit Flipping Environment shows promising results and seamless coordination with the existing HER technique.
|
|
12:30-13:30, Paper TuPS1.10 | Add to My Program |
Design of a Mobile Robot for the Treatment, Reuse and Removal of Manure with Monitoring of Environmental Variables for Poultry Farms |
Velasco, Luis | Pontificia Universidad Católica Del Perú |
Hilario Poma, Javier Alfredo | Pontificia Universidad Católica Del Perú |
Gonzales, Julio | Pontificia Universidad CatÓlica Del PerÚ |
Rodriguez, Laureano | Pontificia Universidad Católica Del Perú |
Cuellar, Francisco | Pontificia Universidad Catolica Del Peru |
Keywords: Autonomous Vehicle Navigation, Computer Vision for Other Robotic Applications, Environment Monitoring and Management
Abstract: The objective of this work is to develop a semi-autonomous robot for the treatment, reuse and removal of poultry manure. With this robot, it’s expected to increase the productive capacity of each farm up to an additional of 20% during each productive cycle and reduce the exposure of the operators to possible sources of infections (due to the manure). The robot is designed for carving the manure bed in all the poultry warehouse implemented with a path following controller and computer vision algorithms. Likewise, the robot is equipped with a system that sprays medicine and sensors connected to the wireless system in order to get information about important environment parameters (temperature, luminosity, humidity, ammoniac and pressure).
|
|
12:30-13:30, Paper TuPS1.11 | Add to My Program |
Autonomous Human-Aware Navigation in Dense Crowds |
Yao, Xinjie | Hong Kong University of Science and Technology |
Zhang, Ji | Carnegie Mellon University |
Oh, Jean | Carnegie Mellon University |
Keywords: Autonomous Vehicle Navigation, Social Human-Robot Interaction, Collision Avoidance
Abstract: In densely populated environments, human-aware navigation is critical for autonomous robots where interaction is expected. Human-aware navigation is challenging given the constraints of human comfort and social rules. Traditional trajectory prediction-based methods are unreliable in dense crowds because the history of individual motion is often not precisely attainable. Other approaches such as reinforcement learning fail to address the naturalness aspect in socially compliant navigation. We present an autonomous navigation system capable of operating in dense crowds and complying with latent social rules. The underlying system incorporates a deep neural network to mimic human motion patterns. In experiments, our robot vehicle is able to drive naturally in a densely populated area (10+ people in a 10mx20m area).
|
|
12:30-13:30, Paper TuPS1.12 | Add to My Program |
Graph Element Networks: A Flexible Model for Robotic Applications |
Alet, Ferran | MIT |
Jeewajee, Adarsh K. | MIT |
Bauza Villalonga, Maria | Massachusetts Institute of Technology |
Rodriguez, Alberto | Massachusetts Institute of Technology |
Kaelbling, Leslie | MIT |
Lozano-Perez, Tomas | MIT |
Keywords: Big Data in Robotics and Automation, AI-Based Methods, Grasping
Abstract: We present Graph Element Networks, an architecture that uses Graph Neural Networks (GNNs) as a flexible model for physical systems. In contrast to previous work, nodes in the GNNs are not tied to objects or object subparts and are instead used to model computational meshes where the number and position of the nodes can be chosen by the user. We show this has three particularly useful applications for robotics: first, we can choose the runtime complexity of the network at test time by using more or less nodes, trading off compute with model accuracy. Second, GENs can be used for end-effector design by putting a mesh on top of the gripper, predicting the probability of grasps being successful and then back-propagating this probability through the GEN to optimize the gripper mesh so that grasping is more likely to be successful. This could open the avenue of automatically designing new grippers in simulation, custom to the objects they are manipulating. Finally, we show that GENs can be used as a flexible spatial memory.
|
|
12:30-13:30, Paper TuPS1.13 | Add to My Program |
Automated Single-Particle Micropatterning System Using Dielectrophoresis |
Huang, Kaicheng | The Hong Kong Polytechnic University |
Mills, James K. | University of Toronto |
Abu Ajamieh, Ihab | University of Toronto |
Cui, Zhenxi | The Hong Kong Polytechnic University |
Lai, Jiewen | The Hong Kong Polytechnic University |
Chu, Henry | The Hong Kong Polytechnic University |
|
|
12:30-13:30, Paper TuPS1.14 | Add to My Program |
Design of a Bipedal Hopping Robot with Morphable Inertial Tail for Agile Locomotion |
An, Jiajun | The Chinese University of Hong Kong |
Chung, Tsz Yin | The Chinese University of Hong Kong |
Au, K. W. Samuel | The Chinese University of Hong Kong |
Keywords: Biologically-Inspired Robots, Legged Robots, Mechanism Design
Abstract: Animals often use their external appendages (such as tails, limbs) to achieve spectacular maneuverability, energy efficient locomotion, and robust stabilization to large perturbations, which may not be easily attained in the existing legged robots. Their appendages, particularly, the tails are very compact, light, highly dexterous with a large range of motion (ROM). Animals can also curl and straighten up their tails in less than one-tenth of a second to facilitate rapid adjustment for the moment of inertia (IoM) and the Center of Mass (CoM). Most of the existing robotic tail designs still lack dexterity, output force, dynamic response. And there is no viable and compact solution to adjust the tail IoM rapidly and effectively. This project aims to address the current limitations in existing tailed robotic systems, in particular, we develop a bipedal tailed hopping robot to investigate how a large ROM morphable inertial tail can enable agile locomotion.
|
|
12:30-13:30, Paper TuPS1.15 | Add to My Program |
Continuous Neural Control Based on Integration of BCI and Adaptive Controller for Steering a Vehicle |
Shi, Haonan | Beijing Institute of Technology |
Bi, Luzheng | Beijing Institute of Technology |
Keywords: Brain Machine Interfaces, Human-Centered Automation, Human Factors and Human-in-the-Loop
Abstract: In this paper, we propose a continuous neural control method by integrating an electroencephalography (EEG)-based brain-computer interfaces (BCIs) with an adaptive assistant controller for steering a vehicle. The assistant controller includes the fuzzy logic module and adaptive law. Experimental results show that the proposed method has better performance and less workload to users than the existing methods.
|
|
12:30-13:30, Paper TuPS1.16 | Add to My Program |
Single-Hand Movement Direction Decoding from EEG Signals under Opposite-Hand Movement Distraction |
Wang, Jiarong | Beijing Institude of Technology |
Bi, Luzheng | Beijing Institute of Technology |
Fei, Weijie | Beijing Institute of Technology |
Keywords: Brain Machine Interfaces, Human-Centered Robotics, Human Factors and Human-in-the-Loop
Abstract: Decoding human motor intention from electroencephalograms (EEG) signals is of great value for human-machine collaboration. Existing studies are concentrated on single-hand motion decoding from EEG signals given the opposite hand is maintained still. However, in many real human-machine interaction systems, it is unrealistic to require operators to keep the opposite hand still all the time. In this paper, we investigated how to decode right-hand movement direction from EEG signals under a left-hand movement distraction. Experimental results show the feasibility of single-hand movement direction decoding from EEG signals under the opposite-hand movement distraction.
|
|
12:30-13:30, Paper TuPS1.17 | Add to My Program |
Static Analysis on the Modular Detachable Climbing Robot for All Wall-To-Wall Transitions |
Park, Changmin | RoDEL |
Lee, Jiseok | Hanyang University |
Ryu, Sijun | Hanyang University |
Seo, TaeWon | Hanyang University |
Keywords: Climbing Robots, Cellular and Modular Robots, Field Robots
Abstract: In this paper, static analysis of the wall to wall conversion is performed and tested and verified. The condition that the force of the front part of the module should be 0 or more was confirmed and the optimum tail force was calculated to satisfy the condition. In addition, although it was not possible to switch the inner wall of two modules in the existing paper, what was possible in this paper was verified.
|
|
12:30-13:30, Paper TuPS1.18 | Add to My Program |
Obstacle Overcoming on a Façade: Novel Design of a Rotating Leg Mechanism |
Lee, Youngjoo | Hanyang University |
Yoo, Sungkeun | Seoul National University |
Seo, Myoungjae | Hanyang University |
Kim, Jongwon | Seoul National University |
Kim, Hwa Soo | Kyonggi University |
Seo, TaeWon | Hanyang University |
Keywords: Climbing Robots, Field Robots, Mechanism Design
Abstract: This late-breaking poster paper presents a novel design of a wheel mechanism for obstacle overcoming on a façade. We design a 1 DOF rotating leg with different length spokes. It assembled designated angle on the tapered disk. This leg mechanism have an spatial efficiency of the design layout for a robot, also reduce any chance of interference between legs with other parts of robot. The design, mechanism of leg wheel and movement of overcoming obstacle are going to be shared in the followings.
|
|
12:30-13:30, Paper TuPS1.19 | Add to My Program |
Design of a Novel Leg for a Small Tree Climbing Robot Driven by Shape Memory Alloy |
Ishibashi, Keitaro | Waseda University |
Takanishi, Atsuo | Waseda University |
Ishii, Hiroyuki | Waseda University |
Keywords: Climbing Robots, Legged Robots, Mechanism Design
Abstract: Our long-term objective is to develop a small tree climbing robot for ecological surveys to protect endangered birds and mammals nesting in trees. When investigators climb trees, there is the danger of a crash. In addition, they can disturb their nesting because animals can abandon their nests with such disturbance stimuli: approaching a larger object, approaching more quickly, and louder noises. Therefore, we deem it necessary to develop a small robot that climbs trees quietly and slowly on behalf of the investigators and acts as a tree camera. We developed a small tree climbing robot driven only by shape memory alloy (SMA) actuators, which are extremely silent, as a prototype. It has six legs, and the fore and middle legs each have an antagonistic drive mechanism consisting of an SMA actuator and spring. Since the robot has six legs, the leg should be lightweight to make the robot small. Moreover, the legs must have spring elements for the antagonistic mechanism. In this paper, we present the design of a multifunctional light leg for a small tree climbing robot. We developed a stainless-steel wire leg. It was designed by modeling each component with rigid links. Its weight is 0.25 [g]. It has three coil parts. One is a spring element for roll rotation, the others are holes for passing the SMA actuator. The elasticity of the stainless-steel wire itself is used as a spring element for yaw rotation. The end of the leg is processed into a needle shape and plays the role of a nail. Thus, by processing one wire and giving various functions, the number of parts is reduced, and weight reduction is realized. In addition, the leg is reinforced with UV resin so that the through holes do not function as spring elements. To confirm effectiveness of the reinforcement, we reinforced with beams instead of UV resin and simulated the deformation of the leg during roll and yaw rotation. In both roll and yaw rotations, the reinforced one showed a behavior closer to the model. As a result, the robot climbs a cedar tree tilted up to 30 [deg]. The minimum climbing speed at that time was 0.18 [mm/s]. The reason why the robot failed to climb a surface inclined at more than 30 [deg] is that the tree contact position of the leg shifts in the positive z axis direction. This deviation occurs because the force direction of the SMA actuator changes slightly depending on how the leg is lifted. However, it is difficult to control the force direction. Therefore, a novel plate leg is presently designed to limit the direction of deformation by twisting a part of the thin plate and to eliminate the displacement.
|
|
12:30-13:30, Paper TuPS1.20 | Add to My Program |
IMU-Based Spectrogram Approach with Deep Convolutional Neural Networks for Gait Classification |
Nguyen, Mau Dung | University of Science & Technology |
Mun, Kyung-Ryoul | National University of Singapore |
Jung, Dawoon | Korean Institute of Science and Technology |
Park, Mina | Korea Institute of Science and Technology (KIST) |
Kim, Jinwook | Korea Institute of Science and Technology |
Keywords: Cognitive Human-Robot Interaction, Sensor-based Control, Deep Learning in Robotics and Automation
Abstract: We propose a wearable sensor-based gait classification system. Our approach assumes that multiple IMU sensors attached to various body parts can capture the gait characteristics that are used to predict whether the subject has foot abnormalities or athletic performance. We first transform raw sensor signals into spectrogram images by using the Short-time Fourier transform (STFT) and feed this visual representation to four deep CNN networks, i.e., the newly proposed network, Resnet18, Resnet50, and Densenet121. The IMU data were acquired from 7 sensors attached to the pelvis, thighs, shanks, and feet while 29 semi-professional athletes (AT group), 19 normal participants (N group), and 21 participants with foot abnormalities (AB group) walked on a 20-m straight path. We investigated classification accuracy according to the number and location of attached sensors to optimize performance. In addition, we adopted IMU-based spectrogram approaches with deep CNN models for the gait classification with two different types of input layouts. Our experimental results show that in the case of using a single sensor, the best classification performance was achieved on the pelvic spectrograms. The accuracy increases significantly when we used the combinations of multiple sensors. The classification performance is remarkably increased in the stacked input layout than the flat one. Although the proposed model is simpler than the three existing well-known networks, it shows a comparable classification result. Our proposed network that only used a single IMU sensor data at pelvis can predict the subject groups with an accuracy of 97.58% even without requiring hand-craft extraction and selection of features. Moreover, when we combined the pelvis and two feet sensors data and produced the stacked input layout, the accuracy was increased up to 98.19%. Consequently, practical applications could use two IMU sensors attached in two shoes and one IMU sensor in a belt worn around the waist to help ambient intelligence easily recognize people’s gait group. Obviously, this use of a small number of sensors located on the human body could allow us to save power consumption and make applications more realistic. We made confusion matrices showing the number of correct and incorrect prediction of the proposed model in accordance with the types of input layout with different sensor combinations. No cases are reported, which incorrectly predict N or AT group to AB group, despite some confusions between the N group and the AT group. We plan to generalize our approach to predicting various health status information of people living in the ambient intelligent environment.
|
|
12:30-13:30, Paper TuPS1.21 | Add to My Program |
Artificial Intelligent Navigation Technology for a Robotic Vacuum Cleaner in an Indoor Environment |
Noh, DongKi | LG Electronics Inc |
Kim, Jung-Hwan | LG Electronics |
Yang, Wonkeun | LG Electronics |
Eoh, Gyuho | LG Electronics |
Lee, Minho | LG Electronics |
Yim, Byungdoo | LG Electronics |
Shim, Inbo | LG Electronics Inc |
Cho, Ilsoo | LG Electronics Inc |
Baek, Seung-Min | LG Electronics |
Keywords: Collision Avoidance, Visual Learning, Visual-Based Navigation
Abstract: We introduce AI technique in our new Robotic Vacuum Cleaner (RVC), named CodeZeroTM Hom-Bot. Hom-Bot is able to detect obstacles with a camera and learn them through state-of-the-art AI techniques. When Hom-Bot encounters pre-trained obstacles on the path, the robot can easily avoid them and keep its own cleaning task. In addition, Hom-Bot can recognize a known place in the cleaning environments with visual place recognition technique.
|
|
12:30-13:30, Paper TuPS1.22 | Add to My Program |
External Force Estimation of Human-Cooperative Robot During Object Manipulation Using Recurrent Neural Network |
Hanafusa, Misaki | Tokyo Denki University |
Ishikawa, Jun | Tokyo Denki University |
Keywords: Compliance and Impedance Control, Motion Control, Physical Human-Robot Interaction
Abstract: The proposed external force estimator calculates the net external force from the difference between the force/torque generated by the joint and the force/torque nominally needed for the robot motion using the trained recurrent neural network (RNN) inverse dynamics model. The experimental setup consists of a robot, PA10-7C (Mitsubishi Heavy Industries) and a force sensor (NITTA) installed at between the robot and the manipulated object. The object is an aluminum baton (hereinafter referred to as the baton), on which a weight movable on the linear guide is installed to apply the external force during the baton manipulation. In the experiment, a person touches the baton manipulated by the robot. Then, it is checked if the compliant motion to the external force calculated by the proposed method is properly generated. The experimental result showed that the robot gives way to the person and absorbs the collision force according to a desired mechanical impedance model. The advantages of the proposed external force estimator are summarized as follows: - Even while the robot dynamically manipulates the object, it is possible to estimate the net external force in distinction from the manipulation force, - The values of physical parameters of the robot are not explicitly needed to obtain inverse dynamics model, - If the friction force and/or the parameter error, which are difficult to be modeled, have reproducibility, a model as a part of the inverse dynamics model can be acquired.
|
|
12:30-13:30, Paper TuPS1.23 | Add to My Program |
Untethered Quadrupedal Hopping on a Trampoline |
Wang, Boxing | College of Control Science and Engineering, Zhejiang University, |
Zhou, Chunlin | Zhejiang University |
Wu, Jun | Zhejiang University |
Keywords: Compliant Joint/Mechanism, Legged Robots, Motion Control
Abstract: Considering the hardware complexity of torque-control motors, position-control motors are usually used in small-size quadruped robots and for prototype testing. However, due to the high gear ratio of position-control motors, elastic components are needed to achieve high dynamic motions including jumping or running. Instead of designing compliant legs, we propose using elastic surfaces (such as a trampoline) as energy storages to help a rigid-leg quadruped to jump. Thus, the jumping control algorithms could be developed first, then used to test further compliant legs design, and a jumping robot with elastic legs could be finally built. In this poster, we present the algorithms we used to achieve stable quadrupedal hopping on a trampoline. Experimental results are also breifly illustrated.
|
|
12:30-13:30, Paper TuPS1.24 | Add to My Program |
Target Classification and Prediction of Unguided Rocket Trajectories Using Deep Neural Networks |
Kim, Minwoo | UNIST |
Park, Bumsoo | UNIST |
Oh, Hyondong | UNIST |
Keywords: Computer Vision for Automation, AI-Based Methods, Aerial Systems: Mechanics and Control
Abstract: This paper deals with classification and prediction of low-altitude rocket targets using deep neural networks. Conventionally, model-based methods such as the Kalman filter are widely used. However, they can lack robustness due to unexpected situations and noisy sensor measurements. To address this issue, this study proposes the use of various data-driven methods which are powerful on situations where no model is available. Specifically, three types of neural networks are used: DNN (deep neural network), CNN (convolutional neural network) and RNN (recurrent neural network). To verify the benefit and robustness of the proposed algorithms, comparisons with the model-based method are performed on several scenarios
|
|
12:30-13:30, Paper TuPS1.25 | Add to My Program |
Robot-Assisted Composite Manufacturing Using Deep Learning and Multi-View Computer Vision |
Djavadifar, Abtin | University of British Columbia |
Graham-Knight, John Brandon | University of British Columbia |
Körber, Marian | German Aerospace Center |
Najjaran, Homayoun | University of British Columbia |
Keywords: Computer Vision for Automation, Deep Learning in Robotics and Automation, AI-Based Methods
Abstract: This poster introduces an automated wrinkle detection method on semi-finished fiber products in the aerospace manufacturing industry. Machine learning, computer vision techniques, and evidential reasoning are combined to detect wrinkles during the draping process of fibre-reinforced materials with an industrial robot.
|
|
12:30-13:30, Paper TuPS1.26 | Add to My Program |
Ambiguity Poses Estimation for Objects with Symmetry |
Staszak, Rafal | Poznan University of Technology |
Belter, Dominik | Poznan University of Technology |
|
|
12:30-13:30, Paper TuPS1.27 | Add to My Program |
Contamination Detection and Classification for an Automated Façade Cleaning Operation |
Lee, Jiseok | Hanyang University |
Park, Garam | Hanyang Unviersity |
Hong, Jooyoung | Seoul National University |
Kim, Hwa Soo | Kyonggi University |
Seo, TaeWon | Hanyang University |
Keywords: Computer Vision for Automation, Computer Vision for Other Robotic Applications, Object Detection, Segmentation and Categorization
Abstract: Recently, a lot of high-rise buildings have been built and the need for an unmanned exterior wall cleaning robot is increasing. Therefore, a system for detecting various façade contamination is required to robot just like as workers do. However, existing surface contamination detecting systems can only detect certain type of contaminant or not compact enough to be attached in mobile platforms. In this paper, we try to solve this problem with machine vision system through CNN and image processing method. This makes it possible to detect object, area and particle type contaminants using only one photograph
|
|
12:30-13:30, Paper TuPS1.28 | Add to My Program |
Autonomous Photogrammetry Process for Managing Stockpile Inventory with Unmanned Aerial Vehicle |
Lim, Seungho | POSCO |
Kim, Hyungjin | Graduate Institute of Ferrous Technology, POSTECH |
Keywords: Computer Vision for Automation, Object Detection, Segmentation and Categorization, Inventory management
Abstract: With the goal of specializing in the manufacturing of the world’s best steels, POSCO focuses on enhancing its competitive edge from raw material to final product. Especially, we have the world’s largest integrated steel mill with more than 2,700 hectares and it makes us interested in geospatial information technology including the inventory management system of ore and coal stockpiles. Huge raw material yard over 350 hectares has been managed based on the naked eye of many skilled operator in the past. On that account, the measurement accuracy was not ensured and there was a possibility of safety risk during field survey. Thus commercial unmanned aerial vehicle, in other words, drones are deployed for highly accurate surveys of stockpiles, Generally, drone can acquire a number of aerial images using auto-grid flight plan with a given mission. However, it is necessary to automate the image process after flight in order to increase the frequency of use, because the general aerial survey is professional and cumbersome task. One of the major challenges is automatic geolocation because manual correlation of ground control point (GCPs) in hundreds of photos with a high-resolution is very rigorous and tiresome. Thus we implement the auto-GCP algorithm to save time on manual marker placement and to match geolocation with more precision [2]. In addition, it is considered how many GCPs have to be required, as well as where and how they have to be measured considering the flying height, imager size, field of view (FOV), and overlap rate. Another challenge is that it is difficult to classify variety of stockpiles from a lot of facilities around the raw material yard in steelworks. The boom of stacker and reclaimer passes over the top of the stockpile and it makes occlusion shadows. Belt conveyors, concrete retaining walls, and embankment near stockpiles also complicate the automatic classification. Thus, the smoothing filter and triangulated digital terrain models (DTM) on specified criterion are implemented. In addition, because the terrains on which these stockpiles lie are not flat, the ground elevation is compensated [3]. The above whole process is summarized in Fig. 1. As a result, the proposed autonomous photogrammetry process can reflect high geometric fidelity of stockpile and calculate the volume of hundreds of stockpiles within an accuracy of 3% in four hours, respectively.
|
|
12:30-13:30, Paper TuPS1.29 | Add to My Program |
Autonomous Detection of PV Panels Using Unmanned Aerial Vehicles |
Ismail, Hesham | DEWA |
Al Jasmi, Nawal | DEWA |
Quadir, Jabirul | DEWA |
Bandyopadhyay, Akash | Amity Univeristy |
Salim, Rufaidah | Amity University |
Keywords: Computer Vision for Other Robotic Applications, Energy and Environment-Aware Automation, Aerial Systems: Perception and Autonomy
Abstract: An autonomous system for the detection of solar panels using Unmanned Aerial Vehicles(UAVs) is proposed in this paper. There has been a huge growth in the solar energy field. In Dubai alone there been a massive group of solar farms providing a huge amount of electricity. If the inspection of the PV panels is done completely manually this would require a large amount of manpower and resources not to mention a lot of time. The aim is to have a system in which a drone flies over fixed waypoints that have been set previously based on GPS data. The drone will take images and capture videos which will be used to detect the panels.
|
|
12:30-13:30, Paper TuPS1.30 | Add to My Program |
Co-Simulation of Mechanical Systems with Hydraulic Actuators |
Peiret, Albert | McGill University |
Gonzalez, Francisco | University of a Coruna |
Kovecses, Jozsef | McGill University |
Teichmann, Marek | CMLabs Simulations Inc |
Keywords: Contact Modelling, Simulation and Animation, Hydraulic/Pneumatic Actuators
Abstract: Co-simulation can be used to couple subsystems that present different time-scales, such as hydraulics. However, numerical stability of the co-simulation setup can be compromised by discontinuities and time delays of the coupling variables. Here, we use a reduced-order model of the mechanical subsystem based on the effective mass at the interface in order to obtain a prediction of the coupling variables, which allows for larger communication steps and increases the stability of co-simulation setups.
|
|
12:30-13:30, Paper TuPS1.31 | Add to My Program |
Meta-World: A Benchmark and Evaluation for Multi-Task and Meta Reinforcement Learning |
Yu, Tianhe | Stanford University |
Quillen, Deirdre | Google |
He, Zhanpeng | University of Southern California |
Julian, Ryan | University of Southern California |
Hausman, Karol | University of Southern California |
Levine, Sergey | UC Berkeley |
Finn, Chelsea | UC Berkeley |
Keywords: Deep Learning in Robotics and Automation, Performance Evaluation and Benchmarking
Abstract: Meta-reinforcement learning algorithms can enable robots to acquire new skills much more quickly, by leveraging prior experience to learn how to learn. However, much of the current research on meta-reinforcement learning focuses on task distributions that are very narrow. For example, a commonly used meta-reinforcement learning benchmark uses different running velocities for a simulated robot as different tasks. When policies are meta-trained on such narrow task distributions, they cannot possibly generalize to more quickly acquire entirely new tasks. Therefore, if the aim of these methods is enable faster acquisition of entirely new behaviors, we must evaluate them on task distributions that are sufficiently broad to enable generalization to new behaviors. In this paper, we propose an open-source simulated benchmark for meta-reinforcement learning and multi-task learning consisting of 50 distinct robotic manipulation tasks, with the aim of making it possible to develop algorithms that generalize to accelerate the acquisition of entirely new, held-out tasks. We evaluate 6 state-of-the-art meta-reinforcement learning and multi-task learning algorithms on these tasks. Surprisingly, while each task and its variations (e.g., with different object positions) can be learned with reasonable success, these algorithms struggle to learn with multiple tasks at the same time, even with as few as nine distinct training tasks. Our analysis and open-source environments pave the way for future research in multi-task learning and meta-learning that can enable meaningful generalization, thereby unlocking the full potential of these methods.
|
|
12:30-13:30, Paper TuPS1.32 | Add to My Program |
Synergy-Based Control for Multi-Fingered Hands Using Selected Joint Spaces |
Higashi, Kazuki | Osaka University |
Ozawa, Ryuta | Meiji University |
Nagata, Kazuyuki | National Inst. of AIST |
Wan, Weiwei | Osaka University |
Harada, Kensuke | Osaka University |
Keywords: Dexterous Manipulation, Multifingered Hands
Abstract: This paper proposes subsynergy which provides a synergy-based control method for multi-fingered hands under selected joint spaces. Subsynergy is a synergy composed of subsets of fingers or joints needed for performing a specific task. By using subsynergy, we can perform several different dexterous tasks by controlling high DOF multi-fingered hand with lower dimensional inputs compared to than conventional synergies.
|
|
12:30-13:30, Paper TuPS1.33 | Add to My Program |
Redundant Resolution Method of an Underwater Manipulation for Disturbance Rejection |
Moon, Yecheol | Hanyang University |
Bae, Jangho | Seoul National University |
Jin, Sangrok | Pusan National University |
Kim, Jongwon | Seoul National University |
Seo, TaeWon | Hanyang University |
Keywords: Dual Arm Manipulation, Marine Robotics, Force Control
Abstract: This paper describes an efficient collaborative method of underwater robot-manipulator. We used the weighted pseudo-inverse of the manipulator's collaborative controller to derive the most efficient control method and experiment.
|
|
12:30-13:30, Paper TuPS1.34 | Add to My Program |
Experimental Study on the Parameters of High-Pressure Water-Jet Cleaning on a Facade |
Yoon, Dupyo | Hanyang University |
Lee, Youngjoo | Hanyang University |
Kwon, Daesung | Hanyang University |
Park, Changmin | RoDEL |
Seo, Myoungjae | Hanyang University |
Seo, TaeWon | Hanyang University |
Keywords: Field Robots, Mechanism Design, Robotics in Hazardous Fields
Abstract: This late-breaking poster paper presents the experimental study on the parameters of high-pressure water cleaning on a facade. The cleaning performance parameters, such as water pressure, spray angle and spraying distance, were optimized by using Taguchi method. The performance was evaluated by using the image evaluation method. Also, we looked over the reaction force on nozzle tip and impact force on the specimens to know the effect on building surface.
|
|
12:30-13:30, Paper TuPS1.35 | Add to My Program |
Extrinsic Calibration of Thermal IR Camera and mmWave Radar by Exploiting Depth from RGB-D Camera |
Yoon, SungHo | KAIST (Korea Advanced Institute of Science and Technology) |
Kim, Ayoung | Korea Advanced Institute of Science Technology |
Keywords: Field Robots, Sensor Fusion, Localization
Abstract: Despite the wide utility, RGB cameras and Light Detection and Ranging (LiDAR) have been reported to be vulnerable in low-visibility environments under fire or smoke. To Solve this issue, we introduce a sensor system that consists of a thermal infrared (IR) camera and an mmWave radar. In doing so, extrinsic calibration between two sensors is required, and 14-bit temperature and sparse range measurement from the radar are challenging to calibrate. We propose a solution to this multimodal calibration method by using an RGB-D sensor as an intermedium for the depth based optimization. As a validation for the relative pose estimation, we present the qualitative result by projecting the radar’s depth on the thermal camera frame.
|
|
12:30-13:30, Paper TuPS1.36 | Add to My Program |
Biomimetic Wrinkled MXene Pressure Sensors towards Collision-Aware Robots |
Cai, Catherine | National University of Singapore |
Ren, Hongliang | Faculty of Engineering, National University of Singapore |
Keywords: Force and Tactile Sensing, Soft Sensors and Actuators, Biomimetics
Abstract: The use of surgical robots in the field of minimally invasive neurosurgical procedures can offer several benefits and advantages. However, the lack of force sensing hinders and limits their use in such procedures. Equipping surgical robots with pressure sensors can enhance robot-environment interaction by enabling collision awareness and enhance human-robot interactions by providing surgeons the necessary force feedback for safe tissue manipulation. With the emergence of soft robotics in biomedical applications, the attached pressure sensors are required to be flexible and stretchable in order to comply with the mechanically dynamic robotic movements and deformations. Inspired by the multi-dimensional wrinkles of Shar-Pei dog’s skin, we have fabricated a flexible and stretchable piezoresistive pressure sensor consisting of MXene electrodes with biomimetic topographies.
|
|
12:30-13:30, Paper TuPS1.37 | Add to My Program |
A Force-Controlled Robotic Wrist Module for the Macro-Micro Manipulation of Industrial Robots |
Liu, Yen-Chun | National Cheng Kung University |
Chang, Yu-Hsiang | National Cheng Kung University |
Lan, Chao-Chieh | National Cheng Kung University |
Keywords: Force Control, Cellular and Modular Robots, Industrial Robots
Abstract: Industrial robots have multiple degrees-of-freedom (DOFs) and high dexterity but they currently have two mechanical challenges. First, they have excessive joint accelerations when encountering paths near the singular positions. Second, the large link inertia and gearbox friction make the contact force control of the robot end-effector difficult no matter where the torque sensors are placed. A macro-micro robot is proposed in this paper to address these two challenges. An existing 6-axis robot serves as the macro robot and a 2-axis wrist module is serially attached to the end of the macro robot to form a macro-micro robot. The micro robot is compact with high torque density and internal force sensing capability. The introduced redundant DOFs at the distal end of the macro robot allow more accurate force control and smoother path tracking. Simulation and experiment results are provided to show the advantages of the proposed design. We expect that this macro-micro robot can be used in applications where smooth motion or accurate contact force control is required.
|
|
12:30-13:30, Paper TuPS1.38 | Add to My Program |
Fault-Tolerant Force Tracking for a Multi-Legged Robot |
Cheah, Wei | The University of Manchester |
Watson, Simon | University of Manchester |
Lennox, Barry | The University of Manchester |
Keywords: Force Control, Legged Robots, Motion and Path Planning
Abstract: The control of a legged robot's contact forces to lie within their friction cone is vital to avoid slippage during the robot's locomotion. However, a nominally actuated leg loses its ability to track the desired contact forces, when a fault develops, such as the locking of a joint. Current literature on fault tolerant control on multi-legged robots focuses on modifying the robot's gait but do not address the force tracking problem even with legs that have a force sensor attached to the foot. This research proposes two additional modules to the general architecture of a multi-legged robot to modulate the robot's pose such that all the feet are guaranteed ground contact and force tracking. Preliminary results using the Gazebo simulator shows that force tracking on all legs, including the faulty leg, is achievable using the proposed approach.
|
|
12:30-13:30, Paper TuPS1.39 | Add to My Program |
Admittance Control Based on Stiffness Ellipse for Collision Force Control of Object Manipulation |
Oikawa, Masahide | Saitama University |
Kutsuzawa, Kyo | Saitama University |
Sakaino, Sho | University of Tsukuba |
Tsuji, Toshiaki | Saitama University |
Keywords: Force Control, Motion Control, Contact Modelling
Abstract: This study proposes an admittance control method based on the stiffness ellipse. Some of assembly tasks in factories are quite difficult to automate, because the discontinuous transitions between contact and noncontact phases generate issues. In the contact phase, a control approach is required that handles assembly parts with compliance, while a control method that follows the trajectory in the noncontact phase. To achieve an assembly task, it is often necessary to switch the control when the contact occurs, while such a controller design often degrades the performance owing to the switchover. This study solves this problem by adjusting the stiffness ellipse by admittance control. By tilting the stiffness ellipse, the external force can be induced in a desired direction during contact, without switching the control or adjusting the position/force command. Using the proposed method, tasks including contact and noncontact phases can be achieved with continuous control. Experimental results on a drawing task and simulation results on drawing and peg-in-hole tasks show the advantage of the proposed method.
|
|
12:30-13:30, Paper TuPS1.40 | Add to My Program |
A Linear Series Elastic Actuator for Accurate Force and Impedance Control with High Torque-To-Rotor-Inertia Ratios |
Lee, Yu-Shen | National Cheng Kung University |
Huang, Yan-Lin | National Cheng Kung University |
Lan, Chao-Chieh | National Cheng Kung University |
Keywords: Force Control, Soft Sensors and Actuators, Flexible Robots
Abstract: A series elastic actuator (SEA) combines an actuator in series with an elastic spring. By controlling the deformation of the elastic spring, an SEA provides more accurate force and impedance control than conventional rigid actuators. SEAs are ideal for robots and machines that need to interact safely with human or the environment. The majority of existing SEAs uses brushless or brushed DC motors as the actuators. The advantages of using step motors as the actuators of SEAs have not received enough attention. Step motors have much higher torque-to-rotor-inertia ratios than other DC motors. Hence they can provide better stability and high-speed accuracy of force control while maintaining lightweight. When rotor position feedback is used, step motors can achieve smooth and ultra-accurate dynamic force response. This paper develops the dynamic model of a linear series elastic step motor and presents its prototype. Forward and inverse force/impedance tracking control responses will be provided to show the advantages of the series elastic step motor. It is expected that the method presented here can offer a better actuator selection of SEAs when higher torque-to-rotor-inertia ratios are required.
|
|
12:30-13:30, Paper TuPS1.41 | Add to My Program |
Generating Coordinated Reach-Grasp Motions with Neural Networks |
Chong, Eunsuk | University of California, Los Angeles |
Park, Jinhyuk | Seoul National University |
Kim, Hyungmin | Korea Institute of Science and Technology |
Park, Frank | Seoul National University |
Keywords: Grasping, Learning from Demonstration
Abstract: We propose a framework for generating coordinated reaching and grasping motions of a robotic hand-arm system by learning from human demonstration. Our framework consists of the three stages. The first stage is movement type determination. A mixture version of the conditional restricted Boltzmann machine (CRBM) lies at the core of this stage, which is called the convolutional implicit mixture of CRBMs (conv-imCRBM). To select a mixture mode for the next stage, a movement type is determined by the trained conv-imCRBM given the target object pose and the initial state of hand-arm system. The second stage is trajectory generation given the position and orientation of the target object. In this stage, a control variable l is added to the original CRBM model for encoding target object pose. K-step iterative Gibbs sampling is used in the inference process for the trajectory generation, and then control variable l is updated with gradient descent method to minimize the distance between the target object and robot hand. The final stage is to maximize grasp quality by fine-tuning the trajectory from the previous stage. Movement trajectory is modified to ensure the force closure condition by optimizing the minimum singular value of a grasp matrix as a grasp quality measure.
|
|
12:30-13:30, Paper TuPS1.42 | Add to My Program |
Curiosity Driven Exploration for Classification in the Dark Using Tactile Sensing |
Tickell, Blake | University of California, Berkeley |
Mudigonda, Mayur | UC Berkeley |
Agrawal, Pulkit | UC Berkeley |
Keywords: Haptics and Haptic Interfaces, Deep Learning in Robotics and Automation, Force and Tactile Sensing
Abstract: Combining information from partial observations across multiple sensory modalities to execute goal directed actions is a key aspect of human intelligence. In order to build agents with similar capabilities, we consider the problem of identifying a target object from a drawer using tactile exploration only. We use a simulated anthropomorphic hand with continuous controls and high dimensional motors. The agent is able to interact with several objects in a box by touching them. The trial is a success if the agent is able to successfully identify the target object. Success at this task requires both the construction of a sequence of controls used to drive the hand's motors and learning to discriminate objects from the haptics signal received during an interaction. Formulated as a reinforcement learning problem, we compare the performance of various policies at this task, introducing intrinsic curiosity and classification-based extrinsic reward as a method for learning useful exploration policies in the domain of classification.
|
|
12:30-13:30, Paper TuPS1.43 | Add to My Program |
Behavior Change Based on Stiffness for Haptic Interface |
Ozeki, Tomoe | Gifu University |
Mouri, Tetsuya | Gifu University |
Keywords: Haptics and Haptic Interfaces, Recognition, Perception for Grasping and Manipulation
Abstract: Touch sensation of the finger is very important information for human behavior. The goal of this study is to investigate whether social perception and behavior are influenced by the touch sense of human finger using haptic interface device. In this paper as the pretest of a haptic interface device, experiments are tested whether a human finger attached to a real soft rubber ball or hard wire ball is affected on social perception and behavior of people. Experimental results show that information of a reaction force acted on a human finger has no influence on social perception, and show that the hardness of object has a few influence on social behavior.
|
|
12:30-13:30, Paper TuPS1.44 | Add to My Program |
A Tactile Stimulation System for Robot-Assisted Hand Rehabilitation |
Chen, Jiazhou | Xi'an Jiaotong University |
Li, Min | Xi'an Jiaotong University |
Bo, He | Xi'an Jiaotong University |
Xu, Guanghua | School of Mechanical Engineering, Xi'an Jiaotong University |
Yao, Wei | Strathclyde University |
Keywords: Haptics and Haptic Interfaces, Rehabilitation Robotics, Soft Robot Applications
Abstract: This paper introduces a tactile stimulation system for robot-assisted hand rehabilitation. The tactile stimulation was synchronized with the motion of the hand exoskeleton. Robot-assisted hand trainings with and without tactile stimulation were compared in an experiment involving healthy human subjects. The experimental results confirmed the increase of attention level of human subjects during the hand training process.
|
|
12:30-13:30, Paper TuPS1.45 | Add to My Program |
View Sharing to Enhance Driving Safety through Vehicle-To-Vehicle Communication |
Tran, Duy | Oklahoma State University |
Liu, Fangyao | Oklahoma State University |
Albrecht, Daniel | Oklahoma State University |
Sheng, Weihua | Oklahoma State University |
Keywords: Human Detection and Tracking, Object Detection, Segmentation and Categorization, Computer Vision for Transportation
Abstract: Pedestrians, the most vulnerable part in the transportation system, are more likely to get hurt because of careless drivers. Some Advanced Driver Assistance Systems (ADAS), like adaptive cruise control and pedestrian automatic emergency braking, are mature enough to be adopted in real-world vehicles. However, there is not an effective method to detect moving objects or pedestrians blocked by other vehicles. This paper aims to tackle this problem by sharing the view between vehicles through Vehicle-to-Vehicle (V2V) communication. The overview of view sharing is illustrated in Figure 1. The camera in vehicle 1 detects the pedestrian whose bounding box location is then transformed to vehicle 2’s coordinate and displayed on its windshield. The flowchart of the view sharing system is shown in Figure 2. We detect pedestrians via the video stream from a camera placed inside vehicle 1, v1. The detected pedestrian can be located on the image plane as a 2D vector. On the other hand, to share the view, we need to estimate the real-world location (3D vector) of the pedestrian with respect to the vehicles. Therefore we need conduct coordinate transformation, which calls for camera calibration. Based on that, we can estimate the real-world location of the pedestrian with respect to vehicle v1. Then we transform the 3D-location from v1’s coordinate to v2's coordinate. Finally, a coordinate transformation is applied to estimate the corresponding 2D-pixel location of the pedestrian on the windshield of v2. We conducted experiments on a custom-built assisted driving platform which includes two driving simulators running on two computers. Behind the driver seat of simulator 1, we mounted a camera which is connected to another computer to detect pedestrians. In this way, we can distribute the computation load among the three computers. The three computers communicate and share data with each other via UDP communication. Figure 3 shows the bounding box of the detected pedestrian on the simulator 2. To evaluate the detection performance, we calculated the Intersection-over-Union (IoU) ratio between the detected bounding box and the ground truth one. The IoU ratio varies with different locations of vehicles and the average IoU is 70%. In addition, we also estimated the real-world error by comparing the ground truth location (gx, gy) with the estimated real-world location of the pedestrian with respect to v2. The average errors in both the x and y directions range from 0m to 1m.
|
|
12:30-13:30, Paper TuPS1.46 | Add to My Program |
Structured Classification of Locomotion Modes for Wearable Robot Control |
Narayan, Ashwin | National University of Singapore |
Yu, Haoyong | National University of Singapore |
|
|
12:30-13:30, Paper TuPS1.47 | Add to My Program |
Human Interactive Motion Planning for Shared Teleoperation |
Lee, Kwang-Hyun | Korea University of Technology&Education |
Ryu, Jee-Hwan | Korea Univ. of Tech. and Education |
Pruks, Vitalii | Korea Uneversity of Technology and Education |
Keywords: Human-Centered Automation, Motion and Path Planning, Telerobotics and Teleoperation
Abstract: The current state of the art of robot motion planning in a cluttered unstructured environment is subject to fail. However, introducing human intuition into the motion planning pipeline improves the performance of the planning algorithm. In this research, we propose a user interface to record a suggestion of a trajectory to ensure the convergence of the planning algorithm and reduce the computational time. Simulation results demonstrate the superior performance of the proposed motion planning method.
|
|
12:30-13:30, Paper TuPS1.48 | Add to My Program |
Motion Direction Decoding of Upper Limb from EEG Signals with a Cognitive Distraction Task |
Fei, Weijie | Beijing Institute of Technology |
Bi, Luzheng | Beijing Institute of Technology |
Wang, Jiarong | Beijing Institude of Technology |
Keywords: Human-Centered Robotics, Human Factors and Human-in-the-Loop, Human-Centered Automation
Abstract: Motion intention estimation is critical to active human-robot collaboration. However, existing studies on electroencephalograms (EEG)-based motion intention estimation are all focused on motion intention decoding from EEG signals when subjects perform a motion task by using a single hand or foot, given no perception, cognitive, and motor distraction tasks. However, for healthy people, in many cases, they need to complete a motion task given a distraction task. In this paper, we propose a decoding model of the motion direction of the upper limb from EEG signals with a cognitive distraction task. Experimental results suggest that the decoding accuracy of the proposed model is 90%, close to that without a cognitive distraction task.
|
|
12:30-13:30, Paper TuPS1.49 | Add to My Program |
Improved Energy Efficiency Via Parallel Elastic Elements for the Straight-Legged Vertically-Compliant Robot SLIDER |
Wang, Ke | Imperial College London |
Saputra, Roni Permana | Imperial College London |
Foster, James Paul | Imperial College London |
Kormushev, Petar | Imperial College London |
Keywords: Humanoid and Bipedal Locomotion, Compliant Joint/Mechanism, AI-Based Methods
Abstract: Most state-of-the-art bipedal robots are designed to be as anthropomorphic as possible, and therefore possess articulated legs with compliant elements in the joints to increase energy efficiency. Whilst this facilitates smoother, human-like locomotion, there are implementation issues that make walking with straight or near-straight legs difficult. Many robots have to move with a constant bend in the legs to avoid a singularity occurring at the knee joints or to keep the center of mass at a constant height for control purposes. The actuators must constantly work to maintain this stance, which can result in the complete negation of any energy-saving techniques employed. Furthermore, vertical compliance disappears when the leg is straight and the robot undergoes high-energy events such as impacts from running and jumping, as energy travels through the fully extended joints to the hips. We attempt to improve energy efficiency by attaching bungee cords as elastic elements in parallel to the legs of a novel, knee-less biped robot. Bayesian Optimization is utilized to find the optimal vertical motion behavior that achieves the largest reduction in energy consumption, taking into account the dynamics of the bungee cords as well as friction and other unmodelled effects. Our experiments find up to a 15% energy saving compared to the robot configuration without parallel elastic elements.
|
|
12:30-13:30, Paper TuPS1.50 | Add to My Program |
Towards a General Framework for Generating Stable and Flexible Locomotion Skills |
André, João | Universidade Do Minho |
Tateo, Davide | Politecnico Di Milano |
Santos, Cristina | University of Minho |
Peters, Jan | Technische Universität Darmstadt |
Keywords: Humanoid and Bipedal Locomotion, Deep Learning in Robotics and Automation
Abstract: Moving in dynamics and challenging environments is one of the key issues of robot locomotion. Most of the previous work on this area has focused on the task of solving specific problems, such as efficient learning of compact motor primitives, exploitation of bio-inspired motion models, or fall detection. We propose a general framework for learning general locomotion skills, and we propose a methodology to learn skills that are flexible and adapts to current sensory measurement and environment characteristics. To achieve high performances and generalization we build our framework on top of the previous works on Dynamic Motion Primitives and Synergies and improve their generalization capabilities with state of the art Deep Reinforcement Learning approaches.
|
|
12:30-13:30, Paper TuPS1.51 | Add to My Program |
Joint Offset Optimization of Hip Joints in Humanoid Robots |
Kim, Jihun | Chung-Ang University |
Yang, Jaeha | Chung-Ang Univesity |
Yang, Seung Tae | Chung-Ang University |
Lee, Giuk | Chung-Ang University |
Keywords: Humanoid and Bipedal Locomotion, Mechanism Design, Simulation and Animation
Abstract: This paper presents a Taguchi method optimizing hip joints in humanoid robots in terms of joint offsets. By using MuJoCo simulation, the effects of each offset condition were analyzed for the sum of the RMS power of three hip joints during target motion. The optimized model was compared to the initial model in terms of the RMS power of the sum of total hip joints to verify the improved energy efficiency.
|
|
12:30-13:30, Paper TuPS1.52 | Add to My Program |
Whole-Body Postural Control Approach Based on Multiple ZMP Evaluation in Humanoid Robots |
Garcia-Haro, Juan Miguel | Carlos III University of Madrid |
Martinez, Santiago | Universidad Carlos III De Madrid |
Oña, Edwin Daniel | University Carlos III of Madrid |
Victores, Juan G. | Universidad Carlos III De Madrid |
Balaguer, Carlos | Universidad Carlos III De Madrid |
Keywords: Humanoid Robots, Motion Control
Abstract: This article presents a Whole-Body Postural Control (WBPC) approach for a waiter humanoid robot. This approach is based on the concept of a multi-ZMP evaluation system to control body and transported object stability. Both controllers were developed independently, avoiding cross-linked perturbations. These controllers that deal with this complexity were proposed in previous research. The first one was an improvement for balance control of the body (locomotion) and, the second one was a method to apply classic body balance concepts for transporting objects on a tray (manipulation).
|
|
12:30-13:30, Paper TuPS1.53 | Add to My Program |
Obstacle Climbing by a Humanoid Robot Using Standing Jump Motion |
Ahn, DongHyun | Kookmin University |
Cho, Baek-Kyu | Kookmin University |
Keywords: Humanoid Robots, Optimization and Optimal Control, Climbing Robots
Abstract: This paper describes a method to generate a jump trajectory that will enable a humanoid robot to climb an obstacle. For a robot to climb an obstacle by jumping without colliding with the obstacle, the robot should jump higher than the height of the obstacle and move to the desired horizontal position. Therefore, to perform such a dynamic motion, the trajectory needs to consider the capacity of the actuator, impact force at landing, friction force, zero moment point, and so on. This process was defined as a nonlinear optimization problem, and we successfully generated the optimal jump trajectory. This trajectory was verified using ROK-3, an adult-sized humanoid robot.
|
|
12:30-13:30, Paper TuPS1.54 | Add to My Program |
SpineBot: Pneumatically Actuated Muscle |
Lee, Amos Wei Lun | Singapore Institute of Manufacturing Technology (SIMTech) |
Quek, Zhan Fan | Singapore Institute of Manufacturing Technology |
Short, Joel Stephen | Singapore Institute of Manufacturing Technology |
Tao, Pey Yuen | SIMTech |
Keywords: Hydraulic/Pneumatic Actuators, Flexible Robots, Soft Robot Materials and Design
Abstract: In this project, we present a 6 degrees of freedom hybrid robotics design with both hard and soft components. The robot is constructed with lightweight material and actuated with low pressure air, making it safe for human interaction and operation. The proposed design is modular and scalable.
|
|
12:30-13:30, Paper TuPS1.55 | Add to My Program |
A Methodology for Formulating and Exploiting Innovative Technologies for Collaborative Robots in a Manufacturing Setting |
Parizi, M. Shahab | Blue Ocean Robotics |
Macovetchi, Ana Maria | Blue Ocean Robotics |
Kirstein, Franziska | Blue Ocean Robotics |
Keywords: Industrial Robots, Human-Centered Robotics, Formal Methods in Robotics and Automation
Abstract: This paper presents a methodology that enables the exploitation of innovative technologies for collaborative robots through user involvement from the beginning of product development. The methodology will be applied in the EU-funded project CoLLaboratE that focuses on how industrial robots learn to collaborate with human workers in order to perform new manufacturing tasks. The presented methodology is preliminary and will be improved during the project runtime.
|
|
12:30-13:30, Paper TuPS1.56 | Add to My Program |
XL-Laser: Large-Scale Cable-Driven Laser Cutting/Engraving Robot |
Chan, Ngo Foon | Chinese University of Hong Kong |
Cheng, Hung Hon | The Chinese University of Hong Kong |
Chan, Yuen Shan | The Chinese University of Hong Kong, Mechanical and Automation E |
Lau, Darwin | The Chinese University of Hong Kong |
Keywords: Intelligent and Flexible Manufacturing, Tendon/Wire Mechanism, Parallel Robots
Abstract: XL-laser is a 6 DoFs spatial portable cable-driven parallel Robot (CDR) mounted with a 5.5W laser head to the end-effector. The leaser head powered by lithium battery emits high power visible light laser. It can cut and engrave on various materials. With the advantage of CDR, the XL-laser can also be scaled to cut large work-piece. It has been made with different sizes from centimeters to several meters. The 6-Dofs PoCaBot also allows cutting on an curved or inclined surface.
|
|
12:30-13:30, Paper TuPS1.57 | Add to My Program |
Establishing Safer Human-Vehicle Visual Interaction at Night |
Hirayama, Takatsugu | Nagoya University |
Maeda, Takashi | Nagoya University |
Liu, Hailong | Nagoya University |
Morales Saiki, Luis Yoichi | Nagoya University |
Akai, Naoki | Nagoya University |
Murase, Hiroshi | Nagoya University |
|
|
12:30-13:30, Paper TuPS1.58 | Add to My Program |
Towards Learning Trajectory Segmentation through Semi-Supervised Learning |
Urain De Jesus, Julen | TU Darmstadt |
Tateo, Davide | Politecnico Di Milano |
Peters, Jan | Technische Universität Darmstadt |
Keywords: Learning from Demonstration
Abstract: Learning how to segment complex trajectories in meaningful simpler trajectories have become lately an important research topic for many fields from Computer Vision to Reinforcement Learning. In Robotics, trajectory segmentation could be applied for Imitation Learning. Robots would be able to split human demonstrations for a particular task in semantically meaningful segments and so, build a library of simple actions. While both supervised and unsupervised methods have been proposed, supervised methods are limited to a defined number of actions and so, unsupervised methods have shown better results adapting to new actions. We identified two main limitations of unsupervised methods for segmentation. On one side, the trajectories tend to be segmented based on some heuristics defined by the user. On the other side, these segments use to be clustered based on some similarity measurements defined by the user. In order to deal with these limitations, we have developed a semi-supervised learning procedure in which we will learn how to both segment and cluster the trajectories based on human demonstrations.
|
|
12:30-13:30, Paper TuPS1.59 | Add to My Program |
A Kernelized Approach for Learning and Adapting Symmetric Positive Definite Profiles |
Abu-Dakka, Fares | Aalto University |
Keywords: Learning from Demonstration, Compliance and Impedance Control
Abstract: In many robot control problems, data are found in form of symmetric positive definite (SPD) matrices, e.g. stiffness and damping in impedance controllers, manipulability ellipsoids in tracking control, etc. Such data encapsulate specific geometry characteristics which imply special treatments. In this context, we introduce a new learning-from-demonstration platform that can learn SPD profiles (parametrized using Cholesky factorization or Riemannian metrics) from expert demonstrations and adapt to new situations. This approach uses kernelized movement primitives (KMP) whose predictions allow for the retrieval of proper SPD matrices. The proposed approach has been validated in simulations by learning and adapting stiffness ellipsoids.
|
|
12:30-13:30, Paper TuPS1.60 | Add to My Program |
Optimized Locomotion for Energy-Efficient Quadrupedal Robot Over Rough Terrain |
Chen, Lu | The Chinese University of Hong Kong, Shenzhen |
Sun, Caiming | The Chinese University of Hong Kong, Shenzhen |
Zhang, Aidong | The Chinese University of Hong Kong, Shenzhen |
Keywords: Legged Robots, Motion and Path Planning, Optimization and Optimal Control
Abstract: This paper addresses the problem of achieving energy-efficient locomotion in statically stable walking quadrupedal robot. Conventionally, both of the locomotion planner and locomotion controller of a quadrupedal robot do not consider energy efficiency, as a result, the energy consumption is usually quite huge. This paper focuses on maximizing locomotion efficiency in two ways: Firstly, a COG (center of gravity) trajectory optimizer is used to control the trunk kept a constant distance to its supporting surface. As a result, the COG fluctuations are significantly suppressed. Secondly, a foothold planner is used to select footholds which can generate postures for the stance legs with minimized sum of joint torque. Both of the COG trajectory optimizer and foothold planner are based on the real-time terrain map information. Our approach brings some constraints in locomotion, but results in lower energy consumption. To this end, we present a novel energy optimization approach that enables the robot to locomote efficiently. We experimentally validate our approach with the quadrupedal robot Pegasus using statically stable gait by autonomously traversing typical terrains such flat surface, slope and stairs. The experimental results indicate that our approach has a potential for achieving energy-efficient locomotion.
|
|
12:30-13:30, Paper TuPS1.61 | Add to My Program |
Pose-Graph Based Indoor Navigation Test for Unmanned Underwater Vehicle Navigation |
Lee, Yeongjun | Korea Research Institute of Ships and Ocean Engineering |
Jung, Jongdae | Korea Research Institute of Ships and Ocean Engineering |
Choi, Hyun-Taek | Korea Institute of Oceans Science and Technology |
Keywords: Localization, Mapping, Autonomous Vehicle Navigation
Abstract: This paper presents a preliminary experimental result of testing pose-graph based indoor navigation for the purpose of application to unmanned underwater vehicles (UUV). To verify the usefulness and to estimate the performance of the pose-graph technique, we conduct an indoor test using a ground robot kobuki [1]. In this experiment, the robot travels indoor at an office and at a corridor. This allows creating a pose-graph of robot path continuously and building a map using a light detection and ranging (LIDAR) sensor at the front of the robot, simultaneously.
|
|
12:30-13:30, Paper TuPS1.62 | Add to My Program |
Magnetic Sensor Based Probe for Microrobot Detection and Localization |
Kroubi, Tarik | University Mouloud Mammeri of Tizi-Ouzou, Algeria |
Belharet, Karim | Hautes Etudes d'Ingénieur - HEI Campus Centre |
Bennamane, Kamal | University Mouloud Mammeri , TiziOuzou |
Keywords: Localization, Micro/Nano Robots, Automation at Micro-Nano Scales
Abstract: The rapidity developing field of nanomedicine can significantly impact human disease therapy. Magnetic drug targeting using magnetic microrobots offers a potential solution. To perform the technique of non-invasive drug delivery with high precision, we develop a magnetic sensor based probe for microrobot detection and localization
|
|
12:30-13:30, Paper TuPS1.63 | Add to My Program |
Manipulation Planning with Soft Orientation Constraints Based on Composite Configuration Space |
Wang, Jiangping | Hangzhou Dianzi University |
Liu, Shirong | Hangzhou Dianzi University |
Zhang, Botao | Hangzhou Dianzi University |
Wu, Qiuxuan | Hangzhou Dianzi University |
Yu, Changbin (Brad) | The Australian National University |
Keywords: Manipulation Planning, Motion and Path Planning, Kinematics
Abstract: This poster proposes an efficient and probabilistic complete planning algorithm to address the motion planning problem with soft constraints for spherical wrist manipulators. First, we present a novel configuration space termed Composite Configuration Space ("Composite Space" for short), which is composed of joint space and task space. Then, a collision-free path is generated in the Composite Space by the Rapidly-exploring Random Trees (RRT) algorithm. Finally, the planned path based on the Composite Space is mapped into the corresponding joint-space path by a local planner. The local planner utilizes the analytical inverse kinematics (IK) solver of spherical wrist joints to calculate the joint configurations, thus the proposed planner is characterized by high efficiency and no numerical iteration. This approach can effectively improve the smoothness of the end-effector and joint trajectories. The effectiveness of the proposed algorithm is demonstrated on the Willow Garage's PR2 simulation platform in a wide range of orientation-constrained scenarios.
|
|
12:30-13:30, Paper TuPS1.64 | Add to My Program |
Real-Time Sampling-Based Optimization on FPGA for Accurate Grid Map Merging in Embedded Robotic Systems |
Lee, Heoncheol | Kumoh National Institute of Technology |
Lee, Seung-Hwan | Kumoh National Institute of Technology |
Keywords: Mapping, Multi-Robot Systems, Optimization and Optimal Control
Abstract: This paper addresses the real-time optimization problem of grid map merging in embedded robotics systems. When sampling-based optimization such as particle swarm optimization (PSO) is applied to solve the problem, it should be accelerated to satisfy the real-time requirements of embedded robotic systems. This paper proposes a new variant of the PSO conducted on a field-programmable gate array (FPGA) and can reduce computation times by paralleling the computation blocks based on hardware resources on a FPGA. The proposed method was implemented with synthesizable hardware description languages and evaluated by post-layout simulations through FPGA development tools.
|
|
12:30-13:30, Paper TuPS1.65 | Add to My Program |
Stair Environment Mapping and Walk-Able Plane Detecting Algorithm for Quadrupedal Robot's Locomotion |
Woo, Seungjun | Sungkyunkwan University |
Moon, Hyungpil | Sungkyunkwan University |
Keywords: Mapping, Visual-Based Navigation, Legged Robots
Abstract: In this work-in-progress report, we present a mapping method of stairs for a quadruped robot based on point-cloud measurements and stair's geometric property. Because of the sensor noise from the distance sensor, the mapping result from the conventional point-cloud processing mapping algorithm like occupancy mapping is highly affected by the noise. As a result, the mapping methods, which are only based on sensor measurement, can not be fully reliable. Unlike the other mobile robots that use the map information of their surroundings for path planning or localization, the quadrupedal robot uses the map information for actual contact. Therefore, accurate map data is required for the robot's locomotion. We overcome the sensor noise by representing the mapping by the plane base. Each step of the stairs is represented in its representative plane model, which is calculated by fusing measurement and estimation plane model. The suggested stair mapping algorithm can be used for locomotion and localization of quadrupedal robot on the stairs. The coverage of the algorithm is not limited only for a quadrupedal robot but any robot that tries to navigate on the stairs.
|
|
12:30-13:30, Paper TuPS1.66 | Add to My Program |
Design, Modelling and Adaptive Control of a Novel Autonomous Underwater Vehicle Equipped with Vectored Thrusters |
Jisen, Li | CUHK(SZ) |
Sun, Caiming | The Chinese University of Hong Kong, Shenzhen |
Zhang, Jiaming | The Chinese University of Hong Kong, Shenzhen |
Zhang, Aidong | The Chinese University of Hong Kong, Shenzhen |
Keywords: Marine Robotics, Model Learning for Control, Kinematics
Abstract: A large portion of the earth surface is covered by ocean, most of which still remains unexplored due to the complex underwater environment. To explore the complex underwater world, autonomous underwater vehicle(AUV) with both high speed and high maneuverability is required. To fulfill the requirement, an AUV which is equipped with six fixed thrusters and two vectored thrusters is designed. By changing the configuration of the thrusters, it can achieve movement state transformation. The design, simulation and experimental results will be shown in this poster.
|
|
12:30-13:30, Paper TuPS1.67 | Add to My Program |
Hovering Control of a TTURT with Thrust Vector Decomposition Technique |
Bak, Jeongae | Seoul National University |
Moon, Yecheol | Hanyang University |
Jin, Sangrok | Pusan National University |
Kim, Jongwon | Seoul National University |
Seo, TaeWon | Hanyang University |
Keywords: Marine Robotics, Motion Control, Sensor-based Control
Abstract: This late-breaking poster paper presents hovering control of an underwater robot with tilting thruster. The tilting thruster mechanism can implement six-degree-of-freedom (DOF) motion with only four thrusters, but tilting motion makes the system nonlinear. In order to solve the nonlinear problem, nonlinear thrust vector is appropriately decomposed, then we added some constraints with extra degrees of freedom using null space approach. We designed PD controller obtaining thrust forces and continuous tilting angles and applied it to the robot system. We verify improving the hovering ability of the robot by simulation.
|
|
12:30-13:30, Paper TuPS1.68 | Add to My Program |
Preliminary Study for Developing a Vision-Based Detection System of Unmanned Surface Vessels |
Park, Jeonghong | KRISO |
Lee, Yeongjun | Korea Research Institute of Ships and Ocean Engineering |
Park, Jin-Yeong | Korea Research Institute of Ships & Ocean Engineering |
Kim, Kihun | KRISO |
Son, Namsun | Korea Research Institute of Ships and Ocean Engineering |
Keywords: Marine Robotics, Surveillance Systems, Field Robots
Abstract: This paper describes the framework for a vision-based detection system of unmanned surface vessels (USVs) with minimizing human intervention. In order to compensate each other's drawbacks due to characteristics of each camera, the system consists of an electro-optical (EO) camera, an infrared (IR) camera, and panorama cameras for omnidirectional situational awareness. For object identification and recognition, a systematic procedure using image processing and convolutional neural network (CNN) approaches is framed. As preliminary study, the performance and practical feasibility of the proposed approach demonstrates using field test dataset.
|
|
12:30-13:30, Paper TuPS1.69 | Add to My Program |
Design and Analysis of the All-In-One Actuation Module with Multi-Sensors |
Park, Dongil | Korea Institute of Machinery and Materials (KIMM) |
Kim, Hwi-su | Korea Institute of Machinery & Materials |
Park, Jongwoo | Korea Institue of Machinery & Materials |
Park, Chanhun | KIMM |
Kim, Byung-in | Korea Institute of Machinery & Materials |
Keywords: Mechanism Design, Cooperating Robots
Abstract: The cooperative robot is one of the innovative robot which leads industrial automation by combining human skill with the strength, speed, repeatability and precision of robot in the human-robot coexistence environment. Many researches have been studied about the actuation modules for the cooperative robot in which a motor, a reducer, an encoder and sensors are integrated for the advantages of manufacturing, maintenance and flexibility. In the paper, a new concept of the actuation module with multi sensors is proposed for the safety and maintenance and is optimally designed through the integrated analysis of structural analysis and thermal analysis.
|
|
12:30-13:30, Paper TuPS1.70 | Add to My Program |
The Combination Function for Multi-Leg Modular Robot, Bio-Mimicked from Ant’s Behavior |
Yeoh, Chin Ean | Kyungpook National University |
Kim, Tae-hyun | Kyungpook National University |
Lee, Sang-Ryong | KNU |
Hak, Yi | Kyungpook National University |
Keywords: Mechanism Design, Legged Robots, Gripper and Other End-Effectors
Abstract: This paper presents the combining function of the multi-legged modular robot, which is bio-inspired from ants. To this end, the pair of hook-link structure and holder are proposed as the coupling mechanisms with the ant’s claws structure. The steps of the robot combination using the proposed coupling mechanisms is exhibited. As a result, imitating the ant’s structure allows it to achieve the prosperity of merging the main concept of a modular robot, legged robot and, swarm robot into the developed robot for this project.
|
|
12:30-13:30, Paper TuPS1.71 | Add to My Program |
3-DOF Manipulator Design for a Slender-Shaped Wide End-Effector |
Park, Garam | Hanyang Unviersity |
Hong, Jooyoung | Seoul National University |
Lee, Jiseok | Hanyang University |
Kim, Jongwon | Seoul National University |
Seo, TaeWon | Hanyang University |
Keywords: Mechanism Design, Parallel Robots, Field Robots
Abstract: This paper introduces a 3-degree-of-freedom manifold composed of three linear actuators. The proposed mechanism has a workspace suitable for facade cleaning and can compensate horizontal position due to disturbance in gondola-based exterior wall cleaning. Position, velocity kinematic and Jacobian based singularity analysis is presented, and kinematic variables are defined to extend singularity free workspace. This study can be applied to robot arm for facade cleaning in the future
|
|
12:30-13:30, Paper TuPS1.72 | Add to My Program |
Design of a Variable Counterbalance Mechanism to Minimize Required Torque of Robot Arm |
Kim, Hwi-su | Korea Institute of Machinery & Materials |
Park, Dongil | Korea Institute of Machinery and Materials (KIMM) |
Park, Chanhun | KIMM |
Keywords: Mechanism Design, Robot Safety, Physical Human-Robot Interaction
Abstract: In recent years, with increasing requirements of human–robot co-existence, safety has become one of the most important issues in the field of robotics. The counterbalance mechanism (CBM) can be a suitable solution to address this issue as it can ensure intrinsic safety and minimize the required torque of actuators. In this paper, we propose a variable counter -balance mechanism (VCBM) that improves upon the CBM; the proposed mechanism can adjust the counterbalancing torque according to the weight of the payload attached at the end effector of a robot arm such as a gripper. Therefore, the torque required to support not only the self-weight of the robot arm but also any additional external load can be effectively compensated for by the proposed VCBM.
|
|
12:30-13:30, Paper TuPS1.73 | Add to My Program |
IRonCub: Towards Aerial Humanoid Robotics |
Pucci, Daniele | Italian Institute of Technology |
Fiorio, Luca | Istituto Italiano Di Tecnologia |
Traversaro, Silvio | Istituto Italiano Di Tecnologia |
Nava, Gabriele | Istituto Italiano Di Tecnologia |
L'Erario, Giuseppe | Istituto Italiano Di Tecnologia |
Mohamed, Hosameldin Awadalla Omer | Italian Institute of Technology |
Bergonti, Fabio | Italian Institute of Technology |
Benenati, Emilio | Istituto Italiano Di Tecnologia |
Metta, Giorgio | Istituto Italiano Di Tecnologia (IIT) |
Keywords: Mobile Manipulation
Abstract: Science fiction has long inspired pioneers of new areas of Engineering. When imagination meets the current needs of civil society, creative thinking then often gets real, and projects aiming at breakthroughs for advancing the scientific state of the art are put in place. Robotics is a scientific field that has always been driven by visionary applications of Engineering, often receiving impetus from the human will of having extended locomotion and interaction capacities. iRonCub breaks the ground for unifying Manipulation, Aerial, and Terrestrial Locomotion by implementing Aerial Humanoid Robotics. Aerial humanoid robots can then fly, walk, manipulate, and transport objects in the surrounding environment, thus being pivotal for disaster response and opening new branches of applications for humanoid robots. By doing so, aerial humanoid robots overcome the lack of terrestrial locomotion of aerial manipulators and extend the locomotion capabilities of humanoid robots to the flight case.
|
|
12:30-13:30, Paper TuPS1.74 | Add to My Program |
A Stabilization Analysis of Omni-Mobile Manipulator with 4K Camera |
Quan, Chenghao | Gwangju Institute of Science and Technology (GIST) |
Kim, Jiyong | GIST |
Hong, Yohan | Gwangju Institute of Science and Technology |
Kim, Mun Sang | GIST |
Keywords: Mobile Manipulation, Service Robots, Performance Evaluation and Benchmarking
Abstract: In this paper, we present a stabilization analysis of omni-mobile manipulator with 4K camera. For the image stabilization evaluation, we have adopted the inter-frame transformation fidelity (ITF) which have been used in the image stabilization literature. The experimental results show that the proposed approach has a high potential to be applicable to the image stabilization control in robot-based camera productions.
|
|
TuBT1 Regular session, L1-R1 |
Add to My Program |
3D Vision and Pose Estimation |
|
|
Chair: Xiang, Zhiyu | Zhejiang University |
Co-Chair: Schwertfeger, Sören | ShanghaiTech University |
|
14:45-15:00, Paper TuBT1.1 | Add to My Program |
ESKO6d - a Binocular and RGB-D Dataset of Stored Kitchen Objects with 6d Poses |
Richter-Klug, Jesse | Universität Bremen |
Wellhausen, Constantin | Universität Bremen |
Frese, Udo | Universität Bremen |
Keywords: Object Detection, Segmentation and Categorization, Visual Learning, Computer Vision for Other Robotic Applications
Abstract: We present a new dataset with the goal of advancing the state-of-the-art in object pose estimation especially for stored porcelain and glass crockery in kitchen scenes. Specifically, the ESKO6d (EASE Stored Kitchen Objects with 6d poses) dataset features texture-less, glossy or glassy ordinary used objects which were naturally stored in a cupboard, drawer or dishwasher. There is a large degree of occlusion being the specific challenge in these scenes. Each scene was recorded in video sequences by two cameras (RGB-D (Kinect) and binocular) within multiple setup stages. The dataset contains an RGB-D image or binocular RGB image plus stereo-matched depth image as well as 6d pose ground truth and instance segmentation. Our dataset contains twelve stored object scenes with a combined amount of 47 video sequences captured by each camera, resulting in over 17k annotated Kinect images and more than 42k annotated stereo images showing around 50 different objects. The ground truth annotation is precise to 3:5mm ADD (details see paper). The dataset can be accessed under http://www.informatik.uni-bremen.de/esko6d-dataset. Besides the concrete dataset we propose a method of ground truth pose measurement based on an external 3d tracking system that allows on the one hand to precisely measure the object’s pose inside a tight packed storage and on the other hand to obtain the object pose in several images with just one manual measurement.
|
|
15:00-15:15, Paper TuBT1.2 | Add to My Program |
Pose Estimation for Omni-Directional Cameras Using Sinusoid Fitting |
Kuang, Haofei | ShanghaiTech University |
Xu, Qingwen | ShanghaiTech University |
Long, Xiaoling | ShanghaiTech University |
Schwertfeger, Sören | ShanghaiTech University |
Keywords: Omnidirectional Vision, Localization, SLAM
Abstract: We propose a novel pose estimation method for geometric vision of omni-directional cameras. On the basis of the regularity of the pixel movement after camera pose changes, we formulate and prove the sinusoidal relationship between pixels movement and camera motion. We use the improved Fourier-Mellin invariant (iFMI) algorithm to find the motion of pixels, which was shown to be more accurate and robust than the feature-based methods. While iFMI works only on pin-hole model images and estimates 4 parameters (x, y, yaw, scaling), our method works on panoramic images and estimates the full 6 DoF 3D transform, up to an unknown scale factor. For that we fit the motion of the pixels in the panoramic images, as determined by iFMI, to two sinusoidal functions. The offsets, amplitudes and phase-shifts of the two functions then represent the 3D rotation and translation of the camera between the two images. We perform experiments for 3D rotation, which show that our algorithm outperforms the feature-based methods in accuracy and robustness. We leave the more complex 3D translation experiments for future work.
|
|
15:15-15:30, Paper TuBT1.3 | Add to My Program |
Region-Wise Polynomial Regression for 3D Mobile Gaze Estimation |
Su, Dan | City University of Hong Kong |
Li, You-Fu | City University of Hong Kong |
Chen, Hao | City University of Hong Kong |
Keywords: Computer Vision for Other Robotic Applications, Cognitive Human-Robot Interaction
Abstract: In the context of mobile gaze tracking techniques, a 3D gaze point can be calculated as the middle point between two 3D visual axes. To infer gaze directions and eyeball positions, a nonlinear optimization problem is typically formulated to minimize the angular disparities between the training gaze directions and prediction ones. Nonetheless, the experimental results reported by some previous works show that this kind of approaches are very likely to yield large prediction errors hence considered less useful for human-machine interactions. In this study, we aim to address this widespread issue in three aspects. At first, instead of using a global regression model, a simple local polynomial model is proposed to back-project a pupil center onto its corresponding visual axis. Based on the Leave-One-Out cross-validation criterion, the partition structure is automatically learned in the process of resolving a homography-like relationship. Secondly, a good starting point for nonlinear-optimization is obtained by the image eyeball center, which can be estimated by systematic parallax errors. Meanwhile, it is necessary to add the suitable constraints for 3D eye positions. Otherwise, the optimization may end up with trivial solutions, i.e., faraway eye positions. Thirdly, we explore a strategy for designing the spatial distribution of calibration points in a principled manner. The experiment results demonstrate that an encouraging gaze estimation accuracy can be achieved by our proposed framework for both the normal vision and eyewear users.
|
|
15:30-15:45, Paper TuBT1.4 | Add to My Program |
Camera Pose Estimation Based on PnL with a Known Vertical Direction |
Lecrosnier, Louis | IRESEEM |
Boutteau, Rémi | IRSEEM |
Vasseur, Pascal | Université De Rouen |
Savatier, Xavier | Irseem Ea 4353 |
Fraundorfer, Friedrich | Graz University of Technology |
Keywords: Computational Geometry, Computer Vision for Other Robotic Applications, RGB-D Perception
Abstract: In this paper, we address the problem of camera pose estimation using 2D and 3D line features, also known as PnL (Perspective-n-Line) with a known vertical direction. The minimal number of line correspondences required to estimate the complete camera pose is 3 (P3L) in the general case, yielding to a minimum of 8 possible solutions. Prior knowledge of the vertical direction, such as provided by common sensors (e.g. Inertial Measurement Unit, or IMU), reduces the problem to a 4 Degree of Freedom (DoF) problem and outputs a single solution. We benefit this fact to decouple the remaining rotation estimation and the translation estimation and we present a two-fold contribution: (1) we present a linear formulation of the PnL problem in Plücker lines coordinates with a known vertical direction, including a Gauss-Newton-based orientation and location refinement to compensate IMU sensor noise. (2) we propose a new efficient RANdom SAmple Consensus (RANSAC) scheme for both feature pairing and outliers rejection based solely on rotation estimation from 2 line pairs. This greatly diminishes the computational cost compared to a RANSAC3 or RANSAC4 scheme. We evaluate our algorithms on synthetic data and on our own real dataset. Experimental results show state of the art results in term of accuracy and runtime, when facing 2D noise, 3D noise and vertical direction sensor noise.
|
|
15:45-16:00, Paper TuBT1.5 | Add to My Program |
3D Reconstruction by Single Camera Omnidirectional Multi-Stereo System |
Chen, Shuya | Zhejiang University |
Xiang, Zhiyu | Zhejiang University |
Zou, Nan | Zhejiang University |
Chen, Yiman | Zhejiang University |
Qiao, Chengyu | Zhejiang University |
Keywords: Omnidirectional Vision, Computer Vision for Automation
Abstract: Omnidirectional catadioptric systems are popular in robotic applications thanks to their large field of view. For 3D scene reconstruction in a single shot, usually two different catadioptric cameras are needed. More cameras may contribute to better reconstruction while larger mounting space and higher power cost are required. In this paper, a single camera multi-stereo catadioptric system with vertical and horizontal baseline structure is proposed. It features achieving multi-pair of central or non-central omnidirectional stereos in a compact manner. To make the 3D reconstruction process general and adaptive to various types of system configurations, a flexible calibration and reconstruction algorithm pipeline is presented. The algorithm features approximating the system into multiple central sub-cameras and carrying out the stereo matching in a spherical representation. In addition, an effective 3D point cloud fusion algorithm is proposed to optimize the reconstruction results from multiple stereo pairs. The experiment carried out with synthetic and real data verified the feasibility and effectiveness of our system.
|
|
16:00-16:15, Paper TuBT1.6 | Add to My Program |
Efficient Environment Guided Approach for Exploration of Complex Environments |
Butters, Daniel Benjamin | University College London |
Jonasson, Emil T. | UK Atomic Energy Authority |
Stuart-Smith, Robert | University College London |
Pawar, Vijay Manohar | University College London |
Keywords: Range Sensing, Motion and Path Planning, Computer Vision for Other Robotic Applications
Abstract: Remote inspection of a complex environment is a difficult, time consuming task for human operators to perform. The need to manually avoid obstacles whilst considering other performance factors i.e. time taken, joint effort and information gained represents significant challenges to continuous operation. This paper proposes an autonomous robotic solution for exploration of an unknown, complex environment using a high DoF robot arm with an eye in hand depth sensor. The main contribution of this work is a new strategy to find the next best view by evaluating frontier regions of the map to maximise coverage, in contrast to many current approaches which densely sample joint or workspace configurations of the robot. Multiple utility functions were evaluated that showed different behaviours. Our results indicated that the presented algorithm can explore an arbitrary environment efficiently while optimising various performance criteria based on the utility function chosen, application constraints and the desires of the user.
|
|
TuBT2 Regular session, L1-R2 |
Add to My Program |
Deep Learning for Computer Vision |
|
|
Chair: Cheng, Hong | University of Electronic Science and Technology |
Co-Chair: Ang Jr, Marcelo H | National University of Singapore |
|
14:45-15:00, Paper TuBT2.1 | Add to My Program |
Siamese Convolutional Neural Network for Sub-Millimeter-Accurate Camera Pose Estimation and Visual Servoing |
Yu, Cunjun | Nanyang Technological University |
Cai, Zhongang | Nanyang Technological University |
Pham, Hung | Nanyang Technological University |
Pham, Quang-Cuong | NTU Singapore |
Keywords: Deep Learning in Robotics and Automation, Computer Vision for Automation, Visual Servoing
Abstract: Visual Servoing (VS), where images taken from a camera typically attached to the robot end-effector are used to guide the robot motions, is an important technique to tackle robotic tasks that require a high level of accuracy. We propose a new neural network, based on a Siamese architecture, for highly accurate camera pose estimation. This, in turn, can be used as a final refinement step following a coarse VS or, if applied in an iterative manner, as a standalone VS on its own. The key feature of our neural network is that it outputs the relative pose between any pair of images, and does so with sub-millimeter accuracy. We show that our network can reduce pose estimation errors to 0.6 mm in translation and 0.4 degrees in rotation, from initial errors of 10 mm / 5 degrees if applied once, or of several cm / tens of degrees if applied iteratively. The network can generalize to similar objects, is robust against changing lighting conditions, and to partial occlusions (when used iteratively). The high accuracy achieved enables tackling low-tolerance assembly tasks downstream: using our network, an industrial robot can achieve 97.5% success rate on a VGA-connector insertion task without any force sensing mechanism.
|
|
15:00-15:15, Paper TuBT2.2 | Add to My Program |
INFER: INtermediate Representations for FuturE PRediction |
Srikanth, Shashank | International Institute of Information Technology, Hyderabad |
Ansari, Junaid Ahmed | International Institute of Information Technology, Hyderabad |
Ramesh Kumar, Karnik Ram | IIIT Hyderabad |
Sharma, Sarthak | International Institute of Information Technology, Hyderabad |
Jatavallabhula, Krishna Murthy | International Institute of Information Technology Hyderabad |
Krishna, Madhava | IIIT Hyderabad |
Keywords: Deep Learning in Robotics and Automation, Computer Vision for Automation
Abstract: In urban driving scenarios, forecasting future trajectories of surrounding vehicles is of paramount importance. While several approaches for the problem have been proposed, the best-performing ones tend to require extremely detailed input representations (e.g. image sequences). As a result, such methods do not generalize to datasets they have not been trained on. In this paper, we propose intermediate representations that are particularly well-suited for future prediction. As opposed to using texture (color) information from images, we condition on semantics and train an autoregressive model to accurately predict future trajectories of traffic participants (vehicles). We demonstrate that semantics provide a significant boost over techniques that operate over raw pixel intensities/disparities. Uncharacteristic of state-of-the-art approaches, our representations and models generalize across different sensing modalities (stereo imagery, LiDAR, a combination of both), and also across completely different datasets, collected across several cities, and also across countries where people drive on opposite sides of the road (left-handed vs right-handed driving). Additionally, we demonstrate an application of our approach in multi-object tracking (data association). To foster further research in transferable representations and ensure reproducibility, we release all our code and data.
|
|
15:15-15:30, Paper TuBT2.3 | Add to My Program |
End-To-End Driving Model for Steering Control of Autonomous Vehicles with Future Spatiotemporal Features |
Wu, Tianhao | University of Electronic Science and Technology of China |
Luo, Ao | University of Electronic Science and Technology of China |
Huang, Rui | University of Electronic Science and Technology of China |
Cheng, Hong | University of Electronic Science and Technology |
Zhao, Yang | University of Electronic Science and Technology of China |
Keywords: Deep Learning in Robotics and Automation, Computer Vision for Other Robotic Applications, AI-Based Methods
Abstract: End-to-end deep learning has gained considerable interests in autonomous driving vehicles in both academic and industrial fields, especially in decision making process. One critical issue in decision making process of autonomous driving vehicles is steering control. Researchers has already trained different artificial neural networks to predict steering angle with front-facing camera data stream. However, existing end-to-end methods only consider the spatiotemporal relation on a single layer and lack the ability of extracting future spatiotemporal information. In this paper, we propose an end-to-end driving model based on Convolutional Long Short-Term Memory (Conv-LSTM) neural network with a Multi-scale Spatiotemporal Integration (MSI) module, which aiming to encode the spatiotemporal information from different scales for steering angle prediction. Moreover, we employ future sequential information to enhance spatiotemporal features of the end-to-end driving model. We demonstrate the efficiency of proposed end-to-end driving model on the public Udacity dataset with comparison of some existing methods. Experimental results show that the proposed model has better performances than other existing methods, especially in some complex scenarios. Furthermore, we evaluate the proposed driving model on a real-time autonomous vehicle, and results show that the proposed driving model is able to predict the steering angle with high accuracy compared to skilled human driver.
|
|
15:30-15:45, Paper TuBT2.4 | Add to My Program |
PointAtrousNet: Point Atrous Convolution for Point Cloud Analysis |
Pan, Liang | National University of Singapore |
Wang, Pengfei | National University of Singapore |
Chew, Chee Meng | National University of Singapore |
Keywords: Deep Learning in Robotics and Automation, Computer Vision for Other Robotic Applications
Abstract: In this paper, we propose a permutation-invariant architecture - PointAtrousNet (PAN), which focuses on exploiting multi-scale local geometric details for point cloud analysis. Inspired by Atrous Convolution in image domains, we propose the Point Atrous Convolution (PAC) operation. Our PAC can effectively enlarge the receptive field of filters without introducing more parameters or increasing computation amount. In particular, we propose a novel Point Atrous Spatial Pyramid Pooling (PASPP) module to explicitly exploit neighboring contextual information at multiple scales. Moreover, local geometric details are captured by constructing neighborhood graphs in metric and feature spaces. Experimental results show that our PAN achieves state-of-the-art performance on various point cloud inference applications.
|
|
15:45-16:00, Paper TuBT2.5 | Add to My Program |
A Convolutional Network for Joint Deraining and Dehazing from a Single Image for Autonomous Driving in Rain |
Sun, Hao | National University of Singapore |
Ang Jr, Marcelo H | National University of Singapore |
Rus, Daniela | MIT |
Keywords: Deep Learning in Robotics and Automation, Computer Vision for Transportation
Abstract: In this paper, we focus on a rain removal task from a single image of the urban street scene for autonomous driving in rain. We develop a Convolutional Neural Network which takes a rainy image as input, and directly recovers a clean image in the presence of rain streaks, atmospheric veiling effect (haze, fog, mist) caused by distant rain streak accumulation. We propose a synthetic dataset containing images of urban street scenes with different rain intensities, orientations and haziness levels for training and evaluation. We evaluate our method quantitatively and qualitatively on the synthetic data. Experiments show that our model outperforms state-of-the-art methods. We also test our method qualitatively on the real-world data. Our model is fast and it takes 0.05s for an image of 1024 x 512. Our model can be seamlessly integrated with existing image-based high-level perception algorithms for autonomous driving in rain. Experiment results show that our deraining method improves semantic segmentation and object detection largely for autonomous driving in rain.
|
|
16:00-16:15, Paper TuBT2.6 | Add to My Program |
Improving Learning-Based Ego-Motion Estimation with Homomorphism-Based Losses and Drift Correction |
Wang, Xiangwei | Tongji University/ Carnegie Mellon University |
Maturana, Daniel | Carnegie Mellon University |
Yang, Shichao | Carnegie Mellon University |
Wang, Wenshan | Shanghai Jiao Tong University, Research Institute of Robotics |
Chen, Qijun | Tongji University |
Scherer, Sebastian | Carnegie Mellon University |
Keywords: Deep Learning in Robotics and Automation, Computer Vision for Transportation
Abstract: Visual odometry is an essential problem for mobile robots. Traditional methods for solving VO mostly utilize geometric optimization. While capable of achieving high accuracy, these methods require accurate sensor calibration and complicated parameter tuning to work well in practice. With the rise of deep learning, there has been increased interest in the end-to-end, learning-based methods for VO, which have the potential to improve robustness. However, learning-based methods for VO so far are less accurate than geometric methods. We argue that one of the main issues is that the current ego-motion estimation task is different from other problems where deep learning has been successful such as object detection. We define a novel cost function for learning-based VO considering the mathematical properties of the group homomorphism. In addition to the standard L2 loss, we incorporate losses based on the identity, inverse and closure properties of SE(3) rigid motion. Furthermore, we propose to reduce the VO drift by estimating the drivable regions using semantic segmentation and incorporate this information into a pose graph optimization. Experiments on KITTI datasets show that the novel cost function can improve ego-motion estimation compared to the state-of-the-art and the drivable region-based correction further reduces the VO drift.
|
|
TuBT3 Regular session, L1-R3 |
Add to My Program |
Learning and Adaptive Systems II |
|
|
Chair: Gao, Yue | Shanghai JiaoTong University |
Co-Chair: Wang, Wenxue | Shenyang Institute of Automation, CAS |
|
14:45-15:00, Paper TuBT3.1 | Add to My Program |
Meta-Learning for Multi-Objective Reinforcement Learning |
Chen, Xi | KTH |
Ghadirzadeh, Ali | KTH Royal Institute of Technology, Aalto University |
Björkman, Mårten | KTH |
Jensfelt, Patric | KTH - Royal Institute of Technology |
Keywords: Deep Learning in Robotics and Automation, Learning and Adaptive Systems, AI-Based Methods
Abstract: Multi-objective reinforcement learning (MORL) is the generalization of standard reinforcement learning (RL) approaches to solve sequential decision making problems that consist of several, possibly conflicting, objectives. Generally, in such formulations, there is no single optimal policy which optimizes all the objectives simultaneously, and instead, a number of policies has to be found each optimizing a preference of the objectives. In this paper, we introduce a novel MORL approach by training a meta-policy, a policy simultaneously trained with multiple tasks sampled from a task distribution, for a number of randomly sampled Markov decision processes (MDPs). In other words, the MORL is framed as a meta-learning problem, with the task distribution given by a distribution over the preferences. We demonstrate that such a formulation results in a better approximation of the Pareto optimal solutions in terms of both the optimality and the computational efficiency. We evaluated our method on obtaining Pareto optimal policies using a number of continuous control problems with high degrees of freedom
|
|
15:00-15:15, Paper TuBT3.2 | Add to My Program |
A Comparative Analysis on the Use of Autoencoders for Robot Security Anomaly Detection |
Olivato, Matteo | University of Verona |
Cotugno, Omar | Università La Sapienza Roma |
Brigato, Lorenzo | Department of Computer, Control and Management Engineering, Sapi |
Bloisi, Domenico | University of Basilicata |
Farinelli, Alessandro | University of Verona |
Iocchi, Luca | Sapienza University of Roma |
Keywords: Deep Learning in Robotics and Automation, Learning and Adaptive Systems, Failure Detection and Recovery
Abstract: While robots are more and more deployed among people in public spaces, the impact of cyber-security attacks is significantly increasing. Most of consumer and professional robotic systems are affected by multiple vulnerabilities and the research in this field is just started. This paper addresses the problem of automatic detection of anomalous behaviors possibly coming from cyber-security attacks. The proposed solution is based on extracting system logs from a set of internal variables of a robotic system, on transforming such data into images, and on training different Autoencoder architectures to classify robot behaviors to detect anomalies. Experimental results in two different scenarios (autonomous boats and social robots) show effectiveness and general applicability of the proposed method.
|
|
15:15-15:30, Paper TuBT3.3 | Add to My Program |
Fast and Safe Policy Adaptation Via Alignment-Based Transfer |
Kim, Jigang | Seoul National University |
Choi, Seungwon | Seoul Nat'l University |
Kim, H. Jin | Seoul National University |
Keywords: Deep Learning in Robotics and Automation, Learning and Adaptive Systems, Learning from Demonstration
Abstract: Applying deep reinforcement learning to physical systems, as opposed to learning in simulation, presents additional challenges in terms of sample efficiency and safety. Collecting large amounts of hardware demonstration data is time-consuming and the exploratory behavior of reinforcement learning algorithms may lead the system into dangerous states, especially during the early stages of training. To address these challenges, we apply transfer learning to reuse a previously learned policy instead of learning from scratch. In this paper, we propose a method where given a source policy, policy adaptation is performed via transfer learning to produce a target policy suitable for real-world deployment. For policy adaptation, alignment-based transfer learning is applied to trajectories generated by the source policy and their corresponding safe target trajectories. We apply this method to manipulators and show that the proposed method is applicable to both inter-task and inter-robot transfer whilst considering safety. We also show that the resulting target policy is robust and can be further improved with reinforcement learning.
|
|
15:30-15:45, Paper TuBT3.4 | Add to My Program |
Robotic Tracking Control with Kernel Trick-Based Reinforcement Learning |
Hu, Yazhou | Shenyang Institute of Automation |
Wang, Wenxue | Shenyang Institute of Automation, CAS |
Liu, Hao | Georgia Institute of Technology |
Liu, Lianqing | Shenyang Institute of Automation |
Keywords: Learning and Adaptive Systems, AI-Based Methods
Abstract: Although reinforcement learning is an efficient learning algorithm to solve control problems via interacting with the environment to acquire the optimal control policy, there are still some challenges for reinforcement learning in solving continuous control tasks. Fortunately, the kernel-based methods have the characteristic to deal with continuous control problems via accelerating the speed of convergence and/or improving the convergence precision. What's more, a good reward system can not only speed up the learning processes, but can improve the quality of learning. In this work, a reward function is proposed and a kernel-based reinforcement learning tracking controller is presented for performing tracking control tasks in a robotic manipulator system. A critic system is also introduced to accelerate the speed to find the optimal control policy. The simulation results illustrate that the proposed algorithm can execute tracking control tasks effectively, and the reward function proposed in this work is efficient.
|
|
15:45-16:00, Paper TuBT3.5 | Add to My Program |
Graph-Based Design of Hierarchical Reinforcement Learning Agents |
Tateo, Davide | TU Darmstadt |
Erdenliğ, İdil Su | Politecnico Di Milano |
Bonarini, Andrea | Politecnico Di Milano |
Keywords: Learning and Adaptive Systems, Control Architectures and Programming, Robust/Adaptive Control of Robotic Systems
Abstract: There is an increasing interest in Reinforcement Learning to solve new and more challenging problems, as those emerging in robotics and unmanned autonomous vehicles. To face these complex systems, a hierarchical and multi-scale representation is crucial. This has brought the interest on Hierarchical Deep Reinforcement learning systems. Despite their successful application, Deep Reinforcement Learning systems suffer from a variety of drawbacks: they are data hungry, they lack of interpretability, and it is difficult to derive theoretical properties about their behavior. Classical Hierarchical Reinforcement Learning approaches, while not suffering from these drawbacks, are often suited for finite actions, and finite states, only. Furthermore, in most of the works, there is no systematic way to represent domain knowledge, which is often only embedded in the reward function. We present a novel Hierarchical Reinforcement Learning framework based on the hierarchical design approach typical of control theory. We developed our framework extending the block diagram representation of control systems to fit the needs of a Hierarchical Reinforcement Learning scenario, thus giving the possibility to integrate domain knowledge in an effective hierarchical architecture.
|
|
16:00-16:15, Paper TuBT3.6 | Add to My Program |
Variable Impedance in End-Effector Space: An Action Space for Reinforcement Learning in Contact-Rich Tasks |
Martín-Martín, Roberto | Stanford University |
Lee, Michelle | Stanford University |
Gardner, Rachel | Stanford University |
Savarese, Silvio | Stanford University |
Bohg, Jeannette | Stanford University |
Garg, Animesh | Stanford University |
Keywords: Learning and Adaptive Systems, Deep Learning in Robotics and Automation, Reactive and Sensor-Based Planning
Abstract: Reinforcement Learning (RL) of contact-rich manipulation tasks has yielded impressive results in recent years. While many studies in RL focus on varying the observation space or reward model, few efforts focused on the choice of action space (e.g. joint or end-effector space, position, velocity, etc.). However, studies in robot motion control indicate that choosing an action space that conforms to the characteristics of the task can simplify exploration and improve robustness to disturbances. This paper studies the effect of different action spaces in deep RL and advocates for variable impedance control in end-effector space (VICES) as an advantageous action space for constrained and contact-rich tasks. We evaluate multiple action spaces on three prototypical manipulation tasks: Path Following (task with no contact), Door Opening (task with kinematic constraints), and Surface Wiping (task with continuous contact). We show that VICES improves sample efficiency, maintains low energy consumption, and ensures safety across all three experimental setups. Further, RL policies learned with VICES can transfer across different robot models in simulation, and from simulation to real for the same robot. Further information is available at https://stanfordvl.github.io/vices.
|
|
TuBT4 Regular session, L1-R4 |
Add to My Program |
Award Session II |
|
|
Chair: Asada, Minoru | Osaka University |
Co-Chair: Xiao, Jing | Worcester Polytechnic Institute (WPI) |
|
14:45-15:00, Paper TuBT4.1 | Add to My Program |
Motion Decoupling and Composition Via Reduced Order Model Optimization for Dynamic Humanoid Walking with CLF-QP Based Active Force Control |
Xiong, Xiaobin | California Institute of Technology |
Ames, Aaron | Caltech |
Keywords: Humanoid and Bipedal Locomotion, Humanoid Robots, Legged Robots
Abstract: In this paper, 3D humanoid walking is decoupled into periodic motion and transitional motion, each of which is decoupled into planar walking in the sagittal and lateral plane. Reduced order models (ROMs), i.e. actuated Springloaded Inverted Pendulum (aSLIP) models and Hybrid-Linear Inverted Pendulum (H-LIP) models, are utilized for motion generation on the desired center of mass (COM) dynamics for each type of planar motion. The periodic motion is planned via point foot (underactuated) ROMs for dynamic motion with minimum ankle actuation, while the transitional motion is planned via foot-actuated ROMs for fast and smooth transition. Composition of the planar COM dynamics yields the desired COM dynamics in 3D, which is directly embedded on the humanoid via control Lyapunov function based Quadratic programs (CLF-QPs). Additionally, the ground reaction force profiles of the aSLIP walking are used as desired references for ground contact forces in the CLF-QPs for smooth domain transitions. The proposed framework is realized in an exoskeleton robot in simulation for realizing different walking motion.
|
|
15:00-15:15, Paper TuBT4.2 | Add to My Program |
Early Fusion for Goal Directed Robotic Vision |
Walsman, Aaron | University of Washington |
Bisk, Yonatan | University of Washington |
Gabriel, Saadia | University of Washington |
Misra, Dipendra | Cornell University |
Artzi, Yoav | Cornell University |
Choi, Yejin | University of Washington |
Fox, Dieter | University of Washington |
Keywords: Computer Vision for Other Robotic Applications, Deep Learning in Robotics and Automation, Visual Learning
Abstract: Building perceptual systems for robotics which perform well under tight computational budgets requires novel architectures which rethink the traditional computer vision pipeline. Modern vision architectures require the agent to build a summary representation of the entire scene, even if most of the input is irrelevant to the agent's current goal. In this work, we flip this paradigm, by introducing EarlyFusion vision models that condition on a goal to build custom representations for downstream tasks. We show that these goal specific representations can be learned more quickly, are substantially more parameter efficient, and more robust than existing attention mechanisms in our domain. We demonstrate the effectiveness of these methods on a simulated item retrieval problem that is trained in a fully end-to-end manner via imitation learning.
|
|
15:15-15:30, Paper TuBT4.3 | Add to My Program |
Advanced Autonomy on a Low-Cost Educational Drone Platform |
Eller, Luke | Brown University |
Guerin, Theo | Brown University |
Huang, Baichuan | Brown University |
Warren, Garrett | Brown University |
Yang, Sophie | Brown University |
Roy, Josh | Brown University |
Tellex, Stefanie | Brown |
Keywords: Education Robotics, Aerial Systems: Perception and Autonomy, Sensor Fusion
Abstract: PiDrone is a quadrotor platform created to accompany an introductory robotics course. Students build an autonomous flying robot from scratch and learn to program it through assignments and projects. Existing educational robots do not have significant autonomous capabilities, such as high-level planning and mapping. We present a hardware and software framework for an autonomous aerial robot, in which all software for autonomy can run onboard the drone, implemented in Python. We present an Unscented Kalman Filter (UKF) for accurate state estimation. Next, we present an implementation of Monte Carlo (MC) Localization and FastSLAM for Simultaneous Localization and Mapping (SLAM). The performance of UKF, localization, and SLAM is tested and compared to ground truth, provided by a motion-capture system. Our evaluation demonstrates that our autonomous educational framework runs quickly and accurately on a Raspberry Pi in Python, making it ideal for use in educational settings.
|
|
15:30-15:45, Paper TuBT4.4 | Add to My Program |
Goal-Directed Behavior under Variational Predictive Coding: Dynamic Organization of Visual Attention and Working Memory |
Jung, Minju | Korea Advanced Institute of Science and Technology |
Matsumoto, Takazumi | Okinawa Institute of Science and Technology Graduate University |
Tani, Jun | Okinawa Institute of Science and Technology |
Keywords: Neurorobotics, Deep Learning in Robotics and Automation, Manipulation Planning
Abstract: Mental simulation is a critical cognitive function for goal-directed behavior because it is essential for assessing actions and their consequences. When a self-generated or externally specified goal is given, a sequence of actions that is most likely to attain that goal is selected among other candidates via mental simulation. Therefore, better mental simulation leads to better goal-directed action planning. However, developing a mental simulation model is challenging because it requires knowledge of self and the environment. The current paper studies how adequate goal-directed action plans of robots can be mentally generated by dynamically organizing top-down visual attention and visual working memory. For this purpose, we propose a neural network model based on variational Bayes predictive coding, where goal-directed action planning is formulated by Bayesian inference of latent intentional space. Our experimental results showed that cognitively meaningful competencies, such as autonomous top-down attention to the robot end effector (its hand) as well as dynamic organization of occlusion-free visual working memory, emerged. Furthermore, our analysis of comparative experiments indicated that the introduction of visual working memory and the inference mechanism using variational Bayes predictive coding significantly improved the performance in planning adequate goal-directed actions.
|
|
15:45-16:00, Paper TuBT4.5 | Add to My Program |
Scaling Robot Supervision to Hundreds of Hours with RoboTurk: Robotic Manipulation Dataset through Human Reasoning and Dexterity |
Mandlekar, Ajay Uday | Stanford University |
Booher, Jonathan | Stanford University |
Spero, Max | Stanford University |
Tung, Albert | Stanford University |
Gupta, Anchit | Stanford University |
Zhu, Yuke | Stanford University |
Garg, Animesh | University of Toronto |
Savarese, Silvio | Stanford University |
Fei-Fei, Li | Stanford University |
Keywords: Telerobotics and Teleoperation, Learning from Demonstration, Learning and Adaptive Systems
Abstract: Large, richly annotated datasets have accelerated progress in fields such as computer vision and natural language processing, but replicating these successes in robotics has been challenging. While prior data collection methodologies such as self-supervision have resulted in large datasets, the data can have poor signal-to-noise ratio. By contrast, previous efforts to collect task demonstrations with humans provide better quality data, but they cannot reach the same data magnitude. Furthermore, neither approach places guarantees on the diversity of the data collected, in terms of solution strategies. In this work, we leverage and extend the RoboTurk platform to scale up data collection for robotic manipulation using remote teleoperation. The primary motivation for our platform is two-fold: (1) to address the shortcomings of prior work and increase the total quantity of manipulation data collected through human supervision by an order of magnitude without sacrificing the quality of the data and (2) to collect data on challenging manipulation tasks across several operators and observe a diverse set of emergent behaviors and solutions. We collected over 111 hours of robot manipulation data across 54 users and 3 challenging manipulation tasks in 1 week, resulting in the largest robot dataset collected via remote teleoperation. We evaluate the quality of our platform, the diversity of demonstrations in our dataset, and the utility of our dataset via quantitative and qualitative analysis. For additional results, supplementary videos, and to download our dataset, visit http://roboturk.stanford.edu/realrobotdataset
|
|
16:00-16:15, Paper TuBT4.6 | Add to My Program |
Robot Learning Via Human Adversarial Games |
Duan, Jiali | University of Southern California |
Wang, Qian | University of Southern California |
Pinto, Lerrel Joseph | Carnegie Mellon University |
Kuo, C.-C. Jay | University of Southern California |
Nikolaidis, Stefanos | University of Southern California |
Keywords: Grasping, Cognitive Human-Robot Interaction, Deep Learning in Robotics and Automation
Abstract: Much work in robotics has focused on ``human-in-the-loop'' learning techniques that improve the efficiency of the learning process. However, these algorithms have made the strong assumption of a cooperating human supervisor that assists the robot. In reality, human observers tend to also act in an adversarial manner towards deployed robotic systems. We show that this can in fact improve the robustness of the learned models by proposing a physical framework that leverages perturbations applied by a human adversary, guiding the robot towards more robust models. In a manipulation task, we show that grasping success improves significantly when the robot trains with a human adversary as compared to training in a self-supervised manner.
|
|
TuBT5 Regular session, L1-R5 |
Add to My Program |
Award Session III |
|
|
Chair: Sugano, Shigeki | Waseda University |
Co-Chair: Liu, Yunhui | Chinese University of Hong Kong |
|
14:45-15:00, Paper TuBT5.1 | Add to My Program |
Planning Beyond the Sensing Horizon Using a Learned Context |
Everett, Michael | Massachusetts Institute of Technology |
Miller, Justin | Ford |
How, Jonathan Patrick | Massachusetts Institute of Technology |
Keywords: Deep Learning in Robotics and Automation, Motion and Path Planning, Search and Rescue Robots
Abstract: Last-mile delivery systems commonly propose the use of autonomous robotic vehicles to increase scalability and efficiency. The economic inefficiency of collecting accurate prior maps for navigation motivates the use of planning algorithms that operate in unmapped environments. However, these algorithms typically waste time exploring regions that are unlikely to contain the delivery destination. Context is key information about structured environments that could guide exploration toward the unknown goal location, but the abstract idea is difficult to quantify for use in a planning algorithm. Some approaches specifically consider contextual relationships between objects, but would perform poorly in object-sparse environments like outdoors. Recent deep learning-based approaches consider context too generally, making training/transferability difficult. Therefore, this work proposes a novel formulation of utilizing context for planning as an image-to-image translation problem, which is shown to extract terrain context from semantic gridmaps, into a metric that an exploration-based planner can use. The proposed framework has the benefit of training on a static dataset instead of requiring a time-consuming simulator. Across 42 test houses with layouts from satellite images, the trained algorithm enables a robot to reach its goal 189% faster than with a context-unaware planner, and within 63% of the optimal path computed with a prior map. The proposed algorithm is also implemented on a vehicle with a forward-facing camera in a high-fidelity, Unreal simulation of neighborhood houses.
|
|
15:00-15:15, Paper TuBT5.2 | Add to My Program |
Volumetric Instance-Aware Semantic Mapping and 3D Object Discovery |
Grinvald, Margarita | ETH Zurich |
Furrer, Fadri | ETH Zurich |
Novkovic, Tonci | Autonomous Systems Lab, ETH Zurich |
Chung, Jen Jen | Eidgenössische Technische Hochschule Zürich |
Cadena Lerma, Cesar | ETH Zurich |
Siegwart, Roland | ETH Zurich |
Nieto, Juan | ETH Zürich |
Keywords: RGB-D Perception, Object Detection, Segmentation and Categorization, Mapping
Abstract: To autonomously navigate and plan interactions in real-world environments, robots require the ability to robustly perceive and map complex, unstructured surrounding scenes. Besides building an internal representation of the observed scene geometry, the key insight toward a truly functional understanding of the environment is the usage of higher-level entities during mapping, such as individual object instances. This work presents an approach to incrementally build volumetric object-centric maps during online scanning with a localized RGB-D camera. First, a per-frame segmentation scheme combines an unsupervised geometric approach with instance-aware semantic predictions to detect both recognized scene elements as well as previously unseen objects. Next, a data association step tracks the predicted instances across the different frames. Finally, a map integration strategy fuses information about their 3D shape, location, and, if available, semantic class into a global volume. Evaluation on a publicly available dataset shows that the proposed approach for building instance-level semantic maps is competitive with state-of-the-art methods, while additionally able to discover objects of unseen categories. The system is further evaluated within a real-world robotic mapping setup, for which qualitative results highlight the online nature of the method. Code is available at https://github.com/ethz-asl/voxblox-plusplus.
|
|
15:15-15:30, Paper TuBT5.3 | Add to My Program |
Responsive Joint Attention in Human-Robot Interaction |
Pereira, Andre | KTH Royal Institute of Technology |
Oertel, Catharine | KTH Royal Institute of Technology |
Fermoselle, Leonor | KTH Royal Institute of Technology |
Mendelson, Joe | KTH Royal Institute of Technology |
Gustafson, Joakim | KTH |
Keywords: Social Human-Robot Interaction, Human Factors and Human-in-the-Loop, Entertainment Robotics
Abstract: Joint attention has been shown to be not only crucial for human-human interaction but also human-robot interaction. Joint attention can help to make cooperation more efficient, support disambiguation in instances of uncertainty and make interactions appear more natural and familiar. In this paper, we present an autonomous gaze system that uses multimodal perception capabilities to model responsive joint attention mechanisms. We investigate the effects of our system on people's perception of a robot within a problem-solving task. Results from a user study suggest that responsive joint attention mechanisms evoke higher perceived feelings of social presence on scales that regard the direction of the robot's perception.
|
|
15:30-15:45, Paper TuBT5.4 | Add to My Program |
Deep Dive into Faces: Pose & Illumination Invariant Multi-Face Emotion Recognition System |
Saxena, Suchitra | PES |
Tripathi, Shikha | Faculty of Engineering PES University, Bangalore, India |
T S B, Sudarshan | Faculty of Engineering PES University, Bangalore, India |
Keywords: Recognition, Deep Learning in Robotics and Automation, Computer Vision for Other Robotic Applications
Abstract: One of the advancements in humanization of robots is its ability to recognize human emotions. Facial expression plays a key role in identifying human emotions relative to other cues. In this research, an intelligent network capable of real-time emotion recognition from multiple faces using deep learning technique is presented. The proposed network is based on Convolution Neural Network (CNN) in which three blocks of Convolution layers for feature extraction and two blocks of Dense layers for classification are used. The novelty of this method lies in recognizing emotions from multiple faces simultaneously in real time and its invariance to head pose, illumination and age factor. Most of reported work in literature for multiple faces is for frontal face without illumination variation. The proposed emotion recognition system is deployed on Raspberry Pi3 B+ for human robot interaction applications and achieved an average accuracy of 95.8% in real time.
|
|
15:45-16:00, Paper TuBT5.5 | Add to My Program |
Enthusiastic Robots Make Better Contact |
Saad, Elie | Delft University of Technology (TU Delft) |
Broekens, Joost | Leiden University |
Neerincx, Mark | TNO |
Hindriks, Koen | Vrije Universiteit Amsterdam |
Keywords: Social Human-Robot Interaction, Cognitive Human-Robot Interaction
Abstract: This paper presents the design and evaluation of human-like welcoming behaviors for a humanoid robot to draw the attention of passersby by following a three-step model: (1) selecting a target (person) to engage, (2) executing behaviors to draw the target's attention, and (3) monitoring the attentive response. A computer vision algorithm was developed to select the person, start the behaviors and monitor the response automatically. To vary the robot's enthusiasm when engaging passersby, a waving gesture was designed as basic welcoming behavioral element, which could be successively combined with an utterance and an approach movement. This way, three levels of enthusiasm were implemented: Mild (waving), moderate (waving and utterance) and high (waving, utterance and approach movement). The three levels of welcoming behaviors were tested with a Pepper robot at the entrance of a university building. We recorded data and observation sheets from several hundreds of passersby (N = 364) and conducted post-interviews with randomly selected passersby (N = 28). The level selection was done at random for each participant. The passersby indicated that they appreciated the robot at the entrance and clearly recognized its role as a welcoming robot. In addition, the robot proved to draw more attention when showing high enthusiasm (i.e., more welcoming behaviors), particularly for female passersby.
|
|
16:00-16:15, Paper TuBT5.6 | Add to My Program |
Entropic Risk Measure in Policy Search |
Nass, David | Technische Universität Darmstadt |
Belousov, Boris | Technische Universität Darmstadt |
Peters, Jan | Technische Universität Darmstadt |
Keywords: Probability and Statistical Methods, Learning and Adaptive Systems, Learning from Demonstration
Abstract: With the increasing pace of automation, modern robotic systems need to act in stochastic, non-stationary, partially observable environments. A range of algorithms for finding parameterized policies that optimize for long-term average performance have been proposed in the past. However, the majority of the proposed approaches does not explicitly take into account the variability of the performance metric, which may lead to finding policies that although performing well on average, can perform spectacularly bad in a particular run or over a period of time. To address this shortcoming, we study an approach to policy optimization that explicitly takes into account higher order statistics of the reward function. In this paper, we extend policy gradient methods to include the entropic risk measure in the objective function and evaluate their performance in simulation experiments and on a real-robot task of learning a hitting motion in robot badminton.
|
|
TuBT6 Regular session, L1-R6 |
Add to My Program |
Aerial Robotics II |
|
|
Chair: Saska, Martin | Czech Technical University in Prague |
Co-Chair: Kim, H. Jin | Seoul National University |
|
14:45-15:00, Paper TuBT6.1 | Add to My Program |
Can a Robot Become a Movie Director? Learning Artistic Principles for Aerial Cinematography |
Gschwindt, Mirko | Technical University of Munich |
Camci, Efe | Nanyang Technological University |
Bonatti, Rogerio | Carnegie Mellon University |
Wang, Wenshan | Shanghai Jiao Tong University, Research Institute of Robotics |
Kayacan, Erdal | Aarhus University |
Scherer, Sebastian | Carnegie Mellon University |
Keywords: Aerial Systems: Perception and Autonomy, Deep Learning in Robotics and Automation, Human Factors and Human-in-the-Loop
Abstract: Aerial filming is constantly gaining importance due to the recent advances in drone technology. It invites many intriguing, unsolved problems at the intersection of aesthetical and scientific challenges. In this work, we propose a deep reinforcement learning agent which supervises motion planning of a filming drone by making desirable shot mode selections based on aesthetical values of video shots. Unlike most of the current state-of-the-art approaches that require explicit guidance by a human expert, our drone learns how to make favorable viewpoint selections by experience. We propose a learning scheme that exploits aesthetical features of retrospective shots in order to extract a desirable policy for better prospective shots. We train our agent in realistic AirSim simulations using both a hand-crafted reward function as well as reward from direct human input. We then deploy the same agent on a real DJI M210 drone in order to test the generalization capability of our approach to real world conditions. To evaluate the success of our approach in the end, we conduct a comprehensive user study in which participants rate the shot quality of our methods. Videos of the system in action can be seen at https://youtu.be/qmVw6mfyEmw.
|
|
15:00-15:15, Paper TuBT6.2 | Add to My Program |
Online Trajectory Generation of a MAV for Chasing a Moving Target in 3D Dense Environments |
Jeon, Boseong | Seoul National University |
Kim, H. Jin | Seoul National University |
Keywords: Aerial Systems: Perception and Autonomy, Motion and Path Planning, Surveillance Systems
Abstract: This work deals with a moving target chasing mission of an aerial vehicle equipped with a vision sensor in a cluttered environment. In contrast to obstacle-free or sparse environments, the chaser should be able to handle collision and occlusion together with flight efficiency. In order to tackle these challenges in real-time, we introduce a metric for target visibility and propose a hierarchical chasing planner. In the first phase, we generate a sequence of waypoints and chasing corridors which ensure safety and optimize visibility. In the following phase, the corridors and waypoints are utilized as constraints and objective respectively in quadratic programming from which we complete a dynamically feasible trajectory for chasing. The proposed algorithm is tested in multiple dense environments. The simulator AutoChaser with full code implementation & GUI can be found in https: //github.com/icsl-Jeon/traj_gen_vis and video is available at https://youtu.be/-2d3uDlYR_M.
|
|
15:15-15:30, Paper TuBT6.3 | Add to My Program |
Thermal-Inertial Odometry for Autonomous Flight Throughout the Night |
Delaune, Jeff | Jet Propulsion Laboratory |
Hewitt, Robert | Jet Propulsion Laboratory |
Lytle, Laura | Jet Propulsion Laboratory |
Sorice, Cristina | NASA Jet Propulsion Laboratory |
Thakker, Rohan | Nasa's Jet Propulsion Laboratory, Caltech |
Matthies, Larry | Jet Propulsion Laboratory |
Keywords: Aerial Systems: Perception and Autonomy, Visual-Based Navigation, SLAM
Abstract: Thermal cameras can enable autonomous flight at night without GPS. However, image-based navigation in the thermal infrared spectrum has been researched significantly less than in the visible spectrum. In this paper, we demonstrate closed-loop controlled outdoor flights at night on a quadrotor. Our state estimator can tightly couple inertial data with either thermal images at nighttime, or visual images at daytime. It is integrated in an autonomy framework for motion planning and control, which runs in real time on a standard embedded computer. We analyze thermal-inertial odometry performance extensively from sunset to sunrise, for various thermal non-uniformity levels, and compare it to visual-inertial odometry at daytime.
|
|
15:30-15:45, Paper TuBT6.4 | Add to My Program |
Timepix Radiation Detector for Autonomous Radiation Localization and Mapping by Micro Unmanned Vehicles |
Baca, Tomas | Czech Technical Univerzity in Prague |
Jilek, Martin | Czech Technical University in Prague |
Manek, Petr | Czech Technical University in Prague |
Stibinger, Petr | Czech Technical University in Prague |
Linhart, Vladimir | Czech Technical University in Prague |
Jakubek, Jan | Advacam S.r.o |
Saska, Martin | Czech Technical University in Prague |
Keywords: Aerial Systems: Perception and Autonomy, Robotics in Hazardous Fields, Surveillance Systems
Abstract: A system for measuring radiation intensity and for radiation mapping by a micro unmanned robot using the Timepix detector is presented in this paper. Timepix detectors are extremely small, but powerful 14x14 mm, 256x256 px CMOS hybrid pixel detectors, capable of measuring ionizing alpha, beta, gamma radiation, and heaving ions. The detectors, developed at CERN, produce an image free of any digital noise thanks to per-pixel calibration and signal digitization. Traces of individual ionizing particles passing through the sensors can be resolved in the detector images. Particle type and energy estimates can be extracted automatically using machine learning algorithms. This opens unique possibilities in the task of flexible radiation detection by very small unmanned robotic platforms. The detectors are well suited for the use of mobile robots thanks to their small size, lightweight, and minimal power consumption. This sensor is especially appealing for micro aerial vehicles due to their high maneuverability, which can increase the range and resolution of such novel sensory system. We present a ROS-based readout software and real-time image processing pipeline and review options for 3-D localization of radiation sources using pixel detectors. The provided software supports off-the-shelf FITPix, USB Lite readout electronics with Timepix detectors.
|
|
15:45-16:00, Paper TuBT6.5 | Add to My Program |
Flexible Trinocular: Non-Rigid Multi-Camera-IMU Dense Reconstruction for UAV Navigation and Mapping |
Hinzmann, Timo | Swiss Federal Institute of Technology / ETH Zurich |
Cadena Lerma, Cesar | ETH Zurich |
Nieto, Juan | ETH Zürich |
Siegwart, Roland | ETH Zurich |
Keywords: Aerial Systems: Perception and Autonomy, Sensor Fusion, Range Sensing
Abstract: In this paper, we propose a visual-inertial framework able to efficiently estimate the camera poses of a non-rigid trinocular baseline for long-range depth estimation on-board a fast moving aerial platform. The estimation of the time-varying baseline is based on relative inertial measurements, a photometric relative pose optimizer, and a probabilistic wing model fused in an efficient Extended Kalman Filter (EKF) formulation. The estimated depth measurements can be integrated into a geo-referenced global map to render a reconstruction of the environment useful for local replanning algorithms. Based on extensive real-world experiments we describe the challenges and solutions for obtaining the probabilistic wing model, reliable relative inertial measurements, and vision-based relative pose updates and demonstrate the computational efficiency and robustness of the overall system under challenging conditions.
|
|
16:00-16:15, Paper TuBT6.6 | Add to My Program |
Contact-Based Bridge Inspection Multirotors: Design, Modelling and Control Considering the Ceiling Effect |
Jimenez-Cano, Antonio | University of Seville |
Sanchez-Cuevas, Pedro J | University of Seville |
Grau, Pedro | University of Seville |
Ollero, Anibal | University of Seville |
Heredia, Guillermo | University of Seville |
Keywords: Aerial Systems: Applications
Abstract: This paper presents the design, modelling and control of a multirotor for inspection of bridges with full contact. The paper analyzes the aerodynamic ceiling effect when the aerial robot approaches the bridge surface from below, including its aerodynamic characterization using Computational Fluid Dynamics (CFD). The proposed multirotor design takes the modelled aerodynamic effects into account, improving the performance of the aerial platform in terms of the stability and position accuracy during the inspection. Nonlinear attitude and position controllers to manage the aerodynamic effects are derived and tested. Last, outdoor experiments in a real bridge inspection task have been used to validate the system, as well as, the controller and the aerodynamic characterization. The experiments carried out also include a complete autonomous mission of the aerial platform during a structural assessment application.
|
|
TuBT7 Regular session, L1-R7 |
Add to My Program |
Computer Vision and Applications II |
|
|
Chair: Yang, Guang-Zhong | Imperial College London |
Co-Chair: Bhattacharya, Sourabh | Iowa State University |
|
14:45-15:00, Paper TuBT7.1 | Add to My Program |
3D Canonical Pose Estimation and Abnormal Gait Recognition with a Single RGB-D Camera |
Guo, Yao | Imperial College London |
Deligianni, Fani | Imperial College London |
Gu, Xiao | Imperial College London |
Yang, Guang-Zhong | Imperial College London |
Keywords: Computer Vision for Medical Robotics, Human Detection and Tracking, Recognition
Abstract: Assistive robots play an important role in improving the quality of life of patients at home. Among all the monitoring tasks, gait disorders are prevalent in elderly and people with neurological conditions and this increases the risk of fall. Therefore, the development of mobile systems for gait monitoring at home in normal living conditions is important. Here we present a mobile system that is able to track humans and analyze their gait in canonical coordinates based on a single RGB-D camera. Firstly, view-invariant 3D lower limb pose estimation is achieved by fusing information from depth images along with 2D joints derived in RGB images. Next, both the 6D camera pose and the 3D lower limb skeleton are real-time tracked in a canonical coordinate system based on Simultaneously Localization and Mapping (SLAM). A mask-based strategy is exploited to improve the re-localization of the SLAM in dynamic environments. Abnormal gait is detected by using the Support Vector Machine (SVM) and the Bidirectional Long-Short Term Memory (BiLSTM) network with respect to a set of extracted gait features. To evaluate the robustness of the system, we collected multi-cameras, ground truth data from eight healthy volunteers performing six gait patterns that mimic common gait abnormalities. The results demonstrate that our proposed system can achieve good lower limb pose estimation compared to the ground truth data.
|
|
15:00-15:15, Paper TuBT7.2 | Add to My Program |
Path Planning with Incremental Roadmap Update for Visibility-Based Target Tracking |
Laguna, Guillermo | Iowa State University |
Bhattacharya, Sourabh | Iowa State University |
Keywords: Computational Geometry, Surveillance Systems, Visual Tracking
Abstract: In this paper, we address the visibility-based target tracking problem in which a mobile observer moving along a p-route, which we define as a fixed path for target tracking, tries to keep a mobile target in its field-of-view. By drawing a connection to the {it watchman's route problem}, we find a set of conditions that must be satisfied by the p-route. Then we propose a metric for tracking to estimate a sufficient speed for the observer given the geometry of the environment. We show that the problem of finding the p-route on which the observer requires minimum speed is computationally intractable. We present a technique to find a p-route on which the observer needs at most twice the minimum speed to track the intruder and a reactive motion strategy for the observer.
|
|
15:15-15:30, Paper TuBT7.3 | Add to My Program |
Camera Exposure Control for Robust Robot Vision with Noise-Aware Image Quality Assessment |
Shin, Ukcheol | KAIST(Korea Advanced Institute of Science and Technology) |
Park, Jinsun | KAIST |
Shim, Gyu Min | KAIST |
Rameau, Francois | KAIST, RCV Lab |
Kweon, In So | KAIST |
Keywords: Computer Vision for Other Robotic Applications, Computer Vision for Automation
Abstract: In this paper, we propose a noise-aware exposure control algorithm for robust robot vision. Our method aims to capture best-exposed images, which can boost the performance of various computer vision and robotics tasks. For this purpose, we carefully design an image quality metric that captures complementary quality attributes and ensures light-weight computation. Specifically, our metric consists of a combination of image gradient, entropy, and noise metrics. The synergy of these measures allows the preservation of sharp edges and rich texture in the image while maintaining a low noise level. Using this novel metric, we propose a real-time and fully automatic exposure and gain control technique based on the Nelder-Mead method. To illustrate the effectiveness of our technique, a large set of experimental results demonstrates the higher qualitative and quantitative performance compared with conventional approaches. Our source code and dataset is available at https://github.com/WookCheolShin/Noise-AwareCameraExposureControl.
|
|
15:30-15:45, Paper TuBT7.4 | Add to My Program |
Visual Domain Adaptation Exploiting Confidence-Samples |
Tang, Song | University of Hamburg |
Ji, Yunfeng | University of Shanghai for Science and Technology |
Lyu, Jianzhi | University of Hamburg |
Mi, Jinpeng | TAMS, University of Hamburg |
Li, Qingdu | University of Shanghai for Science and Technology |
Zhang, Jianwei | University of Hamburg |
Keywords: Object Detection, Segmentation and Categorization, Visual Learning
Abstract: Domain adaptation methods are used to address a problem, in which train scenario (source domain) and test scenario (target domain) are different. The existing methods mainly perform adaptation via reducing domain discrepancy from the view of a probability distribution. However, the idea of probability distribution matching always leads to a complex optimization process. Thereby these methods are difficult to apply in some scenario like online application or fast perception in dynamic environments. In this paper, we propose a new and simple domain adaptation method that utilizes confidence-samples to facilitate the classifier training on the target domain. Here, the confidence-samples are a subset of the target samples, and they have very credibly predicted labels. In order to detect the samples, a Category Similarity Collaborative Representation (CSCR) is first developed, by which the raw labels of all target samples are predicted using the smallest projection error according to the law of category. After this, the confidence score of the raw predicted labels is evaluated by the energy context information of CSCR. Finally, the target samples with a high confidence score are selected. Because of the linearity of CSCR, our method avoids complex optimization for matching the probability distribution. Empirical studies on a standard dataset demonstrate the advantages of our method.
|
|
15:45-16:00, Paper TuBT7.5 | Add to My Program |
Learning Residual Flow As Dynamic Motion from Stereo Videos |
Lee, Seokju | KAIST |
Im, Sunghoon | KAIST |
Lin, Stephen | Microsoft Research |
Kweon, In So | KAIST |
Keywords: Computer Vision for Other Robotic Applications, Computer Vision for Transportation, Computer Vision for Automation
Abstract: We present a method for decomposing the 3D scene flow observed from a moving stereo rig into stationary scene elements and dynamic object motion. This is accomplished with a proposed unsupervised learning framework that jointly estimates the camera motion, optical flow, and 3D motion of moving objects via three cooperating networks for stereo matching, camera motion estimation, and prediction of residual flow, which represents the flow component due to object motion and not from camera motion. Based on rigid projective geometry, the estimated stereo depth is used to guide the camera motion estimation, and the depth and camera motion are used to guide the residual flow estimation. Moreover, we explicitly reason about the 3D scene flow of dynamic objects based on the residual flow and scene depth. Experiments on the KITTI dataset demonstrate the effectiveness of our approach and show that our method outperforms other state-of-the-art algorithms on the optical flow and visual odometry tasks.
|
|
16:00-16:15, Paper TuBT7.6 | Add to My Program |
Grounding Language Attributes to Objects Using Bayesian Eigenobjects |
Cohen, Vanya | Brown University |
Burchfiel, Benjamin | Duke University |
Nguyen, Thao | Brown University |
Gopalan, Nakul | Brown University |
Konidaris, George | Brown University |
Tellex, Stefanie | Brown |
Keywords: Computer Vision for Other Robotic Applications, Deep Learning in Robotics and Automation, Social Human-Robot Interaction
Abstract: We develop a system to disambiguate object instances within the same class based on simple physical descriptions. The system takes as input a natural language phrase and a depth image containing a segmented object and predicts how similar the observed object is to the object described by the phrase. Our system is designed to learn from only a small amount of human-labeled language data and generalize to viewpoints not represented in the language-annotated depth image training set. By decoupling 3D shape representation from language representation, this method is able to ground language to novel objects using a small amount of language-annotated depth-data and a larger corpus of unlabeled 3D object meshes, even when these objects are partially observed from unusual viewpoints. Our system is able to disambiguate between novel objects, observed via depth images, based on natural language descriptions. Our method also enables view-point transfer; trained on human-annotated data on a small set of depth images captured from frontal viewpoints, our system successfully predicted object attributes from rear views despite having no such depth images in its training set. Finally, we demonstrate our approach on a Baxter robot, enabling it to pick specific objects based on human-provided natural language descriptions.
|
|
TuBT8 Regular session, LG-R8 |
Add to My Program |
Autonomous Agents and Robots |
|
|
Chair: Liu, Zhe | The Chinese University of Hong Kong |
Co-Chair: Sun, Yuxiang | Hong Kong University of Science and Technology |
|
14:45-15:00, Paper TuBT8.1 | Add to My Program |
Metric Monocular Localization Using Signed Distance Fields |
Huang, Huaiyang | Hong Kong University of Science and Technology |
Sun, Yuxiang | Hong Kong University of Science and Technology |
Ye, Haoyang | The Hong Kong University of Science and Technology |
Liu, Ming | Hong Kong University of Science and Technology |
Keywords: Automation Technologies for Smart Cities, Computer Vision for Transportation, Service Robots
Abstract: Metric localization plays a critical role in vision-based navigation. For overcoming the degradation of matching photometry under appearance changes, recent research resorted to introducing geometry constraints of the prior scene structure. In this paper, we present a metric localization method for the monocular camera, using the Signed Distance Field (SDF) as a global map representation. Leveraging the volumetric distance information from SDFs, we aim to relax the assumption of an accurate structure from the local Bundle Adjustment (BA) in previous methods. By tightly coupling the distance factor with temporal visual constraints, our system corrects the odometry drift and jointly optimizes global camera poses with the local structure. We validate the proposed approach on both indoor and outdoor public datasets. Compared to the state-of-the-art methods, it achieves a comparable performance with a minimal sensor configuration.
|
|
15:00-15:15, Paper TuBT8.2 | Add to My Program |
Perception As Prediction Using General Value Functions in Autonomous Driving Applications |
Graves, Daniel | Huawei Technologies Canada, Ltd |
Rezaee, Kasra | Huawei Technologies Canada, Ltd |
Scheideman, Sean | University of Alberta |
Keywords: Autonomous Agents, Learning and Adaptive Systems, Deep Learning in Robotics and Automation
Abstract: We propose and demonstrate a framework called perception as prediction for autonomous driving that uses general value functions (GVFs) to learn predictions. Perception as prediction learns data-driven predictions relating to the impact of actions on the agent's perception of the world. It also provides a data-driven approach to predict the impact of the anticipated behavior of other agents on the world without explicitly learning their policy or intentions. We demonstrate perception as prediction by learning to predict an agent's front safety and rear safety with GVFs, which encapsulate anticipation of the behavior of the vehicle in front and in the rear, respectively. The safety predictions are learned through random interactions in a simulated environment containing other agents. We show that these predictions can be used to produce similar control behavior to an LQR-based controller in an adaptive cruise control problem as well as provide advanced warning when the vehicle behind is approaching dangerously. The predictions are compact policy-based predictions that support prediction of the long term impact on safety when following a given policy. We analyze two controllers that use the learned predictions in a racing simulator to understand the value of the predictions and demonstrate their use in the real-world on a Clearpath Jackal robot and an autonomous vehicle platform.
|
|
15:15-15:30, Paper TuBT8.3 | Add to My Program |
Experience Reuse with Probabilistic Movement Primitives |
Stark, Svenja | Technical University Darmstadt |
Peters, Jan | Technische Universität Darmstadt |
Rueckert, Elmar | University of Luebeck |
Keywords: Autonomous Agents, Learning from Demonstration, Robot Safety
Abstract: Acquiring new robot motor skills is cumbersome, as learning a skill from scratch and without prior knowledge requires the exploration of a large space of motor configurations. Accordingly, for learning a new task, time could be saved by restricting the parameter search space by initializing it with the solution of a similar task. We present a framework which is able of such knowledge transfer from already learned movement skills to a new learning task. The framework combines probabilistic movement primitives with descriptions of their effects for skill representation. New skills are first initialized with parameters inferred from related movement primitives and thereafter adapted to the new task through relative entropy policy search. We compare two different transfer approaches to initialize the search space distribution with data of known skills with a similar effect. We show the different benefits of the two knowledge transfer approaches on an object pushing task for a simulated 3-DOF robot. We can show that the quality of the learned skills improves and the required iterations to learn a new task can be reduced by more than 60% when past experiences are utilized.
|
|
15:30-15:45, Paper TuBT8.4 | Add to My Program |
SeqLPD: Sequence Matching Enhanced Loop-Closure Detection Based on Large-Scale Point Cloud Description for Self-Driving Vehicles |
Liu, Zhe | The Chinese University of Hong Kong |
Suo, Chuanzhe | The Chinese University of Hong Kong |
Zhou, Shunbo | The Chinese University of Hong Kong |
Xu, Fan | Shanghai Jiao Tong University |
Wei, Huanshu | Chinese University of Hong Kong |
Chen, Wen | The Chinese University of Hong Kong |
Wang, Hesheng | Shanghai Jiao Tong University |
Liang, Xinwu | Shanghai Jiao Tong University |
Liu, Yunhui | Chinese University of Hong Kong |
Keywords: Factory Automation
Abstract: Place recognition and loop-closure detection are main challenges in the localization, mapping and navigation tasks of self-driving vehicles. In this paper, we solve the loop-closure detection problem by incorporating the deep-learning based point cloud description method and the coarse-to-fine sequence matching strategy. More specifically, we propose a deep neural network to extract a global descriptor from the original large-scale 3D point cloud, then based on which, a typical place analysis approach is presented to investigate the feature space distribution of the global descriptors and select several super keyframes. Finally, a coarse-to-fine strategy, which includes a super keyframe based coarse matching stage and a local sequence matching stage, is presented to ensure the loop-closure detection accuracy and real-time performance simultaneously. Thanks to the sequence matching operation, the proposed approach obtains an improvement against the existing deep-learning based methods. Experiment results on a self-driving vehicle validate the effectiveness of the proposed loop-closure detection algorithm.
|
|
15:45-16:00, Paper TuBT8.5 | Add to My Program |
Belief Space Metareasoning for Exception Recovery |
Svegliato, Justin | University of Massachusetts Amherst |
Wray, Kyle | Alliance Innovation Lab Silicon Valley |
Witwicki, Stefan | Alliance Innovation Laboratory |
Biswas, Joydeep | University of Massachusetts Amherst |
Zilberstein, Shlomo | University of Massachusetts |
Keywords: Failure Detection and Recovery, Autonomous Agents, Autonomous Vehicle Navigation
Abstract: Due to the complexity of the real world, autonomous systems use decision-making models that rely on simplifying assumptions to make them computationally tractable and feasible to design. However, since these limited representations cannot fully capture the domain of operation, an autonomous system may encounter unanticipated scenarios that cannot be resolved effectively. We first formally introduce an introspective autonomous system that uses belief space metareasoning to recover from exceptions by interleaving a main decision process with a set of exception handlers. We then apply introspective autonomy to autonomous driving. Finally, we demonstrate that an introspective autonomous vehicle is effective in simulation and on a fully operational prototype.
|
|
16:00-16:15, Paper TuBT8.6 | Add to My Program |
IVOA: Introspective Vision for Obstacle Avoidance |
Rabiee, Sadegh | University of Massachusetts Amherst |
Biswas, Joydeep | University of Massachusetts Amherst |
Keywords: Failure Detection and Recovery, Collision Avoidance, Visual-Based Navigation
Abstract: Vision, as an inexpensive yet information rich sensor, is commonly used for perception on autonomous mobile robots. Unfortunately, accurate vision-based perception requires a number of assumptions about the environment to hold -- some examples of such assumptions, depending on the perception algorithm at hand, include purely lambertian surfaces, texture-rich scenes, absence of aliasing features, and refractive surfaces. In this paper, we present an approach for introspective vision for obstacle avoidance (IVOA) -- by leveraging a supervisory sensor that is occasionally available, we detect failures of stereo vision-based perception from divergence in plans generated by vision and the supervisory sensor. By projecting the 3D coordinates where the plans agree and disagree onto the images used for vision-based perception, IVOA generates a training set of reliable and unreliable image patches for perception. We then use this training dataset to learn a model of which image patches are likely to cause failures of the vision-based perception algorithm. Using this model, IVOA is then able to predict whether the relevant image patches in the observed images are likely to cause failures due to vision (both false positives and false negatives). We empirically demonstrate with extensive real-world data from both indoor and outdoor environments, the ability of IVOA to accurately predict the failures of two distinct vision algorithms.
|
|
TuBT9 Regular session, LG-R9 |
Add to My Program |
Social Human-Robot Interaction II |
|
|
Chair: Yamane, Katsu | Honda |
Co-Chair: Basilico, Nicola | University of Milan |
|
14:45-15:00, Paper TuBT9.1 | Add to My Program |
Can a Social Robot Encourage Children's Self-Study? |
Maeda, Risa | Kyoto University |
Even, Jani | Kyoto University |
Kanda, Takayuki | Kyoto University |
Keywords: Social Human-Robot Interaction, Education Robotics, Robot Companions
Abstract: This paper presents a robot behavioral model designed to support children during self-study. In particular, we want to investigate how a robot could increase the time children keep concentration. The behavioral model was developed by ob- serving children during self-study and by collecting information from experienced tutors through interviews. After observing the children, we decided to consider three states corresponding to different levels of concentration. The child can be smoothly performing the task (“learning” state), encountering some difficulties (“stuck” state) or distracted (“distracted” state). The behavioral model was designed to increase the time spent concentrating on the task by implementing adequate behaviors for each of these three states. These behaviors were designed using the advices collected during the interview survey of the experienced tutors. A self-study system based on the proposed behavior model was implemented. In this system, a small robot sits on the table and encourages the child during self-study. An operator is in charge of determining the state of the child (Wizard of Oz) and the behavioral model triggers the appropriate behaviors for the different states. To demonstrate the effectiveness of the proposed behavioral model, a user study was conducted: 22 children were asked to solve problems alone and to solve problems with the robot. The children spent significantly (p = 0.024) more time in the “learning” state when studying with the robot.
|
|
15:00-15:15, Paper TuBT9.2 | Add to My Program |
Perception of Pedestrian Avoidance Strategies of a Self-Balancing Mobile Robot |
Lo, Shih-Yun | Carnegie Mellon University |
Yamane, Katsu | Honda |
Sugiyama, Ken-ichiro | Honda |
Keywords: Social Human-Robot Interaction, Human-Centered Robotics, Collision Avoidance
Abstract: Mobile robots moving in crowded environments have to navigate among pedestrians safely. Ideally, the way the robot avoids the pedestrians should not only be physically safe but also perceived safe and comfortable. Despite the rich literature in collision-free crowd navigation, limited research has been conducted on how humans perceive robot behaviors in the navigation context. In this paper, we implement three local pedestrian avoidance strategies inspired by human avoidance behaviors on a self-balancing mobile robot and evaluate their perception in a human-robot crossing scenario through a large-scale user study with 98 participants. The study reveals that the avoidance strategies positively affect the participants' perception of the robot's safety, comfort, and awareness to different degrees. Furthermore, the participants perceive the robot as more intelligent, friendly and reliable in the last trial than in the first even with the same strategy.
|
|
15:15-15:30, Paper TuBT9.3 | Add to My Program |
A Deep Learning Approach for Multi-View Engagement Estimation of Children in a Child-Robot Joint Attention Task |
Hadfield, Jack | National Technical University of Athens |
Chalvatzaki, Georgia | National Technical University of Athens |
Koutras, Petros | National Technical University of Athens |
Khamassi, Mehdi | Cnrs / Upmc |
Tzafestas, Costas S. | ICCS - Inst of Communication and Computer Systems |
Maragos, Petros | National Technical University of Athens |
Keywords: Social Human-Robot Interaction, RGB-D Perception, Deep Learning in Robotics and Automation
Abstract: In this work we tackle the problem of child engagement estimation while children freely interact with a robot in a friendly, room-like environment. We propose a deep-based multi-view solution that takes advantage of recent developments in human pose detection. We extract the child's pose from different RGB-D cameras placed elegantly in the room, fuse the results and feed them to a deep neural network trained for classifying engagement levels. The deep network contains a recurrent layer, in order to exploit the rich temporal information contained in the pose data. The resulting method outperforms a number of baseline classifiers, and provides a promising tool for better automatic understanding of a child's attitude, interest and attention while cooperating with a robot. The goal is to integrate this model in next generation social robots as an attention monitoring tool during various Child Robot Interaction (CRI) tasks both for Typically Developed (TD) children and children affected by autism (ASD).
|
|
15:30-15:45, Paper TuBT9.4 | Add to My Program |
Evaluating the Acceptability of Assistive Robots for Early Detection of Mild Cognitive Impairment |
Luperto, Matteo | Università Degli Studi Di Milano |
Romeo, Marta | University of Plymouth |
Lunardini, Francesca | Visa |
Basilico, Nicola | University of Milan |
Abbate, Carlo | Fondazione IRCCS Ca' Granda, Ospedale Maggiore Policlinico, Mila |
Jones, Ray | Mr |
Cangelosi, Angelo | University of Plymouth |
Ferrante, Simona | Politecnico Di Milano |
Borghese, N. Alberto | University of Milano |
Keywords: Social Human-Robot Interaction, Service Robots, Human-Centered Robotics
Abstract: The employment of Social Assistive Robots (SARs) for monitoring elderly users represents a valuable gateway for at-home assistance. Their deployment in the house of the users can provide effective opportunities for early detection of Mild Cognitive Impairment (MCI), a condition of increasing impact in our aging society, by means of digitalized cognitive tests. In this work, we present a system where a specific set of cognitive tests is selected, digitalized, and integrated with a robotic assistant, whose task is the guidance and supervision of the users during the completion of such tests. The system is then evaluated by means of an experimental study involving potential future users, in order to assess its acceptability and identify key directions for technical improvements.
|
|
15:45-16:00, Paper TuBT9.5 | Add to My Program |
A Method for Guiding a Person Combining Robot Movement and Projection |
Tamai, Aki | Hiroshima City University |
Ikeda, Tetsushi | Hiroshima City University |
Iwaki, Satoshi | Hiroshima City University |
Keywords: Social Human-Robot Interaction, Service Robots, Robot Companions
Abstract: This paper proposes a new method to guide a person along with a route when a mobile robot explains an exhibition. In previous work, there were problems that people overtake the robot and go out of the guide route, and that people move to positions that interferes with the movement of the robot. In response to these problems, a conventional solution is to frequently instruct people to move by using voice, displays, etc., and it was not a comfortable for people who are guided. In this paper, we focus on a mobile robot with a projection function and propose a method of controlling human positions without explicit instructions by combining movement and projection of a robot. We introduce three basic guide behaviors combining projection and movement, each controlling a person and a robot into different positional relationship according to the situation of guiding. Experiments of basic guide behaviors showed that both the existence of the robot and the position of the projected image are combined to affect the movement of the guided person. The result indicates that by selecting appropriate guide behavior according to the situation of guiding, the robot can guide a person effectively while controlling the position of the person.
|
|
TuBT10 Regular session, LG-R10 |
Add to My Program |
SLAM II |
|
|
Chair: Song, Dezhen | Texas A&M University |
Co-Chair: Nieto, Juan | ETH Zürich |
|
14:45-15:00, Paper TuBT10.1 | Add to My Program |
Free-Space Features: Global Localization in 2D Laser SLAM Using Distance Function Maps |
Millane, Alexander James | ETH Zurich |
Oleynikova, Helen | ETH Zürich |
Nieto, Juan | ETH Zürich |
Siegwart, Roland | ETH Zurich |
Cadena Lerma, Cesar | ETH Zurich |
Keywords: SLAM, Localization
Abstract: In many applications, maintaining a consistent map of the environment is key to enabling robotic platforms to perform higher-level decision making. Detection of already visited locations is one of the primary ways in which map consistency is maintained, especially in situations where external positioning systems are unavailable or unreliable. Mapping in 2D is an important field in robotics, largely due to the fact that man-made environments such as warehouses and homes, where robots are expected to play an increasing role, can often be approximated as planar. Place recognition in this context remains challenging: 2D lidar scans contain scant information with which to characterize, and therefore recognize, a location. This paper introduces a novel approach aimed at addressing this problem. At its core, the system relies on the use of the distance function for representation of geometry. This representation allows extraction of features which describe the geometry of both surfaces and free-space in the environment. We propose a feature for this purpose. Through evaluations on public datasets, we demonstrate the utility of free-space in the description of places and show an increase in localization performance over a state-of-the-art descriptor extracted from only surface geometry.
|
|
15:00-15:15, Paper TuBT10.2 | Add to My Program |
Eigen-Factors: Plane Estimation for Multi-Frame and Time-Continuous Point Cloud Alignment |
Ferrer, Gonzalo | Skolkovo Institute of Science and Technology |
Keywords: SLAM, Mapping, Localization
Abstract: In this paper, we introduce the Eigen-Factor (EF) method, which estimates a planar surface from a set of point clouds (PCs), with the peculiarity that these points have been observed from different poses, i.e. the trajectory described by a sensor. We propose to use multiple Eigen-Factors (EFs) or different planes' estimations, that allow to solve the multi-frame alignment over a sequence of observed PCs. Moreover, the complexity of the EFs optimization is independent of the number of points, but depends on the number of planes and poses. To achieve this, a closed-form of the gradient is derived by differentiating over the minimum eigenvalue with respect to poses, hence the name Eigen-Factor. In addition, a time-continuous trajectory version of EFs is proposed. The EFs approach is evaluated on a simulated environment and compared with two variants of ICP, showing that it is possible to optimize over all point errors, improving both the accuracy and computational time. Code has been made publicly available.
|
|
15:15-15:30, Paper TuBT10.3 | Add to My Program |
A Robust Laser-Inertial Odometry and Mapping Method for Large-Scale Highway Environments |
Zhao, Shibo | Northeastern University |
Fang, Zheng | Northeastern University |
Li, Haolai | Northeastern University |
Scherer, Sebastian | Carnegie Mellon University |
Keywords: SLAM, Mapping, Localization
Abstract: In this paper, we propose a novel laser-inertial odometry and mapping method to achieve real-time, low-drift and robust pose estimation in large-scale highway environments. The proposed method is mainly composed of four sequential modules, namely scan pre-processing module, dynamic object detection module, laser-inertial odometry module and laser mapping module. Scan pre-processing module uses inertial measurements to compensate the motion distortion of each laser scan. Then, the dynamic object detection module is used to detect and remove dynamic objects from each laser scan by applying CNN segmentation network. After obtaining the undistorted point cloud without moving objects, the laser-inertial odometry module uses an Error State Kalman Filter to fuse the data of laser and IMU and output the coarse pose estimation at high frequency. Finally, the laser mapping module performs a fine processing step and the "Frame-to-Model" scan matching strategy is used to create a static global map. We compare the performance of our method with two state-of-the-art methods, LOAM and SuMa, using KITTI dataset and real highway scene dataset. Experiment results show that our method performs better than the state-of-the-art methods in real highway environments and achieves competitive accuracy on the KITTI dataset.
|
|
15:30-15:45, Paper TuBT10.4 | Add to My Program |
Degeneracy-Aware Factors with Applications to Underwater SLAM |
Hinduja, Akshay | Carnegie Mellon University |
Ho, Bing-Jui | Carnegie Mellon University |
Kaess, Michael | Carnegie Mellon University |
Keywords: SLAM, Mapping, Marine Robotics
Abstract: Simultaneous Localization and Mapping (SLAM) is commonly formulated as an optimization over a graph. A popular approach is the pose graph, which seeks to solve for robots poses that are constrained by pose-to-pose measurements, such as odometry measurements or loop closures. For range sensors, these pose-to-pose constraints can be achieved by performing scan matching techniques, such as Iterative Closest Point (ICP). However, in environments with insufficient or degenerate geometric features, the ICP solution can be unreliable and lead to significant drift in the trajectory of the graph optimization solution. In this paper, we propose a degeneracy-aware approach which has two stages: (1) a degeneracy-aware ICP algorithm and (2) a partially constrained loop closure factor to incorporate the results from (1) into the SLAM pose graph optimization. Our approach performs updates and optimizes both ICP and the pose graph in only the well constrained directions of the state space. These directions are selected on the basis of a dynamic threshold, which updates at each iteration. We apply the proposed algorithm to autonomous underwater mapping with sonar. To evaluate the performance of this algorithm, we conduct experiments in both simulation and real world scenarios, and show the method’s robustness to navigational drift and ability to reject poor loop closures in degenerate environments, which would otherwise degrade the accuracy of the trajectory and the quality of the resulting map.
|
|
15:45-16:00, Paper TuBT10.5 | Add to My Program |
On the Tunable Sparse Graph Solver for Pose Graph Optimization in Visual SLAM Problems |
Chou, Chieh | Texas A&M University |
Wang, Di | Texas A&M University |
Song, Dezhen | Texas A&M University |
Davis, Timothy | Texas A&M University |
Keywords: SLAM, Mapping
Abstract: We report a tunable sparse optimization solver that can trade a slight decrease in accuracy for significant speed improvement in pose graph optimization in visual simultaneous localization and mapping (vSLAM). The solver is designed for devices with significant computation and power constraints such as mobile phones or tablets. Two approaches have been combined in our design. The first is a graph pruning strategy by exploiting objective function structure to reduce the optimization problem size which further sparsifies the optimization problem. The second step is to accelerate each optimization iteration in solving increments for the gradient-based search in Gauss-Newton type optimization solver. We apply a modified Cholesky factorization accelerate the computation. We reuse the decomposition result from last iteration by using Cholesky update/downdate to reduce the repeated computation. We have implemented our solver and tested it with open source data. The experimental results show that our solver can be twice as fast as the counterpart while maintaining a loss of less than 5% in accuracy.
|
|
16:00-16:15, Paper TuBT10.6 | Add to My Program |
Radar SLAM for Indoor Disaster Environments Via Multi-Modal Registration to Prior LiDAR Map |
Park, Yeong Sang | KAIST |
Kim, Joowan | Korea Advanced Institute of Science and Technology (KAIST) |
Kim, Ayoung | Korea Advanced Institute of Science Technology |
Keywords: Range Sensing, Robotics in Hazardous Fields, SLAM
Abstract: This paper presents a localization and mapping algorithm that leverages a radar system in low-visibility environments. We aim to address disaster situations in which prior knowledge of a place is available from CAD or light detection and ranging (LiDAR) maps, but incoming visibility is severely limited. In smoky environments, typical sensors (e.g., cameras and LiDARs) fail to perform reliably due to the large particles in the air. Radars recently attracted attention for their robust perception in low-visibility environments; however, radar measurements’ angular ambiguity and low resolution prevented the direct application to the simultaneous localization and mapping (SLAM) framework. In this paper, we propose registering radar measurements against a previously built dense LiDAR map for localization and applying radar-map refinement for mapping. Our proposed method overcomes the significant density discrepancy between LiDAR and radar with a density-independent point registration algorithm. We validate the proposed method in an environment containing dense fog.
|
|
TuBT11 Regular session, LG-R11 |
Add to My Program |
Medical Robot: Control |
|
|
Chair: Au, K. W. Samuel | The Chinese University of Hong Kong |
Co-Chair: Stoyanov, Danail | University College London |
|
14:45-15:00, Paper TuBT11.1 | Add to My Program |
Setup and Method for Remote Center of Motion Positioning Guidance During Robot-Assisted Surgery |
Smits, Jonas | KU Leuven |
Reynaerts, Dominiek | Division Production Engineering, Machine Design andAutomation, K |
Vander Poorten, Emmanuel B | KU Leuven |
Keywords: Medical Robots and Systems, Physical Human-Robot Interaction, Mechanism Design
Abstract: During robot-assisted surgery, a Remote Center of Motion (RCM) is often implemented to constrain instrument motion through and about a specific point in space. Aligning and re-positioning this point during surgery is not trivial, as this is a defined, yet often non-visualised, point in space. When misaligned with the patient surrounding anatomy is excessively strained, potentially causing post-operative complications. Not being able to safely re-position the RCM for these purposes limits the use of surgical robotics. This work introduces a general approach relying on anatomy-based haptic fixtures to simplify and improve RCM positioning during robot-assisted surgery. The proposed approach is extended with a novel method and mechanism to mechanically implement such fixtures. This is applied for the use case of vitreoretinal surgery, for which purpose a dedicated robotic setup and fixture mechanism was developed. These were used to conduct an initial experimental validation, during which the feasibility of both a virtual and mechanical implementation is reviewed. Initial outcomes suggest that both implementations are feasible. With the currently used impedance-type system, the mechanical implementation is shown to offer at least one order of magnitude stiffness increase when compared to an equivalent virtual implementation.
|
|
15:00-15:15, Paper TuBT11.2 | Add to My Program |
A Reliable Gravity Compensation Control Strategy for dVRK Robotic Arms with Nonlinear Disturbance Forces |
Lin, Hongbin | Chinese University of Hong Kong |
Hui, Chiu-Wai | The Chinese University of Hong Kong |
Wang, Yan | Chinese University of Hong Kong |
Deguet, Anton | Johns Hopkins University |
Kazanzides, Peter | Johns Hopkins University |
Au, K. W. Samuel | The Chinese University of Hong Kong |
Keywords: Medical Robots and Systems, Surgical Robotics: Laparoscopy, Motion Control
Abstract: External disturbance forces caused by nonlinear springy electrical cables in the Master Tool Manipulator (MTM) of the da Vinci Research Kit (dVRK) limits the usage of the existing gravity compensation methods. Significant motion drifts at the MTM tip are often observed when the MTM is located far from its identification trajectory, preventing the usage of these methods for the entire workspace reliably. In this paper, we propose a general and systematic framework to address the problems of the gravity compensation for the MTM of the dVRK. Particularly, high order polynomial models were used to capture the highly nonlinear disturbance forces and integrated with the Multi-step Least Square Estimation (MLSE) framework. This method allows us to identify the parameters of both the gravitational and disturbance forces for each link sequentially, preventing residual error passing among the links of the MTM with uneven mass distribution. A corresponding gravity compensation controller was developed to compen-sate the gravitational and disturbance forces. The method was validated with extensive experiments in the majority of the manipulator’s workspace, showing significant performance enhancements over existing methods. Finally, a deliverable software package in MATLAB and C++ was integrated with dVRK and published in the dVRK community for open-source research and development.
|
|
15:15-15:30, Paper TuBT11.3 | Add to My Program |
Collaborative Needle Insertion with Active Tissue Deformation Control |
Zhong, Fangxun | The Chinese University of Hong Kong |
Wang, Yaqing | The Chinese University of Hong Kong |
Wang, Zerui | The Chinese University of Hong Kong |
Liu, Yunhui | Chinese University of Hong Kong |
Keywords: Medical Robots and Systems, Surgical Robotics: Laparoscopy
Abstract: A major issue occurred during needle insertion into soft tissue is the complex tissue deformation that hinders the minimization of tip-target positioning error. In this letter, we present a new robot control framework to solve target deviation by integrating active deformation control. We first characterize the motion behavior of the desired target under needle-tissue interaction by introducing the needle-induced deformation matrix. Note that the modelling does not require the exact knowledge of tissue or needle insertion properties. The unknown parameters are online updated during the insertion procedure by an adaptive estimator via sensor-based measurement. A closed-loop controller is then proposed for dual-arm robotic execution upon image guidance. The collaboration aims to regulate a feature vector concerning the tip-target alignment to ensure target reachability. The feasibility of the proposed algorithm is studied via simulations and experiments with different biological tissues being tested using the da Vinci Research Kit (dVRK) as the robot platform.
|
|
15:30-15:45, Paper TuBT11.4 | Add to My Program |
Twin Kinematics Approach for Robotic-Assisted Tele-Echography |
Santos, Luís | University of Coimbra |
Cortesao, Rui | University of Coimbra, Institute of Systems and Robotics |
Quintas, João | Instituto Pedro Nunes |
Keywords: Medical Robots and Systems, Telerobotics and Teleoperation, Compliance and Impedance Control
Abstract: This paper discusses a new teleoperation approach for robotic-assisted tele-echography. A teleoperation architecture in the joint space is presented, taking advantage of kinematic similarity between master and slave manipulators. Haptic force feedback is provided based on the slave control command torque, without using force/torque sensing data. The slave manipulator is controlled in the joint space using computed torque techniques and featuring Kalman Active Observers (AOBs). Experiments in a typical telemedice scenario, assess and validate the teleoperation architecture in a clinical context, with a radiologist performing a robotic-assisted abdominal ultrasound examination on a healthy volunteer.
|
|
15:45-16:00, Paper TuBT11.5 | Add to My Program |
Semi-Autonomous Interventional Manipulation Using Pneumatically Attachable Flexible Rails |
D'Ettorre, Claudia | University College of London |
Stilli, Agostino | University College London |
Dwyer, George | University College London |
Neves, Joana B. | University College London |
Tran, Maxine | University College London |
Stoyanov, Danail | University College London |
Keywords: Surgical Robotics: Laparoscopy, Soft Robot Applications, Soft Robot Materials and Design
Abstract: During laparoscopic surgery, tissues frequently need to be retracted and mobilized for manipulation or visualisation. State-of-the-art robotic platforms for minimally invasive surgery (MIS) typically rely on rigid tools to interact with soft tissues. Such tools offer a very narrow contact surface thus applying relatively large forces that can lead to tissue damage, posing a risk for the success of the procedure and ultimately for the patient. In this paper, we show how the use of Pneumatically Attachable Flexible (PAF) rail, a vacuum-based soft attachment for laparoscopic applications, can reduce such risk by offering a larger contact surface between the tool and the tissue. textit{Ex vivo} experiments are presented investigating the short- and long-term effects of different levels of vacuum pressure on the tissue’s surface. These experiments aim at evaluating the best trade-off between applied pressure, potential damage, task duration and connection stability. A hybrid control system has been developed to perform and investigate the organ repositioning task using the proposed system. The task is only partially automated allowing the surgeon to be part of the control loop. A gradient-based planning algorithm is integrated with learning from teleoperation algorithm which allows the robot to improve the learned trajectory. The use of Similar Smooth Path Repositioning (SSPR) algorithm is proposed to improve a learned trajectory based on a known cost function. The results obtained show that a smoother trajectory allows to decrease the minimum level of pressure needed to guarantee active suction during PAF positioning and placement.
|
|
16:00-16:15, Paper TuBT11.6 | Add to My Program |
Optimizing Motion-Planning Problem Setup Via Bounded Evaluation with Application to Following Surgical Trajectories |
Niyaz, Sherdil | University of Washington |
Kuntz, Alan | University of North Carolina at Chapel Hill |
Salzman, Oren | Carnegie Mellon University |
Alterovitz, Ron | University of North Carolina at Chapel Hill |
Srinivasa, Siddhartha | University of Washington |
Keywords: Surgical Robotics: Planning, Motion and Path Planning
Abstract: A motion-planning problem's setup can drastically affect the quality of solutions returned by the planner. In this work we consider optimizing these setups, with a focus on doing so in a computationally-efficient fashion. Our approach interleaves optimization with motion planning, which allows us to consider the actual motions required of the robot. Similar prior work has treated the planner as a black box: our key insight is that opening this box in a simple-yet-effective manner enables a more efficient approach, by allowing us to bound the work done by the planner to optimizer-relevant computations. Finally, we apply our approach to a surgically-relevant motion-planning task, where our experiments validate our approach by more-efficiently optimizing the fixed insertion pose of a surgical robot.
|
|
TuBT12 Regular session, LG-R12 |
Add to My Program |
Gesture, Posture and Facial Expressions |
|
|
Chair: Kerzel, Matthias | Uni Hamburg |
Co-Chair: Frederiksen, Morten Roed | IT-University of Copenhagen |
|
14:45-15:00, Paper TuBT12.1 | Add to My Program |
Generalized Multiple Correlation Coefficient As a Similarity Measurement between Trajectories |
Urain De Jesus, Julen | TU Darmstadt |
Peters, Jan | Technische Universität Darmstadt |
Keywords: Gesture, Posture and Facial Expressions, Cognitive Human-Robot Interaction
Abstract: Similarity distance measure between two trajectories is an essential tool to understand patterns in motion, for example, in Human-Robot Interaction or Imitation Learning. The problem has been faced in many fields, from Signal Processing, Probabilistic Theory field, Topology field or Statistics field. Anyway, up to now, none of the trajectory similarity measurements metrics are invariant to all possible linear transformation of the trajectories (rotation, scaling, reflection, shear mapping or squeeze mapping). Also not all of them are robust in front of noisy signals or fast enough for real-time trajectory classification. To overcome this limitation this paper proposes a similarity distance metric that will remain invariant in front of any possible linear transformation. Based on Pearson’s Correlation Coefficient and the Coefficient of Determination, our similarity metric, the Generalized Multiple Correlation Coefficient (GMCC) is presented like the natural extension of the Multiple Correlation Coefficient. The motivation of this paper is two-fold: First, to introduce a new correlation metric that presents the best properties to compute similarities between trajectories invariant to linear transformations and compare it with some state of the art similarity distances. Second, to present a natural way of integrating the similarity metric in an Imitation Learning scenario for clustering robot trajectories.
|
|
15:00-15:15, Paper TuBT12.2 | Add to My Program |
Fusing Body Posture with Facial Expressions for Joint Recognition of Affect in Child-Robot Interaction |
Filntisis, Panagiotis Paraskevas | National Technical University of Athens |
Efthymiou, Niki | National Technical University of Athens |
Koutras, Petros | National Technical University of Athens |
Potamianos, Gerasimos | University of Thessaly |
Maragos, Petros | National Technical University of Athens |
Keywords: Gesture, Posture and Facial Expressions, Computer Vision for Other Robotic Applications
Abstract: In this paper we address the problem of multi-cue affect recognition in challenging scenarios such as child-robot interaction. Towards this goal we propose a method for automatic recognition of affect that leverages body expressions alongside facial ones, as opposed to traditional methods that typically focus only on the latter. Our deep-learning based method uses hierarchical multi-label annotations and multi-stage losses, can be trained both jointly and separately, and offers us computational models for both individual modalities, as well as for the whole body emotion. We evaluate our method on a challenging child-robot interaction database of emotional expressions collected by us, as well as on the GEMEP public database of acted emotions by adults, and show that the proposed method achieves significantly better results than facial-only expression baselines.
|
|
15:15-15:30, Paper TuBT12.3 | Add to My Program |
Exploring Low-Level and High-Level Transfer Learning for Multi-Task Facial Recognition with a Semi-Supervised Neural Network |
Barros, Pablo | University of Hamburg |
Fliesswasser, Erik | University of Hamburg |
Kerzel, Matthias | Uni Hamburg |
Wermter, Stefan | University of Hamburg |
Keywords: Gesture, Posture and Facial Expressions, Visual Learning
Abstract: Facial recognition tasks like identity, age, gender, and emotion recognition have been extensively tackled in recent years. Their deployment in robotic platforms became necessary for the characterization of most of the non-verbal Human-Robot Interaction (HRI) scenarios. In this regard, deep convolution neural networks have shown to be effective on processing different facial representations but with a high cost: to achieve maximum generalization, they require an enormous amount of task-specific labeled data. This paper proposes a unified semi-supervised deep neural model to address this problem. Our hybrid model is composed of an unsupervised deep generative adversarial network which learns fundamental characteristics of facial representations, and a set of convolution channels that fine-tunes the high-level facial concepts for the recognition of identity, age group, gender, and facial expressions. Our network employs progressive lateral connections between the convolution channels so that they share the high-abstraction particularities of each of these tasks in order to reduce the necessity of a large amount of strongly labeled training data. We propose a series of experiments to evaluate each individual mechanism of our hybrid model, in particular, the impact of the progressive connections on learning the specific facial recognition tasks and we observe that our model achieves a better performance when compared to task-specific models.
|
|
15:30-15:45, Paper TuBT12.4 | Add to My Program |
A Systematic Comparison of Affective Robot Expression Modalities |
Frederiksen, Morten Roed | IT-University of Copenhagen |
Stoy, Kasper | IT University of Copenhagen |
Keywords: Social Human-Robot Interaction, Gesture, Posture and Facial Expressions, Robot Companions
Abstract: This paper provides a survey of the different means of expression employed by robots conveying affective states to human recipients. The paper introduces a model of affective expression modalities (MOAM) that describes and compares the emphasis on specific means of expression and applies it to the surveyed robots. Using the model entails viewing the effect of applied expression modalities in light of how well the robot responds to external stimuli and with attention to how aligned the robot’s means of affective expressions are with the intended working scenario. The model-based survey shows that a majority (85%) of the surveyed robots contain a category with room for additional affective means of expression, and a quarter (25.6%) of the robots use a single or two affective expression modalities to convey affective states. The result of the survey indicates there is an under-researched opportunity in exploring synergies between the different expression modalities to amplify the overall affective impact of a robot.
|
|
15:45-16:00, Paper TuBT12.5 | Add to My Program |
Towards More Realistic Human-Robot Conversation: A Seq2Seq-Based Body Gesture Interaction System |
Hua, Minjie | CloudMinds Technologies Inc |
Shi, Fuyuan | CloudMinds Technologies Inc |
Nan, Yibing | CloudMinds Technologies Inc |
Wang, Kai | CloudMinds Technologies |
Chen, Hao | CloudMinds Technologies Inc |
Lian, Shiguo | CloudMinds Technologies Inc |
Keywords: Natural Machine Motion, Social Human-Robot Interaction, Deep Learning in Robotics and Automation
Abstract: This paper presents a novel system that enables intelligent robots to exhibit realistic body gestures while communicating with humans. The proposed system consists of a listening model and a speaking model used in corresponding conversational phases. Both models are adapted from the sequence-to-sequence (seq2seq) architecture to synthesize body gestures represented by the movements of twelve upper-body keypoints. All the extracted 2D keypoints are firstly 3D-transformed, then rotated and normalized to discard irrelevant information. Substantial videos of human conversations from Youtube are collected and preprocessed to train the listening and speaking models separately, after which the two models are evaluated using metrics of mean squared error (MSE) and cosine similarity on the test dataset. The tuned system is implemented to drive a virtual avatar as well as Pepper, a physical humanoid robot, to demonstrate the improvement on conversational interaction abilities of our method in practice.
|
|
TuBT13 Regular session, LG-R13 |
Add to My Program |
Humanoid and Bipedal Locomotion II |
|
|
Chair: Hur, Pilwon | Texas A&M University |
Co-Chair: Rebula, John | University of Michigan |
|
14:45-15:00, Paper TuBT13.1 | Add to My Program |
Virtual-Mass-Ellipsoid Inverted Pendulum Model and Its Applications to 3D Bipedal Locomotion on Uneven Terrains |
Guan, Kaixuan | The University of Tokyo |
Yamamoto, Ko | University of Tokyo |
Nakamura, Yoshihiko | University of Tokyo |
Keywords: Humanoid and Bipedal Locomotion, Humanoid Robots, Legged Robots
Abstract: It is still an open problem to develop a reduced order model of bipedal walking that closely represents the complex dynamics of humanoid robots. In this paper, we propose control methodologies, removing both the constant CoM height constraint and the constant centroidal angular momentum constraint. We define a capturability criterion. and propose an enhanced intrinsically stable model predict control to fulfill this new capturability criterion. Then the angular momentum can be controlled. The results of simulations using humanoid robot HRP-4 show the proposed methods can improve the stability of bipedal locomotion on uneven terrains.
|
|
15:00-15:15, Paper TuBT13.2 | Add to My Program |
A Robustness Analysis of Inverse Optimal Control of Bipedal Walking |
Rebula, John | University of Michigan |
Schaal, Stefan | MPI Intelligent Systems & University of Southern California |
Finley, James | University of Southern California |
Righetti, Ludovic | New York University |
Keywords: Humanoid and Bipedal Locomotion, Legged Robots, Passive Walking
Abstract: Cost functions have the potential to provide compact and understandable generalizations of motion. The goal of Inverse Optimal Control (IOC) is to analyze an observed behavior which is assumed to be optimal with respect to an unknown cost function, and infer this cost function. Here we develop a method for characterizing cost functions of legged locomotion, with the goal of representing complex humanoid behavior with simple models. To test this methodology we simulate walking gaits of a simple 5 link planar walking model which optimize known cost functions, and assess the ability of our IOC method to recover them. In particular, the IOC method uses an iterative trajectory optimization process to infer cost function weightings consistent with those used to generate a single demonstrated optimal trial. We also explore sensitivity of the IOC to sensor noise in the observed trajectory, imperfect knowledge of the model or task, as well as uncertainty in the components of the cost function used. With appropriate modeling, these methods may help infer cost functions from human data, yielding a compact and generalizable representation of human-like motion for use in humanoid robot controllers, as well as providing a new tool for experimentally exploring human preferences.
|
|
15:15-15:30, Paper TuBT13.3 | Add to My Program |
Periodic Trajectory Planning and Robust Output Zeroing Control for Underactuated Bipedal Robots with Predicted Disturbances |
Takano, Rin | Tokyo Institute of Technology |
Chang, Junho | Tokyo Institute of Technology |
Yamakita, Masaki | Tokyo Inst. of Technology |
Keywords: Humanoid and Bipedal Locomotion, Model Learning for Control, Motion and Path Planning
Abstract: For underactuated bipedal robots, it is important to compensate disturbances which affects the zero dynamics to realize a stable locomotion when we use output zeroing controllers. In order to deal with such disturbances, this paper presents a framework of a periodic trajectory planning and a robust output zeroing controller using a disturbance model learned by Gaussian process regression (GPR). In particular, we propose a control method considering the modeling uncertainty of the disturbance model by using a variance information of GPR model. We show the effectiveness of the proposed method through a numerical simulation of walking control of an underactuated robot model affected by an external force.
|
|
15:30-15:45, Paper TuBT13.4 | Add to My Program |
Learning Footstep Planning on Irregular Surfaces with Partial Placements |
Castro, German | The University of New South Wales |
Sammut, Claude | The University of New South Wales |
Keywords: Humanoid and Bipedal Locomotion, Motion and Path Planning
Abstract: We present two contributions built upon on a previous footstep planner based on the ARA* search. Firstly, we have developed an improved foothold selection method using support polygons, to increase foothold availability in rough terrain. Secondly, we present a footstep classification method using the C5.0 algorithm, that takes advantage of cost similarity between adjacent steps. This is intended to learn feasibility and approximate transition costs for the ARA* planner. These contributions extend capabilities of the planner by increasing footstep availability and allowing to generate more complex plans, without compromising safety.
|
|
15:45-16:00, Paper TuBT13.5 | Add to My Program |
A Robust Biped Locomotion Based on Linear-Quadratic-Gaussian Controller and Divergent Component of Motion |
Kasaei, Seyed Mohammadreza | IEETA / DETI University of Aveiro |
Lau, Nuno | Aveiro University |
Pereira, Artur | University of Aveiro |
Keywords: Humanoid and Bipedal Locomotion, Optimization and Optimal Control
Abstract: Generating robust locomotion for a humanoid robot in the presence of disturbances is difficult because of its high number of degrees of freedom and its unstable nature. In this paper, we used the concept of Divergent Component of Motion (DCM) and propose an optimal closed-loop controller based on Linear-Quadratic-Gaussian to generate a robust and stable walking for humanoid robots. The biped robot dynamics has been approximated using the Linear Inverted Pendulum Model (LIPM). Moreover, we propose a controller to adjust the landing location of the swing leg to increase the withstanding level of the robot against a severe external push. The performance and also the robustness of the proposed controller is analyzed and verified by performing a set of simulations using MATLAB. The simulation results showed that the proposed controller is capable of providing a robust walking even in the presence of disturbances and in challenging situations.
|
|
16:00-16:15, Paper TuBT13.6 | Add to My Program |
Generalized Contact Constraints of Hybrid Trajectory Optimization for Different Terrains and Analysis of Sensitivity to Randomized Initial Guesses |
Chao, Kenneth | Texas A&M |
Hur, Pilwon | Texas A&M University |
Keywords: Humanoid and Bipedal Locomotion, Optimization and Optimal Control
Abstract: To generate a dynamic bipedal walking with foot rolling motion for bipedal robot, hybrid trajectory optimization is capable of planning level walking with great energetic efficiency. However, the direct implementation of this optimization requires different sets of variables to express different active contact constraints, which can be complicated to implement. To simplify the optimization formulation, we propose the generalized contact constraints where the same set of variables are used through all the walking phases. By changing the variable and constraint bounds, different contact constraints for different contact conditions can be generally expressed. The proposed modifications are applied on the bipedal robot AMBER 3, where the optimization results on different terrains are compared and discussed. On the other hand, it is known that a randomized initial guess can be used to solve this optimization, yet its effect on the gaits on different terrains is unclear. As a result, we analyzed the sensitivity of the optimization to a set of randomized initial guesses. The level and downslope walking gaits are also validated via the experiments on AMBER 3.
|
|
TuBT14 Regular session, LG-R14 |
Add to My Program |
Climbing Robots |
|
|
Chair: Nikolakopoulos, George | Luleå University of Technology |
Co-Chair: Li, Peng | Harbin Institute of Technology (ShenZhen) |
|
14:45-15:00, Paper TuBT14.1 | Add to My Program |
Design of an Adhesion-Aware Façade Cleaning Robot |
Muthugala Arachchige, Viraj Jagathpriya Muthugala | Singapore University of Technology and Design |
Vega-Heredia, Manuel | Singapore University of Technology and Design & Universidad Autó |
Ayyalusami, Vengadesh | Singapore University of Technology and Design |
Sriharsha, Ghanta | SUTD |
Elara, Mohan Rajesh | Singapore University of Technology and Design |
Keywords: Climbing Robots, Failure Detection and Recovery, Robot Safety
Abstract: Cleaning requirements of glass façades of high-rise buildings have been significantly increased in recent years due to the growth of the construction industry. The conventional cleaning methods for high-rise buildings require human labor where efficiency, cost, and safety are major concerns. Therefore, robotic systems that can climb and clean glass façades in high-rise buildings have been developed. Capability to attach to a façade surface is one of the critical requirements of a glass façade cleaning robot. Diverse approaches have been proposed for adhesion of cleaning robots to a glass façade. Active vacuum suction mechanisms are widely used for these robots since they provide better controllability to move the robots smartly on the surface. Notably, those are decent for a reconfigurable robot that can transit between window frames. The suction mechanism of a glass façade cleaning robot must provide a reliable suction force to make the robot stay safely and move smartly on a façade. Nevertheless, these suction mechanisms can be failed due to improper fastening between a façade and the mechanism, which may lead to safety and operational issues. Therefore, this paper proposes a novel method to realize the adhesion-awareness of a glass façade cleaning robot. The adhesion-awareness is realized by analyzing The current drawn by the motors attached to the impellers of the vacuum mechanisms. Experiment results validate the capability of the proposed approach in raising the adhesion-awareness of a façade cleaning robot.
|
|
15:00-15:15, Paper TuBT14.2 | Add to My Program |
A Novel Capabilities of Quadruped Robot Moving through Vertical Ladder without Handrail Support |
Saputra, Azhar Aulia | Tokyo Metropolitan University |
Toda, Yuichiro | Okayama University |
Takesue, Naoyuki | Tokyo Metropolitan University |
Kubota, Naoyuki | Tokyo Metropolitan University |
Keywords: Climbing Robots, Legged Robots, Motion Control
Abstract: This paper presents the novel capabilities of a quadruped robot by performing horizontal-vertical-horizontal movement transition through vertical ladder without handrailing supporter. To overcome the proposed problem, we propose a multi-behavior generation model using independent stepping and pose control in the quadruped robot. The model is able to generate appropriate behavior depending on external (3D point clouds) and internal sensors (ground touch sensor, and Inertial Measurement Unit). Posture condition, safe movement area, possible touchpoint, grasping possibility, and target movement are the information that is analyzed from the sensors. There are four options developed in behavior generation, which are, Approaching, Body Placing, Stepping, and Grasping behavior. In order to prove the effectiveness of the proposed algorithm, the model was implemented on the computer simulation and the real application. Before being applied in the real robot, the proposed model is optimized in the computer simulation. Then, the optimized parameter is used for applying in the real robot. As a result, the robot succeeded to move through the ladder without handrail from lower stair to upper stair. From the analysis, the body placing behavior is the most important strategy in the proposed case.
|
|
15:15-15:30, Paper TuBT14.3 | Add to My Program |
Adaptive Vision-Based Control for Rope-Climbing Robot Manipulator |
Sun, Guangli | The Chinese University of Hong Kong |
Li, Xiang | Tsinghua University |
Li, Peng | Harbin Institute of Technology (ShenZhen) |
Yue, Linzhu | Chinese University of Hong Kong |
Zhou, Yang | The Chinese University of Hong Kong |
Liu, Yunhui | Chinese University of Hong Kong |
Keywords: Climbing Robots, Sensor-based Control, Dynamics
Abstract: While the mechanism of Rope-Climbing provides much flexibility, it opens up challenges to the development of the controller for Robotic Manipulator installed on Rope-Climbing robot(RCR), which is called Rope-Climbing Robot Manipulator(RCRM) here. In particular, the deformable nature of the rope results in the vibration to the manipulator and hence affects the positioning of the end effector. In this paper, a new adaptive vision-based controller is proposed for RCRM, which enables the robot to carry out the high-accuracy task under the unknown vibration from the rope. The proposed controller guarantees the performance of the robot in twofold. First, the control problem is directly formulated in the image space such that the exact spatial relationship between the moving base of the manipulator (due to the vibrating rope) and the target (e.g. the wall) is not required. Second, novel adaptation laws are developed to estimate the vibration from the rope online and are cancelled out in the robot control input to stabilize the end effector. The stability of the closed-loop system is rigorously proved with Lyapunov methods, and experimental results are presented to illustrate the performance of the proposed controller.
|
|
15:30-15:45, Paper TuBT14.4 | Add to My Program |
On Model-Based Adhesion Control of a Vortex Climbing Robot |
Andrikopoulos, George | Luleå University of Technology |
Papadimitriou, Andreas | Luleå University of Technology |
Brusell, Angelica | Luleå University of Technology |
Nikolakopoulos, George | Luleå University of Technology |
Keywords: Climbing Robots, Wheeled Robots
Abstract: In this article, the adhesion modeling and control case of a Vortex Climbing Robot (VCR) is investigated against a surface of variable orientations. The critical adhesion force exerted from the implemented Vortex Actuator (VA) and the VCR’s achievable payload are analyzed under 3-DOF rotations of the test surface, while extracted from both geometrical analysis and dynamically-simulated numerical results. A model-based control scheme is later proposed, with the goal of achieving adhesion while the VCR remains immobilized, limiting the power consumption and compensating for disturbances (e.g. moving cables) leading to Center-of-Mass (CoM) changes. Finally, the model-based control scheme is experimentally evaluated, with the VCR prototype on a rotating and moving flat surface. The presented results support the use of the proposed methodology in climbing robots targeting inspection and maintenance of stationary surfaces (flat, curved etc.), as well as future robotic solutions operating on moving structures (e.g. ships, cranes, folding bridges).
|
|
15:45-16:00, Paper TuBT14.5 | Add to My Program |
An Interactive Physically-Based Model for Active Suction Phenomenon Simulation |
Bernardin, Antonin | INSA Rennes |
Duriez, Christian | INRIA |
Marchal, Maud | INSA/INRIA |
Keywords: Contact Modelling, Simulation and Animation
Abstract: While suction cups are widely used in Robotics, the literature is underdeveloped when it comes to the modelling and simulation of the suction phenomenon. In this paper, we present a novel physically-based approach to simulate the behavior of active suction cups. Our model relies on a novel formulation which assumes the pressure exerted on a suction cup during active control is based on constraint resolution. Our algorithmic implementation uses a classification process to handle the contacts during the suction phenomenon of the suction cup on a surface. Then, we formulate a convenient way for coupling the pressure constraint with the multiple contact constraints. We propose an evaluation of our approach through a comparison with real data, showing the ability of our model to reproduce the behavior of suction cups. Our approach paves the way for improving the design as well as the control of robotic actuators based on suction cups such as vaccum grippers.
|
|
TuBT15 Regular session, LG-R15 |
Add to My Program |
Motion and Path Planning II |
|
|
Chair: Yoon, Sung-eui | KAIST |
Co-Chair: Pěnička, Robert | Czech Technical University in Prague |
|
14:45-15:00, Paper TuBT15.1 | Add to My Program |
Coverage Path Planning Using Path Primitive Sampling and Primitive Coverage Graph for Visual Inspection |
Jing, Wei | A*STAR |
Deng, Di | Carnegie Mellon University |
Xiao, Zhe | Institute of High Performance Computing |
Liu, Yong | A*STAR Institute of High Performance Computing |
Shimada, Kenji | Carnegie Mellon University |
Keywords: Motion and Path Planning, Task Planning, Planning, Scheduling and Coordination
Abstract: Planning the path to gather the surface information of the target objects is crucial to improve the efficiency of and reduce the overall cost, for visual inspection applications with Unmanned Aerial Vehicles (UAVs). Coverage Path Planning (CPP) problem is often formulated for these inspection applications because of the coverage requirement. Traditionally, researchers usually plan and optimize the viewpoints to capture the surface information first, and then optimize the path to visit the selected viewpoints. In this paper, we propose a novel planning method to directly sample and plan the inspection path for a camera-equipped UAV to acquire visual and geometric information of the target structures as a video stream setting in complex 3D environment. The proposed planning method first generates via-points and path primitives around the target object by using sampling methods based on voxel dilation and subtraction. A novel Primitive Coverage Graph (PCG) is then proposed to encode the topological information, flying distances, and visibility information, with the sampled via-points and path primitives. Finally graph search is performed to find the resultant path in the PCG to complete the inspection task with the coverage requirements. The effectiveness of the proposed method is demonstrated through simulation and field tests in this paper.
|
|
15:00-15:15, Paper TuBT15.2 | Add to My Program |
Sampling-Based Motion Planning of 3D Solid Objects Guided by Multiple Approximate Solutions |
Vonasek, Vojtech | Czech Technical University in Prague |
Pěnička, Robert | Czech Technical University in Prague |
Keywords: Motion and Path Planning
Abstract: Sampling-based motion planners are often used to solve motion planning problems for robots with many degrees of freedom. These planners explore the related configuration space by random sampling. The well-known issue of the sampling-based planners is the narrow passage problem. Narrow passages are small collision-free regions in the configuration space that are, due to their volume, difficult to cover by the random samples. The volume of the narrow passages can be artificially increased by reducing the size of the robot, e.g., by scaling-down its geometry, which increases the probability of placing the random samples into the narrow passages. This allows us to find an approximate solution (trajectory) and use it as a guide to find the solution for a larger robot. Guiding along an approximate solution may, however, fail if this solution leads through such parts of the configuration space that are not reachable or traversable by a larger robot. To improve this guiding process, we propose to compute several approximate solutions leading through different parts of the configuration space, and use all of them to guide the search for a larger robot. We introduce the concept of disabled regions that are prohibited from the exploration using the sampling process. The disabled regions are defined using trajectories already found in the space being searched. The proposed method can solve planning problems with narrow passages with higher success rate than other state-of-the-art planners.
|
|
15:15-15:30, Paper TuBT15.3 | Add to My Program |
LEGO: Leveraging Experience in Roadmap Generation for Sampling-Based Planning |
Kumar, Rahul | IIT Kharagpur |
Mandalika, Aditya | University of Washington |
Choudhury, Sanjiban | University of Washington |
Srinivasa, Siddhartha | University of Washington |
Keywords: Motion and Path Planning
Abstract: We consider the problem of leveraging prior experience to generate roadmaps in sampling-based motion planning. A desirable roadmap is one that is sparse, allowing for fast search, with nodes spread out at key locations such that a low-cost feasible path exists. An increasingly popular approach is to learn a distribution of nodes that would produce such a roadmap. State-of-the-art is to train a conditional variational auto-encoder (CVAE) on the prior dataset with the shortest paths as target input. While this is quite effective on many problems, we show it can fail in the face of complex obstacle configurations or mismatch between training and testing. We present an algorithm, LEGO, that addresses these issues by training the CVAE with target samples that satisfy two important criteria. Firstly, these samples belong only to bottleneck regions along near-optimal paths that are otherwise difficult-to-sample with a uniform sampler. Secondly, these samples are spread out across diverse regions to maximize the likelihood of a feasible path existing. We formally define these properties and prove performance guarantees for LEGO. We extensively evaluate LEGO on a range of planning problems, including robot arm planning, and report significant gains over both heuristics and learned baselines.
|
|
15:30-15:45, Paper TuBT15.4 | Add to My Program |
Volumetric Tree*: Adaptive Sparse Graph for Effective Exploration of Homotopy Classes |
Kim, Donghyuk | KAIST |
Kang, Mincheul | KAIST |
Yoon, Sung-eui | KAIST |
Keywords: Motion and Path Planning
Abstract: We present volumetric tree*, a hybridization of sampling-based and optimization-based motion planning. Volumetric tree* constructs an adaptive sparse graph with volumetric vertices, hyper-spheres encoding free configurations, using a sampling-based motion planner for a homotopy exploration. The coarse-grained paths computed on the sparse graph are refined by optimization-based planning during the execution, while exploiting the probabilistic completeness of the sampling-based planning for the initial path generation. We also suggest a dropout technique probabilistically ensuring that the sampling-based planner is capable of identifying all possible homotopies of solution paths. We compare the proposed algorithm against the state-of-the-art planners in both synthetic and practical benchmarks with varying dimensions, and experimentally show the benefit of the proposed algorithm.
|
|
15:45-16:00, Paper TuBT15.5 | Add to My Program |
Multilevel Incremental Roadmap Spanners for Reactive Motion Planning |
Ichnowski, Jeffrey | University of North Carolina at Chapel Hill |
Alterovitz, Ron | University of North Carolina at Chapel Hill |
Keywords: Motion and Path Planning
Abstract: Generating robot motions from a precomputed graph has proven to be an effective approach to solving many motion planning problems. After their generation, roadmaps reduce complex motion planning problems to that of solving a graph-based shortest path. However, generating the graph can involve tradeoffs, such as how sparse or dense to make the graph. Sparse graphs may not provide enough options to navigate around a new obstacle or may result in grossly suboptimal motions. Dense graphs may take too long to search and result in an unresponsive robot. In this paper we present an algorithm that generates a graph with multiple sparse levels---the sparsest level can be searched quickly, while the densest level allows for asymptotically optimal motions. With the paired multilevel shortest path algorithm, after the robot computes an initial solution, it can then incrementally refine the shortest-path as time allows. We demonstrate the algorithms on an articulated robot with 8 degrees of freedom, having them compute an initial solution in a fraction of the time required for a full graph search, and subsequently, incrementally refine the solution to the optimal shortest path from the densest level of the graph.
|
|
16:00-16:15, Paper TuBT15.6 | Add to My Program |
MT-RRT: A General Purpose Multithreading Library for Path Planning |
Casalino, Andrea | Politecnico Di Milano |
Zanchettin, Andrea Maria | Politecnico Di Milano |
Rocco, Paolo | Politecnico Di Milano |
Keywords: Motion and Path Planning, Motion Control, Industrial Robots
Abstract: Rapidly Random exploring Trees are popular algorithms in the field of motion planning. A feasible path connecting two different poses is found by incrementally building a tree data structure. They are powerful and flexible, but also computationally intense, requiring thousands of iterations before their termination. The aim of this article is to show the capabilities of MT-RRT, a general purpose library which exploits four different multithreading strategies to speed up the planning process of Rapidly Random exploring Trees. MT-RRT will be proved to significantly reduce the computation time on various benchmarks.
|
|
TuBT16 Regular session, LG-R16 |
Add to My Program |
Grasping II |
|
|
Chair: Manocha, Dinesh | University of Maryland |
Co-Chair: Stuart, Hannah | UC Berkeley |
|
14:45-15:00, Paper TuBT16.1 | Add to My Program |
Generating Grasp Poses for a High-DOF Gripper Using Neural Networks |
Liu, Min | School of Computer, National University of Defense Technology |
Pan, Zherong | The University of North Carolina at Chapel Hill |
Xu, Kai | National University of Defense Technology |
Ganguly, Kanishka | University of Maryland, College Park |
Manocha, Dinesh | University of Maryland |
Keywords: Grasping, Deep Learning in Robotics and Automation, Perception for Grasping and Manipulation
Abstract: We present a learning-based method for representing grasp poses of a high-DOF hand using neural networks. Due to redundancy in such high-DOF grippers, there exists a large number of equally effective grasp poses for a given target object, making it difficult for the neural network to find consistent grasp poses. We resolve this ambiguity by generating an augmented dataset that covers many possible grasps for each target object and train our neural networks using a consistency loss function to identify a one-to-one mapping from objects to grasp poses. We further enhance the quality of neural-network-predicted grasp poses using a collision loss function to avoid penetrations. We use an object dataset that combines the BigBIRD Database, the KIT Database, the YCB Database, and the Grasp Dataset to show that our method can generate high-DOF grasp poses with higher accuracy than supervised learning baselines. The quality of the grasp poses is on par with the groundtruth poses in the dataset. In addition, our method is robust and can handle noisy object models such as those constructed from multi-view depth images, allowing our method to be implemented on a 25-DOF Shadow Hand hardware platform.
|
|
15:00-15:15, Paper TuBT16.2 | Add to My Program |
Robust Grasp Planning Over Uncertain Shape Completions |
Lundell, Jens | Aalto University |
Verdoja, Francesco | Aalto University |
Kyrki, Ville | Aalto University |
Keywords: Grasping, Deep Learning in Robotics and Automation, Perception for Grasping and Manipulation
Abstract: We present a method for planning robust grasps over uncertain shape completed objects. For shape completion, a deep neural network is trained to take a partial view of the object as input and outputs the completed shape as a voxel grid. The key part of the network is dropout layers which are enabled not only during training but also at run-time to generate a set of shape samples representing the shape uncertainty through Monte Carlo sampling. Given the set of shape completed objects, we generate grasp candidates on the mean object shape but evaluate them based on their joint performance in terms of analytical grasp metrics on all the shape candidates. We experimentally validate and benchmark our method against another state-of-the-art method with a Barrett hand on 90000 grasps in simulation and 200 grasps on a real Franka Emika Panda. All experimental results show statistically significant improvements both in terms of grasp quality metrics and grasp success rate, demonstrating that planning shape-uncertainty- aware grasps brings significant advantages over solely planning on a single shape estimate, especially when dealing with complex or unknown objects.
|
|
15:15-15:30, Paper TuBT16.3 | Add to My Program |
Partial Caging: A Clearance-Based Definition and Deep Learning |
Varava, Anastasiia | KTH, the Royal Institute of Technology |
Welle, Michael C. | KTH Royal Institute of Technology |
Mahler, Jeffrey | University of California, Berkeley |
Goldberg, Ken | UC Berkeley |
Kragic, Danica | KTH |
Pokorny, Florian T. | KTH Royal Institute of Technology |
Keywords: Grasping, Deep Learning in Robotics and Automation
Abstract: Caging grasps limit the mobility of an object to a bounded component of configuration space. We introduce a notion of partial cage quality based on maximal clearance of an escaping path. As this is a computationally demanding task even in a two-dimensional scenario, we propose a deep learning approach. We design two convolutional neural networks and construct a pipeline for real-time partial cage quality estimation directly from 2D images of object models and planar caging tools. One neural network, CageMaskNN, is used to identify caging tool locations that can support partial cages, while a second network that we call CageClearanceNN is trained to predict the quality of those configurations. A dataset of 3811 images of objects and more than 19 million caging tool configurations is used to train and evaluate these networks on previously unseen objects and caging tool configurations. Furthermore, the networks are trained jointly on configurations for both 3 and 4 caging tool configurations whose shape varies along a 1-parameter family of increasing elongation. In experiments, we study how the networks' performance depends on the size of the training dataset, as well as how to efficiently deal with unevenly distributed training data. In further analysis, we show that the evaluation pipeline can approximately identify connected regions of successful caging tool placements and we evaluate the continuity of the cage quality score evaluation along caging tool trajectories. Experiments show that evaluation of a given configuration on a GeForce GTX 1080 GPU takes less than 6 ms.
|
|
15:30-15:45, Paper TuBT16.4 | Add to My Program |
Grasping Unknown Objects Based on Gripper Workspace Spheres |
Sorour, Mohamed | University of Montpellier |
Elgeneidy, Khaled | University of Lincoln |
Srinivasan, Aravinda | University of Lincoln, UK |
Hanheide, Marc | University of Lincoln |
Neumann, Gerhard | University of Lincoln |
Keywords: Grasping, Gripper and Other End-Effectors, Kinematics
Abstract: In this paper, we present a novel grasp planning algorithm for unknown objects given a registered point cloud of the target from different views. The proposed methodology requires no prior knowledge of the object, nor offline learning. In our approach, the gripper kinematic model is used to generate a point cloud of each finger workspace, which is then filled with spheres. At run-time, first the object is segmented, its major axis is computed, in a plane perpendicular to which, the main grasping action is constrained. The object is then uniformly sampled and scanned for various gripper poses that assure at least one object point is located in the workspace of each finger, as well as no collision with the object or its table, using computationally inexpensive gripper shape approximation. Our methodology is both time efficient (consumes less than 1.5 seconds in average) and versatile. Successful experiments have been conducted on a simple jaw gripper (Franka Panda gripper) as well as a complex, high Degree of Freedom (DoF) hand (Allegro hand).
|
|
15:45-16:00, Paper TuBT16.5 | Add to My Program |
Optimization Model for Planning Precision Grasps with Multi-Fingered Hands |
Fan, Yongxiang | University of California, Berkeley |
Zhu, Xinghao | University of California, Berkeley |
Tomizuka, Masayoshi | University of California |
Keywords: Grasping, Gripper and Other End-Effectors
Abstract: Precision grasps with multi-fingered hands are important for precise placement and in-hand manipulation tasks. Searching precision grasps on the object represented by point cloud, is challenging due to the complex object shape, high-dimensionality, collision and undesired properties of the sensing and positioning. This paper proposes an optimization model to search for precision grasps with multi-fingered hands. The model takes noisy point cloud of the object as input and optimizes the grasp quality by iteratively searching for the palm pose and finger joints positions. The collision between the hand and the object is approximated and penalized by a series of least-squares. The collision approximation is able to handle the point cloud representation of the objects with complex shapes. The proposed optimization model is able to locate collision-free optimal precision grasps efficiently. The average computation time is 0.50 sec/grasp. The searching is robust to the incompleteness and noise of the point cloud. The effectiveness of the algorithm is demonstrated by experiments.
|
|
16:00-16:15, Paper TuBT16.6 | Add to My Program |
Tunable Contact Conditions and Grasp Hydrodynamics Using Gentle Fingertip Suction (I) |
Stuart, Hannah | UC Berkeley |
Wang, Shiquan | Stanford University |
Cutkosky, Mark | Stanford University |
Keywords: Grasping, Multifingered Hands, Dynamics
Abstract: Gentle suction flow at the fingertips of a compliant hand can enhance object acquisition and increase the robustness of pinch grasps underwater. The approach adds a low-pressure pump and flexible tubes that terminate at the distal phalanges. The light flow rate does not create a powerful suction force, nor does it stir up significant sediment. The method works on porous and rough objects in addition to smooth objects as it does not require forming a seal. It changes contact conditions — normal force and coefficient of friction — and enlarges the acquisition region when grasping free objects under water. A simple hydrodynamic model matches empirical force measurements adequately for incorporation in a dynamic simulation to explore the effects of flow rate and object mass. Simulations and experiments show that effects of fingertip suction flow are most pronounced for acquiring objects on the order of 1 kg or less and when pinching large objects. Gentle suction flow is an effective, versatile and convenient addition for robots that must grasp and manipulate objects under water.
|
|
TuBT17 Regular session, LG-R17 |
Add to My Program |
Micro/Nano Robots II |
|
|
Chair: Kim, MinJun | Southern Methodist University |
Co-Chair: Zhang, Li | The Chinese University of Hong Kong |
|
14:45-15:00, Paper TuBT17.1 | Add to My Program |
Automated Sorting of Rare Cells Based on Autofocusing Visual Feedback in Fluorescence Microscopy |
Bai, Kailun | Beijing Institute of Technology |
Wang, Huaping | Beijig Institute of Technology |
Shi, Qing | Beijing Institute of Technology |
Zheng, Zhiqiang | Beijing Institute of Technology |
Cui, Juan | Beijing Institute of Technology |
Sun, Tao | Beijing Institute of Technology |
Huang, Qiang | Beijing Institute of Technology |
Dario, Paolo | Scuola Superiore Sant'Anna |
Fukuda, Toshio | Meijo University |
Keywords: Micro/Nano Robots, Automation at Micro-Nano Scales, Biological Cell Manipulation
Abstract: The research on rare cells makes a significant contribution to biology research and medical treatment for the application of diagnostic operation as well as prognoses treatment. Therefore, sorting them from heterogeneous mixtures is crucial and valuable. Traditional cell sorting methods featured with poor purity and recovery rate as well as limited flexibility, which are not ideal approaches for rare type. In this paper, we proposed a cell screening method based on automated microrobotic aspiration-and-placement strategy under fluorescence microscope. An innovative autofocusing visual feedback (AVF) method is proposed for precise three-dimensional (3D) locating of target cells. For depth detection, multiple depth from defocus (MDFD) method is adopted to solve symmetry problem and attain an average accuracy of 97.07%. For planar locating, Markov random field (MRF) based locating method is utilized to separate and locate the overlapped cells. The end actuator locating and real-time tracking are performed relying on normalized cross-correlation (NCC) method. Experiential results show that our system collects rare cells (100 cells ml-1) at a speed of 5 cells min-1 with 90% purity and 75% recovery rate, which is valuable for biological and medical application.
|
|
15:00-15:15, Paper TuBT17.2 | Add to My Program |
Continuous Mechanical Indexing of Single Cell Spheroids Using a Robot Integrated Microfluidic Chip |
Sakuma, Shinya | Nagoya University |
Nakahara, Kou | Department of Micro-Nano Mechanical Science and Engineering, Nag |
Arai, Fumihito | Nagoya University |
Keywords: Micro/Nano Robots, Automation at Micro-Nano Scales, Biological Cell Manipulation
Abstract: In single cell/spheroid analyses, flow cytometry plays a key role in high-throughput measurements. Among the considerable numbers of indexes available for a target evaluation, mechanical characteristics such as a Young's modulus have been focused upon as new indexes related to physiological condition. However, a continuous mechanical indexing system has yet to be achieved due to the difficulty in flow control, which brings about a fluctuation of the force sensor probe. In this letter, we propose an automated mechanical indexing system of spheroids. By utilizing a combination of two syringe pumps, we succeeded in positioning the target spheroids without an undesirable pinching. We conducted experiments on the continuous mechanical indexing of single cell spheroids using 26 mesenchymal stem cell (MSC) spheroids. The evaluation results of positioning showed that the error in the position control was 14.1 micrometer with a standard deviation of 87.7 micrometer, which corresponded to a measurement error of the Young's modulus of under 2%. In addition, we succeeded in achieving a measurement throughput of 3.14 spheroids per minute, which is approximately 2 times faster than a manual operation. Finally, we demonstrated the on-chip sorting of single spheroids based on the mechanical index to show the adaptability of the proposed method toward the flow cytometry.
|
|
15:15-15:30, Paper TuBT17.3 | Add to My Program |
3D Micromanipulation of Particle Swarm Using a Hexapole Magnetic Tweezer |
Zhang, Xiao | Southern Methodist University |
Rogowski, Louis | Southern Methodist University |
Kim, MinJun | Southern Methodist University |
Keywords: Micro/Nano Robots, Automation at Micro-Nano Scales, Swarms
Abstract: This article discusses the design, modeling, and application of a powerful hexapole magnetic tweezer system for closed-loop 3D swarm control applications. The system consists of six sharp tapered magnetic poles that are integrated with six electromagnetic coils and mounted on two yokes composed of 3D printed magnetic material. Magnetic field gradients are generated at the sharp tips of the magnetic poles when current is applied through the attached electromagnetic coils. Different combinations of current input can interact with magnetized microparticles to create three-dimensional motion. A closed-loop control algorithm based on image processing and hardware integration through MATLAB was developed to automatically operate external power supplies connected to the magnetic tweezer system. Coordinate system transformation is utilized to transform the tilted actuation coordinates, by virtue of the system hardware configuration, to the measurement coordinates used during experiments and analysis. This magnetic tweezer system has the advantage of a larger working space and higher magnetic field strengths when compared to several other similar designs. The magnetic tweezer system allows for more diverse applications within the microscale, such as microparticle swarm control, cell penetration, and cell therapy. Experimental analysis performed in this article demonstrates the closed-loop navigation of a microparticle swarm moving freely in both 2D and 3D environments. Results show highly consistent trajectories within the swarm with only a few fluctuations due to microflows. This system will keep being updated and optimized to investigate the performance of microparticles in in vivo environments.
|
|
15:30-15:45, Paper TuBT17.4 | Add to My Program |
High-Speed On-Chip Mixing by Micro-Vortex Generated by Controlling Local Jet Flow Using Dual Membrane Pumps |
Kasai, Yusuke | Nagoya University |
Sakuma, Shinya | Nagoya University |
Arai, Fumihito | Nagoya University |
Keywords: Micro/Nano Robots
Abstract: Robot integrated microfluidic chip is a key technology for microscale applications. Recently, the technology has been applied to on-chip mixing which mix solutions on a microfluidic chip because it is a promising tool to analyze not only the chemical reaction with the small sample volume but also the response of cells to environmental changes. However, these conventional mixing methods require the mixing time of millisecond-order due to the difficulty of mixing in the laminar condition of a microchannel whose Reynolds number tends to be low. In this letter, we propose a high-speed on-chip mixing by the micro-vortex generated by controlling local jet flow using dual membrane pumps. First, we confirmed that vortices were successfully generated within 20 us by the local jet flow. The velocity and Reynolds number were analytically estimated as approximately 20 m/s and 1.6x10^3, respectively. Second, we evaluated the response time of the mixing using the micro-vortex. We mixed 200-nm nanobead suspension and the DI water in the velocity of main flow of 1 m/s. By measuring the intensity at the certain observation area, we confirmed that our method successfully mixed solutions and the mixing time was approximately 500 us, whose speed has not been achieved by conventional robot integrated on-chip mixers. Moreover, we confirmed that our system can control the concentration of mixed flow by controlling flow rate ratio of sample and sheath flow. From these results, we confirmed that we achieved high-speed on-demand on-chip mixing by the micro-vortex.
|
|
15:45-16:00, Paper TuBT17.5 | Add to My Program |
Magnetic-Needle-Assisted Micromanipulation of Dynamically Self-Assembled Magnetic Droplets for Cargo Transportation |
Wang, Qianqian | The Chinese University of Hong Kong |
Du, Xingzhou | The Chinese University of Hong Kong |
Ji, Fengtong | The Chinese University of Hong Kong |
Zhang, Li | The Chinese University of Hong Kong |
Keywords: Micro/Nano Robots
Abstract: Dynamic self-assembly is treated as a promising approach for generating a robotic swarm to perform coordinated tasks, and the assembled pattern can be tuned by regulating the energy input. However, location of a dynamically assembled pattern is hard to be determined, especially under global fields, such as magnetic field. In this paper, we report the formation and manipulation of dynamic self-assembled droplets at the air-liquid interface with the assistance of a magnetic needle. Affected by the locally induced field gradient near the needle, reconfigurable assembled droplets are obtained with higher time-efficiency, and the location of the pattern can be determined. The pattern is reversibly tuned to exhibit expansion and shrinkage by adjusting the height of the needle. Assembled droplets are able to be steered via following the needle in a controlled manner. Moreover, cargo is trapped by exploiting the induced rotational flow around the droplets, and it can also be caged into the central area of the pattern and transported to the desired location. The proposed method opens new prospects of using energy-dissipative pattern as an untethered end-effector for microrobotic manipulation.
|
|
16:00-16:15, Paper TuBT17.6 | Add to My Program |
Vision-Based Magnetic Platform for Actuator Positioning and Wireless Control of Microrobots |
Zarrouk, Azaddien | INSA Centre Val De Loire |
Belharet, Karim | Hautes Etudes d'Ingénieur - HEI Campus Centre |
Tahri, Omar | INSA Centre Val-De-Loire |
Keywords: Medical Robots and Systems, Automation at Micro-Nano Scales, Micro/Nano Robots
Abstract: This work presents a method to guide microrobots by positioning a magnetic actuator using hybrid vision system. The used actuator mounted at a robot end-effector creates local maxima of the magnetic field magnitude, which results in an attractive point for microrobots in its influence zone. The hybrid vision system serves to control the actuator position through the robotic platform to make the trapped microrobot undergo a planned trajectory. In first validation results, the particle driving is achieved in open loop with no adjustment with respect to the planned trajectory. Such scheme can be used in the case where the microrobot position can not be measured in real time. The second validation results deal with the case of microrobot positioning by visual servoing. This can be used in the case where recovering microrobot position is possible, for example in eye treatment.
|
|
TuBT18 Regular session, LG-R18 |
Add to My Program |
Localization II |
|
|
Chair: Milford, Michael J | Queensland University of Technology |
Co-Chair: Fontanelli, Daniele | University of Trento |
|
14:45-15:00, Paper TuBT18.1 | Add to My Program |
Robot Localization Via Odometry-Assisted Ultra-Wideband Ranging with Stochastic Guarantees |
Magnago, Valerio | University of Trento |
Corbalán, Pablo | University of Trento |
Picco, Gian Pietro | University of Trento |
Palopoli, Luigi | University of Trento |
Fontanelli, Daniele | University of Trento |
Keywords: Localization, Range Sensing, Sensor Fusion
Abstract: We consider the problem of accurate and high-rate self-localization for a mobile robot. We adaptively combine the speed information acquired by proprioceptive sensors with intermittent positioning samples acquired via ultra-wideband (UWB) radios. These are triggered only if and when needed to reduce the positioning uncertainty, itself modeled by a probabilistic cost function. Our formulation is agnostic w.r.t. the source of uncertainty and enables an intuitive specification of user navigation requirements along with stochastic guarantees on the system operation. Experimental results in simulation and with a real platform show that our approach: i) meets these guarantees in practice; ii) achieves the same accuracy of a fixed periodic sampling but with significantly higher scalability and lower energy consumption; iii) is resilient to errors in UWB estimates, enabling the use of low-accuracy ranging schemes which further improve these two performance metrics.
|
|
15:00-15:15, Paper TuBT18.2 | Add to My Program |
Sparse-3D Lidar Outdoor Map-Based Autonomous Vehicle Localization |
Ahmed, Syed Zeeshan | Institute for Infocomm Research (I2R), A*STAR |
Saputra, Vincensius Billy | National University of Singapore |
Verma, Saurab | Institute of Infocomm Research, Agency for Science, Technology A |
Zhang, Kun | Institute for Infocomm Research (I2R), A*STAR |
Adiwahono, Albertus Hendrawan | I2R A-STAR |
Keywords: Localization, Sensor Fusion, Intelligent Transportation Systems
Abstract: Difficulties in capturing unique structures in the outdoor environment hinders the map-based Autonomous Vehicles (AV) localization performance.Accordingly, this necessitates the use of high resolution sensors to capture more information from the environment. However, this approach is costly and limits the mass deployment of AV. To overcome this drawback, in this paper, we propose a novel outdoor map-based localization method for Autonomous Vehicles in urban environments using sparse 3D lidar scan data. In the proposed method, a Point-to-Distribution (P2D) formulation of the Normal Distributions Transform (NDT) approach is applied in a Monte Carlo Localization (MCL) framework. The formulation improves the measurement model of localization by taking individual lidar point measurements into consideration. Additionally, to apply the localization to scalable outdoor environments, a flexible and efficient map structure is implemented. The experimental results indicate that the proposed approach significantly improves the localization and its robustness in outdoor AV environments, especially with limited sparse lidar data.
|
|
15:15-15:30, Paper TuBT18.3 | Add to My Program |
Mobile Robot Localization with Reinforcement Learning Map Update Decision Aided by an Absolute Indoor Positioning System |
Garrote, Luís Carlos | Institute of Systems and Robotics |
Torres, Miguel | Institute of Systems and Robotics - University of Coimbra |
Barros, Tiago | Institute of Systems and Robotics - University of Coimbra |
Perdiz, João | University of Coimbra |
Premebida, Cristiano | Loughborough University |
Nunes, Urbano J. | Instituto De Sistemas E Robotica |
Keywords: Localization, Sensor Fusion, Mapping
Abstract: This paper introduces a new mobile robot localization solution consisting of two main modules: a Particle-Filter based Localization (PFL) and a Reinforcement-Learning based map updating, integrating relative measurements and absolute indoor positioning sensor (A-IPS) data. Concerning localization using 2D-LiDARs, featureless areas are known to be problematic. To solve this problem a classic PFL approach was modified to incorporate A-IPS position measurements in the prediction and update stages. The localization approach has the particularity of including the possibility of updating the map whenever major modifications are detected in the environment in relation to the current localization map. Due to the random sampling-based nature of the PFL, an associated map update solution is not trivial since small inconsistencies in the estimated pose can lead to erroneous map associations. The proposed method learns to decide by assigning higher rewards the greater is the overlap between the map and the 2D-LIDAR scans, via RL, and then a proper update of the map is achieved. Validation of the proposed pipeline was carried out in a differential drive platform with algorithms developed in ROS. Tests were performed in two scenarios in order to assess the performance of both the localization module and the map update stage. The results show that the proposed localization method offers improvements in relation to known approaches, and consequently suggest promising perspectives for the proposed map update decision framework.
|
|
15:30-15:45, Paper TuBT18.4 | Add to My Program |
GLFP: Global Localization from a Floor Plan |
Wang, Xipeng | Toyota Research Institute |
Marcotte, Ryan | University of Michigan |
Olson, Edwin | University of Michigan |
Keywords: Localization
Abstract: In this paper, we describe a method for global localization in a previously unvisited environment using only a schematic floor plan as a prior map. The floor plan need not be a precision map – it can be the sort of image found in buildings to guide people or aid evacuation. The core idea is to identify features that are stable across both a drawn floor plan and robot point-of-view LIDAR data, for example wall intersections, which appear as corners from overhead and as vertical lines from the ground. We introduce a factor graph-based global localization method that uses these features as landmarks. The detections of such descriptorless features are noisy and often ambiguous. We therefore propose robust data association based on a pairwise measurement consistency check and max- mixtures error model. We evaluate the resulting system in a real-world indoor environment, demonstrating performance comparable to a baseline system that uses a conventional LIDAR-based prior map.
|
|
15:45-16:00, Paper TuBT18.5 | Add to My Program |
Automatic Coverage Selection for Surface-Based Visual Localization |
Mount, James | Queensland University of Technology |
Dawes, Les | Queensland University of Technology |
Milford, Michael J | Queensland University of Technology |
Keywords: Localization
Abstract: Localization is a critical capability for robots, drones and autonomous vehicles operating in a wide range of environments. One of the critical considerations for designing, training or calibrating visual localization systems is the coverage of the visual sensors equipped on the platforms. In an aerial context for example, the altitude of the platform and camera field of view plays a critical role in how much of the environment a downward facing camera can perceive at any one time. Furthermore, in other applications, such as on roads or in indoor environments, additional factors such as camera resolution and sensor placement altitude can also affect this coverage. The sensor coverage and the subsequent processing of its data also has significant computational implications. In this paper we present for the first time a set of methods for automatically determining the trade-off between coverage and visual localization performance, enabling the identification of the minimum visual sensor coverage required to obtain optimal localization performance with minimal compute. We develop a localization performance indicator based on the overlapping coefficient, and demonstrate its predictive power for localization performance with a certain sensor coverage. We evaluate our method on several challenging real-world datasets from aerial and ground-based domains, and demonstrate that our method is able to automatically optimize for coverage using a small amount of calibration data. We hope these results will assist in the design of localization systems for future autonomous robot, vehicle and flying systems.
|
|
16:00-16:15, Paper TuBT18.6 | Add to My Program |
BTEL: A Binary Tree Encoding Approach for Visual Localization |
Le, Huu | Chalmers University of Technology |
Hoang, Tuan | Singapore University of Technology and Design |
Milford, Michael J | Queensland University of Technology |
Keywords: Localization
Abstract: The navigation capability of autonomous and unmanned aerial critically relies on the effectiveness of the underlying localization algorithms. Recent advances in camera technology and vision-based techniques have yielded significant improvements in localization performance, with one critical caveat: all current approaches currently scale at best linearly with the size of the environment with respect to both storage, and consequentially in most approaches, query time. This limitation severely curtails the capability of autonomous systems in a wide range of compute, power, storage, size, weight or cost constrained applications such as drones. More generally and regardless of the absolute hardware constraints, sub-linear storage growth enables either improved scaling in operational envelope, an increase in the sophistication and complexity of information stored per place or a reduction in the computational and storage requirements in comparison to traditional methods. In this work, we present a novel binary tree encoding approach for visual localization which can serve as an alternative for existing quantization and indexing techniques. The proposed tree structure allows us to derive a compressed training scheme that achieves sub-linearity in both required storage and inference time. The encoding memory can be easily configured to satisfy different storage constraints. Moreover, our approach is amenable to an optional sequence filtering mechanism to further improve the localization results, while maintaining the same amount of storage. Our system is entirely agnostic to the front-end descriptors, allowing it to be used on top of recent state-of-the-art image representations. Experimental results show that the proposed method significantly outperforms state-of-the-art approa
|
|
TuBT19 Regular session, LG-R19 |
Add to My Program |
Planning, Scheduling, and Coordination I |
|
|
Chair: Bezzo, Nicola | University of Virginia |
Co-Chair: Bhattacharya, Sourabh | Iowa State University |
|
14:45-15:00, Paper TuBT19.1 | Add to My Program |
Planning in Stochastic Environments with Goal Uncertainty |
Saisubramanian, Sandhya | University of Massachusetts Amherst |
Wray, Kyle | Alliance Innovation Lab Silicon Valley |
Pineda, Luis | University of Massachusetts Amherst |
Zilberstein, Shlomo | University of Massachusetts |
Keywords: Planning, Scheduling and Coordination, Autonomous Agents, AI-Based Methods
Abstract: We present the Goal Uncertain Stochastic Shortest Path (GUSSP) problem-a general framework to model path planning and decision making in stochastic environments with goal uncertainty. The framework extends the stochastic shortest path (SSP) model to dynamic environments in which it is impossible to determine the exact goal states ahead of plan execution. GUSSPs introduce flexibility in goal specification by allowing a belief over possible goal configurations. The unique observations at potential goals helps the agent identify the true goal during plan execution. The partial observability is restricted to goals, facilitating the reduction to an SSP with a modified state space. We formally define a GUSSP and discuss its theoretical properties. We then propose an admissible heuristic that reduces the planning time using FLARES-a start-of-the-art probabilistic planner. We also propose a determinization approach for solving this class of problems. Finally, we present empirical results on a search and rescue mobile robot and three other problem domains in simulation.
|
|
15:00-15:15, Paper TuBT19.2 | Add to My Program |
Adaptive Outcome Selection for Planning with Reduced Models |
Saisubramanian, Sandhya | University of Massachusetts Amherst |
Zilberstein, Shlomo | University of Massachusetts |
Keywords: Planning, Scheduling and Coordination, Autonomous Agents
Abstract: Reduced models allow autonomous robots to cope with the complexity of planning in stochastic environments by simplifying the model and reducing its accuracy. The solution quality of a reduced model depends on its fidelity. We present 0/1 reduced model that selectively improves model fidelity in certain states by switching between using a simplified deterministic model and the full model, without significantly compromising the run time gains. We measure the reduction impact for a reduced model based on the values of the ignored outcomes and use this as a heuristic for outcome selection. Finally, we present empirical results of our approach on three different domains, including an electric vehicle charging problem using real-world data from a university campus.
|
|
15:15-15:30, Paper TuBT19.3 | Add to My Program |
Fast Run-Time Monitoring, Replanning, and Recovery for Safe Autonomous System Operations |
Yel, Esen | University of Virginia |
Bezzo, Nicola | University of Virginia |
Keywords: Planning, Scheduling and Coordination, Autonomous Vehicle Navigation, Aerial Systems: Applications
Abstract: In this paper, we present a fast run-time monitoring framework for safety assurance during autonomous system operations in uncertain environments. Modern unmanned vehicles rely on periodic sensor measurements for motion planning and control. However, a vehicle may not always be able to obtain its state information due to various reasons such as sensor failures, signal occlusions, and communication problems. To guarantee the safety of a system during these circumstances under the presence of disturbance and noise, we propose a novel fast reachability analysis approach that leverages Gaussian process regression theory to predict future states of the system at run-time. We also propose a self/event-triggered monitoring and replanning approach which leverages our fast reachability scheme to recover the system when needed and replan its trajectory to guarantee safety constraints (i.e., the system will not collide with any obstacles). Our technique is validated both with simulations and experiments on unmanned aerial vehicles case studies in cluttered environments under the effect of unknown wind disturbance at run-time.
|
|
15:30-15:45, Paper TuBT19.4 | Add to My Program |
Cooperative Schedule-Driven Intersection Control with Connected and Autonomous Vehicles |
Hu, Hsu-Chieh | Carnegie Mellon University |
Smith, Stephen F. | Carnegie Mellon University |
Goldstein, Richard | CMU |
Keywords: Planning, Scheduling and Coordination, Intelligent Transportation Systems, Automation Technologies for Smart Cities
Abstract: Recent work in decentralized, schedule-driven traffic control has demonstrated the ability to improve the efficiency of traffic flow in complex urban road networks. In this approach, a scheduling agent is associated with each intersection. Each agent senses the traffic approaching its intersection and in real-time constructs a schedule that minimizes the cumulative wait time of vehicles approaching the intersection over the current look-ahead horizon. In this paper, we propose a cooperative algorithm that utilizes both connected and autonomous vehicles (CAV) and schedule-driven traffic control to create better traffic flow in the city. The algorithm enables an intersection scheduling agent to adjust the arrival time of an approaching platoon through use of wireless communication to control the velocity of vehicles. The sequence of approaching platoons is thus shifted toward a new shape that has smaller cumulative delay. We demonstrate how this algorithm outperforms the original approach in a real-time traffic signal control problem.
|
|
15:45-16:00, Paper TuBT19.5 | Add to My Program |
Multirobot Charging Strategies: A Game-Theoretic Approach |
Gao, Tianshuang | Iowa State University |
Bhattacharya, Sourabh | Iowa State University |
Keywords: Planning, Scheduling and Coordination, Multi-Robot Systems
Abstract: This work considers the problem of assigning multiple robots to charging stations in order to minimize the total time required by all robots for the charging operation. We first show that the centralized problem is NP-hard. Then we formulate the charging problem as a non-cooperative game. We propose an algorithm to obtain the pure strategy Nash equilibrium of the non-cooperative game, and show its uniqueness. We investigate the price of anarchy (PoA) of this equilibrium as a function of the number of robots and stations. Next, we leverage our analysis on static charging stations to propose strategies for reducing the total cost when the charging stations are mobile. Finally, we analyze the performance of the strategies proposed for the charging stations through extensive simulation.
|
|
16:00-16:15, Paper TuBT19.6 | Add to My Program |
Toward Model-Based Benchmarking of Robot Components |
Bardaro, Gianluca | Politecnico Di Milano |
El-Shamouty, Mohamed | Fraunhofer IPA |
Fontana, Giulio | Politecnico Di Milano |
Awad, Ramez | Fraunhofer IPA |
Matteucci, Matteo | Politecnico Di Milano |
Keywords: Performance Evaluation and Benchmarking
Abstract: The results of scientific experiments performed by different groups are rarely directly comparable. Efforts such as the European Robotics League propose to the community, in the form of competitions, well documented and stable benchmarks to assess the performance of existing systems. However, benchmarks can be equally useful at design time: the Plug&Bench Benchmark Meta-model provides robot designers with a valuable addition to their toolkit. Additionally, it enables -with benchmark composition- to predict system performance given the benchmark results of individual components.
|
|
TuBT20 Regular session, LG-R20 |
Add to My Program |
Biologically-Inspired Robots II |
|
|
Chair: Kurabayashi, Daisuke | Tokyo Institute of Technology |
Co-Chair: Guo, Shuxiang | Kagawa University |
|
14:45-15:00, Paper TuBT20.1 | Add to My Program |
Lifelong Federated Reinforcement Learning: A Learning Architecture for Navigation in Cloud Robotic Systems |
Liu, Boyi | Chinese Academy of Sciences |
Wang, Lujia | Shenzhen Institutes of Advanced Technology |
Liu, Ming | Hong Kong University of Science and Technology |
Keywords: Biologically-Inspired Robots, Legged Robots, Climbing Robots
Abstract: This paper was motivated by the problem of how to make robots fuse and transfer their experience so that they can effectively use prior knowledge and quickly adapt to new environments. To address the problem, we present a learning architecture for navigation in cloud robotic systems: Lifelong Federated Reinforcement Learning (LFRL). In the work, we propose a knowledge fusion algorithm for upgrading a shared model deployed on the cloud. Then, effective transfer learning methods in LFRL are introduced. LFRL is consistent with human cognitive science and fits well in cloud robotic systems. Experiments show that LFRL greatly improves the efficiency of reinforcement learning for robot navigation. The cloud robotic system deployment also shows that LFRL is capable of fusing prior knowledge. In addition, we release a cloud robotic navigation-learning website to provide the service based on LFRL: www.shared-robotics.com.
|
|
15:00-15:15, Paper TuBT20.2 | Add to My Program |
Study on Elastic Elements Allocation for Energy-Efficient Robotic Cheetah Leg |
Borisov, Ivan | ITMO University |
Kulagin, Ivan | University of Information Technologies, Mechanics and Optics (IT |
Larkina, Anastasiya | ITMO University, Saint-Peterburg |
Egorov, Artem | ITMO University, Saint Petersburg, Russia |
Kolyubin, Sergey | ITMO University |
Stramigioli, Stefano | University of Twente |
Keywords: Biologically-Inspired Robots, Legged Robots, Simulation and Animation
Abstract: The biomimetic approach in robotics is promising: nature has found many good solutions through millions of years of evolution. However, creating a design that enables fast and energy-efficient locomotion remains a major challenge. This paper focuses on the development of a full leg mechanism for a fast and energy-efficient 4-legged robot inspired by a cheetah morphology. In particular, we analyze how the allocation of flexible elements and their stiffness affects the cost of transport and peak power characteristics for vertical jumps and a galloping motion. The study includes the femur and full leg mechanism's locomotory behavior simulation, capturing its interaction with the ground.
|
|
15:15-15:30, Paper TuBT20.3 | Add to My Program |
A Novel Small-Scale Turtle-Inspired Amphibious Spherical Robot |
Xing, Huiming | Beijing Institute of Technology |
Guo, Shuxiang | Kagawa University |
Shi, Liwei | Beijing Institute of Technology |
Xihuan, Hou | Beijing Institude of Technology |
Liu, Yu | Beijing Institute of Technology |
Liu, Huikang | Beijing Institute of Technology |
Hu, Yao | Beijing Institute of Technology |
Xia, Debin | Beijing Institute of Technology |
Li, Zan | Beijing Institute of Technology |
Keywords: Biologically-Inspired Robots, Legged Robots, Mechanism Design
Abstract: This paper describes a novel small-scale turtle-inspired Amphibious Spherical Robot (ASRobot) to accomplish exploration tasks in the restricted environment, such as amphibious areas and narrow underwater cave. A Legged, Multi-Vectored Water-Jet Composite Propulsion Mechanism (LMVWCPM) is designed with four legs, one of which contains three connecting rod parts, one water-jet thruster and three joints driven by digital servos. Using this mechanism, the robot is able to walk like amphibious turtles on various terrains and swim flexibly in submarine environment. A simplified kinematic model is established to analyze crawling gaits. With simulation of the crawling gait, the driving torques of different joints contributed to the choice of servos and the size of links of legs. Then we also modeled the robot in water and proposed several underwater locomotion. In order to assess the performance of the proposed robot, a series of experiments were carried out in the lab pool and on flat ground using the prototype robot. Experiments results verified the effectiveness of LMVWCPM and the amphibious control approaches.
|
|
15:30-15:45, Paper TuBT20.4 | Add to My Program |
Experimental Analysis of the Influence of Olfactory Property on Chemical Plume Tracing Performance |
Shigaki, Shunsuke | Osaka University |
Okajima, Kei | Yokohama National University |
Sanada, Kazushi | Yokohama National University |
Kurabayashi, Daisuke | Tokyo Institute of Technology |
Keywords: Biologically-Inspired Robots, Localization, Calibration and Identification
Abstract: In this study, we investigated the relationship between the performance of chemical plume tracing (CPT) and odor sensing property. Tracking of chemical plumes plays an important role because it facilitates the identification of an odor source. Conventional research has focused on the development of CPT algorithms, whereas the influence of the performance odor sensors on CPT performance has not been investigated. Therefore, in this study, we first compared the olfactory characteristics of an insect (silkworm moth, Bombyx mori), which has high CPT performance and the characteristics of an artificial odor sensor. In particular, we focused on and compared the recovery time of the two types of sensors, which plays an important role in the acquisition of odor information. As a result, it was determined that the recovery time for the insect olfactory sensor was 10 times faster than that of the artificial odor sensor. We also experimentally evaluated the effect of this difference in the recovery time on CPT performance. CPT experiments using silkworm moths and a robot revealed that there was a correlation between the CPT performance and sensor recovery time. As such, it was demonstrated that it is necessary to improve not only the algorithm but also the sensor recovery time improve the CPT performance.
|
|
15:45-16:00, Paper TuBT20.5 | Add to My Program |
Efficient Quadrupedal Walking Via Decentralized Coordination Mechanism between Limbs and Neck |
Fukuhara, Akira | Tohoku University |
Suzuki, Shura | Tohoku University |
Kano, Takeshi | Tohoku University |
Ishiguro, Akio | Tohoku University |
Keywords: Biologically-Inspired Robots, Legged Robots
Abstract: An optional arm unit for a quadruped robot, similar to the neck of an animal, facilitates an improved performance in multifunctional tasks such as object manipulation and patrolling. However, the increase in body weight due to the neck unit will exacerbate the energy issues for the locomotion of legged robots. This study presents a minimal mechanism for efficient walking, inspired by the nodding behaviors of quadruped mammals, via a decentralized coordination of the neck and limbs. The results of simulations in two dimensions (2D) suggest that, to achieve efficient quadrupedal walking, the mechanism of bilateral sensory feedback between the neck and the limbs plays an essential role for the quadruped robots' coordination of the neck and limbs in response to the physical situation of the body parts.
|
|
16:00-16:15, Paper TuBT20.6 | Add to My Program |
Effects of a Bio-Mimicked Flapping Path on Propulsion Efficiency of Two-Segmental Fish Robots |
Abedinzadeh Shahri, Majid | University of Tehran |
Rouhollahi, Ali | University of Tehran, Faculty of Engineering, School of Electric |
Nili Ahmadabadi, Majid | University of Tehran |
Keywords: Biologically-Inspired Robots, Marine Robotics, Biomimetics
Abstract: Having an appropriate flapping path to yield efficient propulsion is an interesting issue in fish robotics. In most works, especially two-segmental structures, the flapping motion is limited to sinusoidal functions. In this paper, to cope with the aforementioned limitation, a conceptual nonsinusoidal path is proposed. The proposed flapping path and the conventional one, both are optimized for a sample fish robot. According to some simulation results, it is shown that if a proper actuator is employed to generate both optimized paths, the proposed approach yields more propulsion efficiency. Furthermore, it is discussed that our method can better imitate fish muscle output power. Finally, through experiments, some practical issues are considered.
|
|
TuCT1 Regular session, L1-R1 |
Add to My Program |
RGB-D Perception |
|
|
Chair: Vincze, Markus | Vienna University of Technology |
Co-Chair: Manocha, Dinesh | University of Maryland |
|
16:45-17:00, Paper TuCT1.1 | Add to My Program |
Directional TSDF: Modeling Surface Orientation for Coherent Meshes |
Splietker, Malte | University of Bonn |
Behnke, Sven | University of Bonn |
Keywords: RGB-D Perception, Computer Vision for Automation, Object Detection, Segmentation and Categorization
Abstract: Real-time 3D reconstruction from RGB-D sensor data plays an important role in many robotic applications, such as object modeling and mapping. The popular method of fusing depth information into a truncated signed distance function (TSDF) and applying the marching cubes algorithm for mesh extraction has severe issues with thin structures: not only does it lead to loss of accuracy, but it can generate completely wrong surfaces. To address this, we propose the directional TSDF — a novel representation that stores opposite surfaces separate from each other. The marching cubes algorithm is modified accordingly to retrieve a coherent mesh representation. We further increase the accuracy by using surface gradient-based ray casting for fusing new measurements. We show that our method outperforms state-of-the-art TSDF reconstruction algorithms in mesh accuracy.
|
|
17:00-17:15, Paper TuCT1.2 | Add to My Program |
Monocular Depth Estimation in New Environments with Absolute Scale |
Roussel, Tom | KU Leuven |
Tuytelaars, Tinne | KU Leuven |
Van Eycken, Luc | KU Leuven - University of Leuven, Department of Electrical Engin |
Keywords: RGB-D Perception, Deep Learning in Robotics and Automation, Computer Vision for Automation
Abstract: In this work we propose an unsupervised training method that finetunes a single image depth estimation CNN towards a new environment. The network, which has been pretrained on stereo data, only requires monocular input for finetuning. Unlike other unsupervised methods, it produces depth estimations with absolute scale – a feature that is essential for most practical applications, yet has mostly been overlooked in the literature. First, we show how our method allows adapting a network trained on one dataset (Cityscapes) to another (KITTI). Next, by splitting KITTI in subsets, we show the sensitivity of pretrained models to a domain shift. We then demonstrate that, by finetuning the model using our method, it is possible to improve the performance on the target subset, without using stereo or any form of groundtruth depth and with preservation of the correct absolute scale.
|
|
17:15-17:30, Paper TuCT1.3 | Add to My Program |
Frustum ConvNet: Sliding Frustums to Aggregate Local Point-Wise Features for Amodal 3D Object Detection |
Wang, Zhixin | South China University of Technology |
Jia, Kui | South China University of Technology |
Keywords: RGB-D Perception, Object Detection, Segmentation and Categorization
Abstract: In this work, we propose a novel method termed emph{Frustum ConvNet (F-ConvNet)} for amodal 3D object detection from point clouds. Given 2D region proposals in an RGB image, our method first generates a sequence of frustums for each region proposal, and uses the obtained frustums to group local points. F-ConvNet aggregates point-wise features as frustum-level feature vectors, and arrays these feature vectors as a feature map for use of its subsequent component of fully convolutional network (FCN), which spatially fuses frustum-level features and supports an end-to-end and continuous estimation of oriented boxes in the 3D space. We also propose component variants of F-ConvNet, including an FCN variant that extracts multi-resolution frustum features, and a refined use of F-ConvNet over a reduced 3D space. Careful ablation studies verify the efficacy of these component variants. F-ConvNet assumes no prior knowledge of the working 3D environment and is thus dataset-agnostic. We present experiments on both the indoor SUN-RGBD and outdoor KITTI datasets. F-ConvNet outperforms all existing methods on SUN-RGBD, and at the time of submission it outperforms all published works on the KITTI benchmark. Code has been made available at:{url{https://github.com/zhixinwang/frustum-convnet}.}
|
|
17:30-17:45, Paper TuCT1.4 | Add to My Program |
Recurrent Convolutional Fusion for RGB-D Object Recognition |
Loghmani, Mohammad Reza | Vienna University of Technology |
Planamente, Mirco | Italian Institute of Technology |
Caputo, Barbara | Sapienza University |
Vincze, Markus | Vienna University of Technology |
Keywords: RGB-D Perception, Recognition, Visual Learning
Abstract: Providing robots with the ability to recognize objects like humans has always been one of the primary goals of robot vision. The introduction of RGB-D cameras has paved the way for a significant leap forward in this direction thanks to the rich information provided by these sensors. However, the robot vision community still lacks an effective method to synergically use the RGB and depth data to improve object recognition. In order to take a step in this direction, we introduce a novel end-to-end architecture for RGB-D object recognition called recurrent convolutional fusion (RCFusion). Our method generates compact and highly discriminative multi-modal features by combining RGB and depth information representing different levels of abstraction. Extensive experiments on two popular datasets, RGB-D Object Dataset and JHUIT-50, show that RCFusion significantly outperforms state-of-the-art approaches in both the object categorization and instance recognition tasks. In addition, experiments on the more challenging Object Clutter Indoor Dataset confirm the validity of our method in the presence of clutter and occlusion. The code is publicly available at: https://github.com/MRLoghmani/rcfusion.
|
|
17:45-18:00, Paper TuCT1.5 | Add to My Program |
Piecewise Rigid Scene Flow with Implicit Motion Segmentation |
Goerlitz, Andreas | University of Siegen |
Geiping, Jonas | University of Siegen |
Kolb, Andreas | University of Siegen |
Keywords: RGB-D Perception, Visual Tracking, Object Detection, Segmentation and Categorization
Abstract: In this paper, we introduce a novel variational approach to estimate the scene flow from RGB-D images. We regularize the ill-conditioned problem of scene flow estimation in a unified framework by enforcing piecewise rigid motion through decomposition into rotational and translational motion parts. Our model crucially regularizes these components by an L_0 ``norm'', thereby facilitating implicit motion segmentation in a joint energy minimization problem. Yet, we also show that this energy can be efficiently minimized by a proximal primal-dual algorithm. By implementing this approximate L_0 rigid motion regularization, our scene flow estimation approach implicitly segments the observed scene of into regions of nearly constant rigid motion. We evaluate our joint scene flow and segmentation estimation approach on a variety of test scenarios, with and without ground truth data, and demonstrate that we outperform current scene flow techniques.
|
|
18:00-18:15, Paper TuCT1.6 | Add to My Program |
3D Deformable Object Manipulation Using Deep Neural Networks |
Hu, Zhe | City University of Hong Kong |
Han, Tao | City University of Hong Kong |
Sun, Peigen | City University of Hong Kong |
Pan, Jia | University of Hong Kong |
Manocha, Dinesh | University of Maryland |
Keywords: Dual Arm Manipulation, Perception for Grasping and Manipulation
Abstract: Due to its high dimentionality, deformable object manipulation is a challenging problem in robotics. In this paper, we present a deep neural network based controller to servo-control the position and shape of deformable objects with unknown deformation properties. In particular, a multi-layer neural network is used to map between the robotic end-effector's movement and the object's deformation measurement using an online learning strategy. In addition, we introduce a novel feature to describe deformable objects' deformation used in visual-servoing. This feature is directly extracted from the 3D point cloud rather from the 2D image as in previous work. In addition, we perform simultaneous tracking and reconstruction for the deformable object to resolve the partial observation problem during the deformable object manipulation. We validate the performance of our algorithm and controller on a set of deformable object manipulation tasks and demonstrate that our method can achieve effective and accurate servo-control for general deformable objects with a wide variety of goal settings. Experiment videos are available at https://sites.google.com/view/mso-deep.
|
|