| |
Last updated on September 10, 2024. This conference program is tentative and subject to change
Technical Program for Tuesday August 27, 2024
|
TuAT1 |
Room T1 |
Personalising Robots Behaviour in Social Human-Robot Interactions I |
Special Session |
Chair: Andriella, Antonio | Artificial Intelligence Research Institute, CSIC |
Co-Chair: Merino, Luis | Universidad Pablo De Olavide |
|
10:15-10:30, Paper TuAT1.1 | |
Human-Like Social Learning for Social Robots: A Systematic Review (I) |
|
Burkart, Diana (Karlsruhe Institute of Technology (KIT)), Bruno, Barbara (Karlsruhe Institute of Technology (KIT)) |
Keywords: Social Learning and Skill Acquisition Via Teaching and Imitation, Machine Learning and Adaptation, Cognitive Skills and Mental Models
Abstract: Social learning is a learning paradigm aiming to make the process of teaching new skills to an artificial intelligent agent (such as a robot) as close as possible to the process we employ when teaching skills to other humans. Equipping robot companions with the ability to learn from social interactions with their users would enable naive users to effortlessly teach skills to robots and support personalization and long-term use. Aiming to support research on the field, in this article we present a systematic literature review of human-like social learning for social robots, with a specific focus on works employing real robots and real human-robot social interactions. The review examines relevant papers from perspectives including the type of robot used, the task to be learned, and the method and assessment metrics employed, and allows for the identification of open research avenues.
|
|
10:30-10:45, Paper TuAT1.2 | |
Social Attributes in a Handshake Robot Adaptive to Human Shaking Motion Using a CPG Controller (I) |
|
Yamasaki, Kakeru (Kyushu Institute of Technology), Shibata, Tomohiro (Kyushu Institute of Technology), Henaff, Patrick (Université De Lorraine, CNRS, INRIA, LORIA, F-54000 Nancy, Fra) |
Keywords: Applications of Social Robots, Creating Human-Robot Relationships, Social Touch in Human–Robot Interaction
Abstract: In the field of human-robot interaction(HRI), there are two types of human-robot coordination, one is physical coordination, which refers to the direct coordination between the robot and the human who interacts, and the other is social coordination, which refers to the human/robot coordination through emotional and cognitive states. In particular, social coordination can be an important factor in considering the social position of robots for the social implementation of technology. A handshake is usually the first interaction between humans that involves physical contact, and it is the first opportunity for social exchanges. In this paper, a handshaking robot that physically synchronizes with human handshaking motions is controlled with a Rowat-Selverston CPG (Central Pattern Generator) neuron oscillator. We compared a handshake robot adapted to a passive control method using simple impedance control and a cooperative and active control method using the developed CPG controller, and investigated how the control method affects the robot's sociodemographic attributes as perceived by humans. Changes in social attributes were evaluated in a 14-subject experiment using the Robotic Social Attributes Scale (RoSAS), which measures the robot's social attributes. The experimental results showed that CPG-based control methods improved the social attribute of being competent and comfortable with the robot. Based on the results of these studies, we propose guidelines for future research in the field of handshake robotics, which will contribute to the development of social robots in the future.
|
|
10:45-11:00, Paper TuAT1.3 | |
Identifying Socio-Emotional Features with a Mediator Robot (I) |
|
Cooper, Sara (Honda Research Institute Japan), Gomez, Randy (Honda Research Institute Japan Co., Ltd), Szapiro, Deborah (University of Technology Sydney), Merino, Luis (Universidad Pablo De Olavide) |
Keywords: Child-Robot Interaction, Affective Computing, Applications of Social Robots
Abstract: In this paper, we identify a set of socio-emotional cues and signals that are promoted by a tabletop social mediator robot in the context of a school setting. The robot adopts different roles to enhance such socio-emotional features, consequently aiding their identification by a structured annotation system. Various socio-emotional signals were observed for different robot roles, as well as different cues (gaze, speech). Future work will analyze cultural nuances as it expands the pilot to more schools worldwide.
|
|
11:00-11:15, Paper TuAT1.4 | |
Evaluation of Preference on Context-Aware Utterances Based on Personality Traits Using a Conversational Android Robot System (I) |
|
Fukui, Ryota (Tokyo University of Science), Yuguchi, Akishige (Tokyo University of Science), Matsumoto, Yoshio (AIST), Okadome, Yuya (Tokyo University of Science) |
Keywords: Linguistic Communication and Dialogue, Androids, Creating Human-Robot Relationships
Abstract: With recent technological improvements, the performance of conversational robots such as context-aware dialogue is expected to be dramatically enhanced. While context-aware utterance is considered one of the crucial functions of conversational robots, only a few studies investigate the relationship between context awareness and user preference. As is the case with human social relationships, human-robot interactions are under the effect of user's attributes such as personality, and thus considering these factors of users is essential to continue interacting with robots. In this paper, we evaluate the preference for context awareness based on personality trait effects in an open-domain dialogue using a conversational android robot system. Two types of utterances, context-aware and context-free, are implemented in the robot system, and the participants talk with and evaluate our conversational system. The preference for two systems and the personality traits of each participant are collected. We analyze the relationship between preferences and personalities. The results suggest that the participants' preference for whether an utterance is context-aware or context-free is distinguished by some personality traits.
|
|
TuAT2 |
Room T2 |
Virtual and Augmented Tele-Presence Environments I |
Regular Session |
Chair: Recchiuto, Carmine Tommaso | University of Genova |
|
10:15-10:30, Paper TuAT2.1 | |
Designing Augmented Reality Robot Guidance Interactions through the Metaphors of Re-Embodiment and Telepresence |
|
Zhu, Yifei (Colorado School of Mines), Brush, Colin (Colorado School of Mines), Williams, Tom (Colorado School of Mines) |
Keywords: Virtual and Augmented Tele-presence Environments, Assistive Robotics
Abstract: Robots deployed into real-world task-based environments may need to provide assistance, troubleshooting, and on-the-fly instruction for human users. While previous work has considered how robots can provide this assistance while co-located with human teammates, it is unclear how robots might best support users once they are no longer co-located. We propose the use of Augmented Reality as a medium for conveying long-distance task guidance from humans’ existing robot teammates, through Augmented Reality facilitated Robotic Guidance (ARRoG). Moreover, because there are multiple ways that a robot might project its identity through an Augmented Reality Head Mounted Display, we identify two candidate designs inspired by existing interaction patterns in the human-robot interaction (HRI) literature (re-embodiment-based and telepresence-based identity projection designs), present the results of a design workshop to explore how these designs might be most effectively implemented, and the results of a human-subject study intended to validate these designs
|
|
10:30-10:45, Paper TuAT2.2 | |
Approach of Team Avatar-Hubo to the ANA Avatar XPRIZE Finals |
|
KIM, BaekSeok (University of Nevada, Las Vegas), Kassai, Nathan (University of Nevada, Las Vegas), Castrejon, Zahir (University of Nevada Las Vegas), Kosanovic, Nicolas (University of Louisville), Chagas Vaz, Jean (University of Louisville), Oh, Paul Y. (University of Nevada, Las Vegas (UNLV)) |
Keywords: Virtual and Augmented Tele-presence Environments, Embodiment, Empathy and Intersubjectivity, Novel Interfaces and Interaction Modalities
Abstract: This paper presents Team Avatar-Hubo's Avatar system designed for telepresence, along with our strategy to fulfill the testing task requirements in the ANA Avatar XPRIZE Finals. Our system is built upon the foundation of DRC-Hubo and incorporates a VR Motion Tracker system. During the ANA Avatar XPRIZE Finals evaluation, our proposed Avatar system achieved the 11th Place Award among a pool of 17 Finalists. We provide comprehensive insights into our Avatar System, including its Hardware-Software Architecture, capabilities in Auditory-Vision, manipulation, Texture perception, locomotion, and Interface. Furthermore, we extensively discuss the key lessons we learned during our participation in the competition.
|
|
10:45-11:00, Paper TuAT2.3 | |
Advancing Behavior Generation in Mobile Robotics through High-Fidelity Procedural Simulations |
|
Kich, Victor Augusto (University of Tsukuba), Bottega, Jair Augusto (UFSM), Steinmetz, Raul (Universidade Federal De Santa Maria), Grando, Ricardo (Federal University of Rio Grande), Yorozu, Ayanori (University of Tsukuba), Ohya, Akihisa (University of Tsukuba) |
Keywords: Virtual and Augmented Tele-presence Environments, Motion Planning and Navigation in Human-Centered Environments, Machine Learning and Adaptation
Abstract: This paper introduces YamaS, a simulator integrating Unity3D Engine with Robotic Operating System for robot navigation research and aims to facilitate the development of both Deep Reinforcement Learning (Deep-RL) and Natural Language Processing (NLP). It supports single and multi-agent configurations with features like procedural environment generation, RGB vision, and dynamic obstacle navigation. Unique to YamaS is its ability to construct single and multi-agent environments, as well as generating agent's behaviour through textual descriptions. The simulator's fidelity is underscored by comparisons with the real-world Yamabiko Beego robot, demonstrating high accuracy in sensor simulations and spatial reasoning. Moreover, YamaS integrates Virtual Reality (VR) to augment Human-Robot Interaction (HRI) studies, providing an immersive platform for developers and researchers. This fusion establishes YamaS as a versatile and valuable tool for the development and testing of autonomous systems, contributing to the fields of robot simulation and AI-driven training methodologies.
|
|
11:00-11:15, Paper TuAT2.4 | |
Immersive Control of a Quadruped Robot with Virtual Reality Eye-Wear |
|
Yousefi, Ali (University of Genova), Betta, Zoe (University of Genova), Mottola, Giovanni (University of Genoa), Recchiuto, Carmine Tommaso (University of Genova), Sgorbissa, Antonio (University of Genova) |
Keywords: Virtual and Augmented Tele-presence Environments, Novel Interfaces and Interaction Modalities, Degrees of Autonomy and Teleoperation
Abstract: This work describes an immersive control system for a quadruped robot, designed to track the head movements of the operator wearing a virtual reality eye-wear, while also utilizing joystick commands for locomotion control. The article details the implemented closed-loop velocity control approach and the locomotion task specifications. The proposed method has been implemented on a Spot robot from Boston Dynamics, with Meta Quest 2 virtual reality system. Evaluation of the approach involved a user study, where participants engaged in immersive control of the quadruped robot within an indoor experimental environment and provided feedback through standardized questionnaires. Pairwise comparison of the resulting data revealed significant advantages for the proposed immersive control system over a standard remote controller, with enhanced performance observed in the second trial of using the control system. However, participants lacking experience with virtual reality systems reported increased distress symptoms following the experiment.
|
|
TuAT3 |
Room T3 |
Affective Computing |
Regular Session |
Chair: Zhao, Fangyun | University of Wisconsin-Madison |
|
10:15-10:30, Paper TuAT3.1 | |
Evaluation of Social Robots with Social Signals in Public Spaces |
|
Schiffmann, Michael (TH Köln University of Applied Sciences), Richert, Anja (University of Applied Sciences Cologne) |
Keywords: Affective Computing, Applications of Social Robots, Evaluation Methods
Abstract: Evaluating the users` subjective impression of social robots with questionnaires in field trials is time-consuming, and it is impossible to analyze every interaction as users cannot always participate for various reasons. In social robotics, social signals (e.g., gestures, facial expressions, body language) can be used to control and adapt the robot's behavior to improve interaction. Social signals should allow a statement to be made about how the user feels in a situation or what attitude or opinion they have about something. This paper focuses on finding out to what extent social signals can contribute to an automatic evaluation of the subjective user satisfaction without the need for user questionnaires and solely through social signals. For this purpose, it is first relevant whether social signals occur and if they can be used in the present use case. For this purpose, a field test was carried out in the entrance area of a city administration, in which the social robot Furhat was used as a point for service and information. Four interactions were recorded on video with the consent of the users, and a post-questionnaire on satisfaction was collected for each case. A qualitative video analysis was used to examine the interactions concerning potential social signals displayed by the users. The results suggest that the direction of view and the head orientation are the body parts that are mostly moved in the interaction. All users interacted with the system with relatively little movement of the rest of the body. The results suggest that it is not recommended to focus only on social signals alone.
|
|
10:30-10:45, Paper TuAT3.2 | |
Sigh!!! There Is More Than Just Faces and Verbal Speech to Recognize Emotion in Human-Robot Interaction |
|
Maharjan, Rahul Singh (The University of Manchester), Romeo, Marta (Heriot-Watt University), Cangelosi, Angelo (University of Manchester) |
Keywords: Affective Computing, Non-verbal Cues and Expressiveness, Machine Learning and Adaptation
Abstract: Understanding human emotions is paramount for effective human-human interactions. As technology advances, social robots are increasingly being developed with the capability to discern and respond to human emotions, with the ultimate aim of providing assistance and companionship. However, existing research on emotion recognition for human-robot interaction predominantly focuses on facial expressions or verbal speech, neglecting other potential mediums of emotional expression. In this study, we shed light on the significance of considering various forms of emotional expression, mainly non-verbal vocalization known as vocal bursts, which have been overlooked in the context of emotion modeling for human-robot interaction. Vocal bursts, characterized by brief and intense vocal utterances, represent a rich source of emotional cues that can significantly enhance the capabilities of social robots in understanding and responding to human emotions. Driven by the increasing interest in vocal bursts within speech and affective computing research, we propose a baseline model for affective vocal burst recognition that can outperform large audio models. The proposed baseline model achieves weighted F1 scores of 0.606, 0.342, and 0.287 on 10, 24, and 30 emotion classes, respectively. Additionally, we identify challenges that must be addressed to enhance affective vocal burst recognition for human-robot interaction. Code available at github.com/rahullabs/Sigh
|
|
10:45-11:00, Paper TuAT3.3 | |
The Feeling of Kawaii Toward a Robot's Head-Tilting Motion: Effects of Speed, Direction, and Accompanying Hand Gestures |
|
Shiomi, Masahiro (ATR), Kato, Yuina (Nara Women's University), Nittono, Hiroshi (Osaka University), Anzai, Emi (Nara Women's University), Saiwaki, Naoki (Nara Women's University) |
Keywords: Affective Computing, Non-verbal Cues and Expressiveness, Robot Companions and Social Robots
Abstract: Head-tilting motions are social signals that play a key role in communication. In human-robot interaction, researchers have found that this motion increases the observer's feelings of kawaii toward a robot. Kawaii is a Japanese word meaning cute and adorable, and the feeling of kawaii refers to such affectionate, tender feelings toward an object. The feeling of kawaii has proven to be crucial for designing socially acceptable robots. However, the impact of the head-tilting speed of social robots on the observer's kawaii feelings remains unclear because previous research mainly used static images. This study determined the optimal head-tilting speed and direction to deepen the observer's feeling of kawaii toward a robot. Our findings found that participants reported feeling more kawaii toward a robot when it tilted its head to its right within 0.5 or 1 second. Our results also showed that gestures associated with head tilting are critical for robot behavior design.
|
|
11:00-11:15, Paper TuAT3.4 | |
Analysis of Heart-To-Heart Communication with Robot Using Transfer Entropy |
|
Sato, Moe (Osaka University), Minato, Takashi (RIKEN), Funayama, Tomo (Osaka University), Sumioka, Hidenobu (ATR), Sakai, Kurima (ATR), Mikata, Ryusuke (ATR), Ishiguro, Hiroshi (Osaka University), Akane, Kikuchi (KiQ Co., Ltd.,), Kaito, Sakuma (Apples and Oranges Co., Ltd.,), Kazuya, Horibe (Osaka University) |
Keywords: Androids, Affective Computing, Creating Human-Robot Relationships
Abstract: Human robot interaction studies have investigated how various non-verbal expressions of robots can enhance feelings of familiarity and trust in users' and establish good relationships between users and robots. In the field of art, this approach has been taken one step further by creating robot artworks and demonstrations through which people feel as if they are experiencing heart-to-heart communication with a robot. Although robotic behavior that conveys such a feeling is useful for human coexistence and providing a sense of security and trust, no engineering methodology can yet achieve such behavior. This study attempts to explain what kind of robot behavior is connected to such feelings by analyzing demonstrations of such a robot based on information theory. We found that the intensity of these feelings can be partially explained by transfer entropy between humans and the robot's body movements. We expect this research to clarify how to design a robot's behavior so that it can provide heart-to-heart communication with people as well as how to construct a robot that can build solid relationships with people in daily life.
|
|
TuAT4 |
Room T4 |
Human Factors and Ergonomics I |
Regular Session |
Chair: Liu, Rui | Kent State University |
Co-Chair: Arevalo Arboleda, Stephanie | Technische Universität Ilmenau |
|
10:15-10:30, Paper TuAT4.1 | |
An Exploratory Study on the Impact of Varying Levels of Robot Control on Presence in Robot-Mediated Communication |
|
Arevalo Arboleda, Stephanie (Technische Universität Ilmenau), Fischedick, Söhnke Benedikt (Ilmenau University of Technology), Diao, Chenyao (TU-Ilmenau), Richter, Kay (Ilmenau University of Technology), Gross, Horst-Michael (Ilmenau University of Technology), Raake, Alexander (Ilmenau University of Technology, Audiovisual Technology Group) |
Keywords: Social Presence for Robots and Virtual Humans, Storytelling in HRI, Human Factors and Ergonomics
Abstract: Telepresence robots can enhance communication experiences by providing a sense of physical presence, embodiment and may evoke co-presence. In spite of that, telepresence robots have not made it fully to consumer markets. In this paper, we investigate how different levels of controlling a telepresence robot (teleoperation, shared control, and no control) influence presence. To this aim, we conducted a study (N=45) where participants were evenly distributed to one of the robot control conditions. The task involved navigating an unknown room and listening to stories told by a person co-located with the robot. We collected subjective impressions of presence using the temple presence inventory and performed a thematic content analysis on a post-experiment interview. Our results suggest nuances in perceived presence under different levels of robot control after performing a thematic content analysis. Co-presence can be experienced during teleoperation and shared control, and teleoperation may evoke negative sentiments if it does not provide enough spatial information during navigation. However, our results did not point to significant differences in spatial or social presence. We consider that these findings encourage further discussions on how presence is perceived in robot-mediated communication.
|
|
10:30-10:45, Paper TuAT4.2 | |
All Too White? Effects of Anthropomorphism on the Stereotypical Perception of Robot Color |
|
Barenbrock, Julia (University of Applied Sciences Ruhr West), Eimler, Sabrina C. (Hochschule Ruhr West, University of Applied Sciences), Helgert, André (University of Applied Sciences Ruhr West), Straßmann, Carolin (University of Applied Sciences Ruhr West) |
Keywords: Human Factors and Ergonomics, Anthropomorphic Robots and Virtual Humans, Creating Human-Robot Relationships
Abstract: Looking at state-of-the-art robots, the majority have a white surface. This raises the questions of why robots are designed predominantly white, and how surface color affects the robot’s social perception. This not only reinforces human stereotypes but also influences the acceptance and usage of robots. Accordingly, this online study with a 2 (black vs. white) x 3 (Pepper robot vs. Temi robot vs. Kuka robot) within-subjects design (N = 100) investigated the effect on racist stereotypical perception and the role of anthropomorphism using implicit (Implicit Association Test) and explicit (Robotic Social Attributes Scale) measures and questionnaires on behavioral intentions. Our results show that robot color influences the implicit perception of robots but rarely affects explicit perception and behavior measures. Moreover, no effects of the robots' anthropomorphism on racist perceptions were found. As more effects of anthropomorphism on our outcome variables were found than for robot color, we assume that robot color activates stereotypes in terms of color rather than race.
|
|
10:45-11:00, Paper TuAT4.3 | |
Large Language Model Driven Interactive Learning for Real-Time Cognitive Load Prediction in Human-Swarm Systems |
|
Zang, Wenshuo (Kent State University), Hu, Mengsha (Kent State University), Liu, Rui (Kent State University) |
Keywords: Human Factors and Ergonomics, Cognitive and Sensorimotor Development
Abstract: The rapid advancements of drones have demonstrated the versatility and promising potential of human-swarm systems (HSS) across various domains. However, human performance within these systems may be impaired by factors such as limited domain knowledge and mental stress, often leading to cognitive overload and hindering the efficiency and effectiveness of human-swarm teaming. Consequently, the accurate monitoring of cognitive load levels is crucial for optimizing HSS performance. To address the challenges of existing measurement methods, which are often expensive, time-consuming, or lack real-time capabilities, we propose a Large Language Model driven cognitive load prediction framework. This framework integrates comprehensive task context, domain knowledge, and behavior analysis to provide fast and cost-effective predictions in complex scenarios. By leveraging the capabilities of Large Language Models and employing reinforcement learning to model the cognitive load generation, our framework aims to offer real-time insights into human-related factors causing high cognitive load and predict cognitive levels over time, ultimately enhancing the performance of HSS teaming.
|
|
11:00-11:15, Paper TuAT4.4 | |
Improving Transparency in Human-Collective Visualizations |
|
Bhagat Smith, Joshua (Oregon State University), Baskaran, Prakash (Oregon State University), Adams, Julie (Oregon State University) |
Keywords: Human Factors and Ergonomics, Cooperation and Collaboration in Human-Robot Teams
Abstract: Quantifying transparency requires evaluating the transparency embedded in the various system design elements to determine how they impact one another and influence human-collective interactions. Prior work demonstrated limitations of an abstract collective interaction. Interface designs to address these limitations and improve human-collective interaction transparency were evaluated for a sequential best-of-N decision-making task with four collectives, each consisting of 200 individual entities. The Informed and Simple visualizations’ predictive progress bars improved transparency and the overall human-collective team performance.
|
|
TuAT5 |
Room T5 |
Machine Learning and Adaptation I |
Regular Session |
Co-Chair: Nunes, Urbano J. | Instituto De Sistemas E Robotica |
|
10:15-10:30, Paper TuAT5.1 | |
Deep Learning-Based Adaptation of Robot Behaviour for Assistive Robotics |
|
Stolarz, Michał Mikołaj (Hochschule Bonn-Rhein-Sieg, Sankt Augustin), Romeo, Marta (Heriot-Watt University), Mitrevski, Alex (Hochschule Bonn-Rhein-Sieg), Plöger, Paul G. (Hochschule Bonn Rhein Sieg) |
Keywords: Machine Learning and Adaptation, Assistive Robotics, Robots in Education, Therapy and Rehabilitation
Abstract: Robot behaviour models in socially assistive robotics are typically trained using high-level features, such as a user's engagement, such that inaccuracies in the feature extraction can have a significant effect on a robot's subsequent performance. In this paper, we study whether a behaviour model can be meaningfully represented using an end-to-end approach, where multimodal input, concretely visual data and activity information, is directly processed by a neural network. This paper concretely analyses the different building blocks of such a model, such that the aim is to identify a suitable architecture that can meaningfully combine the different modalities for guiding a robot's behaviour. We conduct the analysis in the context of a sequence learning game, such that we compare different vision-only models that are then combined with an activity processing network into a joint multimodal model. The results of our evaluation on a dedicated dataset from the sequence learning game demonstrate that a multimodal end-to-end behaviour model has potential for assistive robotics — we report an F1 score of around 0.88 across different dataset-based test scenarios — but the real-life transferability strongly depends on whether the data is diverse enough for capturing meaningful variations in real-world scenarios, such as users being at different distances from a robot.
|
|
10:30-10:45, Paper TuAT5.2 | |
Two-Stream Architecture with Contrastive and Self-Supervised Attention Feature Fusion for Error-Related Potentials Classification |
|
Garrote, Luís Carlos (Institute of Systems and Robotics, University of Coimbra), Perdiz, João (University of Coimbra), Yasemin, Mine (ISR-UC), Pires, Gabriel (University of Coimbra), Nunes, Urbano J. (Instituto De Sistemas E Robotica) |
Keywords: Machine Learning and Adaptation, Assistive Robotics
Abstract: Error-related potentials (ErrPs) extracted from electroencephalographic signals hold potential for application in Brain-Machine Interfaces, in contexts such as robot teleoperation or shared control in assistive platforms. Due to difficulties in signal classification, in part caused by its non-stationary and noisy nature, their use has not been fully realized yet. This work proposes a new approach to ErrP classification based on a two-stream deep learning architecture with three training stages. Its first stage is a self-supervised autoencoder architecture with a multi-head attention layer providing relevant latent features. The second stage comprises a supervised contrastive learning approach considering two backbone networks, where one inherits weights from the first stage and the other is updated by considering the feature embeddings distribution. The final stage comprises supervised classification, where the two backbones are fused and used to classify the input EEG signal. At the end of the three stages, a data-driven two-stream ErrP model is obtained. Twenty-five variants of the proposed approach using the Deep Convolutional Network, Shallow Convolutional Network and EEGNet backbones were tested in an ablation study and benchmarked against a large number of classical classification methods, using data from the BNCI dataset intended to assess cross subject generalization capabilities. The proposed approach obtained the best results overall, highlighting the approach's capabilities in capturing relevant representations of the EEG signal.
|
|
10:45-11:00, Paper TuAT5.3 | |
Discrimination and Prediction of Soft Surfaces for Cloth Classification |
|
Duran Jimenez, Raul Ariel (Kyutech), Suppaadirek, Natchanon (Kyushu Institute of Technology), Shibata, Tomohiro (Kyushu Institute of Technology) |
Keywords: Machine Learning and Adaptation, Assistive Robotics
Abstract: The sense of touch plays an indispensable role in human perception by enabling the discernment of various objects based on their textures, roughness, and shapes. In the realm of robotics, touch sensors play a vital role in the manipulation and control of objects. By providing information on mechanical properties, these sensors facilitate the visualization of grip, shear, and pressure force at contact points. However, handling deformable objects such as clothing poses a significant challenge. While some research relies on visual sensors for the detection and classification of clothing types, this research focuses on leveraging tactile sensors for the discrimination and classification of deformable surfaces in real-time. The aim is centered on soft surfaces, with a specific focus on examining seven different types of materials. The system we present focuses on signal processing and machine learning techniques for cloth recognition, and classification in real time. The system integrates a soft touch sensor fingertip into a robotic arm to achieve these objectives. The current study has developed a classification model leveraging Long Short-Term Memory (LSTM) networks. The objective of this model is to accurately classify seven distinct types of clothing by analyzing their texture features.
|
|
11:00-11:15, Paper TuAT5.4 | |
GraspPC: Generating Diverse Hand Grasp Point Clouds on Objects |
|
Megyeri, Ava (Wright State University), Wiederhold, Noah (Clarkson University), Kyrarini, Maria (Santa Clara University), Banerjee, Sean (Wright State Univeristy), Banerjee, Natasha Kholgade (Wright State University) |
Keywords: Machine Learning and Adaptation, Detecting and Understanding Human Activity, Cooperation and Collaboration in Human-Robot Teams
Abstract: We present GraspPC, an approach to perform learning-based synthesis of multiple human hand grasps as point clouds from point clouds of objects. GraspPC benefits human-robot handover approaches by providing hypotheses of human grasp on objects to inform robotic manipulation algorithms on how to bias robotic grasp for safe handover. Existing learning-based approaches to conduct hand grasp prediction require datasets to contain annotated articulated hand models, making them difficult to train on datasets that lack hand model annotations. GraspPC treats the problem of hand point cloud generation from object point clouds as a set-to-set translation problem. We contribute a Transformer architecture to synthesize point clouds via GraspPC. To generate diverse hand grasps, we generate multiple object-dependent queries and train the network using a winner-takes-gradient strategy. We show results of diverse grasps by training and testing on a variety of real-world datasets. We demonstrate how human grasps generated by GraspPC can be used to filter robotic grasp candidates to inform human-robot handover. Our code is available at [link on acceptance].
|
|
TuAT6 |
Room T6 |
Motion Planning and Navigation in Human-Centered Environments I |
Regular Session |
Co-Chair: Oursland, Jacob | South Dakota School of Mines and Technology |
|
10:15-10:30, Paper TuAT6.1 | |
Why Did My Robot Choose This Path? Explainable Path Planning for Off-Road Navigation |
|
Eder, Matthias (Graz University of Technology), Steinbauer-Wagner, Gerald (Graz University of Technology) |
Keywords: Motion Planning and Navigation in Human-Centered Environments, Cooperation and Collaboration in Human-Robot Teams
Abstract: In the field of off-road navigation, where traditional maps often fall short, intuitive and efficient path planning is essential for autonomous off-road vehicles. Navigating in off-road terrain poses unique challenges, requiring innovative solutions for users to understand and trust path suggestions made by an autonomous system. In this paper, we explore the integration of Explainable AI into off-road navigation systems to better understand the complexity of off-road environments. Our research introduces a method tailored to generate contextual explanations for chosen paths using terrain features, environmental factors, and robot capabilities. By combining inverse optimization techniques with shortest path algorithms, our approach aims to answer the question "Why is path p* recommended over path p', which was expected by the user?" These explanations aim to shed light on the process of a robot's path planning task, focusing on elevation changes, terrain obstacles, and optimal path choices, thus improving the user's understanding of the chosen paths. A short user study evaluating the provided explanations generated in different off-road environments validates the effectiveness of our explanation algorithm and shows that it contributes to understanding the planning process of off-road navigation systems.
|
|
10:30-10:45, Paper TuAT6.2 | |
But Where Are You Going?! Motion Is What Is Most Important for Real-World Co-Present Mobile Robots |
|
Fraune, Marlena (New Mexico State University), Knepper, Ross (--) |
Keywords: Motion Planning and Navigation in Human-Centered Environments, Cooperation and Collaboration in Human-Robot Teams
Abstract: Mobile robots are being introduced to industrial workplaces in roles that require copresence with humans. To develop effective robots that do not negatively impact humans, including their subjective experience and ability to get their work done, we must understand humans’ needs for working near these robots. In this paper, we examine what human workers need from copresent robots’ motion during work at a large warehouse. To do so, report and synthesize findings about robot motion from across five studies (e.g., focus group, observation, experiment). Results indicate that workers were most focused on robot movement, including consistency, distance, prioritizing people, and indicating when it sensed people. Researchers and practitioners can use these findings to prioritize what aspects of mobile robots to develop to improve human worker experiences around robots and team efficiency.
|
|
10:45-11:00, Paper TuAT6.3 | |
An Invariant Extended Kalman Filter for IMU-UWB Sensor Fusion |
|
Oursland, Jacob (South Dakota School of Mines and Technology), Mehrabian, Mohammadreza (South Dakota School of Mines and Technology) |
Keywords: Motion Planning and Navigation in Human-Centered Environments, Degrees of Autonomy and Teleoperation
Abstract: Orientation estimation is crucial for the successful operation of robots in autonomous control, enabling effective navigation, environmental interaction, and precise task execution. Obtaining a low-error orientation estimate promptly is essential for ensuring safe and successful autonomous robot operation. However, existing approaches, such as the Error-State Kalman Filter (ESKF), suffer from slow convergence, leading to unstable flights and increased collision risks, particularly when initial heading alignment errors occur. We introduce an Invariant Extended Kalman Filter (InEKF) for inertial measurement unit (IMU) and ultra-wideband (UWB) sensor fusion that addresses the slow convergence issue in orientation estimation of the Crazyflie Bolt drone. We demonstrate significant improvements over the reference ESKF, achieving a RMS heading error of 0.08 radians (4.6 degrees) compared to the ESKF's 1.34 radians (77 degrees), a 94% reduction in error, when starting from a 180 degree initial heading error, while maintaining similar accuracy levels in position and velocity estimates.
|
|
11:00-11:15, Paper TuAT6.4 | |
Fusion of Inertial Sensor Suit and Monocular Camera for 3D Human Pelvis Pose Estimation |
|
Popescu, Mihaela (University of Bremen), Shinde, Kashmira (German Aerospace Center (DLR)), Sharma, Proneet (DFKI GmbH), Gutzeit, Lisa (University of Bremen), Kirchner, Frank (University of Bremen) |
Keywords: Motion Planning and Navigation in Human-Centered Environments, Detecting and Understanding Human Activity, Multi-modal Situation Awareness and Spatial Cognition
Abstract: In real-world scenarios, robots come closer to humans in many applications, sharing the same workspace or even manipulating the same objects. To ensure safe and intuitive collaboration, it is crucial to have an accurate knowledge of the human's 3D position in space, which should be estimated with high precision, high frequency and low latency. However, individual sensors such as inertial measurement units (IMUs) or cameras cannot meet all requirements for reliable human pose estimation under conditions such as long operating times, large distances and occlusions. In this study, we highlight the limitations of different visual pose methods and present a fused approach for real-time estimation of the 3D position of the human pelvis using machine learning-based visual pose from a monocular camera and an IMU sensor suit. The multimodal fusion is based on the Invariant Extended Kalman filter (InEKF) on Lie Groups, which fuses drift-free visual poses with high-frequency inertial measurements in a loosely-coupled manner. The evaluation is performed on a recorded dataset of multiple subjects performing various experimental scenarios. The results show that the fused approach can increase the accuracy and robustness of the estimates, taking a step closer towards smooth human-robot collaboration.
|
|
TuBT1 |
Room T1 |
Personalising Robots Behaviour in Social Human-Robot Interactions II |
Special Session |
Chair: Rossi, Silvia | Universita' Di Napoli Federico II |
Co-Chair: Andriella, Antonio | Artificial Intelligence Research Institute, CSIC |
|
11:30-11:45, Paper TuBT1.1 | |
Exploring the Potential of a Robot-Assisted Frailty Assessment System for Elderly Care (I) |
|
Civit, Aniol (Universitat Politècnica De Catalunya), Andriella, Antonio (Artificial Intelligence Research Institute, CSIC), Antonio, Maite (Institut Català D'Oncologia), Javierre, Casimiro (Universitat De Barcelona), Boqué, Concepción (Institut Català D'Oncologia), Alenyà, Guillem (Institut De Robòtica I Informàtica Industrial CSIC-UPC) |
Keywords: Applications of Social Robots, Medical and Surgical Applications, Assistive Robotics
Abstract: Frailty assessment plays a pivotal role in providing older adults care. However, the current process is time-consuming and only measures patients' completion time for each test. This paper introduces a set of algorithms to be used in robots to autonomously perform frailty assessments. In doing so we aim at reducing therapists' burden and provide additional frailty-related metrics that can enhance the effectiveness of diagnosis. We conducted a pilot study with 22 elderly participants and compared our system's performance with that of medical professionals to assess its precision. The results demonstrate that our approach achieved performances close to that of its human counterpart. This research represents an important step forward in the integration of social robotics in healthcare, offering potential benefits for patient care and clinical decision-making.
|
|
11:45-12:00, Paper TuBT1.2 | |
Are Emotions Important? a Study on Social Distances for Path Planning Based on Emotions (I) |
|
Mizaridis, Vasileios (University of Naples Federico II), Vigni, Francesco (Interdepartmental Center for Advances in Robotic Surgery - ICARO), Arampatzis, Efstratios (Noosware B.V), Rossi, Silvia (Universita' Di Napoli Federico II) |
Keywords: Non-verbal Cues and Expressiveness, Social Intelligence for Robots, User-centered Design of Robots
Abstract: This study explores the complex dynamics between humans and robots, focusing on how emotional states influence proxemics. We conducted a user study using a standard mobile robot to investigate whether emotions elicited from a loudspeaker, affect human perception of robot proximity. Based on previous research on Human-Human Interaction (HHI), we analysed participants’ responses to robots displaying different behaviours. Participants observed the robot’s approach while experiencing positive or negative emotions. Our findings suggest that emotional states induced by external stimuli can affect participants’ perception of robot proximity. In detail, the results indicate that while comfortable stopping distances were unaffected by participants’ emotional state, individuals who experienced positive emotions judged the same proxemics distance used while performing an avoidance behaviour to be more acceptable compared to the case of negative emotions. This study describes the extent to which our emotions can alter the perception of robot behaviours, ultimately affecting our acceptance of these novel social agents.
|
|
12:00-12:15, Paper TuBT1.3 | |
Too Close to You? a Study on Emotion-Adapted Proxemics Behaviours (I) |
|
Vigni, Francesco (Interdepartmental Center for Advances in Robotic Surgery - ICARO), Maglietta, Dimitri (University of Naples Federico II), Rossi, Silvia (Universita' Di Napoli Federico II) |
Keywords: Non-verbal Cues and Expressiveness, Social Intelligence for Robots, Motivations and Emotions in Robotics
Abstract: The development of social robots with advanced conversational capabilities aiming to engage humans in natural interactions has recently surged. This paper investigates the dynamic aspect of Human-Robot Interaction (HRI), focusing on the regulation of interpersonal distance based on human emotion. Through a user study with a humanoid robot, we explore how participants perceive and respond to rule-based versus randomly generated robot behaviours in adjusting interpersonal space during an unconstrained conversation. Results suggest that participants perceive the robot using rule-based behaviours as more socially competent and adaptable to human behaviour and emotions compared to the random ones. These findings highlight the importance of considering subtle nonverbal cues and adapting robot behaviour based on human emotions to improve the quality of HRI, and consequently facilitate the successful integration of human natural nuances in robots.
|
|
12:15-12:30, Paper TuBT1.4 | |
LEMMA: Learning Language-Conditioned Multi-Robot Manipulation |
|
Gong, Ran (UCLA), Gao, Xiaofeng (Amazon), Gao, Qiaozi (Amazon), M Shakiah, Suhaila (Amazon), Thattai, Govind (Amazon), Sukhatme, Gaurav (University of Southern California) |
|
TuBT2 |
Room T2 |
Virtual and Augmented Tele-Presence Environments II |
Regular Session |
Chair: Ravankar, Ankit A. | Tohoku University |
Co-Chair: Biswas, PRADIPTA | Indian Institute of Science |
|
11:30-11:45, Paper TuBT2.1 | |
A B-Spline Approach for Improved Environmental Awareness in Virtual Walking System Using Avatar Robot |
|
Miuccio, Alessandra (Politecnico Di Milano), Manríquez-Cisterna, Ricardo (Tohoku University), Ravankar, Ankit A. (Tohoku University), Salazar Luces, Jose Victorio (Tohoku University), Hirata, Yasuhisa (Tohoku University), Rocco, Paolo (Politecnico Di Milano) |
Keywords: Virtual and Augmented Tele-presence Environments, Novel Interfaces and Interaction Modalities, Degrees of Autonomy and Teleoperation
Abstract: Recent advancements in medical science and technology have led to a remarkable increase in the lifespan of the global elderly population. However, this demographic often struggles with mobility issues, leading to a sedentary lifestyle fueled by concerns over their physical capabilities. Traditional treadmill gait training, although beneficial, often becomes monotonous and lacks the real-world feedback necessary for engaging and effective rehabilitation. Addressing this gap, research into virtual walking systems utilizing avatar robots has gained traction. Despite the progress, several challenges remain where the systems prioritize visual feedback without considering the crucial need for alerting users to potential dangers and obstacles. This lack of comprehensive environmental awareness and feedback undermines both user engagement and safety. To address this problem, this paper proposes an algorithm that employs B-splines for precise free space detection, integrated with a safety stop mechanism for an avatar robot. This novel approach enhances user awareness of their surroundings through a sophisticated Graphic User Interface (GUI) that leverages Augmented Reality (AR) technology. By superimposing free space boundaries and warning messages directly onto a real-time camera feed, the system provides an intuitive and immersive navigation aid. The efficacy of our proposed GUI was rigorously tested across a series of realistic scenarios, comparing teleoperation control performance with and without the augmented interface. Our findings reveal that our GUI markedly enhances user safety and navigational effectiveness, fostering a deeper understanding of and interaction with the surrounding environment, thereby redefining user experience in virtual mobility assistance.
|
|
11:45-12:00, Paper TuBT2.2 | |
UnRealTHASC – a Cyber-Physical XR Testbed for Underwater Real-Time Human Autonomous Systems Collaboration |
|
Surve, Sushrut (Cornell University), Guo, Jia (Cornell University), Menezes, Jovan (Cornell University), Tate, Connor (University of West Florida, Institute for Human and Machine Cogn), Jin, Yiting (Cornell University), Walker, Justin (Cornell University), Ferrari, Silvia (Cornell University) |
Keywords: Virtual and Augmented Tele-presence Environments, Novel Interfaces and Interaction Modalities, Monitoring of Behaviour and Internal States of Humans
Abstract: Research on underwater human-robot teaming is particularly promising because of complementary sensory skills that may overcome unique challenges associated with the combination of limited connectivity and low visibility. Nevertheless, testing human-robot collaboration under water, especially in complex, real-world scenarios, poses severe safety, cost, and time constraints that significantly hinder academic research in this space. This paper presents a novel cyber-physical extended-reality (XR) testbed, referred to as Underwater Real-Time Human Autonomous Systems Collaboration (UnRealTHASC), designed to enable human-robot interactions in simulated photorealistic environments with mixed real and virtual wearables, and advanced autonomous underwater sensors in-the-loop. Novel sensor interfaces are designed to integrate real and virtual sensors for measuring physiological and cognitive human states underpinning decision-making abilities. Physics-based human and robot motion models are developed along with new sensor simulations in order to capture the couplings between underwater behaviors and perception based on measurements from optical and sonar sensors. Real-time data acquisition pipelines are created to access and share data from both real and virtual sensors and robots, such that new methods for online planning and collaboration may be tested via human-in-the-loop demonstrations.
|
|
12:00-12:15, Paper TuBT2.3 | |
Human(s) on the Loop Demand Aware Robot Scheduling: A Mixed Reality Based User Study |
|
SANDULA, AJAY KUMAR (Indian Institute of Science, Bengaluru), M, Rajatsurya (Indian Institute of Science, Bangalore), Ghose, Debasish (Indian Institute of Science), Biswas, PRADIPTA (Indian Institute of Science) |
Keywords: Virtual and Augmented Tele-presence Environments, User-centered Design of Robots, Novel Interfaces and Interaction Modalities
Abstract: Scheduling tasks and subtasks for multiple heterogeneous robots poses a significant challenge, particularly in scenarios where human supervision is essential. The complexity is further compounded when considering human factors in the scheduling process, especially in demand-aware task scheduling environments(tasks are generated based on external demand). The primary goal of this research is to understand human decision-making and how it affects their scheduling process, and we undertook two studies. In the first study, we employed a mixed reality based user study to explore how human perception of the scheduling environment influences task scheduling and facilitates personalized resource allocation. Our findings indicate that human task schedulers exhibit enhanced performance when assisted by autonomous agents, compared to scenarios with limited autonomy in robotic systems. To explore the impact of robot planning on human decision-making and its effects on task scheduling and collision-aware scheduling capabilities, we conducted the second study. This study employed a mixed reality-based warehouse environment, where two users controlled different robots with shared objectives. The findings suggest that human operators exhibited improved collision-aware scheduling without compromising their demand-aware scheduling capabilities when visual aids such as collision cones were incorporated.
|
|
12:15-12:30, Paper TuBT2.4 | |
Immersive Teleoperation of Collaborative Robot for Remote Vital Monitoring in Healthcare |
|
Vishwakarma, Himanshu (Indian Institute of Science), Sampath, Ananyan (All India Institute of Medical Sciences, Bhopal), Vinay Krishna Sharma, Vinay Krishna (Indian Institute of Science), Biswas, PRADIPTA (Indian Institute of Science) |
Keywords: Virtual and Augmented Tele-presence Environments, User-centered Design of Robots, Novel Interfaces and Interaction Modalities
Abstract: Healthcare robotics has emerged as an innovative system with vast potential across various domains. This paper explores the development and application of collaborative robots for teleoperation, which facilitate safe and remote medical general examinations of patients. The study aims to help medical staff to provide vital sign check-ups remotely using legacy medical devices. We proposed an immersive media tool that was integrated with augmented reality (AR) and virtual reality (VR) interfaces to enhance the reach of medical professionals for vital signs check-ups. The medical cobot was teleoperated through AR and VR interfaces. A qualitative user study was used to compare the AR and VR interface systems within the loop of healthcare professionals.
|
|
TuBT3 |
Room T3 |
Social Touch in Human–Robot Interaction |
Regular Session |
Chair: Maurelli, Francesco | Constructor University |
Co-Chair: Nikolovska, Kristina | Constructor University |
|
11:30-11:45, Paper TuBT3.1 | |
Learning Human-Robot Handshaking Preferences for Quadruped Robots |
|
Chappuis, Alessandra (Ecole Polytechnique Fédérale De Lausanne), Bellegarda, Guillaume (EPFL), Ijspeert, Auke (EPFL) |
Keywords: Social Touch in Human–Robot Interaction, Creating Human-Robot Relationships
Abstract: Quadruped robots are showing impressive abilities to navigate the real world. If they are to become more integrated into society, social trust in interactions with humans will become increasingly important. Additionally, robots will need to be adaptable to different humans based on individual preferences. In this work, we study the social interaction task of learning optimal handshakes for quadruped robots based on user preferences. While maintaining balance on three legs, we parameterize handshakes with a Central Pattern Generator consisting of an amplitude, frequency, stiffness, and duration. Through 10 binary choices between handshakes, we learn a belief model to fit individual preferences for 25 different subjects. Our results show that this is an effective strategy, with 76% of users feeling happy with their identified optimal handshake parameters, and 20% feeling neutral. Moreover, compared with random and test handshakes, the optimized handshakes have significantly decreased errors in amplitude and frequency, lower Dynamic Time Warping scores, and improved energy efficiency, all of which indicate robot synchronization to the user's preferences
|
|
11:45-12:00, Paper TuBT3.2 | |
Unified Understanding of Environment, Task, and Human for Human-Robot Interaction in Real-World Environments |
|
Yano, Yuga (Kyushu Institute of Technology), Mizutani, Akinobu (Kyushu Institute of Technology), Fukuda, Yukiya (Kyushu Institute of Technology), Kanaoka, Daiju (Kyushu Institute of Technology), Ono, Tomohiro (Toyota Motor Corporation), Tamukoh, Hakaru (Kyushu Institute of Technology) |
Keywords: Social Touch in Human–Robot Interaction, Motion Planning and Navigation in Human-Centered Environments, Social Intelligence for Robots
Abstract: To facilitate human-robot interaction (HRI) tasks in real-world scenarios, service robots must adapt to dynamic environments and understand the required tasks while effectively communicating with humans. To accomplish HRI in practice, we propose a novel indoor dynamic map, task understanding system, and response generation system. The indoor dynamic map optimizes robot behavior by managing an occupancy grid map and dynamic information, such as furniture and humans, in separate layers. The task understanding system targets tasks that require multiple actions, such as serving ordered items. Task representations that predefine the flow of necessary actions are applied to achieve highly accurate understanding. The response generation system is executed in parallel with task understanding to facilitate smooth HRI by informing humans of the subsequent actions of the robot. In this study, we focused on waiter duties in a restaurant setting as a representative application of HRI in a dynamic environment. We developed an HRI system that could perform tasks such as serving food and cleaning up while communicating with customers. In experiments conducted in a simulated restaurant environment, the proposed HRI system successfully communicated with customers and served ordered food with 90% accuracy. In a questionnaire administered after the experiment, the HRI system of the robot received 4.2 points out of 5. These outcomes indicated the effectiveness of the proposed method and HRI system in executing waiter tasks in real-world environments.
|
|
12:00-12:15, Paper TuBT3.3 | |
Soothing Sensations: Enhancing Interactions with a Socially Assistive Robot through Vibrotactile Heartbeats |
|
Borgstedt, Jacqueline (University of Glasgow), Macdonald, Shaun (University of Glasgow), Marky, Karola (Ruhr-Universität Bochum), Pollick, Frank (University of Glasgow), Brewster, Stephen (University of Glasgow) |
Keywords: Social Touch in Human–Robot Interaction, Robot Companions and Social Robots, Novel Interfaces and Interaction Modalities
Abstract: Physical interactions with socially assistive robots (SARs) positively affect user wellbeing. However, haptic experiences when touching a SAR are typically limited to perceiving the robot’s movements or shell texture, while other modalities that could enhance the touch experience with the robot, such as vibrotactile stimulation, are under-explored. In this exploratory qualitative study, we investigate the potential of enhancing human interaction with the PARO robot through vibrotactile heartbeats, with the goal to regulate subjective wellbeing during stressful situations. We conducted in-depth one-on-one interviews with 30~participants, who watched three horror movie clips alone, with PARO, and with a PARO that displayed a vibrotactile heartbeat. Our findings show that PARO’s presence and its interactive capabilities can help users regulate emotions through attentional redeployment from a stressor toward the robot. The vibrotactile heartbeat further reinforced PARO’s physical and social presence, enhancing the socio-emotional support provided by the robot and its perceived life-likeness. We discuss the impact of individual differences in user experience and implications for the future design of life-like vibrotactile stimulation for SARs.
|
|
12:15-12:30, Paper TuBT3.4 | |
The Impact of Social Inter-Robot Encounters on User Perception |
|
Nikolovska, Kristina (Constructor University), Pohl, Jan (Technische Universität Dresden), Hommel, Bernhard (Shandong Normal University, Jinan), Kappas, Arvid (Constructor University), Maurelli, Francesco (Constructor University) |
Keywords: Social Touch in Human–Robot Interaction, Social Intelligence for Robots, Creating Human-Robot Relationships
Abstract: In recent years, there has been a growing trend of integrating robots into various dimensions of daily life, to improve user experiences and offer a range of services. This study explores the interactions of robots showing social behavior towards other robots and their impact on human perceptions. We focus on three types of social behavior: social sensitivity, attention-sharing, and helping, aiming to understand how these interactions affect human perceptions of robots. Utilizing Duckiebot mobile robots in carefully crafted experimental setups, participants observed video recordings of these robots' interactions, which either included or excluded each targeted social behavior. The study measured user responses using established scales such as the Mind Attribution Scale (MAS), the Goodspeed Scale, and the Robotic Social Attributes Scale (RoSAS). The results demonstrated that robots displaying social behavior towards other robots were perceived more positively compared to those that did not exhibit such behavior. Specifically, social sensitivity positively impacted animacy, experience, likability, perceived intelligence, safety, and warmth. Attention-sharing improved perceptions of competence, experience, likability, perceived intelligence, and warmth. Additionally, helping behavior positively affected agency, animacy, anthropomorphism, competence, experience, likability, perceived intelligence, safety, and warmth. This research contributes valuable insights into Human-Robot Interaction (HRI), highlighting the significant impact of robots' interactions with each other on user experiences and perceptions. The exploration of social behavior lays a foundation for designing robots that evoke positive responses, fostering smoother integration of robotic technology into various
|
|
TuBT4 |
Room T4 |
Human Factors and Ergonomics II |
Regular Session |
Co-Chair: tong, yanzhang | Cardiff University |
|
11:30-11:45, Paper TuBT4.1 | |
Evaluating Human-Robot Interaction User Experiences in Manufacturing: An Initial Assessment Framework |
|
tong, yanzhang (Cardiff University), Zhang, Qiyuan (Cardiff University), Ji, Ze (Cardiff University) |
Keywords: Human Factors and Ergonomics, Evaluation Methods, User-centered Design of Robots
Abstract: In the manufacturing sector, enhancing user experiences (UX) in Human-Robot Interaction (HRI) and Human-robot collaboration (HRC) are becoming increasingly essential. This study is dedicated to the development of an initial assessment framework specifically designed for evaluating HRI UX in manufacturing. By qualitative semi-structured interviews, we focus on identifying key factors that constitutes UX in manufacturing context. The core contribution of this research is the development of a UX assessment framework tailored for evaluating HRI in manufacturing, include 5 facets of UX. This framework is derived from in-depth analysis of user feedback, providing a structured approach to understanding user interactions with robots. Our findings contribute valuable insights to the HRI discourse in manufacturing.
|
|
11:45-12:00, Paper TuBT4.2 | |
Sequential Transfer Learning-Based Human Decision Making Model for Human-Robot Co-Learning and Insights from User Feedback Analysis |
|
Kumar, Rajul (George Mason University), Som, Vaidehi (University of Pennsylvania), Yao, Ningshi (George Mason University) |
Keywords: Human Factors and Ergonomics, HRI and Collaboration in Manufacturing Environments, Cooperation and Collaboration in Human-Robot Teams
Abstract: For robots to transform work processes and seamlessly integrate into human settings, such as warehouses and assistive roles, they must adeptly manage a wide range of repetitive and complex interactions with various humans. Despite great advances in developing robots capable of modeling and predicting human actions for smooth collaboration, a gap persists on how these prediction models can account for the impact of robot actions on human behavior. In this paper, we introduced a human-to-human, sequential transfer learning-based, data-driven model for human decision-making in co-learning scenarios, where humans and robots mutually learn and observe from each other. We model the problem as a sequence prediction task, employing one-dimensional convolutional neural network (or 1D CNN) and dual-input bidirectional gated recurrent units (or Bi-GRU) models learned on interaction-rich, closed-loop human-robot co-learning scenarios. Our model inherently captures the dynamic irrational strategies humans develop in response to their robot counterparts. Additionally, the transfer learning approach improved the generalizability of our model, enabling it to accommodate diverse human behaviors, ideologies, and personal beliefs with reasonable accuracy. Using user feedback, we presented an in-depth analysis of some of the key factors considered by participants during interactions, which can further help in refining prediction models.
|
|
12:00-12:15, Paper TuBT4.3 | |
Advanced Robots in Healthcare and Their Impact on the Health and Safety of Medical Workers |
|
Heinold, Eva (Federal Institute for Occupational Safety and Health (BAuA)), Rosen, Patricia Helen (Federal Institute for Occupational Safety and Health), Wischniewski, Sascha (Federal Institute for Occupational Safety and Health (BAuA)) |
Keywords: Human Factors and Ergonomics, Medical and Surgical Applications, Assistive Robotics
Abstract: This article reviews recent work on advanced robotic systems in the medical field, with a focus on implications for occupational safety and health. The Human Health and Social Activities sector is currently the research field with the highest activity when it comes to the topic of human-robot interaction. The technology holds the potential to impact occupational safety and health for medical workers. Based on an extensive literature review we categorize and analyze the current research activity in the healthcare sector. This includes a mapping of the type of robotic system used, what tasks are automated, how tasks and jobs are transformed and the possible OSH impact occurs. We identify trends and gaps in research and derive recommendations for the implementation of robotic system in the sector. A total of 13 relevant meta-analysis and systematic reviews were selected and analyzed. The results show a strong focus on the automation of person related tasks via socially assistive robots. Secondly, robotic systems for rehabilitation and mobility are highly present in the sample. The OSH impacts are predominantly psychosocial or organizational. However, there are significant research gaps when it comes ergonomic impact on the medical personal. Our results can be used to guide new research and to design and implement robotic systems with a human centered approach.
|
|
12:15-12:30, Paper TuBT4.4 | |
Interaction of Robot Speed and Distance in Human-Robot Collaboration: Impact on Human Trust, Safety, and Comfort |
|
Büttner, Sebastian Thomas (University of Duisburg-Essen), Alhaji, Basel (Duisburg-Essen University), Katariya, Kaushikbhai (Clausthal University of Technology), Prilla, Michael (University of Duisburg-Essen) |
Keywords: Human Factors and Ergonomics, User-centered Design of Robots
Abstract: Robots are gradually being freed from their safety fences due to the advances in safety features integrated with most new robots. These robots enable a new form of human-robot interaction in which contact is possible. There are two robot-related factors that play a decisive role in the interaction dynamics and human perception in such a case – the motion speed and distance the robot keeps away from the human. Literature indicates that these factors influence whether humans perceive trust, safety, and comfort, which are essential components in the acceptance of robots by their end users. However, although speed and distance were intensively investigated in isolation, little is known about their combined effect. To this end, we conducted an experiment investigating the impact of the industrial robot arm’s approaching speed and stopping distance on the users’ trust, safety, and comfort while they were performing a collaborative task. Our results provide interesting insights into the implications of speed and trust. While we were able to replicate former studies in terms of implications of speed and distance, our data does not show any interaction effect between the two variables. However, our participants’ observations indicated that distance impacts the dependent variables more severely than speed.
|
|
TuBT5 |
Room T5 |
Machine Learning and Adaptation II |
Regular Session |
Chair: Torresen, Jim | University of Oslo |
Co-Chair: Liu, Junjia | The Chinese University of Hong Kong |
|
11:30-11:45, Paper TuBT5.1 | |
Fast LiDAR Upsampling Using Conditional Diffusion Models |
|
Helgesen, Sander (University of Oslo), Nakashima, Kazuto (Kyushu University), Torresen, Jim (University of Oslo), Kurazume, Ryo (Kyushu University) |
Keywords: Machine Learning and Adaptation, Motion Planning and Navigation in Human-Centered Environments
Abstract: The search for refining 3D LiDAR data has attracted growing interest motivated by recent techniques such as supervised learning or generative model-based methods. Existing approaches have shown the possibilities for using diffusion models to generate refined LiDAR data with high fidelity, although the performance and speed of such methods have been limited. These limitations make it difficult to execute in real-time, causing the approaches to struggle in real-world tasks such as autonomous navigation and human-robot interaction. In this work, we introduce a novel approach based on conditional diffusion models for fast and high-quality sparse-to-dense upsampling of 3D scene point clouds through an image representation. Our method employs denoising diffusion probabilistic models trained with conditional inpainting masks, which have been shown to give high performance on image completion tasks. We introduce a series of experiments, including multiple datasets, sampling steps, and conditional masks. This paper illustrates that our method outperforms the baselines in sampling speed and quality on upsampling tasks using the KITTI-360 dataset. Furthermore, we illustrate the generalization ability of our approach by simultaneously training on real-world and synthetic datasets, introducing variance in quality and environments.
|
|
11:45-12:00, Paper TuBT5.2 | |
D3D: Conditional Diffusion Model for Decision-Making under Random Frame Dropping |
|
Xia, Bo (Tsinghua Shenzhen International Graduate School), Luo, yifu (Tsinghua University), Chang, Yongzhe (Tsinghua Shenzhen International Graduate School), Yuan, Bo (Tsinghua University), Li, Zhiheng (Tsinghua University), WANG, xueqian (Center for Artificial Intelligence and Robotics, Graduate School) |
Keywords: Machine Learning and Adaptation
Abstract: The occurrence of frame drops due to issues such as corrupted communications or malfunctioning sensors presents a significant challenge to an agent's decision-making, especially in remote control scenarios. Classical reinforcement learning (RL) usually assumes a continuous data stream without frame drops and relies heavily on online interactions, which is time-consuming, resource-intensive, and often impractical in certain scenarios. Consequently, the performance of RL may deteriorate significantly in face of non-negligible frame drops. To tackle this challenge caused by frame dropping, We propose Conditional underline{bm{text{D}}}iffusion Model for underline{bm{text{D}}}ecision-Making under Random Frame underline{bm{text{D}}}ropping (D3D), an offline algorithm that can effectively enhance performance robustness in frame dropping scenarios. D3D addresses this issue through a two-phase approach: 1) During the policy generation phase, D3D adopts a return-conditional decision diffusion model rather than the temporal difference learning, whose policy is derived using offline datasets of return-labeled trajectories without information loss. 2) When frame dropping occurs during evaluation, D3D seamlessly substitutes the missing state with its corresponding prediction in the horizon made by the diffusion model. Extensive experiments are conducted on MuJoCo and Adroit tasks to validate D3D’s robustness and efficiency. The results demonstrate that D3D consistently outperforms state-of-the-art RL algorithms, especially excelling on tasks featuring severe drop rates.
|
|
12:00-12:15, Paper TuBT5.3 | |
A Deep Learning-Based Grasp Pose Estimation Approach for Large-Size Deformable Objects in Clutter |
|
YU, Minghao (The Chinese University of Hong Kong), Li, Zhuo (The Chinese University of Hong Kong), Li, Zhihao (The Chinese University of Hong Kong), Liu, Junjia (The Chinese University of Hong Kong), Teng, Tao (The Chinese University of Hong Kong & Hong Kong Centre for Logis), Chen, Fei (The Chinese University of Hong Kong) |
Keywords: Machine Learning and Adaptation
Abstract: Deformable objects especially large-size deformable objects grasping is unappreciated but widespread in industrial applications (e.g., clothes recycling). While it encounters several challenges, for example, the existing methods didn't take large-size deformable objects into account, no typical boundary of deformable objects. To solve the challenges, we proposed a grasp detection framework consisting of a self-trained object detection network, an instance segmentation module, and a grasp pose generation pipeline. The experiments were successfully conducted on the industrial laundry mock-up with an 88.9% success ratio. The experiments result indicates the effectiveness of the proposed framework on spatial-constrained large-size deformable objects grasping in clutter.
|
|
12:15-12:30, Paper TuBT5.4 | |
Secure and Efficient Face Recognition Via Supervised Federated Learning |
|
Solomon, Enoch (Virginia State University), Zewoudie, Abraham Woubie (Silo AI), Abdelzaher, Ahmed (Virginia State University), Emiru, Eyael Solomon (University of Trento) |
Keywords: Machine Learning and Adaptation
Abstract: The state-of-the-art face recognition systems frequently rely on training with large image dataset acquired from several users on a single device. However, these databases usually contain sensitive information, which raises substantial privacy concerns. To address this issue, we propose the use of federated learning for a supervised face recognition. Federated learning enables model training while maintaining privacy by leveraging decentralized edge devices. In our method, each edge device trains a model individually, which is subsequently sent to a secure aggregator. We use GAN to introduce different data without sending actual data. The aggregator combines various models into a global one, which is subsequently disseminated to the edge devices. Experiments with CelebA datasets show that federated learning effectively protects privacy while maintaining performance levels.
|
|
TuBT6 |
Room T6 |
Motion Planning and Navigation in Human-Centered Environments II |
Regular Session |
Co-Chair: Park, Jeongeun | Korea University |
|
11:30-11:45, Paper TuBT6.1 | |
Age and Spatial Cue Effects on User Performance for an Adaptable Verbal Wayfinding System |
|
Takahira, Rin (Kobe University), Liu, Chaoran (Riken), Ishi, Carlos Toshinori (RIKEN), Ohkawa, Takenao (Kobe University) |
Keywords: Motion Planning and Navigation in Human-Centered Environments, Linguistic Communication and Dialogue, Cognitive Skills and Mental Models
Abstract: This study aims on developing an interactive verbal wayfinding system that can adapt directions based on the user's context. Rather than simply providing the shortest route like in existing navigation systems, the developed system searches for routes with fewer turns and generates easy-to-understand verbal explanations. A key feature is prioritizing the inclusion of landmark references at potential points of confusion to provide appropriate contextual cues. Base functions for route searching and verbal direction generation were implemented in this study. A subjective experiment was then conducted to evaluate how user age and the number of provided landmark references impact wayfinding performance when using the system's verbal directions. Results revealed an age-dependent interaction: increasing landmarks did not universally improve wayfinding success. The optimal number of landmarks varied across age groups. The findings highlight the need to adapt the level of spatial cue details in verbal directions based on the user's context.
|
|
11:45-12:00, Paper TuBT6.2 | |
EnQuery: Ensemble Policies for Diverse Query-Generation in Preference Alignment of Robot Navigation |
|
de Heuvel, Jorge (University of Bonn), Seiler, Florian (Rheinische Friedrich-Wilhelms-Universität Bonn), Bennewitz, Maren (University of Bonn) |
Keywords: Motion Planning and Navigation in Human-Centered Environments, Machine Learning and Adaptation, Novel Interfaces and Interaction Modalities
Abstract: To align mobile robot navigation policies with user preferences through reinforcement learning from human feedback (RLHF), reliable and behavior-diverse user queries are required. However, deterministic policies fail to generate a variety of navigation trajectory suggestions for a given navigation task. In this paper, we introduce EnQuery, a query generation approach using an ensemble of policies that achieve behavioral diversity through a regularization term. For a given navigation task, EnQuery produces multiple navigation trajectory suggestions, thereby optimizing the efficiency of preference data collection with fewer queries. Our methodology demonstrates superior performance in aligning navigation policies with user preferences in low-query regimes, offering enhanced policy convergence from sparse preference queries. The evaluation is complemented with a novel explainability representation, capturing full scene navigation behavior of the mobile robot in a single plot. Our code is available online at https://github.com/hrl-bonn/EnQuery.
|
|
12:00-12:15, Paper TuBT6.3 | |
YoloTag: Vision-Based Robust UAV Navigation with Fiducial Markers |
|
Raxit, Sourav (University of New Orleans), Singh, Simant Bahadur (The University of New Orleans), Redwan Newaz, Abdullah Al (University of New Orleans) |
Keywords: Motion Planning and Navigation in Human-Centered Environments, Machine Learning and Adaptation, Virtual and Augmented Tele-presence Environments
Abstract: By harnessing fiducial markers as visual landmarks in the environment, Unmanned Aerial Vehicles (UAVs) can rapidly build precise maps and navigate spaces safely and efficiently, unlocking their potential for fluent collaboration and coexistence with humans. Existing fiducial marker methods rely on handcrafted feature extraction, which sacrifices accuracy. On the other hand, deep learning pipelines for marker detection fail to meet real-time runtime constraints crucial for navigation applications. In this work, we propose YoloTag: a real-time fiducial marker-based localization system. YoloTag uses a lightweight YOLO v8 object detector to accurately detect fiducial markers in images while meeting the runtime constraints needed for navigation. The detected markers are then used by an efficient perspective-n-point algorithm to estimate UAV states. However, this localization system introduces noise, causing instability in trajectory tracking. To suppress noise, we design a higher-order Butterworth filter that effectively eliminates noise through frequency domain analysis. We evaluate our algorithm through real-robot experiments in an indoor environment, comparing the trajectory tracking performance of our method against other approaches in terms of several distance metrics.
|
|
12:15-12:30, Paper TuBT6.4 | |
Towards Text-Based Human Search and Approach Using a Robot Dog |
|
Park, Jeongeun (Korea University), Silveira, Jefferson (Queen's University), Pan, Matthew (Queen's University), Choi, Sungjoon (Korea University) |
Keywords: Motion Planning and Navigation in Human-Centered Environments, Machine Learning and Adaptation
Abstract: In this paper, we propose a SOCratic model for Robots Approaching humans based on TExt System (SOCRATES) focusing on the human search and approach based on free-form textual description; the robot first searches for the target user, then the robot proceeds to approach in a human-friendly manner. We present a Socratic human search model that connects large pre-trained foundation models to solve the downstream task of searching for the target person based on textual descriptions. In particular, textual descriptions used for searching are composed of appearance (e.g., wearing a white shirt with black hair) and location clues (e.g., a student that works with robots). Additionally, we propose a hybrid learning-based framework for generating human-friendly robotic motion to approach a person, consisting of a learning-from-demonstration module and a knowledge distillation module utilizing LLMs. We evaluate the search performance of the proposed method in both simulation and real-world environments using the Boston Dynamics Spot robot. Moreover, we evaluate the effectiveness of our proposed framework through analysis involving human participants and investigate the perceived warmth of our system.
|
|
TuCT1 |
Room T1 |
Bridging Trust and Context: Dynamic Interactions in HAI I |
Special Session |
Chair: Fukuchi, Yosuke | Tokyo Metropolitan University |
Co-Chair: Yamada, Seiji | National Institute of Informatics |
|
14:40-16:10, Paper TuCT1.1 | |
Associative Games with Virtual Agents: Unexpectedness Increases Willingness to Continue Dialog (I) (withdrawn from program) |
|
Matsui, Tetsuya (Osaka Institute of Technology) |
|
14:40-16:10, Paper TuCT1.2 | |
Changing Human's Impression of Empathy from Agent by Verbalizing Agent's Position (I) |
|
Tsumura, Takahiro (Toyo University), Yamada, Seiji (National Institute of Informatics) |
Keywords: Anthropomorphic Robots and Virtual Humans, Embodiment, Empathy and Intersubjectivity, Creating Human-Robot Relationships
Abstract: As anthropomorphic agents (AI and robots) are increasingly used in society, empathy and trust between people and agents are becoming increasingly important. A better understanding of agents by people will help to improve the problems caused by the future use of agents in society. In the past, there has been a focus on the importance of self-disclosure and the relationship between agents and humans in their interactions. In this study, we focused on the attributes of self-disclosure and the relationship between agents and people. An experiment was conducted to investigate hypotheses on trust and empathy with agents through six attributes of self-disclosure (opinions and attitudes, hobbies, work, money, personality, and body) and through competitive and cooperative relationships before a robotic agent performs a joint task. The experiment consisted of two between-participant factors: six levels of self-disclosure attributes and two levels of relationship with the agent. The results showed that the two factors had no effect on trust in the agent, but there was statistical significance for the attribute of self-disclosure regarding a person's empathy toward the agent. In addition, statistical significance was found regarding the agent's ability to empathize with a person as perceived by the person only in the case where the type of relationship, competitive or cooperative, was presented. The results of this study could lead to an effective method for building relationships with agents, which are increasingly used in society.
|
|
14:40-16:10, Paper TuCT1.3 | |
User Decision Guidance with Selective Explanation Presentation from Explainable-AI (I) |
|
Fukuchi, Yosuke (Tokyo Metropolitan University), Yamada, Seiji (National Institute of Informatics) |
Keywords: Machine Learning and Adaptation, Social Intelligence for Robots, Cognitive Skills and Mental Models
Abstract: This paper addresses the challenge of selecting explanations for XAI (Explainable AI)-based Intelligent Decision Support Systems (IDSSs). IDSSs have shown promise in improving user decisions through XAI-generated explanations along with AI predictions, and the development of XAI made it possible to generate a variety of such explanations. However, how IDSSs should select explanations to enhance user decision-making remains an open question. This paper proposes X-Selector, a method for selectively presenting XAI explanations. It enables IDSSs to strategically guide users to an AI-suggested decision by predicting the impact of different combinations of explanations on a user’s decision and selecting the combination that is expected to minimize the discrepancy between an AI suggestion and a user decision. We compared the efficacy of X-Selector with two naive strategies (all possible explanations and explanations only for the most likely prediction) and two baselines (no explanation and no AI support) in virtual stock-trading support scenarios. The results suggest the potential of X-Selector to guide users to AI-suggested decisions and improve task performance under the condition of a high AI accuracy.
|
|
14:40-16:10, Paper TuCT1.4 | |
Empirical Investigation of How Robot Head Motion Influences Acceptance of Heatmap-Based XAI: Designing XAI with Social Robot (I) |
|
Maehigashi, Akihiro (Shizuoka University), Fukuchi, Yosuke (Tokyo Metropolitan University), Yamada, Seiji (National Institute of Informatics) |
Keywords: Non-verbal Cues and Expressiveness
Abstract: This study investigated how a robot head motion towards an AI attention heatmap during a visual identification task influences a human user’s trust in eXplainable AI (XAI). The findings revealed that the robot head motion presented in a video increased the user's acceptance of AI-generated results compared to the robot eye gaze displayed in a static image with or without the AI attention heatmap. However, displaying the heatmap improved task performance more than displaying no heatmap with or without the robot. Overall, these results suggest a possibility that showing a robot head motion towards an AI attention heatmap in a movie can serve as an interpretable XAI for visual tasks.
|
|
14:40-16:10, Paper TuCT1.5 | |
Two Is Better Than One: Cultural Differences in the Number of Apologizing Robots in the U.S. and Japan (I) |
|
Shiomi, Masahiro (ATR), Hirayama, Taichi (Doshisha University), Kimoto, Mitsuhiko (Meiji University), Iio, Takamasa (Doshisha University), Shimohara, Katsunori (Doshisha University) |
Keywords: Robot Companions and Social Robots, Applications of Social Robots, Interaction with Believable Characters
Abstract: Apology behavior design is becoming important for social robots that work in daily environments because of their widespread use. Robotics researchers have reported that multiple robots can effectively achieve more acceptance and trust in apology situations. Unfortunately, such effects have only been confirmed in a single country: Japan. Some studies investigated cultural differences in apologies between Japan and other cultures and reported how the former influences the perceived function and meaning of apologies. Therefore, we conducted a web-based survey to investigate whether using multiple robots in apology situations is effective in another country. We compared such perceived feelings as forgiveness and trust toward a robot’s apologies between Japan and the U.S. by using the visual stimuli of one and two robots. The experiment results showed that U.S. people felt that multiple robot apologies are more acceptable than apologies from just one robot, similar to results with Japanese participants. Perceived trust did show a different phenomenon between the two countries.
|
|
14:40-16:10, Paper TuCT1.6 | |
Measuring Algorithm Aversion and Betrayal Aversion to Humans and AI Using Trust Games (I) |
|
Takagi, Hisashi (BeNEXT Solutions Inc), Li, Yang (Nagoya University), Komori, Masashi (Osaka Electro-Communication University), Terada, Kazunori (Gifu University) |
Keywords: Robot Companions and Social Robots, Philosophical Issues in Human-Robot Coexistence, Social Intelligence for Robots
Abstract: Behavioral trust is to entrust resources to the trustee, expecting a high return while accepting the risk of betrayal. Previous studies have demonstrated bias on behavioral trust, that trust is not necessarily rationally to the expected values, and is inconsistent to human vs. non-human counterparts. People on the one hand seem to be aversive to intentional betrayal (Betrayal Aversion), but on the other hand, people avoid being dependent on algorithms rather than human (Algorithm Aversion). Yet, these aversions are not comprehensively investigated. The present study conducted a well-controlled behavioral game experiment that systematically explored the entrusting (risk-taking) behavior when facing a counterpart (human vs. AI) or a natural risk, and further explored the effect of the counterpart's computational ability (intentional for human and algorithmic for AI). Participants (n=284) played a trust game with (a) a human with intentional decisions, (b) an AI with algorithmic decisions, (c) a human with random decisions, or (d) an AI with random decisions whose probability of return was known, as well as a lottery task structurally equivalent to the trust game. Entrusting decisions to different levels of probability of return were measured to compute the Minimum Acceptable Probability (MAP) as a quantitative measure of trust. The results showed that participants were more trusting of a counterpart than a lottery machine, yet this tendency did not differ by counterparts (i.e., human vs. AI counterparts), or between high computational ability or lower (i.e., intentional/algorithmic decision vs. random decisions). The results suggest an overtrust bias, rather than an aversive bias to a counterpart with agency, whether they are AIs or humans, regardless of their intentions.
|
|
TuCT2 |
Room T2 |
Assistive Robotics I |
Regular Session |
Chair: Guo, Yunfei | Futronics (NA) |
|
14:40-16:10, Paper TuCT2.1 | |
Adaptive Control in Assistive Application - a Study Evaluating Shared Control by Users with Limited Upper Limb Mobility |
|
Goldau, Felix Ferdinand (DFKI GmbH), Pascher, Max (TU Dortmund University), Baumeister, Annalies (Frankfurt University of Applied Sciences), Tolle, Patrizia (Deutsch), Gerken, Jens (TU Dortmund University), Frese, Udo (Universität Bremen) |
Keywords: Assistive Robotics, Applications of Social Robots, Robots in Education, Therapy and Rehabilitation
Abstract: Shared control in assistive robotics blends human autonomy with computer assistance, thus simplifying complex tasks for individuals with physical impairments. This study assesses an adaptive Degrees of Freedom control method specifically tailored for individuals with upper limb impairments. It employs a between-subjects analysis with 24 participants, conducting 81 trials across three distinct input devices in a realistic everyday-task setting. Given the diverse capabilities of the vulnerable target demographic and the known challenges in statistical comparisons due to individual differences, the study focuses primarily on subjective qualitative data. The results reveal consistently high success rates in trial completions, irrespective of the input device used. Participants appreciated their involvement in the research process, displayed a positive outlook, and quick adaptability to the control system. Notably, each participant effectively managed the given task within a short time frame.
|
|
14:40-16:10, Paper TuCT2.2 | |
Do Mistakes Matter? Comparing Trust Responses of Different Age Groups to Errors Made by Physically Assistive Robots |
|
Wald, Sasha (Carnegie Mellon University), Puthuveetil, Kavya (Carnegie Mellon University), Erickson, Zackory (Carnegie Mellon University) |
Keywords: Assistive Robotics, Creating Human-Robot Relationships, Robotic Etiquette
Abstract: Trust is a key factor in ensuring acceptable human-robot interaction, especially in settings where robots may be assisting with critical activities of daily living. When practically deployed, robots are bound to make occasional mistakes, yet the degree to which these errors will impact a care recipient's trust in the robot, especially in performing physically assistive tasks, remains an open question. To investigate this, we conducted experiments where participants interacted with physically assistive robots which would occasionally make intentional mistakes while performing two different tasks: bathing and feeding. Our study considered the error response of two populations: younger adults at a university (median age 26) and older adults at an independent living facility (median age 83). We observed that the impact of errors on a users' trust in the robot depends on both their age and the task that the robot is performing. We also found that older adults tend to evaluate the robot on factors unrelated to the robot's performance, making their trust in the system more resilient to errors when compared to younger adults. Code and supplementary materials are available on our project webpage.
|
|
14:40-16:10, Paper TuCT2.3 | |
Contrastive Learning for Body Gesture Detection During Adapted Physical Activity |
|
Carlos Martinez, Juan (Lisv Systems Engineering Laboratory of Versailles), Pennino, Federico (Bologna University), Dubois, Cécile (Versailles Engineering Systems Laboratory-LISV), Monacelli, Eric (LISV, University of Versailles), Gabbrielli, Maurizio (Bologna University) |
Keywords: Assistive Robotics, Machine Learning and Adaptation, Novel Interfaces and Interaction Modalities
Abstract: This paper presents a novel approach to body gesture recognition for powered wheelchair users, leveraging inertial data from wrist-mounted sensors to facilitate movement and enhance autonomy in Adapted Physical Activity (APA). Gesture recognition technology interprets human gestures to allow non-direct communication with devices, enhancing human-machine interaction across various fields. APA fosters inclusion and well-being through tailored physical engagement. Our model not only identifies known gestures with high accuracy, as indicated by a mean Average Precision (mAP) score of 0.92 and a Recall@1 score of 0.983, but also demonstrates the ability to recognize gestures not included in the training set. This research contributes to the field of human-robot interaction by offering a more dynamic and inclusive form of interaction for individuals reliant on powered mobility aids.
|
|
14:40-16:10, Paper TuCT2.4 | |
Adaptive Weight Compensation in Assistive Upper-Limb Exoskeletons: An EMG Analysis |
|
Salman, Nada (University of Paris-Saclay), Benali, Abderraouf (Université De Versailles) |
Keywords: Assistive Robotics, Medical and Surgical Applications, HRI and Collaboration in Manufacturing Environments
Abstract: Abstract— This study explores the integration of adaptive weight compensation within an upper-limb exoskeleton by analyzing muscle activity from electromyography (EMG) signals. We enabled the exoskeleton to translate user-applied forces into precise joint movements by implementing an admittance controller. Through our experimental protocol, we measured EMG signals from the biceps brachii muscle, focusing on changes in two time-domain features -RMS and variance- under varying conditions. Then, we analyzed these changes to evaluate the impact on the muscle when users were assisted in carrying a one-kilogram object, compared to when they were not assisted. Our initial findings indicate that adaptive weight compensation significantly reduces muscle activity and enhances stability during lifting tasks, suggesting potential for real-time EMG-based activation of weight compensation. This research highlights the importance of combining force and EMG data to develop adaptive control laws for assistive robotics, paving the way for personalized Human-Robot Interaction (HRI).
|
|
14:40-16:10, Paper TuCT2.5 | |
Smart Music Therapist 1.0: Rhythmic Auditory Stimulation Integrated Robotic Walker As a Therapeutic Companion for Gait Rehabilitation |
|
Thalahitiya Vithanage, Ranul Helitha Vithanage (University of Moratuwa), Senaratne, Nissanka Arachchi Appuhamilage Nidula Ransika (University of Moratuwa), Welangalle, Pesala Dulkith (University of Moratuwa), Amarasinghe, Ranjith (University of Moratuway), Jayathilaka, Wanasinghe Arachchige Dumith Madushanka (Department of Mechanical Engineering, University of Moratuwa), Jayawardena, Madura (University of Sri Jayewardenepura) |
Keywords: Assistive Robotics, Medical and Surgical Applications, Robot Companions and Social Robots
Abstract: This study investigates the development of a robotic walker for gait rehabilitation that integrates social assistive robotics and principles of Rhythmic Auditory Stimulation (RAS), a form of music therapy. The robotic system is intended to supplement, rather than replace, the work of professional music therapists by enabling therapeutic interventions to continue outside of therapy sessions. Robot-facilitated Music Therapy is an evolutionary step of telerehabilitation, providing patients with greater autonomy while still benefiting from the expertise and guidance of therapists. Contributions of the research include 1) Robot-facilitated Music Therapy through a robotic walker as a platform for gait rehabilitation. 2) Remote Music Therapy through a robotic walker as a telerehabilitation platform for music therapists to remotely monitor and consult with patients. 3) Intelligent companionship integrated with a robotic walker for comprehensive user assistance. The proposed system, tested for its functionalities including personified RAS, social interaction, user monitoring, mobility control, and emergency response yielded promising results for real world application.
|
|
14:40-16:10, Paper TuCT2.6 | |
Analysis of Hesitation Behavior by Human Receiver Toward Facing Motion of Robot During Handover |
|
Goto, Ryoya (University of Tsukuba), Hassan, Modar (University of Tsukuba), Suzuki, Kenji (University of Tsukuba) |
Keywords: Assistive Robotics, Monitoring of Behaviour and Internal States of Humans, Non-verbal Cues and Expressiveness
Abstract: This paper examines whether the robot face-turning to a human during handover can be a behavior cue that supports the human action plan in accomplishing a task. A series of behavior cues observed in interactions, such as Joint action, are understood as social signals, which can be used as cues to help each other make decisions about their own action plans. We conducted temporal analysis and subjective evaluation of the receiver's behavior in an object handover experiment. The results showed that the subjective experience did not significantly change when the robot turned its head toward the receiver, but the time to receive the participant significantly increased. This suggests that the robot's face-turning behavior may have altered the receiver's human prediction and action plan, causing the receiver to hesitate to respond. These findings lead to the basis for constructing a mathematical model of empathic behavior that allows humans and robots to share the intention of action goals in the real world.
|
|
TuCT3 |
Room T3 |
HRI and Collaboration in Manufacturing Environments I |
Regular Session |
Chair: Assunção, Gustavo | Institute of Systems and Robotics - University of Coimbra |
Co-Chair: Celik, Cankan | Middle East Technical University |
|
14:40-16:10, Paper TuCT3.1 | |
Intuitive Force Feedback for Robot-Assisted Load Transport |
|
Celik, Cankan (Middle East Technical University), Ankarali, Mustafa Mert (Middle East Technical University) |
Keywords: HRI and Collaboration in Manufacturing Environments, Assistive Robotics, Innovative Robot Designs
Abstract: This paper introduces a force-guided robotic platform that aims to reduce musculoskeletal injuries in manufacturing and healthcare environments. The platform utilizes force feedback algorithms, system modeling with distinct force and virtual torque processing, and virtual dynamics for seamless control of heavy loads. Similar mobile robot platforms in previous studies were designed for rehabilitation purposes. This study focuses on heavy-duty applications using a cost-effective load cell configuration on a differential drive system. We propose a novel model that simplifies the human-robot interaction analysis by decomposing pushing forces into linear and rotational components represented by a combination of linear and rotational spring-damper systems. These elements are integrated into the dynamics of the mobile robot platform for a comprehensive representation of human-robot interaction. Additionally, we developed a new control method to counteract large force fluctuations and anti-integral windup with dynamic clamping to improve system robustness.
|
|
14:40-16:10, Paper TuCT3.2 | |
Methodology on the Cyber-Physical System Construction for a User-Friendly Smart Clothing Manufacturing Robot System |
|
Kang, Bonggu (Korea Institute of Industrial Technology(KITECH)), Park, Hong-Sun (Korea Institute of Industrial Technology), Lee, Jinmyeong (Korea Institute of Industrial Technology), YUN, JUNGMIN (Korea Institute of Industrial Technology) |
Keywords: HRI and Collaboration in Manufacturing Environments, Computational Architectures, Creating Human-Robot Relationships
Abstract: In clothing manufacturing, automated equipment systems capable of customized small-scale production using robots are receiving much attention. To flexibly produce various types of clothing using such a system, a monitoring and control environment that is friendly to user decision-making is required. Nevertheless, existing equipment systems do not provide such an environment because they are focused on the operation itself, and there are time and financial limitations in establishing a monitoring and control platform environment in equipment systems intended for small-quantity production. This paper verified data collection, simulation-based analysis, and control on an actual clothing production micro-factory system for verifying the operation behavior. Experimentally, we performed the operation verification of the proposed CPS methodology. We confirmed that control of existing scenarios and simulation-based analysis and control of new scenarios are possible. We expect that the proposed method will provide a user-friendly decision-making environment for existing manufacturing facilities and improve the efficiency of the production line.
|
|
14:40-16:10, Paper TuCT3.3 | |
Exploring Continuous Awareness Modelling for Improving Worker Safety and Trust |
|
Solovov, Andrey (University of Coimbra), Zarei, Mohammad (Instituto De Sistemas E Robótica - Universidade De Coimbra), Gautam, Vishal (Institute of Systems and Robotics - University of Coimbra), Ferreira, Bruno (Instituto De Sistemas E Robótica - Universidade De Coimbra), Assunção, Gustavo (Institute of Systems and Robotics - University of Coimbra), Marín Hernández, Antonio (Universidad Veracruzana), Menezes, Paulo (Institute of Systems and Robotics) |
Keywords: HRI and Collaboration in Manufacturing Environments, Cooperation and Collaboration in Human-Robot Teams, Cognitive Skills and Mental Models
Abstract: The Industry 5.0 HRI goals have the potential for significantly reducing production costs, but require factory workers to trust that robots will not cause them physical harm when sharing the workspace in order to be achieved. To this end, we propose a model that adjusts the robot's actions to an estimate of the workers' awareness of its behavior. Its premise is that mutual awareness should be continuous, and if no recent visual contact has been established, then the robot should adapt its range of motion and speed in order to minimize the intersection between the robot's space and the worker's space. On the other hand, if the worker steadily establishes visual contact from times to times, the robot operates under a normal or a faster configuration. After a period of habituation, this awareness model can be further tweaked to continuously improve production times, while improving worker's safety and trust. We present preliminary results of a pilot study with 32 participants, to asses if the existence of this awareness model impacts the outcomes of the combined task. The obtained tepidly optimistic results are the key to build upon in future designs.
|
|
14:40-16:10, Paper TuCT3.4 | |
A Robust Filter for Marker-Less Multi-Person Tracking in Human-Robot Interaction Scenarios |
|
Martini, Enrico (Università Di Verona), Parekh, Harshil (University of Pennsylvania), Peng, Shaoting (University of Pennsylvania), Bombieri, Nicola (University of Verona), Figueroa, Nadia (University of Pennsylvania) |
Keywords: HRI and Collaboration in Manufacturing Environments, Cooperation and Collaboration in Human-Robot Teams, Detecting and Understanding Human Activity
Abstract: Pursuing natural and marker-less human-robot interaction (HRI) has been a long-standing robotics research focus, driven by the vision of seamless collaboration without physical markers. Marker-less approaches promise an improved user experience, but state-of-the-art struggles with the challenges posed by intrinsic errors in human pose estimation (HPE) and depth cameras. These errors can lead to issues such as robot jittering, which can significantly impact the trust users have in collaborative systems. We propose a filtering pipeline that refines incomplete 3D human poses from an HPE backbone and a single RGB-D camera to address these challenges, solving for occlusions that can degrade the interaction. Experimental results show that using the proposed filter leads to more consistent and noise-free motion representation, reducing unexpected robot movements and enabling smoother interaction.
|
|
14:40-16:10, Paper TuCT3.5 | |
Human Planning of Robot Actions through LLM-Guided State Machine Synthesis |
|
Swick, Brennan (The Ohio State University), Donegan, Sean (Air Force Research Laboratory), Gillman, Andrew (Air Force Research Laboratory), Groeber, Michael (The Ohio State University) |
Keywords: HRI and Collaboration in Manufacturing Environments, Cooperation and Collaboration in Human-Robot Teams, User-centered Design of Robots
Abstract: Developments in natural language processing have made it easier than ever to program robots using language. Specifically, Large Language Models (LLMs) capture both language patterns and some embedded knowledge that enables emergent capabilities, such as planning. However, it is not well understood how to best use LLMs to reduce robotic programming effort. We present a system that aims to reduce the burden of robot programming without affecting the safety and observability of the system. This is achieved through a system that: 1) translates human plans to robotic plans via an LLM; 2) displays robotic plans in a state machine; and 3) demonstrates robotic plans in simulation. Our results show that the system can convert natural language to synthesized state machines that achieve goal specifications in 13/30 cases, with two sets of exemplar instructions producing 12 of the 13 state machines. Thus, given proper instructions, the programming burden can be reduced while maintaining safety and observability through verifiable discrete actions, state machine visualizations, and simulated demonstrations. These results suggest that the given instructions have a significant effect on the success of the plan and additional tools need to be added to support directability and corrective feedback from the human before a human-in-the-loop study for a manufacturing scenario should be performed.
|
|
14:40-16:10, Paper TuCT3.6 | |
A Hierarchical Gesture-Guided Framework for Multi-Manipulator Fabrication and Collaboration |
|
Wei, Jiaying (Carnegie Mellon University), Bard, Joshua (Carnegie Mellon University) |
Keywords: HRI and Collaboration in Manufacturing Environments, Degrees of Autonomy and Teleoperation
Abstract: With the rapid development of human-robot collaboration in workspaces, the interest in human-in-the-loop design and fabrication is expanding. In this research, we introduce a modularized gesture-guided framework for multi-manipulator fabrication. Echoing human-centered HRI in robotic construction and design fabrication, we explored different hierarchies of gesture-instructed tasks in a fabrication setting that involves manipulating and operating on materials. This gesture-guided framework for multi-manipulator fabrication contains instruction types at different control levels, making both gesture sets and assigned instructions modular and easily accessible for diverse tasks. Based on the flexibility of the high-level instruction pipeline, we also implemented a preliminary reachability-guided solver specifically for multi-manipulators in fabrication scenarios, improving the success rate for mutually constrained multi-goal assignment in the baseline Multi-Robot System (MRS).
|
|
TuCT4 |
Room T4 |
Non-Verbal Cues and Expressiveness I |
Regular Session |
Co-Chair: Gjaci, Ariel | University of Genoa |
|
14:40-16:10, Paper TuCT4.1 | |
What Kinds of Facial Self-Touches Strengthen Expressed Emotions? |
|
Kawano, Junya (Doshisha University), Kimoto, Mitsuhiko (Meiji University), Iio, Takamasa (Doshisha University), Shiomi, Masahiro (ATR) |
Keywords: Non-verbal Cues and Expressiveness, Anthropomorphic Robots and Virtual Humans, Affective Computing
Abstract: As virtual spaces continue to attract more and more attention, the development of virtual agents is also gaining momentum. Agents capable of expressing emotions are being developed, and various methods exist through which they express such feelings. Although much research has focused on the implementation of human actions in agents, no research has yet implemented facial self-touch for agents used in emotional expressions, even though such transmission of human emotions is a critical tool for them. In this study, we developed a system that enriches agents' emotional expressions by implementing self-touches in virtual agents. From an evaluation of our system, we found that it enhances the emotions of virtual agents. The naturalness of their emotional expressions varied depending on the emotion.
|
|
14:40-16:10, Paper TuCT4.2 | |
Eyes on the Game: Deciphering Implicit Human Signals to Infer Human Proficiency, Trust, and Intent |
|
Hulle, Nikhil (University of Colorado, Boulder), Aroca-Ouellette, Stephane (University of Colorado Boulder), Ries, Anthony (DEVCOM ARL), Brawer, Jake (University of Colorado, Boulder), von der Wense, Katharina (Johannes Gutenberg University Mainz), Roncone, Alessandro (University of Colorado Boulder) |
Keywords: Non-verbal Cues and Expressiveness, Monitoring of Behaviour and Internal States of Humans, Social Intelligence for Robots
Abstract: Effective collaboration between humans and AIs hinges on transparent communication and alignment of mental models. However, explicit, verbal communication is not always feasible in many scenarios. Under such circumstances, human-human teams often depend on implicit, nonverbal cues to glean important information about their teammates such as future intentions and expertise, thereby bolstering team alignment and adaptability. Among these cues a human’s actions in the environment and their visual attention are the most salient. In this paper, we propose a novel method to combine eye gaze data and behavioral data and evaluate their respective predictive power for human proficiency, trust, and intentions. We first collect a dataset of paired eye gaze and gameplay data in the fast-paced collaborative “Overcooked” environment. We then train variations of our models on this dataset to compare how the predictive powers differ between gaze data, gameplay data, and their combination. We additionally compare our method to prior works that aggregate eye gaze data and demonstrate how these aggregation methods can substantially reduce the predictive ability of eye gaze relative to our approach. Our results indicate that although there are certain situations where eye gaze data or gameplay data yield superior results, the model that integrates both types of data consistently outperforms all baselines. This work paves the way for developing intuitive and responsive agents that can efficiently adapt themselves to new and unseen teammates.
|
|
14:40-16:10, Paper TuCT4.3 | |
Legibot: Generating Legible Motions for Service Robots Using Cost-Based Local Planners |
|
Amirian, Javad (Sorbonne University), Abrini, Mouad (Sorbonne University), Chetouani, Mohamed (Sorbonne University) |
Keywords: Non-verbal Cues and Expressiveness, Motion Planning and Navigation in Human-Centered Environments, Applications of Social Robots
Abstract: With the increasing presence of social robots in various environments and applications, there is an increasing need for these robots to exhibit socially-compliant behaviors. Legible motion, characterized by the ability of a robot to clearly and quickly convey intentions and goals to the individuals in its vicinity, through its motion, holds significant importance in this context. This will improve the overall user experience and acceptance of robots in human environments. In this paper, we introduce a novel approach to incorporate legibility into local motion planning for mobile robots. This can enable robots to generate legible motions in real-time and dynamic environments. To demonstrate the effectiveness of our proposed methodology, we also provide a robotic stack designed for deploying legibility-aware motion planning in a social robot, by integrating perception and localization components.
|
|
14:40-16:10, Paper TuCT4.4 | |
Attentiveness: A Key Factor in Fostering Affective and Cognitive Trust with Non-Humanoid Robots |
|
Manor, Adi (Reichman University), parush, avi (Israel Institute of Technology - Technion), Erel, Hadas (Media Innovation Lab, Interdisciplinary Center Herzliya) |
Keywords: Non-verbal Cues and Expressiveness, Motivations and Emotions in Robotics, Robot Companions and Social Robots
Abstract: Affective trust and cognitive trust are fundamental elements in human-robot interactions. They impact robots’ acceptance, the level of engagement, and the tendency to rely on robots in various contexts. Opposite effects are observed when the interaction with a robot is unreliable and untrustworthy. In this study, we examined the possibility of enhancing both aspects of trust by manipulating the robot’s level of attentiveness to the participant. We focused on robotic attentiveness since it can be easily applied even to highly simple non-humanoid robots, positioning it as a method for enhancing trust for robots with different morphologies. Specifically, we evaluated whether minimal attentive robotic gestures can enhance affective and cognitive aspects of trust and whether inattentive robotic behavior can decrease them. Quantitative and qualitative results indicated that the robot’s attentiveness impacted both aspects of trust. Participants in the Attentive Robot condition reported higher affective and cognitive trust scores, smaller interpersonal distance, and a higher number of participants reported that the robot would ”be there for them”, in comparison to the Inattentive Robot and Baseline conditions. Our findings suggest that an attentive robotic behavior, can support human affective and cognitive trust and enhance human-robot interaction.
|
|
14:40-16:10, Paper TuCT4.5 | |
Labeling Sentences with Symbolic and Deictic Gestures Via Semantic Similarity |
|
Gjaci, Ariel (University of Genoa), Recchiuto, Carmine Tommaso (University of Genova), Sgorbissa, Antonio (University of Genova) |
Keywords: Non-verbal Cues and Expressiveness, Multimodal Interaction and Conversational Skills, Social Intelligence for Robots
Abstract: Co-speech gesture generation on artificial agents has gained attention recently, mainly when it is based on data-driven models. However, end-to-end methods often fail to generate co-speech gestures related to semantics with specific forms, i.e., Symbolic and Deictic gestures. In this work, we identify which words in a sentence are contextually related to Symbolic and Deictic gestures. Firstly, we appropriately chose 12 gestures recognized by people from the Italian culture, which different humanoid robots can reproduce. Then, we implemented two rule-based algorithms to label sentences with Symbolic and Deictic gestures. The rules depend on the semantic similarity scores computed with the RoBerta model between sentences that heuristically represent gestures and sub-sentences inside an objective sentence that artificial agents have to pronounce. We also implemented a baseline algorithm that assigns gestures without computing similarity scores. Finally, to validate the results, we asked 30 persons to label a set of sentences with Deictic and Symbolic gestures through a Graphical User Interface (GUI), and we compared the labels with the ones produced by our algorithms. For this scope, we computed Average Precision (AP) and Intersection Over Union (IOU) scores, and we evaluated the Average Computational Time (ACT). Our results show that semantic similarity scores are useful for finding Symbolic and Deictic gestures in utterances.
|
|
14:40-16:10, Paper TuCT4.6 | |
Two Is Better Than One: Explanations by Multiple Robots Strengthen the Feeling of Kawaii Toward Objects |
|
Kimura, Yuki (Nara Women's University), Anzai, Emi (Nara Women's University), Saiwaki, Naoki (Nara Women's University), Shiomi, Masahiro (ATR) |
Keywords: Non-verbal Cues and Expressiveness, Robot Companions and Social Robots, Affective Computing
Abstract: Using multiple robots in presentation effectively attracts and provides information to people compared to a single robot. Many robotics researchers employed this approach to realize effective robot-presentation systems in real environments. These systems implicitly equalize the amount of speech in conversational content between multiple robots; however, it is not known whether that rule is actually important. For example, the effects of the second robot that provides only a back-channel have not yet been investigated. If we unveil the relationships between the presentation effects and the amount of speech in multiple robot presentations, it is crucial in designing the content of multiple robots' conversations. For this purpose, we conducted a video-based web-survey. Our finding showed that product recommendations by multiple robots make them more positive, and a part of the advantages of equalizing the amount of speech compared to biasing the amount of speech in presentation.
|
|
TuCT5 |
Room T5 |
Computational Architectures and Cognitive Systems |
Regular Session |
Chair: Karydis, Konstantinos | University of California, Riverside |
Co-Chair: Liu, Rui | Kent State University |
|
14:40-16:10, Paper TuCT5.1 | |
Strategies for Moment Compensation in Supernumerary Robotic Limbs Manipulation Tasks |
|
Moon, Chaerim (University of Illinois, Urbana-Champaign), Kim, Joohyung (University of Illinois at Urbana-Champaign) |
Keywords: Cognitive and Sensorimotor Development, Cooperation and Collaboration in Human-Robot Teams, User-centered Design of Robots
Abstract: When designing the motions of wearable robotic systems, physical Human-Robot Interaction (pHRI) is a crucial consideration. This study suggests a coordinated motion planner for Supernumerary Robotic Limbs (SRLs) to minimize the applied moment from the robotic system’s operation to the human body. Given task motion trajectories, moment compensation motions are generated to reduce the asymmetric moment elements while maintaining the other element below a nominal value. The performance of the motion planner is evaluated by the simulation and the hardware experiment. An object pick-and-place scenario is used as a goal task. The simulation results demonstrate that under different conditions, the compensation motions significantly decrease the generated roll and yaw moments when compared to the results without the motions. The hardware experiment validates its practicality as a real-time motion controller of a physical robotic system.
|
|
14:40-16:10, Paper TuCT5.2 | |
Multi and Binary Task Classification for Real and Imagined Motor Movements Using EEG Signals |
|
Cheng, Yibo (Princeton International School of Mathematics and Science), Yadav, Nikhil (St. John's University, Queens, New York) |
Keywords: Cognitive and Sensorimotor Development, Machine Learning and Adaptation
Abstract: Vascular dysfunction, trauma, and other disorders often cause a loss in mobility. Brain-computer-aided prosthetics have the potential to help restore critical motor functions and carry out normal daily activities. Through brain-computer interfaces (BCI), electroencephalogram (EEG) brain signals can be used to generate prosthetic gesture commands and control external devices. However, due to the complexity posed by noise in EEG signals, it can be difficult to extract useful features and classify a wide range of motor tasks. In this study, an EEG Motor Movement and Imagery Dataset collected by the New York State Department of Health using a BCI2000 system, is used to perform detailed motor task classification. This dataset includes both real and imagined motions across six different tasks. Convolutional neural networks (CNN) are built to classify six tasks, ultimately leading to binary classification between real and imagined movements, together with motion and rest. The latter models achieve accuracies of 91.2% and 74.4%, respectively. Our approach has the potential for broader applications where comprehensive classifications are needed to discern multiple prosthetic movement tasks.
|
|
14:40-16:10, Paper TuCT5.3 | |
What Would I Do If...? Promoting Understanding in HRI through Real-Time Explanations in the Wild |
|
Love, Tamlin (Institut De Robòtica I Informàtica Industrial CSIC-UPC), Andriella, Antonio (Artificial Intelligence Research Institute, CSIC), Alenyà, Guillem (Institut De Robòtica I Informàtica Industrial CSIC-UPC) |
Keywords: Cognitive Skills and Mental Models, Multi-modal Situation Awareness and Spatial Cognition, Evaluation Methods
Abstract: As robots become more and more integrated in human spaces, it is increasingly important for them to be able to explain their decisions to the people they interact with. These explanations need to be generated automatically and in real-time in response to decisions taken in dynamic and often unstructured environments. However, most research in explainable human-robot interaction only considers explanations (often manually selected) presented in controlled environments. We present an explanation generation method based on counterfactuals and demonstrate its use in an ``in-the-wild" experiment using automatically generated and selected explanations of autonomous interactions with real people to assess the effect of these explanations on participants' ability to predict the robot's behaviour in hypothetical scenarios. Our results suggest that explanations aid one's ability to predict the robot's behaviour, but also that the addition of counterfactual statements may add some burden and counteract this beneficial effect.
|
|
14:40-16:10, Paper TuCT5.4 | |
Adaptive Environment-Aware Robotic Arm Reaching Based on a Bio-Inspired Neurodynamical Computational Framework |
|
Chatziparaschis, Dimitrios (UC Riverside), Zhong, Shan (University of California, Riverside), Christopoulos, Vassilios (California Institute of Technology), Karydis, Konstantinos (University of California, Riverside) |
Keywords: Computational Architectures, Cognitive and Sensorimotor Development, Social Intelligence for Robots
Abstract: Bio-inspired robotic systems are capable of adaptive learning, scalable control, and efficient information processing. Enabling real-time decision-making for such systems is critical to respond to dynamic changes in the environment. We focus on dynamic target tracking in open areas using a robotic six-degree-of-freedom manipulator with a bird-eye view camera for visual feedback, and by deploying the Neurodynamical Computational Framework (NeuCF). NeuCF is a recently developed bio-inspired model for target tracking based on Dynamic Neural Fields (DNFs) and Stochastic Optimal Control (SOC) theory. It has been trained for reaching actions on a planar surface toward localized visual beacons, and it can re-target or generate stop signals on the fly based on changes in the environment (e.g., a new target has emerged, or an existing one has been removed). We evaluated our system over various target-reaching scenarios. In all experiments, NeuCF had high end-effector positional accuracy, generated smooth trajectories, and provided reduced path lengths compared with a baseline cubic polynomial trajectory generator. In all, the developed system offers a robust and dynamic-aware robotic manipulation approach that affords real-time decision-making.
|
|
14:40-16:10, Paper TuCT5.5 | |
Quantum Exploration-Based Reinforcement Learning for Efficient Robot Path Planning in Sparse-Reward Environment |
|
Huang, Chao (Kent State University), Guo, Yibei (Kent State University), Zhu, Zhihui (Ohio State University), Si, Mei (Rensselaer Polytechnic Institute), Blankenberg, Daniel (Cleveland Clinic), Liu, Rui (Kent State University) |
Keywords: Computational Architectures, Evaluation Methods
Abstract: With the latest developments in sensors, battery, and Artificial Intelligence (AI) technologies, robots can perform missions in unstructured environments (e.g., disaster sites). However, their adaptation speed is still slow due to their limited onboard computing capability and dynamic disruption from environments, leading to inaccurate planning and delayed emergency reactions. Even though reinforcement learning increases robots' exploration speed by fusing effective guidance, the numerous interactions make the learning time-consuming and even risky (e.g., collisions). To fundamentally improve robot adaptation speed, this work seeks help from quantum power. A novel Quantum Exploration based Dreamer model (QED) was developed to facilitate reinforcement learning explorations. QED based on stochastic quantum walker quickly explores environments, evaluates action quality, and obtains a global optimal exploration strategy; then, these high-quality exploration samples will be used to facilitate reinforcement learning speed. A theoretical benefit of QED is facilitating reinforcement learning speed without changing the underlying learning architecture, which makes the proposed QED applicable to general robot learning scenarios. To validate QED effectiveness, a robot path-planning task in an obstacle-dense environment was designed. The number of needed training episodes validated the effectiveness. The results show that QED achieves around ten times faster policy learning compared with Monte Carlo tree-based reinforcement learning methods and vanilla reinforcement learning methods.
|
|
14:40-16:10, Paper TuCT5.6 | |
Open Access NAO (OAN): A ROS2-Based Software Framework for HRI Applications with the NAO Robot |
|
Bono, Antonio (University of Calabria), Brameld, Kenji (TRACLabs), D'Alfonso, Luigi (Università Della Calabria), Fedele, Giuseppe (University of Calabria) |
Keywords: Computational Architectures, Robot Companions and Social Robots
Abstract: This paper presents a new software framework for HRI experimentation with the sixth version of the common NAO robot produced by the United Robotics Group. Embracing the common demand of researchers for better performance and new features for NAO, the authors took advantage of the ability to run ROS2 onboard on the NAO to develop a framework independent of the APIs provided by the manufacturer. Such a system provides NAO with not only the basic skills of a humanoid robot such as walking and reproducing movements of interest but also features often used in HRI such as: speech recognition/synthesis, face and object detention, and the use of Generative Pre-trained Transformer (GPT) models for conversation. The developed code is therefore configured as a ready-to-use but also highly expandable and improvable tool thanks to the possibilities provided by the ROS community. The code repository is: https://github.com/antbono/OAN
|
|
TuCT6 |
Room T6 |
Creating Human-Robot Relationships I |
Regular Session |
|
14:40-16:10, Paper TuCT6.1 | |
How Do Starship Robots Affect Everyday Campus Life? an Exploratory Posting Board Analysis and Interview-Based Study |
|
Schneider, Adeline (Oregon State University), Robinson, Ayan (Oregon State University), Grimm, Cindy (Oregon State University), Fitter, Naomi T. (Oregon State University) |
Keywords: Long-term Experience and Longitudinal HRI Studies, Cognitive Skills and Mental Models, Evaluation Methods
Abstract: The rapid emergence of food delivery robots in public spaces has raised important questions regarding public perceptions and policy creation. One method for addressing these questions is examining the relationship between delivery robots and the communities they already serve. We assessed our university community's experiences with and perceptions of the Starship Technologies robots (robots that currently operate on campus) using two efforts: analysis of online posting board content related to the robots and interviews of campus community members about the robots. Perspectives captured in the online post analysis tended to be negative, while views tended to be positive in the interview results. At the same time, both results showed differing opinions and complexity; one tension that emerged in both efforts, for example, is the potential of the robots to both benefit and impede disability communities on campus. Further, there were fundamental misunderstandings about what data the robots can and do record. This research can help to inform roboticists and policymakers whose work relates to autonomous robots in public spaces.
|
|
14:40-16:10, Paper TuCT6.2 | |
Robot Social Identity Performance Facilitates Contextually-Driven Trust Calibration and Accurate Human Assessments of Robot Capabilities |
|
Stull, Maria (University of Colorado Boulder), Lohrmann, Clare (University of Colorado Boulder), Hayes, Bradley (University of Colorado Boulder) |
Keywords: Social Presence for Robots and Virtual Humans, Cooperation and Collaboration in Human-Robot Teams
Abstract: People struggle to form accurate expectations of robots because we typically associate behavior (and capability) with the physical entity even when there are clear indicators of different software programs dictating behavior at different times. This is a harmful prior, as commercially available, visibly similar robots do not necessarily share any common ground in terms of capability, safety, or behavior. Prior efforts to calibrate people’s expectations of robots have not extended to anchoring on the robot’s control software rather than its embodiment. In this work, we leverage social participation and flexible identity presentation to facilitate coworkers' associations of robot capability with the currently running software rather than physical entity itself. By linking each of a robot's controllers to a social identity, we enable collaborators to more easily differentiate between them. In a human subjects study (n=30), participants who experienced our social identity signal understood differences between the robot's two controllers and prevented an unreliable controller from harming perceptions of the robot's other controller.
|
|
14:40-16:10, Paper TuCT6.3 | |
Vid2RealHRI: Align Video-Based HRI Study Designs with Real-World Settings |
|
Hauser, Elliott (University of Texas at Austin), Chan, Yao-Cheng (University of Texas at Austin), Modak, Sadanand (The University of Texas at Austin), Biswas, Joydeep (University of Texas at Austin), Hart, Justin (University of Texas at Austin) |
Keywords: Evaluation Methods, Non-verbal Cues and Expressiveness, Robotic Etiquette
Abstract: HRI research using autonomous robots in real-world settings can produce results with the highest ecological validity of any study modality, but many difficulties limit such studies’ feasibility and effectiveness. We propose VID2REAL HRI, a research framework to maximize real-world insights offered by video-based studies. The VID2REAL HRI framework was used to design an online study using first-person videos of robots as real-world encounter surrogates. The online study (n = 385) distinguished the within-subjects effects of four robot behavioral conditions on perceived social intelligence and human willingness to help the robot enter an exterior door. A real-world, between-subjects replication (n = 26) using two conditions confirmed the validity of the online study’s findings and the sufficiency of the participant recruitment target (22) based on a power analysis of online study results. The VID2REAL HRI framework offers HRI researchers a principled way to take advantage of the efficiency of video-based study modalities while generating directly transferable knowledge of real-world HRI. Code and data from the study are provided at vid2real.github.io/vid2realHRI.
|
|
14:40-16:10, Paper TuCT6.4 | |
Fairness-Sensitive Policy-Gradient Reinforcement Learning for Reducing Bias in Robotic Assistance |
|
Zhu, Jie (George Washington University), Hu, Mengsha (Kent State University), Zhang, Amy (Kent State University), Jin, Ruoming (Kent State University), Liu, Rui (Kent State University) |
Keywords: Creating Human-Robot Relationships, Ethical Issues in Human-robot Interaction Research, Human Factors and Ergonomics
Abstract: Robots assist humans in various activities, from daily living to collaborative manufacturing. Because they have biased learning sources (e.g., data, demonstrations, human feedback), robots inevitably have discriminatory performance regarding individual differences (e.g., skin color, mobility, appearance); discriminatory performance will undermine robots' service quality, causes request ignorance and response delay, and even cause emotional offenses. Therefore, mitigating biases is critically important for delivering fair robotic services. In this paper, we design a bias-mitigation method -- Fairness-Sensitive Policy Gradient Reinforcement Learning (FSPGRL), to help robots self-identify and correct biased behaviors. FSPGRL identifies bias by examining the abnormal updates along particular gradients and updates the policy network to provide fair decisions. To validate FSPGRL's effectiveness, we designed a human-centered service scenario: a robot serving people in a restaurant. With a user study involving 24 humans and 1,000 service demonstrations, FSPGRL has proven effective in maintaining fairness during robot services.
|
|
14:40-16:10, Paper TuCT6.5 | |
An Exploratory Analysis of the Harmonious Bond between Home Robots and Their Owners in Japan |
|
Ichikura, Aiko (University of Tokyo), Okada, Kei (The University of Tokyo), Inaba, Masayuki (The University of Tokyo) |
Keywords: Creating Human-Robot Relationships, Evaluation Methods
Abstract: In Japan, home robots have become popular and are widely loved. In particular, Sony’s aibo users have formed a big community, and a society where humans and robots live in harmony is becoming a reality. In order for robots to be more accepted in the society, exploratory analysis and explicit description of their bonds are important. However, the scale for measuring relationships with robots has not yet been defined, so a suitable measure was created from the Friendship Function Scale to measure the relationship of friends and the Pet Awareness Scale to measure the attitude toward pets. Using the proposed scale, we investigated the role of home robots, subjective intimacy, and attitudes toward robots among 54 robot owners. The survey found that more than half of them considered robots to be pets or children. In addition, the average value of subjective intimacy was very high to indicate a very high level of attachment to the robot. Furthermore, as a result of factor analysis of the scale items, five factors were extracted: “Comfort,” “Irreplaceability,” “Trust from the robot,” “Trust to the robot,” and “Social life.” At this time, there was no bias in how users felt about comfort and irreplaceability, but there was a bias in how users felt about trust. Additionally, those with the highest level of subjective intimacy felt significantly more “trusted by the robot” than those with the other level of intimacy, suggesting that recognition of trust from the robot may be an important bond-forming strategy. In the future, we will aim to form further bonds by investigating relationships with robots that play various roles, as well as in regions other than Japan.
|
|
14:40-16:10, Paper TuCT6.6 | |
Action Over Words: Predicting Human Trust in AI Partners through Gameplay Behaviors |
|
Jafari Meimandi, Kiana (University of Virginia), Bolton, Matthew L. (University of Virginia), Beling, Peter A. (Virginia Tech) |
Keywords: Creating Human-Robot Relationships, Human Factors and Ergonomics, Social Touch in Human–Robot Interaction
Abstract: In the burgeoning field of human-AI interaction, trust emerges as a cornerstone because many think that it is critical to the effectiveness of collaboration and the acceptance of AI systems. Traditional methods of assessing trust have predominantly relied on self-reported measures, requiring participants to articulate their perceptions and attitudes through questionnaires. However, these explicit methods may not fully capture the nuanced dynamics of trust, especially in real-time and complex interaction environments. This paper introduces an innovative approach to evaluating trust in human-AI teams, pivoting from the conventional reliance on verbal or written feedback to analyzing gameplay behaviors as implicit indicators of trust levels. Utilizing the Overcooked-AI environment, our study explores how participants' interactions with AI agents of varying performance levels can reveal underlying trust mechanisms without a single query posed to the human players. This approach not only bypasses the efficiency challenges posed by repetitive and lengthy trust assessment methods, but also provides insights comparable to them. We highlight the potential of non-verbal cues and action patterns as reliable trust indicators by comparing the predictive accuracies of questionnaire-based models with those derived from gameplay behavior analysis. Furthermore, our findings suggest that these implicit measures can be integrated into adaptive systems and algorithms for real-time trust calibration in human-agent teaming settings. This shift towards an action-oriented trust assessment challenges existing paradigms and opens new avenues for understanding and enhancing human-AI collaboration.
|
|
TuLBR |
Room T10 - Poster Room |
Late Breaking Reports Session I |
Interactive Session |
|
16:10-17:10, Paper TuLBR.1 | |
Co-Design of a Robot with Incarcerated Youth |
|
Alves-Oliveira, Patrícia (Amazon Lab126), Cusworth Walker, Sarah (University of Washington), Cakmak, Maya (University of Washington) |
|
|
16:10-17:10, Paper TuLBR.2 | |
Development of a Robotic Fingertip-Shaped Hybrid Tactile Sensor Module for Multi-Sensory Detection During Robot Manipulation |
|
Jang, Jin-seok (Korea Research Institute of Standards and Science), Bok, Bo-Gyu (Korea Research Institute of Standards and Science), Kim, Min-Seok (Korea Research Institute of Standards and Science) |
Keywords: Applications of Social Robots, User-centered Design of Robots, HRI and Collaboration in Manufacturing Environments
Abstract: We have developed a robot fingertip-shaped hybrid tactile sensor module that can be attached to robotic hands. The tactile sensor module consists of a dermis with an integrated three-axis force sensor and a temperature sensor, and an epidermis with an integrated pressure sensor. The three-axis force sensor detects various physical quantities (such as three-directional forces and vibrations) by attaching four commercial semiconductor strain gauges to a specially designed metal mechanism. The pressure sensor is composed of a total of 64 high-performance pressure cells in an 8 x 8 array, covering an area of approximately 2 cm x 3 cm, which allows for high spatial resolution in detecting pressure distribution. The temperature sensor uses a commercial thermistor capable of measuring from -10 ℃ to 80 ℃ to sense the temperature of objects. In conclusion, experiments have demonstrated that the developed robot fingertip module can detect multiple senses using the internal three-axis force, pressure, and temperature sensors.
|
|
16:10-17:10, Paper TuLBR.3 | |
Force Profiling of a Shoulder Bidirectional Fabric-Based Pneumatic Actuator for a Pediatric Exosuit |
|
Ayazi, Mehrnoosh (University of California, Riverside), Sahin, Ipsita (University of California, Riverside), Mucchiani, Caio (University of California Riverside), Kokkoni, Elena (University of California, Riverside), Karydis, Konstantinos (University of California, Riverside) |
Keywords: Assistive Robotics, Child-Robot Interaction, User-centered Design of Robots
Abstract: This paper presents a comprehensive analysis of the contact force profile of a single-cell bidirectional soft pneumatic actuator, specifically designed to aid in the abduction and adduction of the shoulder for pediatric exosuits. The actuator was embedded in an infant-scale test rig featuring two degrees of freedom: an actuated revolute joint supporting shoulder abduction/adduction and a passive (but lockable) revolute joint supporting elbow flexion/extension. Integrated load cells and an encoder within the rig were used to measure the force applied by the actuator and the shoulder joint angle, respectively. The actuator's performance was evaluated under various anchoring points and elbow joint angles. Experimental results demonstrate that optimal performance, characterized by maximum range of motion and minimal force applied on the torso and upper arm, can be achieved when the actuator is anchored at two-thirds the length of the upper arm, with the elbow joint positioned at a 90-degree angle. The force versus pressure and joint angle graphs reveal nonlinear and hysteresis behaviors. The findings of this study yield insights about optimal anchoring points and elbow angles to minimize exerted forces without reducing the range of motion.
|
|
16:10-17:10, Paper TuLBR.4 | |
Getting Ready: An Intelligent Robot System for Assisting with Activities of Daily Living |
|
Tognotti, Matthew (Santa Clara University), Kyrarini, Maria (Santa Clara University) |
Keywords: Assistive Robotics, Robot Companions and Social Robots
Abstract: In recent years, mobile manipulators have shown their potential to support people with activities of daily living (ADLs). This paper proposed an intelligent robotic system that can assist people in preparing for ADLs, such as getting ready for work, shopping, or going to the doctor. The proposed system uses a rule-based approach to enable the robot to learn tasks from humans via speech commands and to execute the learned tasks. An experimental setup is developed to evaluate the proposed system in a pilot study. The preliminary results from the pilot study demonstrate the usefulness of the proposed robotic system.
|
|
16:10-17:10, Paper TuLBR.5 | |
More Self-Regulation and Self-Esteem through Social Robots? |
|
Balalic, Sanja (Hanze University of Applied Sciences), Tolboom, Jos (SLO, the Netherlands Institute for Curriculum Development), sanderman, robbert (University of Groningen), degens, nick (University of Arts Utrecht), Hagedoorn, Mariët (University of Groningen) |
Keywords: Assistive Robotics, Robot Companions and Social Robots, User-centered Design of Robots
Abstract: The ideal of staying home till the end of their lives is not realistic for every older adult. However, even while dwelling in nursing homes, they can maintain a certain degree of self-regulation. Deployment of social robots might play a role in supporting this. In this qualitative study, seven residents of a nursing home were followed, while hosting Maatje, a small humanoid social robot, programmed in a personalized manner to give reminders, provide entertainment and answer some basic questions. A robot like Maatje seems to be a useful assistant to activate residents of a nursing home and also seems to make a positive contribution to self-esteem and sense of self-regulation.
|
|
16:10-17:10, Paper TuLBR.6 | |
EiIBM_Robot: An Emotionally Supportive Robot for Delivering Interpretation Bias Modification Exercises |
|
Huang, Shiming (The Hong Kong Polytechnic University), Hoorn, Johan (VU University, Amsterdam) |
Keywords: Assistive Robotics, User-centered Design of Robots, Robots in Education, Therapy and Rehabilitation
Abstract: This paper introduces eiIBM_Robot, an emotionally supportive robot designed to deliver interpretation bias modification (IBM) exercises to individuals with depressive symptoms. The robot incorporates a conversational agent utilizing large language models and a user-centered architecture to enhance user engagement and experience. A preliminary user study (N = 38) revealed positive perceptions of the robot's performance and the effectiveness of the IBM exercises in promoting positive thinking patterns and emotional well-being. The findings highlight the potential of emotionally supportive robots to provide accessible mental health support and the importance of carefully balancing the robot's social presence and perceived role in the intervention.
|
|
16:10-17:10, Paper TuLBR.7 | |
Draw a Robot Task: The Influence of Children's Age and Drawings on Robot Interactions |
|
Howard, Lauren (Franklin and Marshall College), Wilson, Jason (Franklin & Marshall College), Langer, Allison (Temple University), Marshall, Peter (Temple University) |
|
|
16:10-17:10, Paper TuLBR.8 | |
Analysing Explanation-Related Interactions in Collaborative Perception-Cognition-Communication-Action |
|
Roig Vilamala, Marc (Cardiff University), Furby, Jack (Cardiff University), de Gortari Briseno, Julian (University of California, Los Angeles), Srivastava, Mani (UCLA), Preece, Alun (Cardiff University), Fuentes, Carolina (Cardiff University) |
Keywords: Creating Human-Robot Relationships, Cooperation and Collaboration in Human-Robot Teams, Anthropomorphic Robots and Virtual Humans
Abstract: Effective communication is essential in collaborative tasks, so AI-equipped robots working alongside humans need to be able to explain their behaviour in order to cooperate effectively and earn trust. We analyse and classify communications among human participants collaborating to complete a simulated emergency response task. The analysis identifies messages that relate to various kinds of interactive explanation identified in the explainable AI literature. This allows us to understand what type of explanations humans expect from their teammates in such settings, and thus where AI-equipped robots most need explanation capabilities. We find that most explanation-related messages seek clarification in the decisions or actions taken. We also confirm that messages have an impact on the performance of our simulated task.
|
|
16:10-17:10, Paper TuLBR.9 | |
Can Novice Supervisors Determine Successful from Unsuccessful Grasps from Tactile Sensing? |
|
Puente, Karina (Oregon State University), Sadler, Madelyn (Oregon State University), Wilson, Cristina (Oregon State University), Davidson, Joseph (Oregon State University) |
Keywords: Curiosity, Intentionality and Initiative in Interaction, Social Touch in Human–Robot Interaction, Novel Interfaces and Interaction Modalities
Abstract: Underwater manipulation is a challenging problem. To complete tasks requiring dexterous manipulation, the current standard practice is for human operators to remotely teleoperate a manipulator and end-effector, or monitor a semi-autonomous system. The energetic underwater environment (i.e. waves and currents) can cause the robot and objects in the environment to move, and water turbidity can interfere with visual perception of the scene. Turbidity is particularly problematic since cameras are the primary sensing modality operators use to monitor grasp execution. While tactile sensing has been shown to aid human-robot collaborative manipulation in terrestrial applications, it has not been studied for the underwater domain. This paper presents early work to study whether tactile feedback could help operators complete underwater manipulation. We conducted a within-subject in-the-lab user study to determine if the visual presentation of tactile sensing is sufficient to determine if a grasp is successful or unsuccessful. To complete the study, we created a mock interface that presents degraded videos of grasps besides tactile data from commercial off-the-shelf sensors. Twenty novice participants were recruited and asked to observe 36 grasp attempts across four types of objects. Results indicate that participants’ accuracy is better than chance, showing the potential for tactile sensing to increase operator performance during underwater manipulation.
|
|
16:10-17:10, Paper TuLBR.10 | |
Towards a Human Diversity Wheel for Human Robot Interaction |
|
Erle, Lukas (Ruhr West University of Applied Sciences), Timm, Lara (University of Applied Sciences Ruhr West), Kleinhaus, Lara (Ruhr West University of Applied Sciences), Straßmann, Carolin (University of Applied Sciences Ruhr West), Eimler, Sabrina C. (Hochschule Ruhr West, University of Applied Sciences) |
Keywords: Ethical Issues in Human-robot Interaction Research, Evaluation Methods, Human Factors and Ergonomics
Abstract: While extant research on human-robot-interactions (HRI) has dealt with the examination of different human characteristics, quantifying and describing the various char- acteristics of human diversity remains a challenge for HRI research. This in turn often leads to a disregard of human diversity in the design of HRI. Addressing this issue, we conducted a systematic synthesis of existing models on human diversity, culminating in the development of a model we coined the Human Diversity Wheel for Robotic Interactions (HDWRI). The goal of this model is to provide designers and researchers in HRI with an analytical lens to ensure their work considers different human characteristics. To achieve this, we started to conduct expert interviews with HRI researchers to put our model into a practical context. This paper presents the development of our model, preliminary findings of our first interviews, and an outline of future steps.
|
|
16:10-17:10, Paper TuLBR.11 | |
Revolutionizing Qualitative Human-Robot Interaction Research by Using GPT Models for Inductive Category Development |
|
Arlinghaus, Clarissa Sabrina (Bielefeld University), Wulff, Charlotte (Bielefeld University), Maier, Günter W. (Bielefeld University) |
Keywords: Evaluation Methods
Abstract: Coding qualitative data is essential but time-consuming. This late-breaking report presents a new method for developing inductive categories utilizing GPT models. We examined two different GPT models (gpt-3.5-turbo-0125 and gpt-4o-2024-05-03) and three temperature settings (0, 0.5, 1), each with ten repetitions. The generated categories were fairly consistent across settings, although higher temperatures included less relevant aspects. The agreement for GPT-generated category assignments exceeded that of human coders, with the best performance observed at temperature setting 0. Thus, we recommend using a GPT model with the temperature setting 0 to create and assign inductive categories for qualitative data.
|
|
16:10-17:10, Paper TuLBR.12 | |
Evaluating the Impact of Explainability on the Users' Mental Models of Robots Over Time |
|
Gebellí, Ferran (PAL Robotics), Ros, Raquel (PAL Robotics), Lemaignan, Séverin (PAL Robotics), Garrell, Anais (UPC-CSIC) |
Keywords: Evaluation Methods, Long-term Experience and Longitudinal HRI Studies, Curiosity, Intentionality and Initiative in Interaction
Abstract: To evaluate how explanations affect the users' understanding of robots, researchers typically elicit the user's Mental Model (MM) of the robot and then compare it to the robot's actual decision-making and behaviour. However, the user's self-rating of their level of understanding, which we define as ``user-perceived understanding'', is generally not evaluated. Moreover, this evaluation is typically done only once, while robots are often expected to interact with the same users over long periods. In this work, we suggest a framework to analyse the evolution of the mental models over time across the dimensions of completeness and correctness. We argue that the goal of explainability should be two-fold. On one hand, it should help align the user's perceived understanding with the real one. On the other hand, explainability should enhance the completeness of the mental model to a target level, which varies depending on the user type, while also striving for maximum correctness.
|
|
16:10-17:10, Paper TuLBR.13 | |
Explainable Autonomous Mobile Robots: Interface and Socially Aware Learning |
|
Chandra Shekar, Kiruthiga (New York University), Doma, Pranav (New York University), Prashanth, Chinmay (New York University), Subramaniam, Vikram (New York University), Arab, Aliasghar (NYU) |
Keywords: HRI and Collaboration in Manufacturing Environments, Interaction Kinesics, Motion Planning and Navigation in Human-Centered Environments
Abstract: With the growing use of mobile robots in urban settings, their interaction with people in shared spaces is increasing. It's crucial for these robots to navigate in a socially acceptable manner and interact properly with their environment. AI guides robot behavior by interpreting the environment and controlling actions. Poor explainability of robot decisions can lead to mental and physical risks. Proper robot-to-human interfaces can help mitigate these risks. This paper introduces systems and methods for self-explaining mobile robots in common areas using projected information. We analyze data on people's reactions to the robot, with and without self-explainable features, across various university campus settings like maker-spaces, hallways, cafeterias, and study areas. Initial analysis shows improved acceptance of the robot based on task performance and user questionnaires.
|
|
16:10-17:10, Paper TuLBR.14 | |
Autonomous Robotic Transport System Demonstrates Safety and Efficacy in Blood Sample Transport within Emergency Unit |
|
Lewek, Pawel (Medication Adherence Research Centre, Medical University of Lodz), Kardas, Przemyslaw (Medication Adherence Research Centre, Medical University of Lodz), Lojewska, Ewelina (Medical University of Lodz), Timler, Dariusz (Medical University of Lodz), Chiurazzi, Marcello (The BioRobotics Institute, Scuola Superiore Sant'Anna), Ciuti, Gastone (Scuola Superiore Sant'Anna), Granda Sans, Sofia (Inuetum), Votis, Konstantinos (The Centre for Research and Technology Hellas), Lolis, Vasileios (The Centre for Research and Technology Hellas), Flevarakis, Konstantinos (The Centre for Research and Technology Hellas), SORIANO, Angel (R&D Department Robotnik Automation, SLL 46988 Valencia, SPAIN), Millet Pascual-Leone, Marta (The Robotnik Automation S.L), Julia Ros, Raquel (Robotnik), Solaz Estevan, Víctor (Robotnik Automation S.L), Fico, Giuseppe (Universidad Politécnica de Madrid) |
|
|
16:10-17:10, Paper TuLBR.15 | |
Mochi: Designing a Robot with Exaggerated Squash, Stretch, and Shape Alterations |
|
Iimori, Masato (Tsukuba University), Tanaka, Fumihide (University of Tsukuba) |
Keywords: Innovative Robot Designs, Non-verbal Cues and Expressiveness, Motivations and Emotions in Robotics
Abstract: Animation principles are techniques employed in the animation of characters to create the illusion of life. These principles are employed in the field of human-robot interaction. Two fundamental animation principles that contribute to the generation of this illusion are squash and stretch, as well as exaggeration. Nevertheless, no robot has ever been observed to engage in both simultaneously. This research proposes Mochi, a robot that incorporates the aforementioned animation principles. This paper presents a discussion of the concept, design, and motion of the Mochi robot, as well as its current implementation status and prospects for application and improvement.
|
|
16:10-17:10, Paper TuLBR.16 | |
Object-Centric Spatially-Aware Gesture-Based Motion Specification of Robotic Manipulation Systems |
|
Nazir, Rehan (Santa Clara University), Sharma, Manoj (Santa Clara University), Kitts, Christopher (Santa Clara University) |
|
|
TuDT1 |
Room T1 |
Bridging Trust and Context: Dynamic Interactions in HAI II |
Special Session |
Co-Chair: ren, qiaoqiao | AIRO - IDLab - University of Ghent - IMEC |
|
17:10-17:25, Paper TuDT1.1 | |
Automatic Joint Gaze Generation between Local and Remote Persons through Telepresence Robot (I) |
|
Ikoma, Hibiki (Shizuoka University), Takeuchi, Yugo (Shizuoka University) |
Keywords: Virtual and Augmented Tele-presence Environments, Degrees of Autonomy and Teleoperation, Embodiment, Empathy and Intersubjectivity
Abstract: COVID-19 and technological developments have increased opportunities for remote communication. This has led to a focus on ``telepresence avatar robots,'' a technology that makes people feel as if they are sharing the same place, even from remote locations. The telepresence avatar robot takes the approach of supporting communication by generating a realistic sense of conversation that incorporates technologies such as camera motion and eye contact with the speaker. However, this approach does not enable smooth communication because the robot operator cannot recognize the intention of the local worker's behavior, since the information transmitted from the remote location is less than that available in face-to-face communication. In this study, we focus on gaze-following motion as an unconscious human behavior and implement it in a robot to verify whether a robot operator can perceive the intention of a remote worker through this autonomous motion. Eye-tracking is an action in which one tracks the other's gaze in a local conversation, and it is performed unconsciously. The experimental results suggest that this autonomous behavior can generate an estimation of gazing intention. Consequently, this approach may be a useful method for achieving smoother communication.
|
|
17:25-17:40, Paper TuDT1.2 | |
Unveiling Trust Dynamics with a Mobile Service Robot: Exploring Various Interaction Styles for an Agricultural Task (I) |
|
Elias, Alex (University of Lincoln), Galvez Trigo, Maria Jose (Cardiff University) |
Keywords: HRI and Collaboration in Manufacturing Environments, Creating Human-Robot Relationships, Programming by Demonstration
Abstract: As robotics, particularly in agriculture, become more prevalent, understanding the role that different factors play on the trust levels that users have in these robots becomes crucial to facilitate their adoption and integration into the industry. In this paper we present the results of a within-subjects study that included between-subject factors exploring how prior experience with robotics and different interaction styles with a mobile manipulator robot may affect trust levels in said robot before and after the completion of an agriculture-related manipulation task. The results show that interacting with the robot helps improve trust levels, particularly for those without prior experience with robotics, who present a higher trust improvement score, and that an interaction style involving physical human-robot interaction (pHRI), more specifically Learning by Demonstration, was favoured versus less direct interaction styles. We found that incorporating Text-to-Speech (TTS) can be a good design choice when trying to improve trust, and that the improvement score for trust before and after interaction with the robot was significantly higher for older age groups, with these participants being more conservative with their reported trust level before the interaction. Overall, these results offer insights into different interaction styles and their effect on trust levels for an agriculture-related manipulation task, and open the door to future work exploring further interaction styles and task variations.
|
|
17:40-17:55, Paper TuDT1.3 | |
The Effect of Robot Pose and Distance on Pedestrian and Observer Comfort During Passing (I) |
|
Ohnishi, Fumiya (Keio University), Takahashi, Masaki (Keio University) |
Keywords: Cooperation and Collaboration in Human-Robot Teams, User-centered Design of Robots, Motion Planning and Navigation in Human-Centered Environments
Abstract: To facilitate the integration of autonomous mobile robots into environments shared with humans, ensuring the comfort of pedestrians and observers (individuals observing the interaction from a third-party perspective) is crucial. This study explores the correlation between pedestrian and observer comfort and the robot's distance and pose angle during pedestrian encounters. The findings indicate that when a robot passes a pedestrian after completing its avoidance motion, it enhances the comfort of both pedestrians and observers. In addition, when a robot maintains a large distance from pedestrians while passing by, it improves pedestrian comfort. Furthermore, designing the robot's pose during passing so that it appears to move away from the pedestrian from the observer's perspective further increases the observer's comfort.
|
|
17:55-18:10, Paper TuDT1.4 | |
No More Mumbles: Enhancing Robot Intelligibility through Speech Adaptation |
|
ren, qiaoqiao (AIRO - IDLab - University of Ghent - IMEC), Hou, Yuanbo (Ghent University), Botteldooren, Dick (Ghent University), Belpaeme, Tony (Ghent University) |
|
TuDT2 |
Room T2 |
Assistive Robotics II |
Regular Session |
Chair: Liu, Rui | Kent State University |
Co-Chair: Guo, Yibei | Kent State University |
|
17:10-17:25, Paper TuDT2.1 | |
Accessibility-Aware Reinforcement Learning for Inclusive Robotic Navigation |
|
Hu, Mengsha (Kent State University), Wu, Yunhuan (Carnegie Mellon University), Guo, Yibei (Kent State University), Wu, Yi (Paris-Saclay University), Shi, Chao (Binghamton University), Xu, Lei (Kent State University), Liu, Rui (Kent State University) |
Keywords: Assistive Robotics, Motion Planning and Navigation in Human-Centered Environments, Robot Companions and Social Robots
Abstract: Robotic navigation aids humans in essential scenarios, including airport customer boarding, commercial meeting receptions, and disaster site navigation. While individuals benefit from robotic services, these services pose accessibility challenges for people with disabilities. Robots may face limitations in mobility, vision, hearing, and cognitive reasoning, hindering disabled individuals' access to these services. As robotic services become increasingly prevalent, ensuring they are inclusive and accessible is crucial, accommodating human disabilities with adaptations such as slower movements, standing support, and hazard detection. To address this issue, this study introduces a novel Accessibility-Aware Reinforcement Learning model (ARL). ARL extracts disability-related information from human behavioral observations using a context-embedding neural network. It then adjusts robot motions to provide inclusive assistance to individuals with disabilities. To assess the effectiveness of this approach, four assisting services in navigation scenarios ('lead', 'wait', 'assist', and 'protect') and three application scenarios involving navigating elderly individuals in a museum, through traffic, and across a building patio were designed. The results validate ARL's efficacy in enabling robots to make accessibility-aware decisions, thereby enhancing their support for individuals with disabilities.
|
|
17:25-17:40, Paper TuDT2.2 | |
Co-Designing an Accessible Quadruped Navigation Assistant |
|
Doore, Stacy A. (Colby College), Trikasemsak, Narit (Colby College), Gillespie, Alexandra (Colby College), Giudice, Andrea (Pine Tree Guide Dog Users (PTGDU)), Hata, Rayna (Carnegie Mellon University) |
Keywords: Assistive Robotics, User-centered Design of Robots, Cooperation and Collaboration in Human-Robot Teams
Abstract: While there is no replacement for the learned expertise, devotion, and social benefits of a guide dog, there are scenarios in which a robot navigation assistant could be helpful for individuals with blindness or low vision (BLV). This case study investigated the potential for an industrial quadruped robot to perform guided navigation tasks based on a co-design model. The research was informed by a guide dog handler with over 30 years of experience of non-visual navigation. In order to communicate spatial information between the human-robot team, two interface prototypes were created and pilot tested: a voice-based app and a flexible, responsive guide handle. The pilot user study consisted of sighted participants and our BLV co-designer, who completed simple navigation tasks and a post-study survey about the prototype functionality and their trust in the robot. All participants successfully completed the navigation tasks and demonstrated that the interface prototypes were able to pass spatial information between the human and the robot. Findings of this exploratory study will help to inform human-robot teaming and collaboration. Future work will include expanding the voice-based app to allow the robot to directly communicate obstacles to the handler, adding haptic navigation signals to the handle design, and expanding the user study to include a larger sample of experienced guide dog handlers.
|
|
17:40-17:55, Paper TuDT2.3 | |
Enhancing Functional and Extra Motor Abilities: A Focus Group Study on the Re-Design of an Extra-Robotic Finger |
|
Hendriks, Sjoerd (Chalmers University of Technology), Hasanen, Basma (Khalifa University of Science and Technology), Afzal, Hafiz Malik Naqash (UNSW Sydney), Hussain, Irfan (Khalifa University), Obaid, Mohammad (Chalmers University of Technology) |
Keywords: Assistive Robotics, User-centered Design of Robots
Abstract: The utilization of extra-robotic limbs to regain and extend motor abilities poses challenges. Nonetheless, the evidence suggesting the extension of these abilities without functional interference holds promise for the advancement of extra-robotic fingers or limbs. To gain insights into design improvement from diverse perspectives, we conducted a focus group study utilizing a prototype supernumerary robotic finger (SRF). The main purpose of the study is to gain knowledge on the current prototype, and then use the outcomes in a future co-designing process of SRF having the potential to be used as an assistive and augmented tool. Fifteen healthy subjects with an engineering background participated in the current study. We analyzed the collected data through a thematic analysis process to generate key design recommendations. These recommendations contain suggestions for designing for personal and social acceptance, adaptability, and multi-modal interaction. The results serve as a subset of future subsequent studies to be carried out with other stakeholders of the prototype, including patients, clinicians, and designers. Future work focuses on using the recommendations in a co-design process to inform the prototyping of SRF technologies supporting and extending motor abilities in upper limbs.
|
|
17:55-18:10, Paper TuDT2.4 | |
Enhancing Wheelchair Mobility: Virtual Training System Integrating Lateral Trunk Motion |
|
Callupe Luna, Jhedmar Jhonatan (UVSQ University of Paris-Saclay), Bougherara, Selsabil (Université De Versailles Saint-Quentin-En-Yvelines), Monacelli, Eric (LISV, University of Versailles), Hirata, Yasuhisa (Tohoku University) |
Keywords: Assistive Robotics, Virtual and Augmented Tele-presence Environments, Robots in Education, Therapy and Rehabilitation
Abstract: Following an accident or motor impairment from illness, individuals often experience reduced mobility. When lower limb function is affected, individuals may lose the ability to walk and rely on a wheelchair. Despite maintaining upper limb mobility, emotional and societal factors may reduce motivation to use them. The Volting project aims to address this by developing a wheelchair with improved mobility to help users achieve their maximum potential. To overcome the limitations of traditional wheelchair controls, a device called WISP was developed, allowing control of Volting through trunk movements. This not only frees the upper limbs but also encourages users to engage their bodies in mobility. Initial experiments with a professional wheelchair dancer showed rapid adaptation to WISP usage. However, gradual implementation is recommended for non-athletic wheelchair users due to the risk of injury from unfamiliar movements. Therefore, we present a training system for WISP utilization, combining a virtual environment with haptic feedback to aid learning and adjust difficulty levels. Preliminary results demonstrate significant improvements in task completion time, collisions, and spatial errors after three sessions in our virtual environment. Additionally, trunk movements decreased, indicating more efficient adaptation to WISP use. As the next step, we plan to conduct trials with individuals with disabilities in our virtual training system, followed by integrating WISP with Volting to explore enhanced mobility.
|
|
TuDT3 |
Room T3 |
Ethical Issues in Human-Robot Interaction Research II |
Regular Session |
Chair: Wang, Yixiao | Georgia Institute of Technology |
|
17:10-17:25, Paper TuDT3.1 | |
Robots among Us: An Exploratory Study on the Situated Ethics of Human-Robot Interactions in a Public Library |
|
Chan, Kok Hui Jeffrey (Singapore University of Technology and Design), Jiang, Zhuoqun (Singapore University of Technology and Design), Wang, Yixiao (Georgia Institute of Technology) |
Keywords: Ethical Issues in Human-robot Interaction Research, Storytelling in HRI, Robot Companions and Social Robots
Abstract: Social robots are rarely confined to a task-specific or problem-specific work envelope. For instance, an assistive robot working in a senior center can incidentally become a companion or even a trusted confidante. Increasingly operating in public spaces, social robots can be easily caught in mundane yet morally ambivalent urban situations that are often dynamic, complex, and unpredictable, presenting novel factors and agencies that can quickly exceed the scope of any projected human-robot interaction. Anticipating how a robot should respond in these situations may mitigate interaction failures. To improve foresight in these situations, we propose the frame of “Situated Ethics for Urban Encounters” (SEUE) and probe this frame through 18 mundane yet morally ambivalent scenarios of human-robot interaction in a public library using the speed-dating research method with storyboards. Our findings suggest: (i) There are ambiguous but morally important interactions even in a public library; (ii) Unlike humans, robots are expected to do many things at once and some actions may risk deception; (iii) The importance of children’s rights in these situations; (iv) A need for creatively opportunistic interaction when responding to the ethical challenges of the situation; and (v) Residual risks exist even in a perceived ethical response on the part of the robot. We contribute to empirical HRI ethics by demonstrating how the speed-dating technique in a participatory design setting could be used to unpack, interpret, discuss, and elicit possible HRI responses in urban situations, which are expected to only become more important as robots show up in cities.
|
|
17:25-17:40, Paper TuDT3.2 | |
Rust in Peace: Life-Cycle Management of Companion Robots and Implications for Human Users |
|
Haring, Kerstin Sophie (University of Denver), Laity, Weston (University of Denver) |
Keywords: Ethical Issues in Human-robot Interaction Research, User-centered Design of Robots, Robot Companions and Social Robots
Abstract: The interaction between humans and robots transcends mere utility, venturing into the realm of emotional bonds and perceived lifespans. Through a review of the literature and poignant case studies, this article reveals the intricate language and sentiments expressed by individuals as they navigate the loss of their mechanical companions. This exploration is relevant to understanding the complex social and emotional relationships that can form between humans and robots, emphasizing the need for thoughtful consideration of the life-cycle management of social robots. This research also underscores a significant gap in existing literature and industrial practices: the end-of-life phase of robots and the consequential emotional toll on humans.
|
|
17:40-17:55, Paper TuDT3.3 | |
HRI Wasn't Built in a Day: A Call to Action for Responsible HRI Research |
|
Spitale, Micol (University of Cambridge), Stower, Rebecca (KTH), Parreira, Maria Teresa (Cornell University), Yadollahi, Elmira (KTH), Leite, Iolanda (KTH Royal Institute of Technology), Gunes, Hatice (University of Cambridge) |
Keywords: Ethical Issues in Human-robot Interaction Research
Abstract: In recent years, the awareness of the academy around responsible research has notably increased. For instance, with advances in machine learning and artificial intelligence, recent efforts have been made to promote ethical, fair, and inclusive AI and robotics. However, the field of human-robot-interaction (HRI) is seemingly lagging behind these practices. To better understand if and to what extent HRI is incentivizing researchers to engage in responsible research, we conducted an exploratory review of the publishing guidelines for the most popular HRI conference venues. We identified 18 conferences which published at least 7 HRI papers in 2022. From these, we discuss four themes relevant to conducting responsible HRI research in line with the Responsible Research and Innovation framework: ethical and human participant considerations, transparency and reproducibility, accessibility and inclusion, and plagiarism and LLM use. We identify several gaps and room for improvement within HRI regarding responsible research. Finally, we establish a call to action to provoke novel conversations among HRI researchers about the importance of conducting responsible research within emerging fields like HRI.
|
|
TuDT4 |
Room T4 |
Non-Verbal Cues and Expressiveness II |
Regular Session |
|
17:10-17:25, Paper TuDT4.1 | |
Take It! Exploring Cartesian Features for Expressive Arm Motion |
|
Challa, Ramya (Oregon State University), Sanchez, Luke (Oregon State University), Wilson, Cristina (Oregon State University), Knight, Heather (Oregon State University) |
Keywords: Non-verbal Cues and Expressiveness, Robot Companions and Social Robots, Robots in art and entertainment
Abstract: In this work, a controlled user study explores four Cartesian spatial trajectories (high, low, direct, and from the side arcs), with three unique objects (flower, dagger, water). In mathematics, the Cartesian planes of horizontal, vertical, and forward/back are widely used for path-planning, much like any animal seeking to traverse the surface of the earth. Gravity naturally pulls objects down, and extra force must be exerted to push them upwards. In Laban notation, a dance annotation language, actors and dancers are similarly taught to think of table, door, and wheel planes (direct Cartesian correlates). What is the impact of these cardinal directions on interpretation of robot handovers? In these results, we find that direct paths, water, and flowers are generally seen positively, with daggers and high paths as least friendly. There is also a statistically significant interaction between path and object on robot predictability ratings. In terms of nuance, we also find that elegance is best predicted by path shape. These findings expand prior features of mobile robot expression to arcing planes, finding the up/down degree of freedom to significantly impact human interpretation and experience. Future work can continue such investigations grounded in real world tasks such as assistive robotics (passing a blanket or tissue), or service (retrieving a ticket for a bus).
|
|
17:25-17:40, Paper TuDT4.2 | |
Decoding Engagement: The Role of Closeness Cues in Human-Robot Interactions |
|
Loos, Kira (Bielefeld University), Brandt, Mara (Bielefeld University), Vollmer, Anna-Lisa (Bielefeld University) |
Keywords: Non-verbal Cues and Expressiveness, Social Intelligence for Robots, Monitoring of Behaviour and Internal States of Humans
Abstract: This work examines the critical role of nonverbal cues in initiating successful human-robot interactions (HRI), focusing on the development of an algorithm capable of decoding subtle cues that indicate a human's intention to interact with a robot. The foundation for the analysis is the concept of closeness which relates to the individual feeling of intimacy and comfort, and indicates the engagement potential in diverse settings and interactions. This paper introduces a novel algorithm designed to quantify a person's willingness to interact based on analyzing the displayed closeness on a series of images. The algorithm's efficacy is evaluated through an empirical study involving 20 participants, generating a substantial dataset of 130 video sequences. These sequences were analyzed to assess the algorithm's ability to predict communication initiation intentions, with findings suggesting a significant distinction between participants engaged in interaction with a robot and those instructed to ignore it. The results underscore the algorithm's potential in facilitating more nuanced and socially sustainable HRI, laying the groundwork for future advancements in the field. This research contributes to the growing body of work on social robotics, emphasizing the importance of integrating nonverbal cue analysis for enhancing robots' interactive capabilities and fostering more meaningful human-robot connections.
|
|
17:40-17:55, Paper TuDT4.3 | |
Safe Spot: Exploring Perceived Safety of Dominant vs Submissive Quadruped Robots |
|
Hashimoto, Nanami (Delft University of Technology), Hagens, Emma (Delft University of Technology), Zgonnikov, Arkady (Delft University of Technology), Lupetti, Maria Luce (Politecnico Di Torino) |
Keywords: Non-verbal Cues and Expressiveness, Social Presence for Robots and Virtual Humans, User-centered Design of Robots
Abstract: Unprecedented possibilities of quadruped robots have driven much research on the technical aspects of these robots. However, the social perception and acceptability of quadruped robots so far remain poorly understood. This work investigates whether the way we design quadruped robots' behaviors can affect people’s perception of safety in interactions with these robots. We designed and tested a dominant and submissive personality for the quadruped robot (Boston Dynamics Spot). These were tested in two different walking scenarios (head-on and crossing interactions) in a 2x2 within-subjects study. We collected both behavioral data and subjective reports on participants' perception of the interaction. The results highlight that participants perceived the submissive robot as safer compared to the dominant one. The behavioral dynamics of interactions did not change depending on the robot's appearance. Participants' previous in-person experience with the robot was associated with lower subjective safety ratings but did not correlate with the interaction dynamics. Our findings have implications for the design of quadruped robots and contribute to the body of knowledge on the social perception of non-humanoid robots. We call for a stronger standing of felt experiences in human-robot interaction research.
|
|
17:55-18:10, Paper TuDT4.4 | |
“One Soy Latte for Daniel”: Visual and Movement Communication of Intention from a Robot Waiter to a Group of Customers |
|
Hong, Seung Chan (Monash Universitiy), Tian, Leimin (Monash University), Cosgun, Akansel (Monash University), Kulic, Dana (Monash University) |
Keywords: Curiosity, Intentionality and Initiative in Interaction, Non-verbal Cues and Expressiveness, Applications of Social Robots
Abstract: Service robots are increasingly employed in the hospitality industry for delivering food orders in restaurants. However, in current practice the robot often arrives at a fixed location for each table when delivering orders to different patrons in the same dining group, thus requiring a human staff member or the customers themselves to identify and retrieve each order. This study investigates how to improve the robot’s service behaviours to facilitate clear intention communication to a group of users, thus achieving accurate delivery and positive user experiences. Specifically, we conduct user studies (N=30) with a Temi service robot as a representative delivery robot currently adopted in restaurants. We investigated two factors in the robot’s intent communication, namely visualisation and movement trajectories, and their influence on the objective and subjective interaction outcomes. A robot personalising its movement trajectory and stopping location in addition to displaying a visualisation of the order yields more accurate intent communication and successful order delivery, as well as more positive user perception towards the robot and its service. Our results also showed that individuals in a group have different interaction experiences.
|
|
TuDT5 |
Room T5 |
Programming by Demonstration |
Regular Session |
Chair: Rosa, Francesco | University of Salerno |
Co-Chair: Welle, Michael C. | KTH Royal Institute of Technology |
|
17:10-17:25, Paper TuDT5.1 | |
A Hybrid Approach of No-Code Robot Programming for Agile Production: Integrating Finger-Gesture and Point Cloud |
|
Halim, Jayanto (Fraunhofer Institute for Machine Tools and Forming Technology), Eichler, Paul (Fraunhofer Institute for Machine Tools and Forming Technology IW), Krusche, Sebastian (Fraunhofer IWU), Bdiwi, Mohamad (Fraunhofer Institute for Machine Tools and Forming Technology IW), Ihlenfeldt, Steffen (TU Dresden) |
Keywords: Programming by Demonstration, HRI and Collaboration in Manufacturing Environments, Cooperation and Collaboration in Human-Robot Teams
Abstract: Industrial robot programming necessitates specialized expertise and significant time commitment, particularly for small-batch productions. In response to the escalating demand for production agility, novel approaches have emerged in intuitive robot programming. These inventive systems, rooted in diverse conceptual frameworks, are designed to expedite the deployment of robot systems. A prominent innovation in this domain is adopting no-code robot programming through finger-based gestures. A robot program can be generated by capturing and tracking non-expert users' finger movements and gestures, converting 3D coordinates into an executable robot programming language. However, accurately determining finger positions for 3D coordinates and precise geometrical features presents an ongoing challenge. In pursuit of heightened trajectory precision and reducing more significant effort for the users, we propose a hybrid methodology that amalgamates finger-gesture programming with point cloud data. This synergistic integration demonstrates promising outcomes, substantiating its potential to facilitate the precise and adaptive generation of robot paths within real industrial scenarios.
|
|
17:25-17:40, Paper TuDT5.2 | |
Improving Learning from Visual Demonstration Methods by Target Localization |
|
Foggia, Pasquale (University of Salerno), Rosa, Francesco (University of Salerno), Vento, Mario (University of Salerno) |
Keywords: Programming by Demonstration, Machine Learning and Adaptation
Abstract: This paper presents a novel approach to multi-task visual-guided imitation learning. Upon evaluating the current state-of-the-art method, we observed its capability to replicate the intent of the demonstrator, but with the flaw of manipulating incorrect objects. To address this issue, our study introduces a new approach based on the assumption that explicitly addressing task-relevant problems, such as target object localization, can enhance system performance. Our validation shows that the proposal overtakes the leading method thanks to its ability to drive the robot motion towards the target object.
|
|
17:40-17:55, Paper TuDT5.3 | |
A Robotic Skill Learning System Built Upon Diffusion Policies and Foundation Models |
|
Ingelhag, Nils (KTH Royal Institute of Technology), Munkeby, Jesper (KTH Royal Institute of Technology), van Haastregt, Jonne (KTH Royal Institute of Technology), Varava, Anastasiia (KTH, the Royal Institute of Technology), Welle, Michael C. (KTH Royal Institute of Technology), Kragic, Danica (KTH) |
Keywords: Programming by Demonstration, Social Learning and Skill Acquisition Via Teaching and Imitation, Cooperation and Collaboration in Human-Robot Teams
Abstract: In this paper, we build upon two major recent developments in the field, Diffusion Policies for visuomotor manipulation and large pre-trained multimodal foundational models to obtain a robotic skill learning system. The system can obtain new skills via the behavioral cloning approach of visuomotor diffusion policies given teleoperated demonstrations. Foundational models are being used to perform skill selection given the user's prompt in natural language. Before executing a skill the foundational model performs a precondition check given an observation of the workspace. We compare the performance of different foundational models to this end and give a detailed experimental evaluation of the skills taught by the user in simulation and the real world. Finally, we showcase the combined system on a challenging food serving scenario in the real world. Videos of all experimental executions, as well as the process of teaching new skills in simulation and the real world, are available on the project's website.
|
|
17:55-18:10, Paper TuDT5.4 | |
A Biologically Inspired Program-Level Imitation Approach for Robots: Proof-Of-Concept |
|
Aliasghari, Pourya (University of Waterloo), Ghafurian, Moojan (University of Waterloo), Nehaniv, Chrystopher (University of Waterloo), Dautenhahn, Kerstin (University of Waterloo) |
Keywords: Social Learning and Skill Acquisition Via Teaching and Imitation, Programming by Demonstration, Computational Architectures
Abstract: For social robots to succeed in places such as homes, they must learn new skills from various people and act in a manner desirable to different users. We introduce a novel biologically inspired approach for robot learning through program-level imitation, inspired by the way primates, including humans, understand and perform complex actions. Our approach enables robots to discover the hierarchical structure of tasks by identifying sequential regularities and sub-goals from diverse human demonstrations. To do so, human-provided demonstrations, which can be obtained by a robot through different modalities (such as kinesthetic teaching, behavioural observation, and verbal instruction), are processed by an algorithm that discovers multiple possibilities for arranging observed sub-goals to achieve a final goal. Prior to acting, the available sequences are evaluated based on user-defined criteria, through mental simulation of the task by the robot, to find the optimal sequence of actions. As a proof-of-concept, we implemented our system on an iCub humanoid robot and present here how our method allowed the robot to adapt its action sequences for task execution when starting the task from different states, incorporating user preference for finishing the task as fast as possible. Our envisaged system is meant to accommodate variations in human teaching styles and is expected to help a robot perform tasks with greater flexibility and efficiency. This work contributes by proposing a framework for robots to learn from humans at an abstract level, opening the way to more adaptable and intelligent robotic assistants in everyday tasks.
|
|
TuDT6 |
Room T6 |
Personalities and Behaviors |
Regular Session |
Co-Chair: Nardelli, Alice | University of Genoa |
|
17:10-17:25, Paper TuDT6.1 | |
Introducing a Note of Levity to Human-Robot Interaction with Dialogs Containing Irony, Sarcasm and Jocularity |
|
Sievers, Thomas (University of Lübeck), Russwinkel, Nele (Universität Zu Lübeck) |
Keywords: Personalities for Robotic or Virtual Characters, Social Touch in Human–Robot Interaction, Cooperation and Collaboration in Human-Robot Teams
Abstract: Dialogs between humans and social robots are often fact-oriented and not very mood-dependent, at least on the part of the robot. This raises the question of what influence the tonality of the utterances has on the perceived personality of the robot and how this influences the user's further expectations. One element that is not necessarily expected from robots in conversation is humor, especially the use of irony and sarcasm. We examined the effect of a conversation with cheerfully ironic or sullenly sarcastic remarks by the robot in contrast to utterances in a more seriously neutral tone. To measure these effects in terms of perceived warmth, competence and possible discomfort in conversation we used the Robotic Social Attributes Scale (RoSAS) on 20 participants with an age ranging from 26 to 58. In addition, we asked the participants for their assessment of the suitability of the robot's dialog style for different areas of application and which dialog style they personally liked best. The dialog parts of the robot were generated with ChatGPT using suitable prompts. Our results showed that the perceived warmth gained highest rates when when the robot hat a cheerfully ironic tone. Discomfort was rated significantly higher when a sullenly sarcastic tone was used by the robot. Perceived competence seemed to be slightly negatively influenced in the sarcastic condition. The results proved that tonality appears to be a very relevant design element for a successful interaction between humans and robots, but must be used carefully to achieve the desired goal.
|
|
17:25-17:40, Paper TuDT6.2 | |
Personality and Memory-Based Framework for Emotionally Intelligent Agents |
|
Nardelli, Alice (University of Genoa), Maccagni, Giacomo (Reply S.p.A), Minutoli, Federico (Reply S.p.A), Sgorbissa, Antonio (University of Genova), Recchiuto, Carmine Tommaso (University of Genova) |
Keywords: Personalities for Robotic or Virtual Characters, Cognitive Skills and Mental Models, Social Intelligence for Robots
Abstract: The goal-directed behavior observed in humans arises from the intricate interplay of various processes, including personality dynamics, emotional responses to others, memory encoding, the anticipation of future actions, and associated hedonic experiences. Integrating these multiple processes characteristic of human intelligence into a robotic framework aims to enhance the human-likeness of artificial agents and facilitate more natural and intuitive interactions with humans. For this purpose, in this paper, we propose a comprehensive psychological and cognitive architecture where, personality, as it happens for humans, not only influences the execution of actions but also shapes internal reactions to human emotions and guides anticipatory decision-making processes tailored to the agent's traits. We demonstrate the framework's effectiveness in generating perceivable synthetic personalities through an experiment involving participants in a dyadic conversation scenario with a digital human, where the digital human's behavior is driven by its assigned personality. The results show that participants accurately perceive the artificial personality displayed by the digital human. We also demonstrate the potential of our robotic framework to bridge the gap between cognitive and psychological agents, as the findings highlight its ability to create a cognitively and emotionally intelligent digital human.
|
|
17:40-17:55, Paper TuDT6.3 | |
Planning with Critical Decision Points: Robots That Influence Humans to Infer Their Strategy |
|
Ghose, Debasmita (Yale University), Lewkowicz, Michal (Yale University), Dong, David (Yale University), Cheng, Andy (Yale University), Doan, Tran (Yale University), Adams, Emma (Yale University), Vázquez, Marynel (Yale University), Scassellati, Brian (Yale) |
Keywords: Monitoring of Behaviour and Internal States of Humans
Abstract: To enable sophisticated interactions between humans and robots in a shared environment, robots must infer the intentions and strategies of their human counterparts. This inference can provide a competitive edge to the robot or enhance human-robot collaboration by reducing the necessity for explicit communication about task decisions. In this work, we identify specific states within the shared environment, which we refer to as Critical Decision Points, where the actions of a human would be especially indicative of their high-level strategy. A robot can significantly reduce uncertainty regarding the human's strategy by observing actions at these points. To demonstrate the practical value of Critical Decision Points, we propose a Receding Horizon Planning (RHP) approach for the robot to influence the movement of a human opponent in a competitive game of hide-and-seek in a partially observable setting. The human plays as the hider and the robot plays as the seeker. We show that the seeker can influence the hider to move towards Critical Decision Points, and this can facilitate a more accurate estimation of the hider's strategy. In turn, this helps the seeker catch the hider faster than estimating the hider's strategy whenever the hider is visible or when the seeker only optimizes for minimizing its distance to the hider.
|
|
17:55-18:10, Paper TuDT6.4 | |
Human Stress Response and Perceived Safety During Encounters with Quadruped Robots |
|
Gupta, Ryan (University of Texas at Austin), Shin, Hyonyoung (The University of Texas at Austin), Norman, Emily (University of Texas at Austin), Stephens, Keri (The University of Texas at Austin), Lu, Nanshu (The University of Texas at Austin), Sentis, Luis (The University of Texas at Austin) |
Keywords: Monitoring of Behaviour and Internal States of Humans
Abstract: Despite the rise of mobile robot deployments in home and work settings, perceived safety of users and bystanders is understudied in the human-robot interaction (HRI) literature. To address this, we present a study designed to identify elements of a human-robot encounter that correlate with observed stress response. Stress is a key component of perceived safety and is strongly associated with human physiological response. In this study a Boston Dynamics Spot and a Unitree Go1 navigate autonomously through a shared environment occupied by human participants wearing multimodal physiological sensors to track their electrocardiography (ECG) and electrodermal activity (EDA). The encounters are varied through several trials and participants self-rate their stress levels after each encounter. The study resulted in a multidimensional dataset archiving various objective and subjective aspects of a human-robot encounter, containing insights for understanding perceived safety in such encounters. To this end, acute stress responses were decoded from the human participants' ECG and EDA and compared across different human-robot encounter conditions. Statistical analysis of data indicate that on average (1) participants feel more stress during encounters compared to baselines, (2) participants feel more stress encountering multiple robots compared to a single robot and (3) participants stress increases during navigation behavior compared with search behavior.
|
| |