Tele-sense

DURATION

2023.09 ~ Present
2 years 4 months
(ongoing)

KEY WORDS

HRI
Telehpatic

Contribution

Team Leader
Haptic System Design
Robot Engineering

Tele-operated avatar robots serve as a physical extension of the operator, facilitating active and meaningful communication with local users. Recognizing that non-verbal cues are critical for shaping user experience, our system is engineered to deliver rich expressive behaviors. It integrates HMD-based eye tracking to map the operator’s dynamic gaze in real-time, a tele-haptics interface specialized for social touch, and an AI-driven voice modulation system to maintain a consistent character persona. By synchronizing these multimodal signals, the system significantly enhances the naturalness of the interaction, thereby increasing user immersion and engagement.

Indendent Research

|

2023.09 - Present

Outcome

Achievement

[Paper] Hoseok Jung, et al. 2026. Affective Robotic Avatar: Enhancing Social Telepresence via Haptics and Expressive Cues. In Proceedings of HCI International 2026 (HCII ‘26). (Invited Full Paper, Session: Frontiers of HCI in XR)

Background

With the increasing adoption of robotic

technologies in healthcare, interest has grown in leveraging robots to support pediatric patients

through rehabilitation and emotional interaction.

Problem

However, most existing teleoperated robots remain focused on simple, conversation-based interactions, with limited research on more complex tasks such as grasping or object handover.

Solution

This project aims to develop a tele-haptic robot system that enables rich and diverse remote interactions between medical professionals and pediatric patients.

Key Interaction

Improve Nonverbal Communication

Paralinguistic
Voice Expression

Eye Gaze

Interaction

Embodied

Gesture Movement

Tele-haptic
Social Touch

User Experience

Paralinguistic Voice Expression

In interactions with a teleoperated robot, we considered that users are not engaging with a human operator, but with a robot character.

Therefore, the operator’s voice is not transmitted as-is; instead, it is transformed to create the experience of the robot character speaking directly.

Eye Gaze Interaction

The remote operator’s gaze was directly mapped onto the robot’s eyes.
As a result, the robot’s eye movements reflect the operator’s real-time visual attention, enabling natural and expressive non-verbal interaction.

Embodied Gesture Movement

The robot’s upper-body movements are controlled through the user’s physical motions, using inverse kinematics and hand tracking.
Both the neck and arms are mapped to the user’s movements, allowing the robot to mirror the operator’s upper-body behavior in real time.

Tele-haptic Social Touch

A haptic interface is used to enable social touch during interaction.
To ensure safety and precision when engaging with a user face-to-face through the robot, torque generated

while grasping objects with the gripper is fed back to the operator as pressure-based haptic feedback.

User Scenario

User Study

Due to the small sample size, no statistical significance testing was conducted. However, the mean score comparison showed no clear differences between gaze conditions in social presence and intimacy, suggesting that eye movement alone is insufficient to substantially alter users’ sense of co-presence or bonding.


In contrast, immersion and emotional sharing scores were higher under the static gaze condition than under the dynamic, human-like gaze condition. This indicates that continuous gaze movement may introduce additional cognitive load in attention-demanding contexts such as counseling, or that slight delays and mechanical inconsistencies in dynamic gaze synchronization may negatively affect emotional engagement.


Overall, these findings suggest that in remote counseling scenarios, stable and consistent gaze behavior may better support user immersion and emotional connection than mimicking human-like eye movements, highlighting that greater behavioral realism does not always lead to improved interaction experiences in robot design.

Role and Responsibility

Hoseok Jung (Research Scientist, SCIGC): Team Leader & Robot Engineering.

Jiyoon Lee (Visual Communication Design, Hongik Univ.): Software Engineering.

Sungmo Lee (Mechanical System Design Eng., Hongik Univ.): Mechanical Design.

Boyoung Kim (Research Professor, George Mason Univ.) User Study Design.
Okkeun Lee (Postdoc, Civil & Environmental Eng., Stanford Univ.): System Evaluation.

© 2026 Hoseok Jung. All Rights Reserved.