Tele-sense

DURATION

2023.09 ~ Present
2 years 4 months
(ongoing)

KEY WORDS

HRI
Telehpatic

RESEARCH

Thelehaptic

Contribution

Team Leader
Engineering
Design Research

CO-WORKER

Jiyoon Lee
Sungmo Lee
Seungwon Hu

Tele-operated avatar robots serve as a physical extension of the operator, facilitating active and meaningful communication with local users. Recognizing that non-verbal cues are critical for shaping user experience, our system is engineered to deliver rich expressive behaviors. It integrates HMD-based eye tracking to map the operator’s dynamic gaze in real-time, a tele-haptics interface specialized for social touch, and an AI-driven voice modulation system to maintain a consistent character persona. By synchronizing these multimodal signals, the system significantly enhances the naturalness of the interaction, thereby increasing user immersion and engagement.

Indendent Research

|

2023.09 - Present

Background

With the increasing adoption of robotic

technologies in healthcare, interest has grown in leveraging robots to support pediatric patients

through rehabilitation and emotional interaction.

Problem

However, most existing teleoperated robots remain focused on simple, conversation-based interactions, with limited research on more complex tasks such as grasping or object handover.

Solution

This project aims to develop a tele-haptic robot system that enables rich and diverse remote interactions between medical professionals and pediatric patients.

Key Interaction

Improve Nonverbal Communication

Paralinguistic
Voice Expression

Eye Gaze

Interaction

Embodied

Gesture Movement

Tele-haptic
Social Touch

User Experience

Paralinguistic Voice Expression

In interactions with a teleoperated robot, we considered that users are not engaging with a human operator, but with a robot character.

Therefore, the operator’s voice is not transmitted as-is; instead, it is transformed to create the experience of the robot character speaking directly.

Eye Gaze Interaction

The remote operator’s gaze was directly mapped onto the robot’s eyes.
As a result, the robot’s eye movements reflect the operator’s real-time visual attention, enabling natural and expressive non-verbal interaction.

Embodied Gesture Movement

The robot’s upper-body movements are controlled through the user’s physical motions, using inverse kinematics and hand tracking.
Both the neck and arms are mapped to the user’s movements, allowing the robot to mirror the operator’s upper-body behavior in real time.

Tele-haptic Social Touch

A haptic interface is used to enable social touch during interaction.
To ensure safety and precision when engaging with a user face-to-face through the robot, torque generated

while grasping objects with the gripper is fed back to the operator as pressure-based haptic feedback.

User Scenario

Reference

Hoseok Jung, et al. 2026. Affective Robotic Avatar: Enhancing Social Telepresence via Haptics and Expressive Cues. In Proceedings of HCI International 2026 (HCII ‘26).
(Invited Full Paper, Session: Frontiers of HCI in XR).

Achievement

HCII '26 Invited Full Paper, Session: Frontiers of HCI in XR

Hoseok Jung

© 2025 Hoseok Jung. All Rights Reserved.