Author: Denis Avetisyan
New research shows robots can accurately gauge human comfort levels during social interaction by analyzing subtle cues in a user’s gaze.

Eye-tracking data reveals that minimal pupil diameter reliably indicates comfort during human-robot proxemics, enabling more adaptive and user-friendly social robots.
Establishing comfortable interpersonal distances is crucial for successful social interaction, yet remains a challenge for robots navigating human spaces. This study, ‘SensHRPS: Sensing Comfortable Human-Robot Proxemics and Personal Space With Eye-Tracking’, investigates how to estimate user comfort during interactions with the humanoid robot Ameca using mobile eye-tracking and machine learning. Contrary to expectations from human-human interaction research, a simple Decision Tree classifier-driven by minimal pupil diameter-most accurately predicted comfort levels. These findings suggest distinct physiological responses to robotic versus human approaches, raising the question of how to best calibrate robot behavior to foster truly natural and comfortable interactions.
Reading Between the Lines: Why Robots Need to Pay Attention
Truly effective human-robot interaction transcends simple task completion, demanding a nuanced understanding of how individuals respond to robotic presence and behavior. Comfort and discomfort aren’t always expressed through explicit statements; instead, they often manifest as subtle shifts in physiology and demeanor. A robot capable of discerning these cues – a fleeting micro-expression, a slight change in posture, or even variations in speech patterns – can dynamically adjust its actions to foster a more positive and productive collaboration. This sensitivity is crucial because a robot that misinterprets a user’s discomfort risks creating a negative experience, hindering the interaction and potentially triggering the unsettling effects of the uncanny valley. Ultimately, successful interaction isn’t about a robot performing tasks, but about it understanding the human it’s working with and adapting to ensure a comfortable and collaborative experience.
Existing approaches to gauging human comfort during robotic interaction frequently overlook the intricate connection between a robot’s behavior and a user’s involuntary physiological shifts. While questionnaires and self-reporting offer some insight, they are susceptible to bias and fail to capture real-time, subconscious responses. Research indicates that subtle changes in pupil diameter – a phenomenon linked to cognitive load, emotional arousal, and attentional focus – can serve as a sensitive indicator of a user’s state. However, current systems struggle to accurately interpret these nuanced signals amidst the complexities of natural human-robot exchanges, often mistaking momentary fluctuations for broader discomfort or disengagement. This limitation hinders the development of truly adaptive robots capable of dynamically adjusting their actions to foster positive and comfortable interactions, ultimately impacting the potential for widespread acceptance and integration of robotic technologies into daily life.
The notorious “uncanny valley” posits that as robots become more human-like, our emotional response shifts from positive empathy to revulsion, a phenomenon hindering seamless human-robot collaboration. Successfully traversing this valley isn’t simply about improving a robot’s physical resemblance to a human; it demands a nuanced understanding of the user’s real-time psychological and physiological state. Research indicates that subtle cues – micro-expressions, changes in heart rate, and even pupil dilation – offer critical insights into whether a robot’s actions are perceived as comforting or unsettling. Precise assessment of these states during interaction allows for dynamic adjustments to a robot’s behavior, enabling it to modulate its movements, vocalizations, and overall presentation to maintain a comfortable and productive relationship with the user, ultimately avoiding the negative emotional response inherent in the uncanny valley.

Beyond Self-Reporting: Quantifying Comfort with Sensors
Biometric sensing, and specifically eye tracking, provides an objective method for quantifying user comfort during human-robot interaction. Traditional comfort assessments rely on subjective self-reporting, which is prone to bias and may not capture nuanced physiological responses. Eye tracking systems utilize infrared illumination and image processing to precisely determine a user’s gaze position and pupillary responses. These data points are then analyzed to derive metrics indicative of cognitive effort and emotional arousal, offering a quantifiable, real-time assessment of user comfort levels during interaction with social robots. This approach moves beyond self-reported data, providing a more reliable and granular understanding of the user experience.
Eye tracking technology provides quantifiable metrics indicative of a user’s cognitive and emotional responses. Specifically, Pupil Diameter variations are directly associated with cognitive load, with increases generally correlating to heightened mental effort or arousal. Simultaneously, Gaze Point Skewness – a measure of the distribution of gaze positions – provides insight into attentional focus and emotional valence; deviations from a central gaze can indicate disengagement, negative affect, or specific points of interest. These parameters are captured via infrared sensors and analyzed using algorithms to provide objective data regarding user state, enabling a nuanced understanding beyond self-reported measures.
The proposed system leverages real-time analysis of biometric data – specifically, metrics derived from eye tracking such as pupil diameter and gaze point skewness – to modulate robot behavior. This adaptive capability is achieved through a feedback loop where changes in a user’s physiological signals trigger adjustments in the robot’s actions, including speech rate, proxemics, and the complexity of social cues. The goal is to minimize indicators of cognitive load or negative emotional response, as determined by established correlations between biometric measurements and user state. This dynamic adaptation aims to maintain an optimal level of user comfort throughout the human-robot interaction, effectively personalizing the robot’s behavior based on individual physiological responses.

Sorting Signal from Noise: Machine Learning and Comfort Prediction
The investigation encompassed four machine learning classifiers – Decision Tree, Random Forest, Support Vector Machine, and Transformer Model – to assess their efficacy in predicting user comfort levels. These algorithms were selected due to their established performance in classification tasks and differing approaches to data analysis; Decision Trees utilize a hierarchical structure, Random Forests employ ensemble learning with multiple decision trees, Support Vector Machines map data to high-dimensional space, and Transformer Models leverage self-attention mechanisms. The selection aimed to provide a comparative analysis of model performance based on the provided dataset, with the goal of identifying the most accurate predictor of user comfort.
Leave-One-Subject-Out Cross-Validation (LOSOCV) was employed as the primary evaluation metric to assess model generalization performance. This technique involves iteratively training the model on data from all participants except one, and then testing its predictive accuracy on the held-out participant’s data. This process is repeated for each participant in the dataset, ensuring that each participant’s data is used for both training and testing exactly once. The resulting performance metrics are then averaged across all participants to provide a robust estimate of the model’s ability to generalize to unseen data, mitigating the risk of overfitting to the specific characteristics of the training set and providing a more realistic assessment of real-world performance.
Evaluation of machine learning models for user comfort prediction using eye-tracking data yielded an F1-score of 0.73 with a Decision Tree model. This indicates a substantial ability to accurately classify comfort levels based on observed eye movements. Comparative analysis reveals that a prior investigation employing a Transformer model achieved a higher F1-score of 0.87; however, it’s important to note that this prior study was conducted with a smaller participant pool. These results suggest the potential for robust comfort prediction, with performance varying based on model selection and sample size.

Putting it All Together: Responsive Robots in the Real World
The integration of a novel comfort prediction system with the social humanoid robot, Ameca, represents a significant step towards creating truly responsive and engaging conversational partners. This system doesn’t simply dictate robotic behavior, but dynamically adjusts Ameca’s interactions based on predicted user comfort levels. By analyzing various factors, the system allows Ameca to subtly modify its gaze, posture, and even the pacing of dialogue, fostering a more positive and productive collaboration. The result is a robot capable of tailoring its responses and physical expressions to individual preferences, moving beyond pre-programmed routines to create a more human-like and comfortable conversational experience for those interacting with it.
Ameca, a social humanoid robot, achieves remarkably fluid and contextually relevant dialogue through the integration of the Gemma 3 language model, efficiently deployed via Ollama. This powerful combination doesn’t operate in isolation; instead, it’s dynamically linked to a comfort prediction system. This system continuously assesses user responses and subtly adjusts Ameca’s behavior – from conversational tone and pacing to gaze direction and even micro-expressions – to optimize the interaction for individual comfort. The result is a robot capable of not just responding to users, but of proactively adapting to create a more positive and engaging conversational experience, moving beyond pre-programmed scripts towards genuinely responsive social interaction.
The integration of a comfort prediction system with the social humanoid robot Ameca enables a dynamic adjustment of interaction styles based on individual user preferences, resulting in demonstrably more positive experiences. Statistical analysis, specifically an ANOVA, confirmed a significant relationship between physical distance and perceived comfort ($p$-value = 0.029, effect size $\eta^2$ = 0.15), indicating that Ameca can effectively modify its proxemics – the use of space – to enhance user comfort levels. This responsiveness suggests a potential for building more natural and engaging human-robot interactions, where the robot proactively adapts to create a more comfortable and positive social dynamic for the user.
The pursuit of seamless human-robot interaction, as detailed in this research on proxemics and comfort estimation, feels predictably fragile. This work, focusing on gaze features like minimal pupil diameter as comfort indicators, is a clever attempt to quantify something inherently subjective. However, one suspects that even the most meticulously calibrated algorithm will eventually encounter a user who defies prediction. As Edsger W. Dijkstra observed, “Simplicity is prerequisite for reliability.” The drive for increasingly complex interaction models-mapping pupil dilation to comfort levels-risks creating a system too brittle to withstand the messy realities of human behavior. The elegance of the approach is undeniable, yet the inevitable encounter with unpredictable human responses looms large; every abstraction, however beautifully engineered, dies in production.
Beyond the Gaze: The Inevitable Complications
The correlation between pupillary response and perceived comfort, as demonstrated, is…neat. It will undoubtedly become a standard feature in the next generation of social robots. However, the field seems remarkably quick to declare victory on metrics easily captured in controlled lab settings. The transition to uncontrolled environments, populated by individuals who demonstrably enjoy being startled, will reveal the limitations of this approach. Comfort, it turns out, is rarely about physiological responses alone, and much more about expectation, cultural context, and the individual’s tolerance for awkwardness.
Future work will inevitably focus on multi-modal integration – combining gaze data with facial expressions, body language, and perhaps even vocal tone. This will, of course, simply shift the problem from one set of noisy signals to another. The true challenge isn’t building a more comprehensive comfort ‘detector’, but accepting that perfect prediction is an illusion. Robots will, at some point, misread signals, invade personal space, and generally behave inappropriately. The art will then be in designing graceful recovery mechanisms, not flawless prevention.
One anticipates a surge in ‘comfort metrics’ benchmarks. These will be carefully constructed to demonstrate progress, while simultaneously obscuring the inherent messiness of human interaction. It’s a familiar pattern. The core problem isn’t whether a robot can sense comfort, but whether anyone will remember what ‘comfort’ actually feels like when reviewing the logs after the inevitable incident.
Original article: https://arxiv.org/pdf/2512.08518.pdf
Contact the author: https://www.linkedin.com/in/avetisyan/
See also:
- Clash Royale Best Boss Bandit Champion decks
- Best Hero Card Decks in Clash Royale
- Clash Royale Witch Evolution best decks guide
- Clash Royale December 2025: Events, Challenges, Tournaments, and Rewards
- Ireland, Spain and more countries withdraw from Eurovision Song Contest 2026
- JoJo’s Bizarre Adventure: Ora Ora Overdrive unites iconic characters in a sim RPG, launching on mobile this fall
- Cookie Run: Kingdom Beast Raid ‘Key to the Heart’ Guide and Tips
- Clash of Clans Meltdown Mayhem December 2025 Event: Overview, Rewards, and more
- ‘The Abandons’ tries to mine new ground, but treads old western territory instead
- Best Builds for Undertaker in Elden Ring Nightreign Forsaken Hollows
2025-12-10 09:55