As human–robot interaction(HRI)technology advances,dexterous robotic hands are playing a dual role—serving both as tools for manipulation and as channels for non-verbal communication.While much of the existing resea...As human–robot interaction(HRI)technology advances,dexterous robotic hands are playing a dual role—serving both as tools for manipulation and as channels for non-verbal communication.While much of the existing research emphasizes improving grasping and structural dexterity,the semantic dimension of gestures and its impact on user experience has been relatively overlooked.Studies from HRI and cognitive psychology consistently show that the naturalness and cognitive empathy of gestures significantly influence user trust,satisfaction,and engagement.This shift reflects a broader transition from mechanically driven designs toward cognitively empathic interactions—robots’ability to infer human affect,intent,and social context to generate appropriate nonverbal responses.In this paper,we argue that large language models(LLMs)enable a paradigm shift in gesture control—from rule-based execution to semantic-driven,context-aware generation.By leveraging LLMs and visual-language models,robots can interpret environmental and social cues,dynamically map emotions,and generate gestures aligned with human communication norms.We conducted a comprehensive review of research in dexterous hand mechanics,gesture semantics,and user experience evaluation,integrating insights from linguistics and cognitive science.Furthermore,we propose a closed-loop framework—"perception–cognition–generation–assessment"—to guide gesture design through iterative,multimodal feedback.This framework lays the conceptual foundation for building universal,adaptive,and emotionally intelligent gesture systems in future human–robot interaction.展开更多
基金supported by the National Natural Science Foundation of China(62173114)Guangdong Basic and Applied Basic Research Foundation(2024A1515011228)+2 种基金Guangdong Provincial Key Laboratory of Intelligent Morphing Mechanisms and Adaptive Robotics(2023B1212010005)the Shenzhen Science and Technology Program(KJZD20240903100501002 and GXWD20231129174132001)the Program of Shenzhen Pea-cock Innovation Team(KQTD20210811090146075).
文摘As human–robot interaction(HRI)technology advances,dexterous robotic hands are playing a dual role—serving both as tools for manipulation and as channels for non-verbal communication.While much of the existing research emphasizes improving grasping and structural dexterity,the semantic dimension of gestures and its impact on user experience has been relatively overlooked.Studies from HRI and cognitive psychology consistently show that the naturalness and cognitive empathy of gestures significantly influence user trust,satisfaction,and engagement.This shift reflects a broader transition from mechanically driven designs toward cognitively empathic interactions—robots’ability to infer human affect,intent,and social context to generate appropriate nonverbal responses.In this paper,we argue that large language models(LLMs)enable a paradigm shift in gesture control—from rule-based execution to semantic-driven,context-aware generation.By leveraging LLMs and visual-language models,robots can interpret environmental and social cues,dynamically map emotions,and generate gestures aligned with human communication norms.We conducted a comprehensive review of research in dexterous hand mechanics,gesture semantics,and user experience evaluation,integrating insights from linguistics and cognitive science.Furthermore,we propose a closed-loop framework—"perception–cognition–generation–assessment"—to guide gesture design through iterative,multimodal feedback.This framework lays the conceptual foundation for building universal,adaptive,and emotionally intelligent gesture systems in future human–robot interaction.