The recent developments in the research area of human-robot interaction and gesture synthesis are pushing the boundaries of what is possible with current technology. There is a notable shift towards leveraging large language models (LLMs) and vision language models (VLMs) to enhance the capabilities of robots, particularly in generating natural and contextually appropriate gestures. This trend is evident in the integration of LLMs for scalable and controllable co-speech gesture synthesis, as well as the use of VLMs for whole-body motion generation from language descriptions. Additionally, there is a growing interest in enhancing motion variation in text-to-motion models by conditioning them on additional modalities such as video clips or images, which allows for the creation of more diverse and realistic motions. The field is also seeing advancements in unsupervised learning methods for vision-language-action models, which can leverage large-scale video data without the need for ground-truth action labels. These innovations collectively aim to make robots more intuitive and responsive in human environments, bridging the gap between human communication and robotic action.
Noteworthy papers include 'LLM Gesticulator: Leveraging Large Language Models for Scalable and Controllable Co-Speech Gesture Synthesis,' which pioneers the use of LLMs in co-speech gesture generation, and 'Harmon: Whole-Body Motion Generation of Humanoid Robots from Language Descriptions,' which demonstrates the potential of VLMs in creating natural and expressive humanoid motions.