Natural language provides an intuitive and accessible way for humans to communicate with robots, fostering more natural and flexible interaction across a range of tasks. This study investigates how effectively users can command a robot using natural language within a simulated environment. By employing Gemini Flash 2.0 as the underlying Large Language Model (LLM) to interpret and translate user prompts into executable plans, we explore both the strengths and limitations of this approach. The experiments evaluated user-generated prompts across multiple predefined tasks, revealing a spectrum of outcomes - from successful task completions to errors such as misinterpretations, spatial failures, and hallucinated behaviors where the robot acted on non-existent information. The results highlight how different communication strategies, combining directive and conversational phrasing, influenced task performance. This work contributes to advancing Human-Robot Interaction (HRI) design by emphasizing the potential of LLM-powered systems while addressing the challenges of ambiguity and error resilience in user-driven command structures.
Olaiya, K., Delnevo, G., Ceccarini, C., Lam, C.-T., Pau, G., Salomoni, P. (2025). Natural Language and LLMs in Human-Robot Interaction: Performance and Challenges in a Simulated Setting. 345 E 47TH ST, NEW YORK, NY 10017 USA : Institute of Electrical and Electronics Engineers Inc. [10.1109/ICHORA65333.2025.11016850].
Natural Language and LLMs in Human-Robot Interaction: Performance and Challenges in a Simulated Setting
Olaiya K.;Delnevo G.;Ceccarini C.;Pau G.;Salomoni P.
2025
Abstract
Natural language provides an intuitive and accessible way for humans to communicate with robots, fostering more natural and flexible interaction across a range of tasks. This study investigates how effectively users can command a robot using natural language within a simulated environment. By employing Gemini Flash 2.0 as the underlying Large Language Model (LLM) to interpret and translate user prompts into executable plans, we explore both the strengths and limitations of this approach. The experiments evaluated user-generated prompts across multiple predefined tasks, revealing a spectrum of outcomes - from successful task completions to errors such as misinterpretations, spatial failures, and hallucinated behaviors where the robot acted on non-existent information. The results highlight how different communication strategies, combining directive and conversational phrasing, influenced task performance. This work contributes to advancing Human-Robot Interaction (HRI) design by emphasizing the potential of LLM-powered systems while addressing the challenges of ambiguity and error resilience in user-driven command structures.| File | Dimensione | Formato | |
|---|---|---|---|
|
Natural Language and LLMs in Human-Robot Interaction Performance and Challenges in a Simulated Setting.pdf
embargo fino al 02/06/2027
Tipo:
Postprint / Author's Accepted Manuscript (AAM) - versione accettata per la pubblicazione dopo la peer-review
Licenza:
Licenza per accesso libero gratuito
Dimensione
520.89 kB
Formato
Adobe PDF
|
520.89 kB | Adobe PDF | Visualizza/Apri Contatta l'autore |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.



