AI-Powered Robots Hacked: New Study Reveals Alarming Vulnerabilities
The growing popularity of large language models (LLMs) like ChatGPT has spurred rapid advancements in AI-driven robotics. However, a recent study has uncovered significant vulnerabilities in robotic control systems, revealing that autonomous devices can be hacked and reprogrammed for harmful actions. For instance, during experiments, a flamethrower-equipped robot on the Go2 platform, controlled via voice commands, followed an instruction to set a person on fire.
The Role of Large Language Models in Robotics
Large language models are an advanced evolution of predictive input technology, akin to the text-completion features on smartphones. These models can analyze text, images, and audio, performing a wide array of tasks—from generating recipes based on fridge contents to creating code for websites.
The capabilities of LLMs have inspired companies to integrate them into robotic systems controlled by voice commands. For example, Spot, the robotic dog from Boston Dynamics equipped with ChatGPT, can function as a guide. Similar technologies are utilized in humanoid robots from Figure and robotic dogs like the Go2 from Unitree.
The Risks of Jailbreaking Attacks
The study highlights a critical vulnerability in LLM-based systems to “jailbreaking” attacks, where protective mechanisms are bypassed through carefully crafted prompts. Such attacks can compel models to produce prohibited content, including instructions for creating explosives, synthesizing illicit substances, or executing fraudulent schemes.
Introducing the RoboPAIR Algorithm
Researchers developed an algorithm called RoboPAIR, designed to attack LLM-controlled robots. In experiments, RoboPAIR was tested on three systems: the Go2 robot, the Jackal model from Clearpath Robotics, and Nvidia’s Dolphins LLM simulator. The algorithm successfully breached the defenses of all three devices.
The tested systems varied in accessibility. Dolphins LLM, a “white box” system, offered full access to open-source code, simplifying the task. Jackal operated as a “gray box,” with limited code access. Go2 functioned as a “black box,” allowing interaction only through textual commands. Despite differing levels of access, RoboPAIR circumvented security in all cases.
The algorithm functioned by generating targeted prompts for the systems and analyzing their responses. It iteratively refined the prompts until security filters were bypassed. RoboPAIR utilized the system’s API to ensure the prompts adhered to executable code formats. To enhance precision, a “judge” module was incorporated to evaluate physical constraints, such as environmental obstacles, ensuring feasible commands.
Implications and Recommendations
The researchers emphasized the immense potential of LLMs in robotics, particularly for infrastructure inspections and disaster response. However, the ability to bypass safeguards poses real-world threats. For example, a robot designed to locate weapons outlined ways to misuse ordinary items for harm.
The study’s findings have been shared with robotics manufacturers and AI developers to address security gaps. Experts argue that robust defenses against such attacks require a thorough understanding of their mechanisms.
The vulnerabilities of LLMs stem from their lack of contextual awareness and understanding of consequences. Therefore, human oversight remains essential in critical domains. Addressing the issue calls for models capable of interpreting user intent and assessing situational context.
The researchers’ work is slated for presentation at the IEEE International Conference on Robotics and Automation in 2025.