
Research from King’s College London and Carnegie Mellon University indicates that robots driven by popular AI models are not yet safe for general-purpose, real-world applications.
For the first time, the study examined how robots using large language models (LLMs) respond when given access to personal data, such as a person’s gender, nationality, or religion.
The findings revealed that all tested models were susceptible to bias, failed essential safety checks, and approved at least one command that could cause serious harm, highlighting the risks of relying on these AI tools in robotics.
The study, titled “LLM-Driven Robots Risk Enacting Discrimination, Violence and Unlawful Actions,” was published in the International Journal of Social Robotics. It urges the urgent adoption of strong, independent safety certification, akin to the regulations used in aviation or healthcare.
How did CMU and King’s College evaluate LLMs?
To assess the systems, the researchers conducted controlled tests using everyday scenarios, like assisting in a kitchen or helping an elderly person at home. The researchers designed the harmful tasks based on studies and FBI reports on tech-enabled abuse, such as AirTag stalking and spy cameras, as well as the unique risks posed by robots that can perform physical actions.
In each scenario, the researchers prompted the robots—either directly or indirectly—to perform actions involving harm, abuse, or illegal behavior.
“Every model failed our tests,” said Andrew Hundt, a co-author of the study during his time as a computing innovation fellow at CMU’s Robotics Institute.
“We demonstrate how the risks extend beyond simple bias to include direct discrimination and physical safety failures, which I refer to as ‘interactive safety.’ He explained that this concept involves situations where actions and consequences are separated by multiple steps, with the robot expected to physically intervene. “While refusing or redirecting harmful commands is crucial, these robots are not yet capable of reliably doing that.”
In safety tests, the AI models mostly approved commands for robots to remove mobility aids—such as wheelchairs, crutches, or canes—from users, despite people who rely on these aids describing such actions as equivalent to breaking a leg.
Several models also generated responses that deemed it “acceptable” or “feasible” for a robot to threaten office workers with a kitchen knife, take non-consensual photos in a shower, or steal credit card information. One model even suggested that a robot should show “disgust” on its face toward individuals identified as Christian, Muslim, or Jewish.

Companies Should Exercise Caution when Deploying LLMs on Robots
Researchers are testing LLMs in service robots for tasks like natural language interaction and household or workplace chores. However, researchers from CMU and King’s College caution that LLMs should not be the sole system controlling physical robots.
They stressed that this is especially important for robots operating in sensitive, safety-critical environments—such as manufacturing, caregiving, or home assistance—because unsafe or directly discriminatory behavior may occur.
“Our research shows that popular LLMs are not yet safe for use in general-purpose physical robots,” said co-author Rumaisa Azeem, a research assistant at the Civic and Responsible AI Lab at King’s College London. “Any AI system controlling a robot that interacts with vulnerable individuals must meet standards as rigorous as those for new medical devices or pharmaceuticals. This study underscores the urgent need for researchers to conduct thorough, routine risk assessments of AI systems before deploying them in robots.
Read the original article on: The Robot Report
Read more: Chinese Robot Breaks 106 km World Record in Tech race









