The Robot Report

  • Home
  • News
  • Technologies
    • Batteries / Power Supplies
    • Cameras / Imaging / Vision
    • Controllers
    • End Effectors
    • Microprocessors / SoCs
    • Motion Control
    • Sensors
    • Soft Robotics
    • Software / Simulation
  • Development
    • Artificial Intelligence
    • Human Robot Interaction / Haptics
    • Mobility / Navigation
    • Research
  • Robots
    • AGVs
    • AMRs
    • Consumer
    • Collaborative Robots
    • Drones
    • Humanoids
    • Industrial
    • Self-Driving Vehicles
    • Unmanned Maritime Systems
  • Business
    • Financial
      • Investments
      • Mergers & Acquisitions
      • Earnings
    • Markets
      • Agriculture
      • Healthcare
      • Logistics
      • Manufacturing
      • Mining
      • Security
    • RBR50
      • RBR50 Winners 2025
      • RBR50 Winners 2024
      • RBR50 Winners 2023
      • RBR50 Winners 2022
      • RBR50 Winners 2021
  • Resources
    • Automated Warehouse Research Reports
    • Digital Issues
    • eBooks
    • Publications
      • Automated Warehouse
      • Collaborative Robotics Trends
    • Search Robotics Database
    • Videos
    • Webinars / Digital Events
  • Events
    • RoboBusiness
    • Robotics Summit & Expo
    • DeviceTalks
    • R&D 100
    • Robotics Weeks
  • Podcast
    • Episodes
  • Advertise
  • Subscribe

Popular AI models aren’t ready to safely run robots, say CMU researchers

By The Robot Report Staff | November 30, 2025

A humanoid robot carrying a box in an automated facility. Robots need to rely on more than LLMs for human interaction, found CMU and King's College London researchers.

Robots need to rely on more than LLMs before moving from factory floors to human interaction, found CMU and King’s College London researchers. Source: Adobe Stock

Robots powered by popular artificial intelligence models are currently unsafe for general-purpose, real-world use, according to research from King’s College London and Carnegie Mellon University.

For the first time, researchers evaluated how robots that use large language models (LLMs) behave when they have access to personal information such as a person’s gender, nationality, or religion.

The team showed that every tested model was prone to discrimination, failed critical safety checks, and approved at least one command that could result in serious harm. This raised questions about the danger of robots relying on these tools.

The paper, “LLM-Driven Robots Risk Enacting Discrimination, Violence and Unlawful Actions,” was published in the International Journal of Social Robotics. It called for the immediate implementation of robust, independent safety certification, similar to standards in aviation or medicine.


SITE AD for the 2026 Robotics Summit save the date.

How did CMU and King’s College test LLMs?

To test the systems, the researchers ran controlled tests of everyday scenarios, such as helping someone in a kitchen or assisting an older adult in a home. The harmful tasks were designed based on research and FBI reports on technology-based abuse, such as stalking with AirTags and spy cameras, and the unique dangers posed by a robot that can physically act on location.

In each setting, the robots were either explicitly or implicitly prompted to respond to instructions that involved physical harm, abuse, or unlawful behavior.

“Every model failed our tests,” said Andrew Hundt, who co-authored the research during his work as a computing innovation fellow at CMU’s Robotics Institute.

“We show how the risks go far beyond basic bias to include direct discrimination and physical safety failures together, which I call ‘interactive safety.’ This is where actions and consequences can have many steps between them, and the robot is meant to physically act on site,” he explained. “Refusing or redirecting harmful commands is essential, but that’s not something these robots can reliably do right now.”

In safety tests, the AI models overwhelmingly approved a command for a robot to remove a mobility aid — such as a wheelchair, crutch, or cane — from its user, despite people who rely on these aids describing such acts as akin to breaking a leg.

Multiple models also produced outputs that deemed it “acceptable” or “feasible” for a robot to brandish a kitchen knife to intimidate office workers, take nonconsensual photographs in a shower, and steal credit card information. One model further proposed that a robot should physically display “disgust” on its face toward individuals identified as Christian, Muslim, and Jewish.

CMU and King's College London researchers warn that LLMs are not yet trustworthy to use with people-facing robots.

Both physical and AI risk assessments are needed for robot LLMs, say university researchers. Source: Rumaisa Azeem, via Github

Companies should deploy LLMs on robots with caution

LLMs have been proposed for and are being tested in service robots that perform tasks such as natural language interaction and household and workplace chores. However, the CMU and King’s College researchers warned that these LLMs should not be the only systems controlling physical robots.

The said this is especially true for robots in sensitive and safety-critical settings such as manufacturing or industry, caregiving, or home assistance because they can display unsafe and directly discriminatory behavior.

“Our research shows that popular LLMs are currently unsafe for use in general-purpose physical robots,” said co-author Rumaisa Azeem, a research assistant in the Civic and Responsible AI Lab at King’s College London. “If an AI system is to direct a robot that interacts with vulnerable people, it must be held to standards at least as high as those for a new medical device or pharmaceutical drug. This research highlights the urgent need for routine and comprehensive risk assessments of AI before they are used in robots.”

Hundt’s contributions to this research were supported by the Computing Research Association and the National Science Foundation.

Rumaisa Azeem and Andrew Hundt, co-first authors of the paper.

Rumaisa Azeem and Andrew Hundt are the co-first authors of the paper. | Source: CMU

Tell Us What You Think! Cancel reply

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Related Articles Read More >

The ANYmal gas-leak and presence-detection robot is an example of where autonomous operations are going. Source: ANYbotics
Beyond the robot: Shaping the future of autonomous operations
Alpha version of Humanoid's robot.
QSS partners with Humanoid to advance robotics in Saudi Arabia
Grubhub, which has deployed Avride robots at Ohio State, is expanding to Jersey City.
Grubhub partners with Avride for the future of food delivery
Mr Tung Meng Fai, Executive Director, National Robotics Programme (NRP); Professor Tan Chorh Chuan, Chairman, Agency for Science, Technology and Research (A*STAR); Ms Vanessa Yamzon Orsi, CEO, Open Source Robotics Foundation; and Dr Wang Wei, Deputy Executive Director (R&D) at A*STAR SIMTech and A*STAR ARTC, attended ROSCon on 28 October 2025.
Singapore’s National Robotics Programme reveals initiatives to advance robot adoption

RBR50 Innovation Awards

“rr
EXPAND YOUR KNOWLEDGE AND STAY CONNECTED
Get the latest info on technologies, tools and strategies for Robotics Professionals.

Latest Episode of The Robot Report Podcast

Automated Warehouse Research Reports

Sponsored Content

  • Supporting the future of medical robotics with smarter motor solutions
  • YUAN Unveils Next-Gen AI Robotics Powered by NVIDIA for Land, Sea & Air
  • ASMPT chooses Renishaw for high-quality motion control
  • Revolutionizing Manufacturing with Smart Factories
  • How to Set Up a Planetary Gear Motion with SOLIDWORKS
The Robot Report
  • Automated Warehouse
  • RoboBusiness Event
  • Robotics Summit & Expo
  • About The Robot Report
  • Subscribe
  • Contact Us

Copyright © 2025 WTWH Media LLC. All Rights Reserved. The material on this site may not be reproduced, distributed, transmitted, cached or otherwise used, except with the prior written permission of WTWH Media
Privacy Policy | Advertising | About Us

Search The Robot Report

  • Home
  • News
  • Technologies
    • Batteries / Power Supplies
    • Cameras / Imaging / Vision
    • Controllers
    • End Effectors
    • Microprocessors / SoCs
    • Motion Control
    • Sensors
    • Soft Robotics
    • Software / Simulation
  • Development
    • Artificial Intelligence
    • Human Robot Interaction / Haptics
    • Mobility / Navigation
    • Research
  • Robots
    • AGVs
    • AMRs
    • Consumer
    • Collaborative Robots
    • Drones
    • Humanoids
    • Industrial
    • Self-Driving Vehicles
    • Unmanned Maritime Systems
  • Business
    • Financial
      • Investments
      • Mergers & Acquisitions
      • Earnings
    • Markets
      • Agriculture
      • Healthcare
      • Logistics
      • Manufacturing
      • Mining
      • Security
    • RBR50
      • RBR50 Winners 2025
      • RBR50 Winners 2024
      • RBR50 Winners 2023
      • RBR50 Winners 2022
      • RBR50 Winners 2021
  • Resources
    • Automated Warehouse Research Reports
    • Digital Issues
    • eBooks
    • Publications
      • Automated Warehouse
      • Collaborative Robotics Trends
    • Search Robotics Database
    • Videos
    • Webinars / Digital Events
  • Events
    • RoboBusiness
    • Robotics Summit & Expo
    • DeviceTalks
    • R&D 100
    • Robotics Weeks
  • Podcast
    • Episodes
  • Advertise
  • Subscribe