Researchers Suggests More Safety Measures Needed for Robots than AI Chatbots

Insider Brief

  • A new study in Science Robotics warns that current AI safety guardrails developed for chatbots may not be sufficient for robots operating around people in real-world environments.
  • Researchers from University of Pennsylvania, Carnegie Mellon University and the University of Oxford found that AI-powered robots can inherit vulnerabilities seen in chatbots, including jailbreak-style attacks, while also introducing physical safety risks through interaction with people and objects.
  • The study proposes a layered safety framework built around clearer AI operating rules, multiple robotic safety checkpoints and training datasets that explicitly teach robots to recognize unsafe situations as AI-enabled systems move into homes, hospitals and other dynamic environments.

Today’s AI safety guardrails may not be enough once robots begin operating around people in the physical world, according to a new study warning that AI-powered machines require far more context-aware safety systems than chatbots.

Researchers from University of Pennsylvania, Carnegie Mellon University and the University of Oxford, report finding that safety techniques developed for chatbots do not translate cleanly to robotics systems that interact directly with the physical environment. The research, backed by the U.S. Defense Advanced Research Projects Agency and National Science Foundation. was published in Science Robotics.

According to the researchers, large AI models controlling robots can inherit the same vulnerabilities seen in chatbots, including “jailbreaking” attacks that bypass safeguards. But unlike chatbots, robotic systems can manipulate physical objects and operate around people, raising the stakes when safety protections fail.

“AI systems can control robots and give them novel capabilities, like responding to nuanced human instructions and adapting to new environments,” noted former CMU postdoctoral fellow and the paper’s first author Alexander Robey. “But it’s also clear that we need to move far beyond current alignment efforts to ensure those systems are compatible with human safety.”

According to the paper, the core problem is context. Chatbots can often reject obviously harmful requests outright, but robots must continuously interpret whether otherwise normal actions could become dangerous depending on the environment and nearby people. Researchers pointed to a task such as pouring hot water into a cup may be harmless in one situation but dangerous in another as an example.

The study proposes a broader safety framework built around three main approaches:

  • Embedding clearer operating rules directly into AI systems
  • Adding multiple safety checkpoints throughout robotic workflows
  • Training AI models on datasets that explicitly teach robots to recognize unsafe situations

The researchers argue that no single safeguard will be sufficient as robots gain more autonomy and adaptability.

The authors said traditional industrial robotics systems operated in relatively predictable environments where predefined shutdown limits were often enough to prevent accidents. AI-enabled robots, however, are increasingly being designed to work in homes, hospitals, warehouses and other dynamic settings where they must process real-time instructions and adapt to changing conditions.

“If robots are going to operate around people in the real world,” noted Zachary Ravichandran, doctoral student in Penn’s General Robotics, Automation, Sensing and Perception (GRASP) Lab and co-author of the paper. “They need comprehensive safeguards that account for context, uncertainty and the possibility that even reasonable instructions can lead to harm.”

Need Deeper Intelligence on the AI Market?

AI Insider's Market Intelligence platform tracks funding rounds, competitive landscapes, and technology trends across the global AI ecosystem in real time. Get the data and insights your organization needs to make informed decisions.

Related Articles

Mind Robotics Announces $400M in New Funding to Expand Industrial Robotics Deployment

Insider Brief Industrial robotics startup Mind Robotics has raised $400 million in new funding led by Kleiner Perkins, bringing total investment in the company to

Amazon Launches Agentic AI Assistant ‘Alexa for Shopping’

Insider Brief Amazon is rolling out a new AI-powered shopping assistant called Alexa for Shopping that combines conversational AI, personalized recommendations and automation tools across

Autonomous Defense Tech Company Anduril Announces $5B Series H Funding Round

Insider Brief Defense technology company Anduril Industries has raised $5 billion in a Series H funding round to expand manufacturing capacity and autonomous defense systems

Stay Updated with AI Insider

Get the latest AI funding news, market intelligence, and industry insights delivered to your inbox weekly.

$ 0 M

Seed round tracked

Gitar — Code Validation

Get the Weekly Briefing

Funding analysis, market intelligence, and industry trends delivered to your inbox every week.

Need bespoke intelligence?

Our team combines real-time data with decades of sector experience to guide your decisions.

Subscribe today for the latest news about the AI landscape