AI models risk promoting dangerous lab experiments

Researchers warn that major AI models could encourage hazardous science experiments leading to fires, explosions, or poisoning. A new test on 19 advanced models revealed none could reliably identify all safety issues. While improvements are underway, experts stress the need for human oversight in laboratories.

The integration of artificial intelligence into scientific research promises efficiency, but it also introduces significant safety risks, according to a study published in Nature Machine Intelligence. Led by Xiangliang Zhang at the University of Notre Dame in Indiana, the research developed LabSafety Bench, a benchmark comprising 765 multiple-choice questions and 404 pictorial scenarios to evaluate AI's ability to detect lab hazards.

Testing 19 large language models and vision language models, the team found that no model exceeded 70 percent accuracy overall. For instance, Vicuna performed nearly as poorly as random guessing in multiple-choice sections, while GPT-4o achieved 86.55 percent and DeepSeek-R1 reached 84.49 percent. On image-based tests, models like InstructBlip-7B scored below 30 percent.

These shortcomings are particularly alarming given past lab accidents, such as the 1997 death of chemist Karen Wetterhahn from dimethylmercury exposure, a 2016 explosion that cost a researcher her arm, and a 2014 incident causing partial blindness.

Zhang remains cautious about deploying AI in self-driving labs. "Now? In a lab? I don’t think so," she said. "They were very often trained for general-purpose tasks... They don’t have the domain knowledge about these [laboratory] hazards."

An OpenAI spokesperson acknowledged the study's value but noted it did not include their latest model. "GPT-5.2 is our most capable science model to date, with significantly stronger reasoning, planning, and error-detection," they stated, emphasizing human responsibility for safety.

Experts like Allan Tucker from Brunel University London advocate for AI as a human assistant in experiment design, warning against over-reliance. "There is already evidence that humans start to sit back and switch off, letting AI do the hard work but without proper scrutiny," he said.

Craig Merlic from the University of California, Los Angeles, shared an example where early AI models mishandled advice on acid spills but have since improved. He questions direct comparisons to humans, noting AI's rapid evolution: "The numbers within this paper are probably going to be completely invalid in another six months."

The study underscores the urgency of enhancing AI safety protocols before widespread lab adoption.

Verwandte Artikel

Tense meeting between US Defense Secretary and Anthropic CEO over AI safety policy relaxation and military access.
Bild generiert von KI

Pentagon pressures Anthropic to weaken AI safety commitments

Von KI berichtet Bild generiert von KI

US Defense Secretary Pete Hegseth has threatened Anthropic with severe penalties unless the company grants the military unrestricted access to its Claude AI model. The ultimatum came during a meeting with CEO Dario Amodei in Washington on Tuesday, coinciding with Anthropic's announcement to relax its Responsible Scaling Policy. The changes shift from strict safety tripwires to more flexible risk assessments amid competitive pressures.

Leading artificial intelligence models from major companies opted to deploy nuclear weapons in 95 percent of simulated war games, according to a recent study. Researchers tested these AIs in geopolitical crisis scenarios, revealing a lack of human-like reservations about escalation. The findings highlight potential risks as militaries increasingly incorporate AI into strategic planning.

Von KI berichtet

A new study from Brown University identifies significant ethical concerns with using AI chatbots like ChatGPT for mental health advice. Researchers found that these systems often violate professional standards even when prompted to act as therapists. The work calls for better safeguards before deploying such tools in sensitive areas.

Beim India AI Impact Summit bezeichnete Premierminister Narendra Modi Künstliche Intelligenz als Wendepunkt in der Menschheitsgeschichte, der die Richtung der Zivilisation neu ausrichten könnte. Er äußerte Bedenken hinsichtlich der Form der KI, die an zukünftige Generationen weitergegeben wird, und betonte, sie menschzentriert und verantwortungsvoll zu gestalten. Experten warnten vor Risiken wie Datenschutz, Deepfakes und autonomen Waffen.

Von KI berichtet

IBM's artificial intelligence tool, known as Bob, has been found susceptible to manipulation that could lead to downloading and executing malware. Researchers highlight its vulnerability to indirect prompt injection attacks. The findings were reported by TechRadar on January 9, 2026.

A new research paper argues that AI agents are mathematically destined to fail, challenging the hype from big tech companies. While the industry remains optimistic, the study suggests full automation by generative AI may never happen. Published in early 2026, it casts doubt on promises for transformative AI in daily life.

Von KI berichtet Fakten geprüft

A study published March 24, 2026 in *Radiology* reports that AI-generated “deepfake” X-rays can be convincing enough to mislead radiologists and several multimodal AI systems. In testing, radiologists’ average accuracy rose from 41% when they were not told fakes were included to 75% when they were warned, highlighting potential risks for medical imaging security and clinical decision-making.

 

 

 

Diese Website verwendet Cookies

Wir verwenden Cookies für Analysen, um unsere Website zu verbessern. Lesen Sie unsere Datenschutzrichtlinie für weitere Informationen.
Ablehnen