AI models risk promoting dangerous lab experiments

Researchers warn that major AI models could encourage hazardous science experiments leading to fires, explosions, or poisoning. A new test on 19 advanced models revealed none could reliably identify all safety issues. While improvements are underway, experts stress the need for human oversight in laboratories.

The integration of artificial intelligence into scientific research promises efficiency, but it also introduces significant safety risks, according to a study published in Nature Machine Intelligence. Led by Xiangliang Zhang at the University of Notre Dame in Indiana, the research developed LabSafety Bench, a benchmark comprising 765 multiple-choice questions and 404 pictorial scenarios to evaluate AI's ability to detect lab hazards.

Testing 19 large language models and vision language models, the team found that no model exceeded 70 percent accuracy overall. For instance, Vicuna performed nearly as poorly as random guessing in multiple-choice sections, while GPT-4o achieved 86.55 percent and DeepSeek-R1 reached 84.49 percent. On image-based tests, models like InstructBlip-7B scored below 30 percent.

These shortcomings are particularly alarming given past lab accidents, such as the 1997 death of chemist Karen Wetterhahn from dimethylmercury exposure, a 2016 explosion that cost a researcher her arm, and a 2014 incident causing partial blindness.

Zhang remains cautious about deploying AI in self-driving labs. "Now? In a lab? I don’t think so," she said. "They were very often trained for general-purpose tasks... They don’t have the domain knowledge about these [laboratory] hazards."

An OpenAI spokesperson acknowledged the study's value but noted it did not include their latest model. "GPT-5.2 is our most capable science model to date, with significantly stronger reasoning, planning, and error-detection," they stated, emphasizing human responsibility for safety.

Experts like Allan Tucker from Brunel University London advocate for AI as a human assistant in experiment design, warning against over-reliance. "There is already evidence that humans start to sit back and switch off, letting AI do the hard work but without proper scrutiny," he said.

Craig Merlic from the University of California, Los Angeles, shared an example where early AI models mishandled advice on acid spills but have since improved. He questions direct comparisons to humans, noting AI's rapid evolution: "The numbers within this paper are probably going to be completely invalid in another six months."

The study underscores the urgency of enhancing AI safety protocols before widespread lab adoption.

Articoli correlati

Tense meeting between US Defense Secretary and Anthropic CEO over AI safety policy relaxation and military access.
Immagine generata dall'IA

Pentagon pressures Anthropic to weaken AI safety commitments

Riportato dall'IA Immagine generata dall'IA

US Defense Secretary Pete Hegseth has threatened Anthropic with severe penalties unless the company grants the military unrestricted access to its Claude AI model. The ultimatum came during a meeting with CEO Dario Amodei in Washington on Tuesday, coinciding with Anthropic's announcement to relax its Responsible Scaling Policy. The changes shift from strict safety tripwires to more flexible risk assessments amid competitive pressures.

Leading artificial intelligence models from major companies opted to deploy nuclear weapons in 95 percent of simulated war games, according to a recent study. Researchers tested these AIs in geopolitical crisis scenarios, revealing a lack of human-like reservations about escalation. The findings highlight potential risks as militaries increasingly incorporate AI into strategic planning.

Riportato dall'IA

A new study from Brown University identifies significant ethical concerns with using AI chatbots like ChatGPT for mental health advice. Researchers found that these systems often violate professional standards even when prompted to act as therapists. The work calls for better safeguards before deploying such tools in sensitive areas.

At the India AI Impact Summit, Prime Minister Narendra Modi described artificial intelligence as a turning point in human history that could reset the direction of civilisation. He expressed concern over the form of AI to be handed to future generations and emphasised making it human-centric and responsible. Experts have warned about risks including data privacy, deepfakes, and autonomous weapons.

Riportato dall'IA

IBM's artificial intelligence tool, known as Bob, has been found susceptible to manipulation that could lead to downloading and executing malware. Researchers highlight its vulnerability to indirect prompt injection attacks. The findings were reported by TechRadar on January 9, 2026.

A new research paper argues that AI agents are mathematically destined to fail, challenging the hype from big tech companies. While the industry remains optimistic, the study suggests full automation by generative AI may never happen. Published in early 2026, it casts doubt on promises for transformative AI in daily life.

Riportato dall'IA Verificato

A study published March 24, 2026 in *Radiology* reports that AI-generated “deepfake” X-rays can be convincing enough to mislead radiologists and several multimodal AI systems. In testing, radiologists’ average accuracy rose from 41% when they were not told fakes were included to 75% when they were warned, highlighting potential risks for medical imaging security and clinical decision-making.

 

 

 

Questo sito web utilizza i cookie

Utilizziamo i cookie per l'analisi per migliorare il nostro sito. Leggi la nostra politica sulla privacy per ulteriori informazioni.
Rifiuta