Scientists in a lab urgently discussing consciousness amid holographic displays of brains, AI, and organoids, highlighting ethical risks from advancing neurotech.
Bild generiert von KI

Scientists say defining consciousness is increasingly urgent as AI and neurotechnology advance

Bild generiert von KI
Fakten geprüft

Researchers behind a new review in Frontiers in Science argue that rapid progress in artificial intelligence and brain technologies is outpacing scientific understanding of consciousness, raising the risk of ethical and legal mistakes. They say developing evidence-based tests for detecting awareness—whether in patients, animals or emerging artificial and lab-grown systems—could reshape medicine, welfare debates and technology governance.

The rapid development of artificial intelligence and neurotechnology is intensifying calls from consciousness researchers to clarify what it means to be conscious—and how to detect it.

In a review published in Frontiers in Science, Prof. Axel Cleeremans of Université Libre de Bruxelles, Prof. Liad Mudrik of Tel Aviv University, and Prof. Anil Seth of the University of Sussex argue that advances in these technologies are moving faster than scientific agreement on how consciousness arises. They describe consciousness in broadly familiar terms—as awareness of the world and of oneself—while noting that science still lacks consensus on how subjective experience emerges from physical processes.

The authors point to ongoing competition among major scientific theories of consciousness, including global workspace approaches, higher-order theories, integrated information theory and predictive processing frameworks. They argue that progress depends in part on developing stronger methods to test these ideas, including “adversarial collaborations” in which proponents of rival theories jointly design experiments intended to distinguish between them.

A key goal, the review argues, is the development of evidence-based tests for consciousness that can be applied beyond healthy adult humans. Such tools could affect clinical care by helping clinicians detect covert awareness in some patients who appear unresponsive, and by refining assessments in conditions such as coma, advanced dementia, and anesthesia—areas that can influence treatment planning and end-of-life decisions.

The review also outlines potential implications for mental health research. The authors argue that a better scientific account of subjective experience could help narrow gaps between findings in animal models and the lived experience of human symptoms, with possible relevance for conditions including depression, anxiety and schizophrenia.

Beyond medicine, the authors say improved ways of identifying consciousness could reshape debates over animal welfare and ethical obligations, influencing practices in research, agriculture and conservation if society gains clearer evidence about which animals are sentient.

They also highlight potential legal consequences. The review notes that neuroscience findings about unconscious influences on behavior could pressure legal systems to revisit how they interpret responsibility and concepts such as mens rea, the mental element traditionally required for criminal liability.

In technology, the authors argue that emerging systems—from advanced AI to brain organoids and brain–computer interfaces—raise new questions about whether consciousness could be created, altered, or convincingly simulated, and what moral and regulatory obligations might follow. Cleeremans warned that unintended creation of consciousness would pose “immense ethical challenges and even existential risk.” Seth said that advances in the science of consciousness are likely to reshape how humans understand themselves and their relationship to both AI and the natural world. Mudrik argued that a clearer understanding of consciousness in animals could transform how humans treat them and other emerging biological systems.

To move the field forward, the authors call for more coordinated, collaborative research that combines careful theory testing with greater attention to phenomenology—the qualities of experience itself—alongside functional and neural measures.

They argue that such work is needed not only to advance basic science, but also to prepare society for the medical, ethical and technological consequences of being able to detect—or potentially create—consciousness.

Was die Leute sagen

Initial reactions on X to the article primarily involve shares and paraphrases emphasizing the urgency of defining consciousness due to advances in AI and neurotechnology. Users highlight ethical risks, the need for scientific tests for awareness, and potential impacts on medicine, law, animal welfare, and rights for machines or lab-grown systems. Sentiments are mostly neutral with some underscoring the unsettling moral implications.

Verwandte Artikel

Realistic depiction of a rhesus macaque in a Princeton lab with brain overlay showing prefrontal cortex assembling reusable cognitive 'Lego' modules for flexible learning.
Bild generiert von KI

Princeton study reveals brain’s reusable ‘cognitive Legos’ for flexible learning

Von KI berichtet Bild generiert von KI Fakten geprüft

Neuroscientists at Princeton University report that the brain achieves flexible learning by reusing modular cognitive components across tasks. In experiments with rhesus macaques, researchers found that the prefrontal cortex assembles these reusable “cognitive Legos” to adapt behaviors quickly. The findings, published November 26 in Nature, underscore differences from current AI systems and could eventually inform treatments for disorders that impair flexible thinking.

A review article by Borjan Milinkovic and Jaan Aru argues that treating the mind as software running on interchangeable hardware is a poor fit for how brains actually compute. The authors propose “biological computationalism,” a framework that ties cognition and (potentially) consciousness to computation that is hybrid, multi-scale, and shaped by energy constraints.

Von KI berichtet

Scientists have outlined three evolutionary stages of consciousness, from basic alarm responses to self-awareness, suggesting it is an ancient trait shared widely across species. New research highlights that birds exhibit forms of sensory perception and self-consciousness similar to mammals, challenging previous assumptions about its origins. This framework, known as the ALARM theory, emphasizes survival and social functions.

Researchers at Nagoya University in Japan have developed miniature brain models using stem cells to study interactions between the thalamus and cortex. Their work reveals the thalamus's key role in maturing cortical neural networks. The findings could advance research into neurological disorders like autism.

Von KI berichtet

Researchers at Rutgers Health have identified how the brain integrates fast and slow processing through white matter connections, influencing cognitive abilities. Published in Nature Communications, the study analyzed data from nearly 1,000 people to map these neural timescales. Variations in this system may explain differences in thinking efficiency and hold promise for mental health research.

A growing number of companies are evaluating the security risks associated with artificial intelligence, marking a shift from previous years. This trend indicates heightened awareness among businesses about potential vulnerabilities in AI technologies. The development comes as organizations prioritize protective measures against emerging threats.

Von KI berichtet Fakten geprüft

An evolutionarily ancient midbrain region, the superior colliculus, can independently carry out visual computations long attributed mainly to the cortex, according to a PLOS Biology study. The work suggests that attention-guiding mechanisms with roots more than 500 million years old help separate objects from backgrounds and highlight salient details.

 

 

 

Diese Website verwendet Cookies

Wir verwenden Cookies für Analysen, um unsere Website zu verbessern. Lesen Sie unsere Datenschutzrichtlinie für weitere Informationen.
Ablehnen