Illustration depicting linguists studying why human language resists compression like computer code, contrasting brain processing with digital efficiency.
Illustration depicting linguists studying why human language resists compression like computer code, contrasting brain processing with digital efficiency.
Larawang ginawa ng AI

Study explores why human language isn’t compressed like computer code

Larawang ginawa ng AI
Fact checked

A new model from linguists Richard Futrell and Michael Hahn suggests that many hallmark features of human language—such as familiar words, predictable ordering and meaning built up step by step—reflect constraints on sequential information processing rather than a drive for maximum data compression. The work was published in Nature Human Behaviour.

Human language is remarkably rich and intricate. From an information-theory standpoint, the same ideas could, in principle, be transmitted in far more compact strings—similar to how computers represent information using binary digits.

Michael Hahn, a linguist at Saarland University in Saarbrücken, Germany, and Richard Futrell of the University of California, Irvine, set out to address why everyday speech does not resemble a tightly compressed digital code. In a paper published in Nature Human Behaviour in November 2025, the researchers present a model in which “natural-language-like” structure arises when communication is constrained by limits on sequential prediction—how much information must be carried forward from what has already been heard to anticipate what comes next.

In that framework, language benefits from patterns that are easy for people to process as a stream. A ScienceDaily summary of the work, citing materials from the University of Osaka, uses examples to illustrate the idea: an invented word such as “gol” for a hybrid concept (half cat and half dog) would be hard to understand because it does not map cleanly onto shared experience, and a scrambled blend like “gadcot” is similarly difficult to interpret. By contrast, “cat and dog” is immediately meaningful.

The researchers also point to word order as a signal that helps listeners reduce uncertainty in real time. The ScienceDaily release highlights the German noun phrase “Die fünf grünen Autos” (“the five green cars”) as an example of how meaning can be built incrementally as each word narrows the set of plausible interpretations. Reordering those words—for example, “Grünen fünf die Autos”—disrupts that predictability and makes comprehension harder.

Beyond explaining why language is not “maximally compressed,” the paper’s discussion connects the findings to machine learning. Futrell and Hahn argue that natural language is structured in a way that makes next-token prediction comparatively easier under cognitive constraints, a point they say is relevant to modern large language models.

Mga Kaugnay na Artikulo

Illustration of a patient undergoing brain monitoring while listening to a podcast, with neural activity layers mirroring AI language model processing.
Larawang ginawa ng AI

Study links step-by-step brain responses during speech to layered processing in large language models

Iniulat ng AI Larawang ginawa ng AI Fact checked

A new study reports that as people listen to a spoken story, neural activity in key language regions unfolds over time in a way that mirrors the layer-by-layer computations inside large language models. The researchers, who analyzed electrocorticography recordings from epilepsy patients during a 30-minute podcast, also released an open dataset intended to help other scientists test competing theories of how meaning is built in the brain.

A new computational analysis of Paleolithic artifacts reveals that humans over 40,000 years ago engraved structured symbols on tools and figurines, indicating early forms of information recording. These signs, found mainly in southwestern Germany, show complexity comparable to the earliest known writing systems that emerged millennia later. Researchers suggest these markings were purposeful, predating formal writing by tens of thousands of years.

Iniulat ng AI

A researcher using the Lean formalisation language has uncovered a fundamental flaw in a influential 2006 physics paper on the two Higgs doublet model. Joseph Tooby-Smith at the University of Bath made the discovery while building a library of verified physics theorems. The original authors have acknowledged the error and plan to issue an erratum.

Northwestern University researchers report they have printed flexible “artificial neurons” that generate realistic electrical spike patterns and can trigger responses in living mouse brain tissue. The team says the work, published April 15 in Nature Nanotechnology, could help advance brain-machine interfaces and more energy-efficient, brain-inspired computing.

Iniulat ng AI Fact checked

Researchers behind a new review in Frontiers in Science argue that rapid progress in artificial intelligence and brain technologies is outpacing scientific understanding of consciousness, raising the risk of ethical and legal mistakes. They say developing evidence-based tests for detecting awareness—whether in patients, animals or emerging artificial and lab-grown systems—could reshape medicine, welfare debates and technology governance.

Gumagamit ng cookies ang website na ito

Gumagamit kami ng cookies para sa analytics upang mapabuti ang aming site. Basahin ang aming patakaran sa privacy para sa higit pang impormasyon.
Tanggihan