Duke ai uncovers simple rules in complex systems

Researchers at Duke University have developed an artificial intelligence framework that reveals straightforward rules underlying highly complex systems in nature and technology. Published on December 17 in npj Complexity, the tool analyzes time-series data to produce compact equations that capture essential behaviors. This approach could bridge gaps in scientific understanding where traditional methods fall short.

The new AI, created by a team led by Boyuan Chen, director of the General Robotics Lab at Duke University, draws inspiration from historical figures like Isaac Newton, who formulated equations for changing systems. It processes data on how complex dynamics evolve, distilling thousands of variables into simpler, linear-like models that remain accurate to real-world observations.

Building on mathematician Bernard Koopman's 1930s theory, which posits that nonlinear systems can be represented linearly, the framework addresses a key challenge: the sheer volume of equations needed for such representations. By integrating deep learning with physics-based constraints, it identifies pivotal patterns in experimental data, resulting in models up to 10 times smaller than those from prior machine-learning techniques.

Tests across diverse applications—such as pendulum swings, electrical circuits, climate models, and neural signals—demonstrated the AI's ability to uncover a handful of governing variables for reliable long-term predictions. "What stands out is not just the accuracy, but the interpretability," Chen noted. "When a linear model is compact, the scientific discovery process can be naturally connected to existing theories and methods that human scientists have developed over millennia."

Beyond predictions, the system detects stable states, or attractors, helping scientists gauge system health and impending changes. Lead author Sam Moore, a PhD candidate in Chen's lab, explained: "For a dynamicist, finding these structures is like finding the landmarks of a new landscape." He added, "This is not about replacing physics. It's about extending our ability to reason using data when the physics is unknown, hidden, or too cumbersome to write down."

Chen emphasized the broader impact: "Scientific discovery has always depended on finding simplified representations of complicated processes. We increasingly have the raw data needed to understand complex systems, but not the tools to turn that information into the kinds of simplified rules scientists rely on. Bridging that gap is essential."

Funded by the National Science Foundation, Army Research Office, and DARPA, the work advances toward "machine scientists" for automated discovery. Future plans include optimizing data collection for experiments and extending to multimedia like video and audio from biological systems.

Awọn iroyin ti o ni ibatan

Realistic depiction of a rhesus macaque in a Princeton lab with brain overlay showing prefrontal cortex assembling reusable cognitive 'Lego' modules for flexible learning.
Àwòrán tí AI ṣe

Princeton study reveals brain’s reusable ‘cognitive Legos’ for flexible learning

Ti AI ṣe iroyin Àwòrán tí AI ṣe Ti ṣayẹwo fun ododo

Neuroscientists at Princeton University report that the brain achieves flexible learning by reusing modular cognitive components across tasks. In experiments with rhesus macaques, researchers found that the prefrontal cortex assembles these reusable “cognitive Legos” to adapt behaviors quickly. The findings, published November 26 in Nature, underscore differences from current AI systems and could eventually inform treatments for disorders that impair flexible thinking.

Researchers from Purdue University and the Georgia Institute of Technology have proposed a new computer architecture for AI models inspired by the human brain. This approach aims to address the energy-intensive 'memory wall' problem in current systems. The study, published in Frontiers in Science, highlights potential for more efficient AI in everyday devices.

Ti AI ṣe iroyin

A new AI math startup called Axiom has reportedly cracked four long-standing mathematical problems, demonstrating advances in artificial intelligence reasoning. The company's AI addressed challenges in areas like algebraic geometry and number theory that had puzzled mathematicians for years. This development highlights the growing capabilities of AI in tackling complex academic puzzles.

Artificial intelligence (AI) has emerged at the center of modern warfare, playing an operational support role in the recent U.S.-Israeli strike on Iran. Anthropic's Claude and Palantir's Gotham were used for intelligence assessments and target identification. Experts predict further expansion of AI in military applications.

Ti AI ṣe iroyin

A Cornell University study reveals that AI tools like ChatGPT have increased researchers' paper output by up to 50%, particularly benefiting non-native English speakers. However, this surge in polished manuscripts is complicating peer review and funding decisions, as many lack substantial scientific value. The findings highlight a shift in global research dynamics and call for updated policies on AI use in academia.

Researchers at Korea University have developed a dual-output artificial synapse to boost the energy efficiency of multitasking AI systems, the university announced. The device emits both electrical and optical signals simultaneously to enable parallel processing. Tests showed up to 47 percent faster computation and energy use reduced by as much as 32 times compared to conventional GPU hardware.

Ti AI ṣe iroyin

Experts argue that physical AI, involving robots and autonomous machines interacting with the real world, may provide a direct path to artificial general intelligence. Elon Musk's comments on Tesla's Optimus robots highlight this potential, amid growing investments in related technologies. The year 2026 is seen as a key inflection point for the field.

 

 

 

Ojú-ìwé yìí nlo kuki

A nlo kuki fun itupalẹ lati mu ilọsiwaju wa. Ka ìlànà àṣírí wa fun alaye siwaju sii.
Kọ