LLVM implements AI policy requiring human oversight

The open-source project LLVM has introduced a new policy allowing AI-generated code in contributions, provided humans review and understand the submissions. This 'human in the loop' approach ensures accountability while addressing community concerns about transparency. The policy, developed with input from contributors, balances innovation with reliability in software development.

LLVM, a foundational collection of compiler and toolchain components used in projects like Clang, Rust, Swift, and the Linux kernel, has adopted a policy on AI tool use in contributions. Published on January 22, 2026, the guidelines permit developers to employ any AI tools but emphasize full accountability for the submitted work.

Under the policy, contributors must disclose the AI tool used, either in the pull request description, commit message, or authorship details. They are required to review and comprehend their submissions, confidently justifying them during reviews and ensuring they merit a maintainer's attention. The rules clarify that violations will be handled according to existing community processes.

The development process involved extensive community engagement. A LLVM member highlighted discrepancies between the project's AI handling, code of conduct, and actual practices, referencing a notable pull request discussed on Hacker News where AI use was admitted post-submission but not initially disclosed.

LLVM maintainer Reid Kleckner spearheaded the effort. His initial draft, inspired by Fedora's AI policy, proposed restrictions such as limiting newcomers to 150 lines of non-test code. After feedback from community meetings and forums, the final version shifted to more explicit requirements, focusing on review readiness and question-answering ability rather than vague ownership clauses.

The updated AI Tool Use Policy is now available on LLVM's documentation site, including examples of acceptable AI-assisted work and violation guidelines. This move aligns LLVM with other open-source initiatives adapting to AI's growing role in development.

관련 기사

Realistic illustration of Linux Foundation executives and AI partners launching Agentic AI Foundation, featuring collaborative autonomous AI agents on a conference screen.
AI에 의해 생성된 이미지

Linux Foundation launches Agentic AI Foundation

AI에 의해 보고됨 AI에 의해 생성된 이미지

The Linux Foundation has launched the Agentic AI Foundation to foster open collaboration on autonomous AI systems. Major tech companies, including Anthropic, OpenAI, and Block, contributed key open-source projects to promote interoperability and prevent vendor lock-in. The initiative aims to create neutral standards for AI agents that can make decisions and execute tasks independently.

Linus Torvalds, creator of the Linux kernel, has criticized efforts to create rules for AI-generated code submissions, calling them pointless. In a recent email, he argued that such policies would not deter malicious contributors and urged focus on code quality instead. This stance highlights ongoing tensions in open-source development over artificial intelligence tools.

AI에 의해 보고됨

The Linux developer community has shifted from debating AI's role to integrating it into kernel engineering processes. Developers now use AI for project maintenance, though questions persist about writing code with it. Concerns over copyright and open-source licensing remain.

Linus Torvalds, the creator of the Linux kernel, has strongly criticized discussions about AI-generated content in kernel documentation. He called talk of 'AI slop' pointless and stupid. The comments highlight ongoing tensions around AI in open-source development.

AI에 의해 보고됨

An AMD vice president has utilized AI to develop a Radeon Linux userland driver written in Python. A senior AI engineer involved in the project stated that he did not open a code editor once during the process. The tool used was Claude Code, which built a Radeon compute driver.

The Linux Foundation, partnering with Meta, released a report at the India AI Impact Summit in New Delhi on February 17, 2026, highlighting how open source AI drives India's economic growth. The study reveals that 76% of Indian startups use open source AI, with the market projected to expand from $6 billion in 2024 to nearly $32 billion by 2031. It also addresses workforce challenges and social applications of the technology.

AI에 의해 보고됨

The release of version 7.0 of the open-source Python library chardet has sparked controversy over whether an AI-assisted rewrite can change its original restrictive license. Maintainer Dan Blanchard used Anthropic's Claude tool to create a faster, MIT-licensed version, but original author Mark Pilgrim argues it violates the LGPL terms. The case highlights emerging legal and ethical questions in AI-generated code.

 

 

 

이 웹사이트는 쿠키를 사용합니다

사이트를 개선하기 위해 분석을 위한 쿠키를 사용합니다. 자세한 내용은 개인정보 보호 정책을 읽으세요.
거부