Kali Linux launches local AI tools for penetration testing

The Kali Linux team has released a guide for running AI-driven penetration testing entirely on local hardware, eliminating cloud dependencies. This setup uses Ollama, 5ire, and MCP Kali Server to enable natural language commands for security tools. Published on March 10, 2026, the guide addresses privacy concerns in sensitive environments.

The Kali Linux team published a new guide on March 10, 2026, as part of its series on large language model (LLM)-driven security tools. This entry focuses on a fully self-hosted stack that processes all AI operations on local hardware, avoiding third-party cloud services. The approach tackles privacy and operational security issues that have limited cloud-based AI in penetration testing.

The setup requires an NVIDIA GPU with CUDA support. The guide uses an NVIDIA GeForce GTX 1060 with 6 GB of VRAM as reference hardware. It involves installing NVIDIA's proprietary drivers, replacing the open-source Nouveau driver, to enable CUDA acceleration. After installation and reboot, the system confirms Driver Version 550.163.01 and CUDA Version 12.4.

Ollama serves as the core LLM engine, acting as a wrapper for llama.cpp to simplify model management. Installed via a Linux AMD64 tarball and set up as a systemd service, it runs in the background. The guide evaluates three models with tool-calling support: llama3.1:8b (4.9 GB), llama3.2:3b (2.0 GB), and qwen3:4b (2.5 GB), all fitting within the 6 GB VRAM limit.

The Model Context Protocol (MCP) integrates the AI with security tools through the mcp-kali-server package, available in Kali repositories. This creates a local Flask server on 127.0.0.1:5000, verifying tools like nmap, gobuster, dirb, and nikto. It supports tasks such as web application testing, CTF challenges, and interactions with platforms like Hack The Box or TryHackMe.

To connect Ollama and MCP, the guide uses 5ire, an open-source AI assistant and MCP client distributed as a Linux AppImage in version 0.15.3. Installed to /opt/5ire/ and configured with a desktop entry, it enables Ollama as the provider and registers mcp-kali-server for tool access.

Validation involved a natural language prompt in 5ire, using qwen3:4b, to scan scanme.nmap.org on ports 80, 443, 21, and 22. The LLM invoked nmap via MCP, delivering structured results offline, with full GPU processing confirmed.

According to the Kali Linux Team, "the full-stack Ollama, mcp-kali-server, and 5ire are open source, hardware-dependent rather than service-dependent, and tunable based on available VRAM." This configuration offers a privacy-preserving option for red teams and researchers in air-gapped or data-sensitive settings.

Awọn iroyin ti o ni ibatan

Illustration of Kali Linux 2025.4 release on a hacker's laptop screen, showcasing new tools and updated desktop in a realistic cybersecurity workspace.
Àwòrán tí AI ṣe

Kali Linux 2025.4 released with new tools and desktop updates

Ti AI ṣe iroyin Àwòrán tí AI ṣe

Kali Linux has released version 2025.4 on December 12, 2025, marking its final update of the year. The release introduces three new hacking tools, significant desktop environment improvements, and enhanced support for Kali NetHunter. It focuses on modernizing the user experience for cybersecurity professionals and ethical hackers.

The open-source project Ollama has announced the release of its version 0.17. This update features enhancements to OpenClaw onboarding. The news was reported by Phoronix.

Ti AI ṣe iroyin

A new tutorial shows how to run large language models and vision-language models locally on the Arduino UNO Q microcontroller. Edge Impulse's Marc Pous has outlined steps using the yzma tool to enable offline AI inference on the board's Linux environment. This approach allows for privacy-focused applications in edge computing.

AI coding agents from companies like OpenAI, Anthropic, and Google enable extended work on software projects, including writing apps and fixing bugs under human oversight. These tools rely on large language models but face challenges like limited context processing and high computational costs. Understanding their mechanics helps developers decide when to deploy them effectively.

Ti AI ṣe iroyin

The b4 kernel development tool for Linux is now internally testing its AI agent designed to assist with code reviews. This step, known as dog-feeding, marks a practical application of the AI feature within the tool's development process. The update comes from Phoronix, a key source for Linux news.

The Linux Foundation has introduced a new instructor-led workshop focused on deploying small language models in various environments. Titled 'Deploying Small Language Models (LFWS307)', the course offers hands-on training across multiple platforms. Enrollment is now open for this live session.

Ti AI ṣe iroyin

Google has introduced a new command-line interface tool for its Workspace suite, aimed at simplifying integration with AI systems like OpenClaw. The tool bundles APIs from products such as Gmail, Drive, and Calendar, though it is not an officially supported product. This release emphasizes ease of use for both human developers and AI agents.

 

 

 

Ojú-ìwé yìí nlo kuki

A nlo kuki fun itupalẹ lati mu ilọsiwaju wa. Ka ìlànà àṣírí wa fun alaye siwaju sii.
Kọ