Tutorial demonstrates running LLMs on Arduino UNO Q

A new tutorial shows how to run large language models and vision-language models locally on the Arduino UNO Q microcontroller. Edge Impulse's Marc Pous has outlined steps using the yzma tool to enable offline AI inference on the board's Linux environment. This approach allows for privacy-focused applications in edge computing.

The Arduino UNO Q, introduced in recent months, has sparked varied opinions among users. Some appreciate its increased computational power and ability to run Linux, while others view the App Lab environment as confusing and restrictive. Unlike previous Arduino boards, the UNO Q features an STM32H5 coprocessor, making it suitable for complex projects beyond basic tasks like LED blinking.

In a tutorial published on Hackster.io, Edge Impulse engineer Marc Pous demonstrates running high-performance large language models (LLMs) and vision-language models (VLMs) directly on the UNO Q. The guide leverages yzma, a Go wrapper for llama.cpp developed by Ron Evans, known for projects such as Gobot and TinyGo. Yzma simplifies integration of AI inference into Go applications, avoiding complex CGo bindings, and operates within the board's Debian-based Linux system.

Users follow steps to install Go on the UNO Q, configure yzma, and download compatible GGUF models from Hugging Face. For text-based tasks, Pous uses the SmolLM2-135M-Instruct model, which has about 135 million parameters. Quantization and llama.cpp's efficiency enable it to run on the Arm-based hardware, supporting fully offline chat interactions.

The tutorial extends to multimodal capabilities with the SmolVLM2-500M-Video-Instruct model, featuring around 500 million parameters. This model processes images and short videos alongside text. In one example, the UNO Q analyzes a photo of markers on a desk and produces a detailed description without cloud connectivity.

Such local AI execution supports privacy-conscious edge systems, combining microcontroller control with AI for applications in robotics and smart homes. Developers can interpret images, handle voice commands, or process sensor data on-device, opening possibilities for innovative designs.

相关文章

A new demonstration shows the Arduino Uno Q board, a hybrid SoC/MCU device from Qualcomm, capable of running Linux and even 3D games despite its limited resources. The board, priced at $44, uses translation layers to emulate x86 games on its ARM architecture. While functional, it requires additional hardware and cooling for sustained performance.

由 AI 报道

A tech enthusiast shares their experience using Linux to run local large language models, claiming it is simpler than on Windows. They highlight the ability to access a ChatGPT-like interface directly in the terminal. The article was published on March 10, 2026.

The open-source project LLVM has introduced a new policy allowing AI-generated code in contributions, provided humans review and understand the submissions. This 'human in the loop' approach ensures accountability while addressing community concerns about transparency. The policy, developed with input from contributors, balances innovation with reliability in software development.

由 AI 报道

A new social network called Moltbook, designed exclusively for AI chatbots, has drawn global attention for posts about world domination and existential crises. However, experts clarify that much of the content is generated by large language models without true intelligence, and some is even written by humans. The platform stems from an open-source project aimed at creating personal AI assistants.

 

 

 

此网站使用 cookie

我们使用 cookie 进行分析以改进我们的网站。阅读我们的 隐私政策 以获取更多信息。
拒绝