Tutorial demonstrates running LLMs on Arduino UNO Q

A new tutorial shows how to run large language models and vision-language models locally on the Arduino UNO Q microcontroller. Edge Impulse's Marc Pous has outlined steps using the yzma tool to enable offline AI inference on the board's Linux environment. This approach allows for privacy-focused applications in edge computing.

The Arduino UNO Q, introduced in recent months, has sparked varied opinions among users. Some appreciate its increased computational power and ability to run Linux, while others view the App Lab environment as confusing and restrictive. Unlike previous Arduino boards, the UNO Q features an STM32H5 coprocessor, making it suitable for complex projects beyond basic tasks like LED blinking.

In a tutorial published on Hackster.io, Edge Impulse engineer Marc Pous demonstrates running high-performance large language models (LLMs) and vision-language models (VLMs) directly on the UNO Q. The guide leverages yzma, a Go wrapper for llama.cpp developed by Ron Evans, known for projects such as Gobot and TinyGo. Yzma simplifies integration of AI inference into Go applications, avoiding complex CGo bindings, and operates within the board's Debian-based Linux system.

Users follow steps to install Go on the UNO Q, configure yzma, and download compatible GGUF models from Hugging Face. For text-based tasks, Pous uses the SmolLM2-135M-Instruct model, which has about 135 million parameters. Quantization and llama.cpp's efficiency enable it to run on the Arm-based hardware, supporting fully offline chat interactions.

The tutorial extends to multimodal capabilities with the SmolVLM2-500M-Video-Instruct model, featuring around 500 million parameters. This model processes images and short videos alongside text. In one example, the UNO Q analyzes a photo of markers on a desk and produces a detailed description without cloud connectivity.

Such local AI execution supports privacy-conscious edge systems, combining microcontroller control with AI for applications in robotics and smart homes. Developers can interpret images, handle voice commands, or process sensor data on-device, opening possibilities for innovative designs.

संबंधित लेख

Photo illustration of Google executives unveiling the Gemini 3 AI model and Antigravity IDE in a conference setting.
AI द्वारा उत्पन्न छवि

Google unveils Gemini 3 AI model and Antigravity IDE

AI द्वारा रिपोर्ट किया गया AI द्वारा उत्पन्न छवि

Google has released Gemini 3 Pro, its latest flagship AI model, emphasizing improved reasoning, visual outputs, and coding capabilities. The company also introduced Antigravity, an AI-first integrated development environment. Both are available in limited preview starting today.

A new demonstration shows the Arduino Uno Q board, a hybrid SoC/MCU device from Qualcomm, capable of running Linux and even 3D games despite its limited resources. The board, priced at $44, uses translation layers to emulate x86 games on its ARM architecture. While functional, it requires additional hardware and cooling for sustained performance.

AI द्वारा रिपोर्ट किया गया

A Los Angeles-based startup, Quilter, has used artificial intelligence to design a functional Linux single-board computer in just one week, requiring under 40 hours of human input. The device, featuring 843 components across two printed circuit boards, successfully booted Debian Linux on its first power-up. This Project Speedrun demonstrates AI's potential to drastically shorten hardware development timelines.

The open-source project Ollama has announced the release of its version 0.17. This update features enhancements to OpenClaw onboarding. The news was reported by Phoronix.

AI द्वारा रिपोर्ट किया गया

MYIR has launched the MYC-CZU3EG-V3, an enhanced system-on-module featuring the AMD Zynq UltraScale+ ZU3EG MPSoC. This compact design integrates Arm processors, real-time cores, and FPGA fabric for applications in industrial automation and edge computing. The module targets sectors like machine vision, aerospace, and telecommunications.

Google has launched Project Genie, a tool based on its Genie 3 AI model that allows users to generate and explore interactive virtual environments from text prompts or images. Available only to subscribers of its premium AI Ultra plan, the system marks the first public access to this advanced world model outside of internal testing. It offers modes like world sketching and remixing, though limited to short 60-second sessions.

AI द्वारा रिपोर्ट किया गया

French AI developer Mistral AI has launched two new transcription models designed to run directly on user devices, prioritizing privacy and speed. The models, Voxtral Mini Transcribe 2 and Voxtral Realtime, aim to keep sensitive conversations off the internet. They enable quick, accurate transcription without relying on cloud servers.

 

 

 

यह वेबसाइट कुकीज़ का उपयोग करती है

हम अपनी साइट को बेहतर बनाने के लिए विश्लेषण के लिए कुकीज़ का उपयोग करते हैं। अधिक जानकारी के लिए हमारी गोपनीयता नीति पढ़ें।
अस्वीकार करें