Offline Real-time AI Camera Assistant (Flutter + LLaMA.cpp)
Project Overview
SmolVLM Flutter App is a real-time, offline AI camera assistant built using Flutter and LLaMA.cpp. It captures live camera frames and sends them to a locally hosted LLaMA multimodal server running SmolVLM-500M, which responds with intelligent, natural language descriptions of the scene.
This project demonstrates how AI and computer vision can run directly on-device, making it useful for accessibility, education, smart agriculture, and robotics โ all without relying on cloud services.
Key Features:
๐ท Captures images using front/back camera with seamless switching
๐ Sends frames to the AI server every few seconds
๐ง Generates real-time smart feedback using SmolVLM-500M model
๐ ๏ธ Fully offline setup โ no internet required
๐งผ Handles UI layout and preview stretching for clean UX
Technologies Used:
Flutter, Dart, Camera plugin, SmolVLM-500M-Instruct-f16.gguf (model), llama-server from LLaMA.cpp, Base64 communication
How It Works:
Flutter app captures camera frames every few seconds.
Encodes the frame as a Base64 image.
Sends the image to a locally running LLaMA server.
Server uses SmolVLM to generate a description.
App displays the AI-generated feedback in real-time.
Use Cases:
๐ Accessibility for visually impaired users
๐ Educational tools for visual recognition
๐ ๏ธ Real-time debugging or documentation assistant