Browsing Tag
ai-on-device
5 posts
Bringing Multimodal Gemma 4 E2B to the Edge: A Deep Dive into LiteRT-LM and Qualcomm QNN
Running Large Language Models (LLMs) on mobile devices used to be a futuristic dream. Today, with Google’s release…
On-Device RAG for App Developers: Embeddings, Vector Search, and Beyond
Giving your Offline AI agent memory — the ability to search and retrieve your private data We’ve explored why offline AI…
Architectural Evolution and Implementation Strategy of the LiteRT CompiledModel API
Executive Summary The proliferation of high-performance machine learning (ML) on edge devices has precipitated a fundamental shift in…
On-Device Function Calling with FunctionGemma
A Practical Guide to preparing and using Google’s smallest function calling model for Mobile and Web Developers Why…
From Zero to Hero: Running Google’s Gemma 3n on Android with LiteRT & Qualcomm QNN
A complete journey from “What is On-Device AI?” to deploying a state-of-the-art 2B LLM on your phone using…