Skip to content

Blog

100k Downloads and the Quiet Philosophy of Local AI on iPhone

April 06, 2026 · 4 min read

Enclave recently crossed 100,000 downloads. In the scale of the App Store, that is a modest number; in the scale of what we care about, it is a quiet signal. Tens of thousands of people—many of them not developers, not ML researchers, not “AI power users”—have chosen an assistant that runs on the device, where prompts and documents can stay local, and where cloud models are optional, not mandatory. This post is not a victory lap. It is a reflection on what that choice might mean for the next decade of personal computing.

Gemma 4 Release: On-Device Models from E2B to 31B

April 02, 2026 · 5 min read

Gemma 4 is Google’s latest open-weight family, announced April 2, 2026: four checkpoints aimed at everything from ultra-mobile browsers to serious local and self-hosted setups. The lineup pairs 2B and 4B effective “E” models (built with the same Per-Layer Embedding ideas that made Gemma 3n practical on phones) with a 31B dense flagship and a 26B mixture-of-experts variant that activates 4B parameters per token. Smaller models carry a 128K context window; the larger two expand to 256K. Gemma 4 is licensed under Apache 2.0—a shift from the Gemma Terms of Use that applied to earlier generations—so downstream use, modification, and redistribution follow a familiar open-source playbook.

LLM Knowledge Distillation Explained for On-Device AI

March 29, 2026 · 7 min read

LLM knowledge distillation is a training technique where a small “student” model learns to imitate a much larger “teacher” model — so the student can run on a phone or laptop while still behaving more like the big model than it could if trained only on raw text. In 2026 you are seeing this idea everywhere: vendors want flagship-quality answers without shipping a 400-billion-parameter file to every device. This article explains how distillation works in plain language, how it differs from quantization and ordinary fine-tuning, and what it means for privacy when your chat runs locally.

KV Cache Explained Simply: Why Chats Need Extra RAM

March 22, 2026 · 7 min read

When you run AI on your phone or Mac, the app downloads a model file (often a few gigabytes). That number is easy to understand. What is harder to see is that while you chat, the app also keeps a growing pile of scratch notes in memory so the model does not have to reread your entire conversation from scratch every time it adds the next word. That scratch space is usually called the KV cache. It is a normal part of how modern language models work — and it is a big reason a long thread can feel fine at first, then get slow or unstable, even when the model file itself never changed.

LLM Quantization Explained: Run Bigger Models on Less RAM

March 15, 2026 · 10 min read

LLM quantization is a compression technique that shrinks a language model’s memory footprint by storing its parameters in lower-precision number formats — for example, using 4 bits per weight instead of 16. This lets you run models that would normally require 14 GB of RAM in under 4 GB, with surprisingly little quality loss. If you have ever wondered how people run 70-billion-parameter models on a laptop, or what “Q4_K_M” means on a Hugging Face download page, this guide explains it from the ground up.

Enclave 1.70: Thinking Support, Live Metrics, and a New Brain

March 12, 2026 · 6 min read

Enclave 1.70 is the most transparent version of the app we have ever shipped. You can now watch your local model think through problems in real time, see exactly how fast it generates text, and run conversations on a new embedded model that delivers noticeably better answers — all completely offline, completely private, on your own device. Here is everything that changed and why it matters.

Qwen 3.5: A Complete Model Family from 0.8B to 397B

March 08, 2026 · 9 min read

Alibaba’s Qwen team just finished rolling out Qwen 3.5 — not one model but a family of eight, spanning from a 0.8-billion-parameter model that fits on a phone to a 397-billion-parameter flagship that trades blows with the best closed-source systems. Every model ships under Apache 2.0, supports 201 languages, handles text, images, and video natively, and uses a new hybrid architecture that replaces most of the standard attention mechanism with something fundamentally more efficient. GGUF quantizations are already on Hugging Face. Here is what the full lineup offers, how the architecture works, and what it means for running serious AI locally.

Tiny Aya: 70 Languages Running Locally on iPhone and Mac

February 28, 2026 · 6 min read

Most “multilingual” AI models treat language support as a checkbox — train heavily on English, sprinkle in some Chinese and Spanish, and list 100+ languages in the marketing copy. Cohere Labs just shipped something different. Tiny Aya, released February 17, is a 3.35-billion-parameter open-weight model that genuinely supports 70+ languages — including Bengali, Hindi, Tamil, Yoruba, and dozens of others that frontier models treat as afterthoughts — and runs locally on an iPhone 17 Pro at 32 tokens per second. GGUF versions are available on Hugging Face right now, making it immediately usable for local inference on iPhone and Mac.

llama.cpp Joins Hugging Face: What It Means for Local AI

February 21, 2026 · 5 min read

llama.cpp, the open-source engine behind nearly every local AI tool in existence, just joined Hugging Face. Georgi Gerganov and the founding ggml.ai team announced on February 20, 2026 that they are moving to Hugging Face as full-time employees — bringing together the model distribution layer (Hugging Face Hub) with the local inference layer (llama.cpp) under one roof. The projects remain fully open-source. Here is what this means for anyone who runs AI on their own hardware.

OpenClaw Personal AI Assistant: 2026 Guide

February 14, 2026 · 5 min read

OpenClaw personal AI assistant is one of the biggest self-hosted AI projects right now, and for good reason: it lets you run a single gateway that connects your existing chat channels to one always-on assistant. Instead of opening a separate app, you can message your AI through WhatsApp, Telegram, Discord, Slack, iMessage, and more. If you want control, flexibility, and a privacy-first setup, OpenClaw is worth understanding.