← All Posts
AI On-Device Mar 24, 2026

Your Phone Can Now Run a 400B AI Model — No Cloud Required

An iPhone 17 Pro just ran a 400 billion parameter language model locally. No API keys. No monthly subscriptions. No data leaving your device.

This happened yesterday. A team called ANEMLL demonstrated running a massive LLM directly on consumer hardware — and the Hacker News community lost it (638 points, 281 comments).

For indie hackers, this isn't just a cool demo. It's a paradigm shift.

What Actually Happened

ANEMLL (Artificial Neural Engine Machine Learning Library) is an open-source project that ports large language models to Apple's Neural Engine. Their latest demo showed a 400B parameter model running inference on an iPhone 17 Pro.

Key specs: The iPhone 17 Pro launched with 50% more RAM and roughly double the inference performance of its predecessor. Prompt processing speed jumped 10x compared to the previous generation.

This isn't vaporware. It's real hardware running real models. And the software stack is open source.

Why This Matters for Builders

Let me break down what this means if you're building products:

Cloud AI (Today)

$0.01-0.15
Per 1K tokens (GPT-4 class)

On-Device AI (Coming)

$0.00
Per token — hardware already bought

Think about that. Every AI-powered feature you ship today has a per-use cost. Every API call is money leaving your pocket. On-device inference eliminates that entirely.

But cost is just the beginning:

The Open Source Stack

ANEMLL provides a complete pipeline:

They're currently on alpha release 0.1.1. Early, but functional.

ANEMLL Project MSA: Memory Sparse Attention

What You Can Build Right Now

Here's where my mind goes as someone who ships products:

1. Privacy-first journaling app. AI-powered mood analysis and insights that never sync to a server. Parents would pay for this for their kids.

2. Offline translation. Travel apps that work in the subway, the plane, the middle of nowhere. The market for this is massive and underserved.

3. Local code assistant. A coding companion that works in restricted environments — government, healthcare, finance — where cloud AI is banned.

4. Smart home controller. Process voice commands locally. No always-on microphone sending data to the cloud.

None of these require a PhD in ML. The ANEMLL stack handles the hard parts. You build the product layer.

The Honest Take

Let's be real about where we are:

Running a 400B model on a phone is impressive but constrained. Context windows are limited. Generation speed isn't going to beat GPT-4 for complex reasoning tasks. The alpha software will have bugs.

But here's what matters: the trajectory is clear. Hardware keeps getting faster. Model architectures keep getting more efficient. Quantization techniques keep squeezing more into less.

A year ago, running a 7B model on a phone was news. Today it's 400B. That's a 57x improvement in one year.

Where do you think we'll be in another year?

What Smart Builders Do

The indie hackers who win aren't the ones waiting for perfect technology. They're the ones building with what's available today and positioning for what's coming tomorrow.

Start experimenting with on-device inference now. Understand the constraints. Build your mental model of what works and what doesn't. When the hardware catches up to your ambitions (and it will), you'll already know exactly what to build.

The cloud AI era isn't ending tomorrow. But the on-device AI era has officially started. And the window for building in this space while it's still early? That window is open right now.

Ship AI Products Faster

I put together a complete setup guide for building with AI — from local models to deployment. Everything I wish I knew when I started.

Get the OpenClaw Ultimate Setup — $29

The exact workflow I use to ship AI-powered products

Source: HN #9 today — 638 points, 281 comments