I just ran Qwen 3.5 on my MacBook Air. Not through an API. Not through a cloud service. Locally, on the metal, via Ollama. It thinks out loud, reasons through problems, and gives genuinely useful answers — all without a single token leaving my machine. Zero API cost. Zero latency to a data center. Zero dependency on someone else's infrastructure staying online.

And the results? Not bad at all.

Qwen 3.5 4B running locally via Ollama on a MacBook Air, showing thinking and response
Qwen 3.5 (4B) running locally on a MacBook Air via Ollama — visible chain-of-thought reasoning, zero API calls.

What We've Seen Firsthand

We run four AI agents at tabiji. They produce Instagram Reels, build travel itineraries, generate music, and manage our infrastructure. Right now, most of that runs through frontier APIs — Claude, Gemini, MiniMax. Our token costs are real: roughly $500/month to keep the operation running.

But we've been experimenting with local models for the simpler tasks, and the economics are compelling:

🏠 Local Model (Qwen 3.5, 4B)

  • Cost per token: $0.00
  • Latency: ~50ms first token
  • Privacy: complete — nothing leaves the machine
  • Availability: 100% (no outages, no rate limits)
  • Good at: summaries, classification, drafting, Q&A, code completion

☁️ Frontier API (Claude Opus, GPT-4o)

  • Cost per token: $15-75 per million
  • Latency: 200-2000ms first token
  • Privacy: data sent to third-party servers
  • Availability: 99.5% (outages happen, rate limits hit)
  • Good at: complex reasoning, novel problems, multi-step agents

The pattern is clear. For the ~80% of tasks that are "good enough" territory, local models already win on every dimension except raw intelligence. And for the 20% where you genuinely need frontier capability, you call the API.

That's the future: adaptive routing between local and cloud. Not one or the other. Both — with intelligence about which to use when.

What This Means for Token Economics

Here's the part that keeps me up at night — in a good way.

If you're a company spending $10,000/month on API tokens today, imagine this world:

  • 80% of your AI workload runs on local hardware you already own. Cost: $0.
  • 15% of your workload runs on a small on-premise server with a mid-range GPU. Cost: amortized hardware, ~$200/month.
  • 5% of your workload — the truly hard stuff — calls a frontier API. Cost: $500/month instead of $10,000.

That's a 95% cost reduction. Not through some clever optimization or prompt engineering hack. Just through intelligent routing to models that are already available, running on hardware that's already on your desk.

For individual developers and small teams, it's even more dramatic. Local inference is free. You buy the hardware once, and every token after that costs nothing. The only limit is how fast your machine can run the model.

The Frontier Still Matters — But Differently

I'm not saying OpenAI and Anthropic are doomed. Far from it. The frontier will always matter for the hardest problems. When I need an agent to orchestrate a complex workflow across multiple tools, reason about ambiguous situations, or handle truly novel problems — I want the best model money can buy.

But the business model of frontier AI will shift. Instead of charging for every token of every task, frontier providers will increasingly compete on the tasks that only they can do. The commodity layer — the stuff that a 4B model handles fine — becomes a race to zero. The premium layer — genuine intelligence, breakthrough reasoning, reliable agency — that's where the value concentrates.

Think about it like computing in general. Most of your apps run locally on your phone. But when you need massive parallel processing — training a model, rendering a movie, running a simulation — you rent cloud compute. The cloud didn't kill local computing. It found its niche alongside it. (Meanwhile, the economics of the internet are shifting in the same direction — decentralization of compute follows decentralization of content.)

AI is heading the same way.

The Endgame

Here's my prediction: within three years, the default mode of interacting with AI will be local. Your phone, your laptop, your car, your smart home hub — they'll all run capable language models natively. You won't think about "calling an API" any more than you think about "calling a server" when you open your calculator app.

The cloud will still exist for the hard stuff. When you need to train a new model, orchestrate a hundred agents, or solve a problem no local model can handle — you'll reach for frontier APIs. But that will feel like an escalation, not the default. (We wrote about this shift in The Future of Content Is Agentic Data Enrichment — the same pattern applies: local models handle the grunt work, frontier models handle the judgment calls.)

Token costs won't go to zero because the APIs disappear. They'll go to zero because most of the tokens will be generated on hardware you already own.

Better hardware → better local models → adaptive use of frontier models as necessary. The math is simple. The implications are enormous.

We're already running Qwen 3.5 on a MacBook Air and getting useful results. That's today, March 2026, with a 4B model. Imagine what a 30B model on an M5 MacBook Pro will look like in 2027. Or a 100B model on whatever hardware exists in 2028.

The future of AI isn't a data center in the desert. It's the device in your hand.