the personal AI infrastructure is real now
Table of content
by Ray Svitla
three things happened this week that don’t look related until you zoom out.
someone at Shopify shipped a mini search engine for markdown files. no cloud, no indexing service, just a CLI binary you point at your notes. 523 stars overnight.
Dawarich hit 1.0 — a self-hosted alternative to Google Timeline. your location history on your server, not theirs. 350 upvotes, people migrating years of location data.
Apple dropped M5 chips with 4× faster LLM prompt processing. r/LocalLLaMA lit up with people running 70B models at 40+ tok/sec on consumer hardware.
and Cursor — the AI coding assistant — hit $2 billion annual revenue. doubled in three months. 60% from corporate customers.
unrelated? not if you’ve been watching the personal AI OS thesis mature from blog posts into product categories.
the infrastructure thesis
here’s what’s converging:
search without surveillance. qmd is 500 lines of typescript that does one thing: searches your markdown files locally. no Algolia account. no Elasticsearch cluster. no API keys. you type qmd "that thing I wrote about agents" and it returns results from your knowledge base. unix philosophy applied to your second brain.
this matters because most “personal knowledge management” tools are cloud services cosplaying as local-first. Notion is a database you rent. Roam is a graph you don’t own. even Obsidian Sync is convenience over sovereignty. qmd says: your notes, your disk, your grep.
location data without Google. Dawarich does the same for your spatial memory. Google Timeline is a surveillance product dressed as a feature. it knows everywhere you’ve been, everyone you’ve met, every pattern in your life. Dawarich gives you the same visualization — heatmaps, routes, timeline — but the data never leaves your server.
people are migrating years of location history. exporting from Google Takeout, importing into Docker containers running on Raspberry Pis. this isn’t paranoia. it’s pragmatism. when the default is “corporation owns your movement data,” self-hosting is the only exit.
AltStack as the package manager for sovereignty. 450+ self-hosted tools with Docker Compose configs. side-by-side comparisons. best-of rankings. this is npm for the post-cloud era. every SaaS category gets a self-hosted equivalent. Calendly → Cal.com. Airtable → NocoDB. Notion → AppFlowy.
the pattern: sovereignty isn’t one tool. it’s a stack. and the stack needs a catalog.
hardware catches up to ambition
meanwhile, Apple dropped M5 Pro/Max with a single claim: 4× faster LLM prompt processing vs M4.
the reddit reaction was instant: people testing Qwen 3.5, LLaMA 3.3, Mistral variants. reports of 40+ tokens/sec on 70B models. on a laptop. no datacenter. no API bill.
this is the ARM race for personal AI. NVIDIA dominates cloud inference. Apple is betting on-device. if M5 delivers on the 4× claim, running your own agent stack becomes practical, not aspirational. the bottleneck shifts from “can my machine handle this?” to “which model do I want?”
and then LMCache ships — an open-source KV cache layer for local LLMs. stores key-value pairs to accelerate generation. benchmarked across multiple backends. 135 stars on GitHub trending.
LMCache is the software multiplier to M5’s hardware multiplier. if you’re running local inference, caching is the difference between “this is slow” and “this is fast enough to build on.”
the convergence: M5 makes local AI viable. LMCache makes it practical. qmd + Dawarich + AltStack make sovereignty composable.
the market says it’s real
and then there’s Cursor.
$2 billion annual revenue run rate. doubled in three months. founded less than five years ago, valued at $29.3B in November. 60% of revenue from corporate customers.
this is the market saying: agentic coding is infrastructure, not novelty.
when enterprise bets billions on AI agents, it’s not experimenting. it’s replacing workflows. thousands of developers using Cursor means thousands of companies where “your coworker is Claude” is the default, not the future.
Cursor’s revenue is validation for every other personal AI OS bet. if coding agents work at scale, why wouldn’t research agents? why wouldn’t writing agents? why wouldn’t orchestration layers for multi-agent swarms?
the thesis: if one category (coding) can sustain a $29B company, the broader category (personal AI infrastructure) is a multi-hundred-billion-dollar market.
the stack is emerging
zoom out.
qmd: search as infrastructure.
Dawarich: location data as yours.
AltStack: sovereignty as a catalog.
M5: hardware acceleration for local AI.
LMCache: software optimization for inference.
Cursor: market validation that agents are real.
none of these projects know each other. they’re built by different people, for different reasons, in different languages. but they’re all solving the same meta-problem: how do you build a personal AI OS that you own?
the pattern:
- data sovereignty — your notes, your location, your files.
- local-first compute — your models, your hardware, no API bills.
- composable tools — unix philosophy for AI workflows.
- market validation — enterprises paying billions proves it’s not a hobby.
this is the week the personal AI OS graduated from concept to product category.
the infrastructure is real. the hardware is shipping. the market is voting with revenue.
what’s left is integration. someone needs to wire qmd, Dawarich, M5, LMCache, and a dozen other tools into a coherent OS. not a startup. not a VC-funded product. an open protocol.
because if your life is a repo, the OS should be too.
Ray Svitla
stay evolving 🐌