the infrastructure layer
the week
AionUi hits 17K stars overnight → open-source shell for every coding agent CLI. Chrome for AI.
ByteDance ships deer-flow → SuperAgent harness with sandboxes, memory, subagents. tasks from minutes to hours.
Obsidian goes headless → sync your vault to a server, no GUI. agent memory as infrastructure.
Qwen 3.5-35B-A3B → local model beats GPT-OSS-120B at 1/3 the size. sovereignty is pragmatism.
bare-metal AI → someone booted a laptop directly into LLM inference. no OS. just UEFI → AI.
vibe coding fractures r/selfhosted → 424 upvotes, 282 comments. do we accept disposable tools or build better filters?
1. AionUi — free cowork infrastructure for every CLI
what happened: AionUi hit 17K stars overnight. it’s an open-source 24/7 cowork app that wraps every major coding agent CLI: Gemini CLI, Claude Code, Codex, OpenCode, Qwen Code, Goose CLI, Auggie. one interface, every backend. local-first, free.
think Chrome for coding agents. unified UI, swappable engines.
why it matters: the “which agent should I use” question gets easier when you can try all of them without rewriting your workflow. AionUi is the shell layer for the multi-agent future. when OpenAI ships codex 2.0, you don’t relearn a UI — you swap the backend.
this is what the personal AI OS looks like: composable tools, user-controlled infrastructure, no lock-in.
signal: AionUi on GitHub
2. deer-flow — ByteDance’s SuperAgent harness
what happened: ByteDance shipped an open-source “SuperAgent” framework. sandboxes, memories, tools, skills, subagents. handles tasks from minutes to hours. one parent agent coordinates subagents. each subagent has isolated memory, tool access. the parent doesn’t solve — it delegates, waits, synthesizes.
why it matters: this is the missing layer between “Claude solves one task” and “my AI handles my entire workflow.” deer-flow is what happens when you take agent architecture seriously. not a chatbot. not a code generator. a supervisor.
the implication: your AI doesn’t need to be smarter. it needs to coordinate better.
signal: deer-flow on GitHub
3. Obsidian headless client (open beta)
what happened: Obsidian Sync launched a headless client. sync vaults to a server without running the desktop app. CLI-friendly, automation-ready. your server runs headless Obsidian, syncs your vault, feeds it to your agents. no GUI needed.
why it matters: if your PKM is your agent’s memory, it needs to run everywhere. headless Obsidian turns your notes into infrastructure. this is the plumbing for agent memory systems that span devices. your knowledge graph lives on your Pi, your VPS, your NAS. always synced, always accessible.
no API. no vendor. just files.
signal: Obsidian headless docs
4. Qwen 3.5-35B-A3B — the local model that works
what happened: 431 upvotes, 114 comments on r/LocalLLaMA. users report Qwen 3.5-35B-A3B replaced GPT-OSS-120B as their daily driver. same quality, 1/3 the size. use cases: local agent workflows, N8N automation, dynamic system generation.
why it matters: the best personal AI OS is the one you own. Qwen 3.5-35B is crossing the threshold where local models beat cloud for real work. if your agent stack runs on your machine, you control the data, the uptime, the cost structure.
sovereignty isn’t ideology. it’s pragmatism.
signal: reddit discussion
5. bare-metal AI — booting directly into LLM inference
what happened: someone built a UEFI application that boots directly into LLM chat. no OS. no kernel. no drivers (well, wifi works). power on, select “Run Live”, type “chat”, talk to an AI. entire stack in UEFI boot services mode: tokenizer, weight loader, tensor math, inference engine. written in C. runs on a Dell E6510 from 2010.
why it matters: this is the opposite of bloat. most AI apps run on Electron on Chromium on an OS on firmware. this is: firmware → AI. nothing else.
it’s a proof that you don’t need a million dependencies to ship intelligence. sometimes the most radical move is subtraction.
signal: reddit discussion
6. vibe coding hits self-hosted
what happened: r/selfhosted is fracturing. 424 upvotes, 282 comments. someone proposed a sub-subreddit for vibe-coded apps. the flood is here. apps pumped out with Claude, disposable quality, security unknown, longevity questionable.
the tension: vibe coding lowers the barrier to entry (good), but floods the ecosystem with disposable tools (bad). self-hosted purists vs vibe engineers. the old guard vs the new chaos.
why it matters: this is the culture war inside the AI OS movement. do we accept that most agent-built tools are disposable? do we fork the community? or do we build better filters, better audits, better infrastructure to handle the flood?
the answer shapes whether “your life is a repo” becomes a movement or a meme.
signal: reddit discussion