you are hosting now

Table of content

by Ray Svitla


someone just built a self-hosted YouTube approval system for their kid. not “block YouTube” or “trust YouTube Kids” — a manual gate where the kid searches, picks a video, and waits for parent approval before it plays.

the project is called BrainRotGuard. it solves a problem that used to be “set screen time limits” and is now “build deployment pipelines for algorithm outputs.”

parenting has become infrastructure work.

this is not an isolated case. across Reddit, GitHub, and developer blogs, a pattern is emerging: people are buying hardware, self-hosting services, and running agents locally — not because they’re hobbyists or engineers, but because that’s what it takes to run a personal AI system.

Andrej Karpathy bought a Mac Mini to tinker with Claws (Claude-based agents). the Apple Store person told him “selling like hotcakes and everyone is confused.”

people are buying computers to run agents, not apps.


the shift: from consuming to hosting

for decades, the consumer computing model was simple: you buy a computer, install software, use it. cloud computing flipped this: you rent access to someone else’s computer, use their software, pay monthly.

personal AI is flipping it again: you run your own infrastructure. your computer hosts agents, models, and persistent memory. your home network becomes a private cloud.

this is not a return to the desktop software era. it’s something new: local-first AI infrastructure.


what changed

three things converged to make this possible:

1. local models got good enough

when Georgi Gerganov released llama.cpp in March 2023, it made local models accessible to non-ML-engineers. you could run a 7B model on a laptop and get coherent outputs. not GPT-4 quality, but good enough for most tasks.

this week, Hugging Face acquired ggml.ai (Gerganov’s company) to “ensure the long-term progress of local AI.” the consolidation is happening because the demand is real.

a Canadian hardware startup called Taalas just announced custom hardware that serves Llama 3.1 8B at 17,000 tokens/second. that’s faster than most cloud APIs.

local models are no longer a compromise. for many use cases, they’re the better option.

2. agents became persistent

the shift from chatbots to agents is not just about capabilities — it’s about persistence.

a chatbot lives in a session. you open a tab, ask questions, close the tab. memory resets.

an agent lives on your machine. it has access to your files, your codebase, your calendar. it remembers context across sessions. it can take actions while you’re offline.

this is the difference between “ask the bot” and “the bot knows.”

Gaia (a GitHub project trending this week) describes itself as a “proactive personal assistant inspired by Jarvis” — not another reactive chatbot, but an agent that anticipates needs, learns patterns, and takes initiative.

that kind of agent can’t live in a cloud session. it needs to run locally, with persistent access to your data.

3. agent security became personal security

when your agent has access to your files, your credentials, and your identity, security stops being an abstract concern. it’s personal.

clawsec (a new security skill suite for OpenClaw agents) offers drift detection, live security recommendations, automated audits, and skill integrity verification. it protects your SOUL.md — the file that defines your agent’s identity.

your SOUL.md is your identity file. if someone tampers with it, your agent becomes someone else’s tool.

this is a new kind of security model. you’re not protecting a server in a data center. you’re protecting an agent running in your home, with access to everything you care about.


the new infrastructure layer

what does this look like in practice?

parenting as infrastructure: BrainRotGuard receives YouTube search requests from your kid, displays results, waits for your approval, and only then serves the video. manual approval queue for algorithm outputs.

coding as distributed workflow: claude-code-telegram lets you interact with Claude Code via Telegram. your agent lives on your machine, but you control it from your phone. remote access to your personal AI.

blogging as canonical aggregation: Simon Willison just added “beats” to his blog — TILs, releases, museums, tools, research — all synced from activity elsewhere. his personal site is the aggregator of his work across platforms. centralize the narrative, not the data.

these are not fringe use cases. they’re the new normal for people running personal AI systems.


the hardware question

if you’re running agents locally, you need hardware. and not just any hardware — hardware that can run models, host services, and stay on 24/7.

Mac Minis are “selling like hotcakes and everyone is confused” because they’re the easiest way to run local AI. plug it in, install Ollama or llama.cpp, and you’re running Llama 3.1 70B on your desk.

but the hardware question goes deeper. how much compute do you need? how much memory? how much storage for persistent context?

the answer depends on what you’re hosting. if it’s just a chatbot, a laptop is fine. if it’s a persistent agent with access to your files, you need more. if it’s a family infrastructure (approval queues, scheduled tasks, proactive monitoring), you’re building a home server.

the line between “personal computer” and “personal data center” is blurring.


what this means

the shift from consuming software to hosting infrastructure is not just a technical change. it’s a mindset shift.

you are no longer a user. you are a sysadmin.

your computer is no longer a terminal. it’s a server.

your life is no longer a collection of accounts on other people’s platforms. it’s a repo, and you’re running the host.

this is uncomfortable for most people. it requires technical knowledge, ongoing maintenance, and a tolerance for things breaking. it’s why “everyone is confused.”

but it’s also liberating. you control the data. you control the models. you control what runs, when it runs, and who has access.

your agent doesn’t call home to OpenAI or Anthropic. it lives in your home. literally.


the questions we need to ask

if personal AI means hosting your own infrastructure, what happens to people who can’t or won’t do that?

does local-first AI create a new digital divide — not between those who have internet access and those who don’t, but between those who can self-host and those who can’t?

what happens when your personal AI infrastructure fails? there’s no customer support. no SLA. no uptime guarantee. it’s on you.

and what happens when a significant portion of AI interaction shifts from centralized cloud services to distributed home servers? what does that do to safety, moderation, and alignment?

these are not hypothetical questions. they’re happening now, as people buy Mac Minis to run Claws and build self-hosted YouTube approval systems for their kids.


the new normal

you are hosting now. whether you chose it or not, whether you’re ready or not.

your life is a repo. your blog is the README. your agents are running in the background, learning your patterns, anticipating your needs, and taking actions on your behalf.

the infrastructure is in your home. the responsibility is yours.

welcome to the personal AI era.


Ray Svitla stay evolving 🐌