Concepts

The mental models that make a personal operating system work. Understand the why before the how.

easy first
easy first
hard first
10 AI Agent Failure Modes: Why Agents Break in Production The documented ways AI agents fail: hallucination cascades, context overflow, tool calling errors, and 7 more. Diagnosis patterns and fixes for each. Agent Checkpointing: Save, Restore, and Rewind Autonomous Work How checkpoint systems enable long-running agent workflows by saving state periodically, allowing recovery from failures and rollback from bad decisions Agent Guardrails: Input/Output Validation for Autonomous Systems How to implement runtime guardrails that validate agent inputs, filter outputs, and enforce business rules. Covers NeMo Guardrails, layered checking, and production patterns. Agent Handoffs: When and How to Transfer Control Learn when to use agent handoffs vs single agents. Includes code patterns from OpenAI Swarm and practical implementation examples. Agent Memory Systems How AI agents implement memory: short-term context, long-term storage, vector retrieval, and the architecture that ties it together. Agent Observability How to implement distributed tracing, logging, and monitoring for AI agents using OpenTelemetry and purpose-built tools like Langfuse and Braintrust. Agentic Design Patterns: ReAct, Reflection, Planning, Tool Use When to use ReAct loops, self-critique, task decomposition, and tool calling in AI agents. A practical pattern library for building effective agent systems. AI Memory Compression Techniques for compressing AI observations into retrievable semantic summaries that fit in context windows Browser Agents AI that clicks, types, and scrolls your browser autonomously to complete web-based tasks Building an AI Second Brain Transform AI from chatbot to persistent knowledge partner Context Rot: When More Tokens Mean Worse Results LLM performance degrades predictably as context windows fill up. Learn why this happens, how to detect it, and practical strategies to maintain output quality. Context Window Management Keep your AI sharp by managing what fits in its working memory Digital Gardens Non-linear personal knowledge sites that grow over time. Not blogs. Notes evolve publicly without pressure to be finished. Episodic Memory for LLM Agents Give AI agents memory of specific past events with temporal context. The missing piece between semantic facts and procedural rules in the CoALA framework. Graph Memory for Personal AI Knowledge graphs track relationships between people, projects, and time that vector databases miss. Build AI memory that understands context across sessions. Human-on-the-Loop Move from approving every AI action to supervising agents that act autonomously, escalating only when confidence drops or risk rises. Hybrid Retrieval: When RAG Meets Long Context Combine RAG retrieval with long-context windows strategically instead of treating them as competing approaches Hybrid Search: Combining Keyword and Semantic Retrieval Vector search misses exact matches. Keyword search misses concepts. Hybrid search with reciprocal rank fusion combines both for personal knowledge bases. Late Chunking: Context-Aware Document Splitting for Better Retrieval Process entire documents through embedding models before splitting to preserve cross-chunk context that traditional chunking destroys Learning in Public Stop lurking. Create learning exhaust: blogs, videos, talks. The fastest way to level up is documenting what you learn where others can find it. Local LLM Runtimes: When to Use Ollama vs vLLM Ollama excels for single-user development with simple setup. vLLM delivers 20x higher throughput for production multi-user deployments. Choose based on your workload. Malleable Software Software as clay you reshape, not appliances you consume MCP Server Composition Connect your AI agent to multiple MCP servers at once, combining calendar, database, files, and search through one protocol Memory Attribution and Provenance Track where AI memories came from, when they were created, and how much to trust them Memory Consolidation and Forgetting How AI agents consolidate short-term observations into long-term storage using sleep-inspired patterns, plus when and what to forget. Model Quantization: Running 70B Models on a Laptop Reduce model precision from 32-bit to 4-bit to run large language models locally. Covers k-quants, GGUF, and choosing the right quantization level. Multi-Agent Knowledge Management When a single AI can't handle your PKM needs, specialized agents working together can automate capture, processing, and synthesis. Personal Search: Searching Your Own Data Build a search engine for your notes, journals, tweets, and memories instead of the web Plan Mode in Claude Code Research before coding to avoid wrong approaches and hallucinations Preference Learning: AI That Adapts to You How AI systems infer your preferences from interactions and adapt without configuration. Covers POPI, Mem0, LaMP benchmarks, and building preference-aware systems. Principles for AI Delegation What to delegate to AI and what to keep human Prompt Engineering for Agent Coding Structure prompts that make AI coding agents 30-40% more effective Prompt Engineering Patterns That Work Five proven patterns to get better results from LLMs: chain-of-thought, few-shot, ReAct, prompt chaining, and role prompting with examples Prompt Injection in Skill Files AI agents read project files like CLAUDE.md or AGENTS.md. These files can contain hidden instructions that hijack the agent. Here's how the attack works. Self-Evolving Agents Build AI agents that improve through structured feedback capture, automated evaluation, and continuous retraining loops Self-Updating Instructions (Procedural Memory) Build AI agents that modify their own operating instructions based on experience, feedback, and observed failures Small Bets Stop betting everything on one startup. Build a portfolio of small projects that compound over time. Subagent Patterns: Parallel, Sequential, Background Three dispatch patterns for delegating work to AI subagents and when to use each one. Terminal-Native AI Coding: Neovim with Claude Code Run Claude Code alongside Neovim in split terminal panes. Keep your vim keybindings while getting AI assistance. No VS Code required. The Architecture of a Personal OS Personal OS architecture: interface, agent, memory, integration, and tool layers. Build your AI system incrementally in 4 weeks The Jagged Frontier AI capabilities have irregular boundaries. Some tasks it nails, others it fails spectacularly. Learn to map the frontier for your work. The Three-Layer Workflow Match your AI tool to the task: tab completion for most work, agents for multi-file changes, reasoning for architecture Tool Routing: How AI Agents Pick Which Function to Call Modern agents route between dozens of tools using semantic matching, LLM-as-router, hierarchical patterns, and fallback chains. Patterns for scoring, selection, and MCP sampling. Tool Use Patterns: How LLMs Call External Tools Function calling, MCP protocol, and ReAct patterns for AI tool use. Learn when to use each approach and how to implement them. Vibe Coding Building software through conversation instead of keystrokes Vision-Based Web Automation: Why Screenshots Are Replacing Selectors How computer vision and multimodal LLMs enable browser agents that see pages like humans instead of parsing brittle DOM structures. Voice-First Note Capture: Whisper to Structured Markdown Use whisper.cpp for local transcription, then LLM post-processing to convert rambling voice memos into structured notes with headers, bullet points, and action items What is a Personal Operating System? What is a personal operating system? Learn how AI agents like Claude Code manage your tasks, memory, calendar, and decisions autonomously

48 concepts