06-reference

indydevdan agent experts self improving

Sun Apr 19 2026 20:00:00 GMT-0400 (Eastern Daylight Time) ·reference ·source: IndyDevDan YouTube ·by IndyDevDan
indy-dev-danagent-expertsself-improving-agentsmeta-agenticsexpertise-filesmental-modelcore-fourcontext-engineeringjust-in-time-contextcustom-agentsharness-engineering

IndyDevDan — Agent Experts: Finally, Agents That ACTUALLY Learn

Why this is in the vault

19-minute lesson where Dan articulates the agent expert pattern — an agent paired with a YAML “expertise file” that the agent reads BEFORE doing work and writes back AFTER doing work, so the agent’s mental model evolves in place. The vault keeps it because it makes one specific architectural claim that nothing else in the IndyDevDan corpus does: the expertise file is NOT a source of truth — it is a working memory file, distinct from code, comments, PRDs, and plans. This is the missing operational answer to the “agents forget” problem that everyone (Anthropic, every harness vendor, the founder) has been gesturing at but not solving cleanly. Pairs directly with the 2026-04-19-indydevdan-cracked-claude-agent-skills piece — skills are the static building blocks; agent experts are skills with a self-improvement loop bolted on. Also closes a loop opened by 2026-04-15-thariq-claude-code-session-management-1m-context — Thariq said context windows degrade with size; the expertise-file pattern is one concrete answer for how a long-running agent system avoids needing a giant context window in the first place.

Core argument

  1. Memory files, prime prompts, sub-agents, and skills all share one defect: they must be MANUALLY updated. This is the structural reason agents stop improving past a certain point — the human is the bottleneck on the agent’s growth. Every other mechanism Dan lists is steerable but not self-steering.
  2. An agent expert is a self-improving template metaprompt. Three-step lifecycle: (a) read expertise file, (b) execute with mental model + validate against the codebase as ground truth, (c) self-improve — append/update the expertise file based on what was learned during the run.
  3. The expertise file is a working mental model, NOT a source of truth. Critical distinction. The CODE is the source of truth. The expertise file is a YAML data structure that captures “what this agent has learned about a specific area of the codebase” — analogous to a senior engineer’s intuition about a system, not the system’s documentation.
  4. Meta-agentics — meta-prompt, meta-agent, meta-skill — is the foundation layer. The system that builds the system. Dan demos a meta-prompt creating a question-with-mermaid-diagrams prompt, a meta-agent creating a planner agent, a meta-skill creating a start-orchestrator skill. Every codebase Dan creates has these. Three times marks a pattern; pattern marks a meta-tool.
  5. Expertise files give just-in-time context, not always-on context. Memory files (CLAUDE.md, AGENTS.md) are GLOBAL FORCED CONTEXT — they always load. Expertise files are LAZY — only the agent that needs the websocket expertise reads the websocket expertise file. This is how a 10-expert codebase doesn’t blow every agent’s context window with 9 files of irrelevant knowledge.
  6. Scale per-expert with parallel agent fan-out. Three websocket experts answering the same question gives higher confidence in the answer than one expert answering once. Compute scaling at the expert level, not just the agent level.
  7. The Core Four (context, model, prompt, tools) still applies — agent experts are just the prompt+context layer specialized. Dan keeps returning to the Core Four as the only abstraction that matters. Skills, custom agents, MCP servers — all decompose to manipulations of these four.
  8. Self-improve must be the THIRD step of the three-step workflow. Plan, build, self-improve. If self-improve is optional or skipped, the agent never learns. The discipline of always running self-improve at the end is what turns a generic agent into an expert.

Mapping against Ray Data Co

Open follow-ups

Sponsorship

The video is structured as lesson seven of Tactical Agentic Coding / Agentic Horizon — Dan explicitly references this as a lesson and the closing pitch reinforces the course as the place to “go deeper.” The technical claims (agent experts, expertise files, meta-agentics, Core Four) stand on their own; they are descriptive of patterns Dan has been demoing publicly and are independently testable. The COURSE-AS-CONCLUSION pipeline is the bias to flag: Dan’s framing implies that mastery requires the paid course. Per RDCO’s discipline: extract operational ideas from the public video (free), do not buy the course, and validate the patterns against our own implementation before declaring them effective. The expertise-file claim specifically should be tested in our skill ecology before being cited as authoritative in any Sanity Check issue.