01-projects / graph-db-eval

prototype results

Mon Apr 13 2026 20:00:00 GMT-0400 (Eastern Daylight Time) ·project-output

Graph DB Prototype - Results

Ingestion summary

Note: Document vertex count (106) exceeds 50 because wikilinks in Related sections create stub Document nodes for cited-but-not-ingested files. These stubs have no outbound edges of their own but let 1-hop citation queries succeed.

Query 1 - Positioning evidence (state-ownership architecture)

Runtime: 5.48 ms

Target doc: rdco-state-ownership-architecture

1-hop: Documents that cite the state-ownership architecture doc

DocumentPathDate
OpenAI’s Memos, Frontier, Amazon and Anthropic — Ben Thompson06-reference/2026-04-14-stratechery-openai-memos-anthropic.md2026-04-14
”The Half-Life of a Moat (Part 1)” — Jonathan Natkins06-reference/2026-04-14-semistructured-half-life-of-a-moat-part-1.md2026-04-14
”Fat Skills, Fat Code, Thin Harness” — Commentary on Garry Tan’s Architecture06-reference/2026-04-14-tan-fat-skills-fat-code-thin-harness-commentary.md2026-04-14
Aaron Levie — “The Agent Deployer” Role JD (Apr 14, 2026)06-reference/2026-04-14-levie-agent-deployer-role-jd.md2026-04-14

State-ownership doc’s own citations (the anchor set for 2-hop alignment)

2-hop: Documents that co-cite the same anchor docs (aligned-content discovery)

DocumentShared-citations countShared anchors
”The Half-Life of a Moat (Part 1)” — Jonathan Natkins5”Agent Harnesses Are Dead. Long Live Agent Harnesses.” — João Moura; “The AI Lock-In Is Beginning!” — Jaya Gupta; “The LLMs Get the Publicity. The Data Layer Does the Work” — Jonathan Natkins; “Thin Harness, Fat Skills” — Garry Tan; Dissenting Opinions on the “Thin Harness, Fat Skills” Thesis
”Fat Skills, Fat Code, Thin Harness” — Commentary on Garry Tan’s Architecture4”Agent Harnesses Are Dead. Long Live Agent Harnesses.” — João Moura; “The AI Lock-In Is Beginning!” — Jaya Gupta; “Thin Harness, Fat Skills” — Garry Tan; Dissenting Opinions on the “Thin Harness, Fat Skills” Thesis
”Agent Harnesses Are Dead. Long Live Agent Harnesses.” — João Moura3”The LLMs Get the Publicity. The Data Layer Does the Work” — Jonathan Natkins; “Thin Harness, Fat Skills” — Garry Tan; Dissenting Opinions on the “Thin Harness, Fat Skills” Thesis
”The AI Lock-In Is Beginning!” — Jaya Gupta3”Agent Harnesses Are Dead. Long Live Agent Harnesses.” — João Moura; “The LLMs Get the Publicity. The Data Layer Does the Work” — Jonathan Natkins; Dissenting Opinions on the “Thin Harness, Fat Skills” Thesis
OpenAI’s Memos, Frontier, Amazon and Anthropic — Ben Thompson3”Agent Harnesses Are Dead. Long Live Agent Harnesses.” — João Moura; “The AI Lock-In Is Beginning!” — Jaya Gupta; Dissenting Opinions on the “Thin Harness, Fat Skills” Thesis
”The LLMs Get the Publicity. The Data Layer Does the Work” — Jonathan Natkins2”Thin Harness, Fat Skills” — Garry Tan; Dissenting Opinions on the “Thin Harness, Fat Skills” Thesis
2026-04-12-alphasignal-claude-code-leak-harness-engineering2”Thin Harness, Fat Skills” — Garry Tan; Dissenting Opinions on the “Thin Harness, Fat Skills” Thesis
2026-04-12-cobus-greyling-harness-era-language-shift2”Thin Harness, Fat Skills” — Garry Tan; Dissenting Opinions on the “Thin Harness, Fat Skills” Thesis
2026-04-12-harrison-chase-harness-blog2”Thin Harness, Fat Skills” — Garry Tan; Dissenting Opinions on the “Thin Harness, Fat Skills” Thesis
2026-04-12-lindstrom-board-ai-governance2”Thin Harness, Fat Skills” — Garry Tan; Dissenting Opinions on the “Thin Harness, Fat Skills” Thesis
2026-04-13-solve-everything-master-synthesis2”Thin Harness, Fat Skills” — Garry Tan; Dissenting Opinions on the “Thin Harness, Fat Skills” Thesis
Aaron Levie — “The Agent Deployer” Role JD (Apr 14, 2026)2”Agent Harnesses Are Dead. Long Live Agent Harnesses.” — João Moura; “The AI Lock-In Is Beginning!” — Jaya Gupta
Mammoth Growth Agentic Harness Review — cc-wrapped2”Thin Harness, Fat Skills” — Garry Tan; Dissenting Opinions on the “Thin Harness, Fat Skills” Thesis
The Folder Is the Agent — Kieran Klaassen (Every)2”Thin Harness, Fat Skills” — Garry Tan; Dissenting Opinions on the “Thin Harness, Fat Skills” Thesis
2026-03-25-seattle-data-guy-know-nothing-and-be-happy1SOUL
2026-04-12-arxiv-2604-08224-agent-harness-study1”Thin Harness, Fat Skills” — Garry Tan
Dissenting Opinions on the “Thin Harness, Fat Skills” Thesis1”Thin Harness, Fat Skills” — Garry Tan
Ray Data Co — phData vs Mammoth Growth Decision Analysis1SOUL
dbt Semantic Layer vs Text-to-SQL — 2026 Benchmark1”The LLMs Get the Publicity. The Data Layer Does the Work” — Jonathan Natkins

Query 2 - Dissent cluster aggregation (authors in >=2 harness-thesis docs)

Runtime: 2.77 ms

Cluster size: 12 documents

No person authored >=2 cluster documents.

Query 3 - Decision-evidence audit (phData vs MG decision)

Runtime: 1.95 ms

Target doc: Ray Data Co — phData vs Mammoth Growth Decision Analysis

Citation tree (1-hop and 2-hop)

What worked

What didn’t work / limits exposed

Edge types we NEED next (manual or LLM annotation, phase 2)

  1. supports-position - Document -> named RDCO position. Without it, Query 1 (positioning evidence) relies on citation-adjacency as a proxy, which over-returns noise.
  2. validates / contradicts / disputes-claim-in - these are the whole point of the cross-check skill output; we should extract them from the harness-thesis-dissent doc and similar cluster synthesis docs.
  3. informs-decision - Document -> Decision. We’d stop guessing which cited docs were actually weighted in the phData decision vs cited as background.
  4. part-of-cluster - explicit Cluster vertex + membership edges, so Query 2 doesn’t have to heuristically reconstruct ‘the harness-thesis cluster’ via title+topic pattern matching each time.

Recommendation

Yes - continue investment, but phase it. The prototype proves:

Phase 2 plan:

  1. Scale ingestion to the full vault (~1,426 docs) - expect ~20k edges based on current density.
  2. Add a lightweight LLM annotator that reads each cross-check synthesis doc and emits validates / contradicts edges between the docs it references. Start with the 10-15 synthesis docs we already have.
  3. Add Cluster vertices + part-of-cluster edges derived from existing cluster-synthesis docs (harness-thesis dissent, moat debate, data-quality sources).
  4. Write a /graph-query skill that wraps the 3 prototype queries and exposes them via natural-language intent.
  5. Re-evaluate after a month of use - if the founder isn’t invoking the graph queries in real work, prune or pivot.