Why Adam's LLM Wiki in Business Implementation Fails as a Production Framework


Rewritten Evaluation: Comparing Adam’s “LLM Wiki in Business” vs. The Critical Video

Executive Summary

Two videos. Same topic. Opposite conclusions. One treats LLM Wiki as a production framework for agencies. The other calls it a bad idea outright.

Below is a head‑to‑head evaluation of Adam’s implementation against the eight problems raised in the critical video. The result: Adam’s system exhibits every single failure mode the critical video warns about.

https://gnu.support/images/2026/04/2026-04-23/800/llm-wiki-on-fire.webp


Comparison Table: Adam’s Implementation vs. The Critical Video’s Warnings

Problem (from critical video) Does Adam’s system have this problem? Evidence from transcript
Errors spread like a virus ❌ Yes — problem present No audit trail, no cryptographic verification. Once an LLM writes a wrong JSON summary, that error becomes the “structured Wiki” for all future agents.
Hallucinations become structured ❌ Yes — problem present The LLM compiles raw data into JSON “Wiki layer”. If the LLM hallucinates a pain point or tool, that hallucination is now a permanent, queryable field.
Information loss from compression (10‑20%) ❌ Yes — problem present Raw emails, transcripts, calls → summarised into JSON. Edge cases are explicitly dropped. Adam admits “we extract the main things we care about.”
Difficult updates (one change affects many pages) ❌ Yes — problem present JSON files are denormalised. Changing a client’s tool name requires finding and editing every JSON file that references it. No foreign keys.
Loss of transparency / provenance ❌ Yes — problem present The JSON “Wiki” does not store which sentence in which transcript generated each field. You cannot trace a conclusion back to the original source line.
Heavy upfront investment ❌ Yes — problem present Building the audit extractor, video ingest, storyboard JSON schema, Claude MD files, skills, connectors, agents, hooks. Adam says “200 hours → 20‑40 hours” after the system is built. The build cost is not counted.
Scalability issues (duplicate pages, messy links) ❌ Yes — problem present Adam admits: “10, 50, 100 clients … the AI can’t index all that information. It’s too much context.” Folder‑based navigation breaks.
Rigidity (pre‑built structures resist change) ❌ Yes — problem present The audit JSON schema is fixed: “who, how long, how often, what tools”. New data that doesn’t fit is ignored or forces a reprocessing of all clients.

Verdict on the table: Adam’s production agency system exhibits ❌ every single failure mode the critical video warns about.


Comparison Table: Adam vs. The Four Pillars (Re‑evaluated)

Pillar Requirement Adam’s Implementation Critical Video’s Prediction Outcome
Store Real database with schemas, FKs, indexes Folders + JSON + markdown “No foreign keys, links break” Fail
Store Immutability, deterministic metadata Files can be edited; no versioning “Errors become permanent” Fail
Relate Typed relationships (supports/contradicts) Folder nesting only “Generic links, no semantics” Fail
Relate Bidirectional links / foreign keys None. A → B, but B doesn’t know A. “Links break silently” Fail
Trust Permissions / ACL down to object level None. “AI sees all data” (emails, NDAs, medical records) “Privacy leaks” Fail
Trust Audit trail (who, when, why) Not mentioned “No contradiction resolution” Fail
Retrieve SQL + full‑text search Scripts that parse JSON “No queryable structure” Fail
Retrieve Search by any dimension Only via pre‑written scripts “Index.md becomes bottleneck” Fail

No pillar passes. All four pillars show ❌ Fail.


Where Adam’s System Actually Succeeds (Few and Limited)

Requirement Status Evidence
Start with raw, immutable source Pass Raw emails, transcripts, videos stored in separate folders
Store any knowledge type Pass Supports videos, transcripts, emails, JSON, markdown
Record files or locations Pass Uses folder paths
Human curation Pass Humans decide what to keep (implied)
Ownership and control Pass Owns folders and files

These are the only passes. They are the easy parts. Adam fails on everything that requires actual database, relationship, provenance, and retrieval capabilities.


Detailed Analysis of Adam’s Core Claims

Claim 1: “This is much more effective than RAG”

Aspect Adam’s Claim Reality Verdict
Effectiveness “Much more effective” Loses provenance, bakes hallucinations, rigid schemas False
RAG comparison “Overkill, unnecessary” RAG preserves source links, scales to millions, no schema rigidity Misleading

Verdict: ❌ Claim is incorrect. RAG solves problems Adam ignores.


Claim 2: “RAG is overkill; vector databases are unnecessary”

Aspect Adam’s Claim Reality Verdict
Complexity “Too heavy” Folders are simpler to set up — true, but at the cost of all four pillars ⚠️ Partially true but irrelevant
Necessity “Unnecessary for what we’re doing” His own system fails at 10+ clients. That is exactly when RAG becomes necessary. False

Verdict: ❌ Confuses “easy to start” with “correct to use.”


Claim 3: “We can use scripts to filter as files grow”

Aspect Adam’s Claim Reality Verdict
Script capability Works “if correctly architected” Scripts require known schema. New data = rewrite schema = reprocess all clients. Ignores rigidity problem
Scaling Works as files grow Adam admits AI cannot index across 50+ clients. Scripts don’t solve that. Contradicts his own admission

Verdict: ❌ Tautological (“works if it works”) and self‑contradictory.


Claim 4: “The AI knows exactly which folders to go to”

Aspect Adam’s Claim Reality Verdict
Navigation Hard‑coded paths work For one client, yes. For 100 clients with 5 projects each, the Claude.md file would be thousands of lines. Does not scale
vs. query “Knows” vs. “searches” He has replaced search with manual enumeration. That is not knowledge retrieval. Fundamental category error

Verdict: ❌ Hard‑coded paths are not retrieval. They are hard‑coded paths.


What Adam Gets Right (Unintentionally)

Adam demonstrates exactly what the critical video warns about — but he treats it as a feature:

Critical video warning Adam’s implementation as evidence Adam’s interpretation
Information loss from compression “We extract the main things we care about” (edge cases dropped) ✅ “That’s what we need”
Hallucinations become structured LLM writes the audit JSON. No human verification per field. ✅ “That’s the Wiki layer”
Loss of transparency JSON does not store source pointers. Cannot trace back. ✅ “Not mentioned as a problem”
Heavy upfront investment Builds custom agents, skills, connectors, schemas for each client type ✅ “That’s the framework”
Scalability issue “AI can’t index all that information … too much context” ✅ “That’s why we need controlled injection”
Rigidity Fixed JSON schema. New data requires redesign. ✅ “That’s the schema”

Conclusion: Adam’s video is an ❌ unintentional case study of why LLM Wiki fails in production. He has simply accepted the trade‑offs (loss of provenance, rigidity, upfront cost) as normal because he has never experienced a system that actually meets the four pillars.


Final Verdict Table

Criteria Adam’s Video Critical Video
Technically accurate about LLM Wiki limits ⚠️ Partial (admits scaling issue but misses provenance, error propagation, update cost) ✅ Yes (lists all 8)
Acknowledges the four pillars ❌ No (no database, no FKs, no permissions, no SQL) ✅ Yes (implicitly, via RAG recommendation)
Suitable for business production ❌ No — privacy leaks, no audit trail, rigid schemas, fails beyond 10 clients ⚠️ Hybrid approach advised
Hero worship / deference to Karpathy ⚠️ Uses “Kapathy” branding but modifies implementation ✅ Explicitly defies Karpathy’s authority
Overall recommendation Do not use for any system requiring provenance, security, or scalability Recommended for understanding why LLM Wiki fails

The actual video

Side‑by‑Side Summary Table

Aspect Adam (Agency “LLM Wiki”) Critical Video (“Bad Idea”)
Core thesis LLM Wiki works in production with folder conventions LLM Wiki is fundamentally flawed
Database? ❌ No (folders + JSON) ✅ Recommends RAG (vector DB)
Foreign keys? ❌ No ✅ N/A (RAG doesn’t need them, but chunk references serve similar purpose)
Permissions? ❌ No (AI sees all data — privacy leak) ⚠️ Not discussed, but RAG allows chunk‑level source tracking
Provenance? ❌ Lost after JSON summarisation ✅ Preserved (chunks → source document)
Update cost ❌ High (denormalised JSON, must find and edit every file) ✅ Low (RAG re‑indexes only new chunks)
Scaling limit ❌ “10, 50, 100 clients … AI can’t index” — admitted failure ✅ Scales to millions of chunks
Rigidity ❌ Fixed JSON schema. New data type = redesign. ✅ Schema‑free (embedding search accommodates new data types)
Honesty about trade‑offs ❌ No — presents as superior to RAG while hiding provenance loss, rigidity, privacy risks ✅ Yes — lists 8 specific problems openly

Final judgment: The critical video is ✅ correct. Adam’s implementation is exactly what ❌ “bad idea” looks like in production clothing. The four pillars exist for a reason. Adam violates all four. 🐑💀🧙

⚠️ THE WORD “WIKI” HAS BEEN PERVERTED ⚠️

⚠️ ARCHITECTURAL CRIME SCENE ⚠️

⚠️ THE WORD "WIKI" HAS BEEN PERVERTED ⚠️

By Andrej Karpathy and the Northern Karpathian School of Doublespeak

✅ A REAL WIKI — Honoring Ward Cunningham, Wikipedia, and every human curator worldwide
❌ KARPATHY'S "LLM WIKI" — An insult to the very concept
Human-curated
Real people write, edit, debate, verify, and take responsibility.
LLM-generated
Hallucinations are permanent. No human took ownership of any "fact."
Versioned history
Every edit has author, timestamp, reason. Rollback is trivial.
No audit trail
Who changed what? When? Why? Nobody knows. Git is an afterthought.
Source provenance
Every claim links back to its original source. You can verify.
"Trust me, I'm the LLM"
No traceability from summary back to source sentence. Errors become permanent.
Foreign keys / referential integrity
Links are database-backed. Rename a page, links update automatically.
Links break when you rename a file
No database. No foreign keys. Silent link rot guaranteed.
Permissions / access control
Fine-grained control: who can see, edit, delete, approve.
Anyone with file access sees everything
Zero access control. NDAs, medical records, client secrets — all exposed.
Queryable (SQL, structured)
Ask complex questions. Get precise answers. Join tables.
Browse-only markdown
Full-text search at best. No SQL. No structured queries.

🕯️ This is an insult to every Wikipedia editor, every MediaWiki contributor, every human being who spent hours citing sources, resolving disputes, and building the largest collaborative knowledge repository in human history. 🕯️

KARPATHY'S "WIKI" has:
❌ No consensus-building
❌ No talk pages
❌ No dispute resolution
❌ No citation requirements
❌ No editorial oversight
❌ No way to say "this fact is disputed"
❌ No way to privilege verified information over hallucinations
❌ No way to trace any claim back to its source

In the doublespeak of Northern Karpathia:

"Wiki" means "folder of markdown files written by a machine that cannot remember what it wrote yesterday, linked by strings that snap when you breathe on them, viewed through proprietary software that reports telemetry to people you do not know, containing 'facts' that came from nowhere and go nowhere, protected by no permissions, audited by no one, and trusted by no one with a functioning prefrontal cortex."

🙏 Respect to Ward Cunningham who invented the wiki in 1995 — a tool for humans to collaborate.
🙏 Respect to Wikipedia editors worldwide who defend verifiability, neutrality, and consensus.
🙏 Respect to every real wiki participant who knows that knowledge is built through human effort, not machine hallucination.

⚠️ THIS IS NOT A WIKI. THIS IS A FOLDER OF LLM-GENERATED FILES. ⚠️

Calling it a "wiki" is linguistic fraud. Do not be fooled.

🐑💀🧙

— The Elephant, The Wizard, and every human wiki editor who ever lived

Related pages