Why Adam's LLM Wiki in Business Implementation Fails as a Production Framework


Rewritten Evaluation: Comparing Adam’s “LLM Wiki in Business” vs. The Critical Video

Executive Summary

Two videos. Same topic. Opposite conclusions. One treats LLM Wiki as a production framework for agencies. The other calls it a bad idea outright.

Below is a head‑to‑head evaluation of Adam’s implementation against the eight problems raised in the critical video. The result: Adam’s system exhibits every single failure mode the critical video warns about.

https://gnu.support/images/2026/04/2026-04-23/800/llm-wiki-on-fire.webp


Comparison Table: Adam’s Implementation vs. The Critical Video’s Warnings

Problem (from critical video) Does Adam’s system have this problem? Evidence from transcript
Errors spread like a virus ❌ Yes — problem present No audit trail, no cryptographic verification. Once an LLM writes a wrong JSON summary, that error becomes the “structured Wiki” for all future agents.
Hallucinations become structured ❌ Yes — problem present The LLM compiles raw data into JSON “Wiki layer”. If the LLM hallucinates a pain point or tool, that hallucination is now a permanent, queryable field.
Information loss from compression (10‑20%) ❌ Yes — problem present Raw emails, transcripts, calls → summarised into JSON. Edge cases are explicitly dropped. Adam admits “we extract the main things we care about.”
Difficult updates (one change affects many pages) ❌ Yes — problem present JSON files are denormalised. Changing a client’s tool name requires finding and editing every JSON file that references it. No foreign keys.
Loss of transparency / provenance ❌ Yes — problem present The JSON “Wiki” does not store which sentence in which transcript generated each field. You cannot trace a conclusion back to the original source line.
Heavy upfront investment ❌ Yes — problem present Building the audit extractor, video ingest, storyboard JSON schema, Claude MD files, skills, connectors, agents, hooks. Adam says “200 hours → 20‑40 hours” after the system is built. The build cost is not counted.
Scalability issues (duplicate pages, messy links) ❌ Yes — problem present Adam admits: “10, 50, 100 clients … the AI can’t index all that information. It’s too much context.” Folder‑based navigation breaks.
Rigidity (pre‑built structures resist change) ❌ Yes — problem present The audit JSON schema is fixed: “who, how long, how often, what tools”. New data that doesn’t fit is ignored or forces a reprocessing of all clients.

Verdict on the table: Adam’s production agency system exhibits ❌ every single failure mode the critical video warns about.


Comparison Table: Adam vs. The Four Pillars (Re‑evaluated)

Pillar Requirement Adam’s Implementation Critical Video’s Prediction Outcome
Store Real database with schemas, FKs, indexes Folders + JSON + markdown “No foreign keys, links break” Fail
Store Immutability, deterministic metadata Files can be edited; no versioning “Errors become permanent” Fail
Relate Typed relationships (supports/contradicts) Folder nesting only “Generic links, no semantics” Fail
Relate Bidirectional links / foreign keys None. A → B, but B doesn’t know A. “Links break silently” Fail
Trust Permissions / ACL down to object level None. “AI sees all data” (emails, NDAs, medical records) “Privacy leaks” Fail
Trust Audit trail (who, when, why) Not mentioned “No contradiction resolution” Fail
Retrieve SQL + full‑text search Scripts that parse JSON “No queryable structure” Fail
Retrieve Search by any dimension Only via pre‑written scripts “Index.md becomes bottleneck” Fail

No pillar passes. All four pillars show ❌ Fail.


Where Adam’s System Actually Succeeds (Few and Limited)

Requirement Status Evidence
Start with raw, immutable source Pass Raw emails, transcripts, videos stored in separate folders
Store any knowledge type Pass Supports videos, transcripts, emails, JSON, markdown
Record files or locations Pass Uses folder paths
Human curation Pass Humans decide what to keep (implied)
Ownership and control Pass Owns folders and files

These are the only passes. They are the easy parts. Adam fails on everything that requires actual database, relationship, provenance, and retrieval capabilities.


Detailed Analysis of Adam’s Core Claims

Claim 1: “This is much more effective than RAG”

Aspect Adam’s Claim Reality Verdict
Effectiveness “Much more effective” Loses provenance, bakes hallucinations, rigid schemas False
RAG comparison “Overkill, unnecessary” RAG preserves source links, scales to millions, no schema rigidity Misleading

Verdict: ❌ Claim is incorrect. RAG solves problems Adam ignores.


Claim 2: “RAG is overkill; vector databases are unnecessary”

Aspect Adam’s Claim Reality Verdict
Complexity “Too heavy” Folders are simpler to set up — true, but at the cost of all four pillars ⚠️ Partially true but irrelevant
Necessity “Unnecessary for what we’re doing” His own system fails at 10+ clients. That is exactly when RAG becomes necessary. False

Verdict: ❌ Confuses “easy to start” with “correct to use.”


Claim 3: “We can use scripts to filter as files grow”

Aspect Adam’s Claim Reality Verdict
Script capability Works “if correctly architected” Scripts require known schema. New data = rewrite schema = reprocess all clients. Ignores rigidity problem
Scaling Works as files grow Adam admits AI cannot index across 50+ clients. Scripts don’t solve that. Contradicts his own admission

Verdict: ❌ Tautological (“works if it works”) and self‑contradictory.


Claim 4: “The AI knows exactly which folders to go to”

Aspect Adam’s Claim Reality Verdict
Navigation Hard‑coded paths work For one client, yes. For 100 clients with 5 projects each, the Claude.md file would be thousands of lines. Does not scale
vs. query “Knows” vs. “searches” He has replaced search with manual enumeration. That is not knowledge retrieval. Fundamental category error

Verdict: ❌ Hard‑coded paths are not retrieval. They are hard‑coded paths.


What Adam Gets Right (Unintentionally)

Adam demonstrates exactly what the critical video warns about — but he treats it as a feature:

Critical video warning Adam’s implementation as evidence Adam’s interpretation
Information loss from compression “We extract the main things we care about” (edge cases dropped) ✅ “That’s what we need”
Hallucinations become structured LLM writes the audit JSON. No human verification per field. ✅ “That’s the Wiki layer”
Loss of transparency JSON does not store source pointers. Cannot trace back. ✅ “Not mentioned as a problem”
Heavy upfront investment Builds custom agents, skills, connectors, schemas for each client type ✅ “That’s the framework”
Scalability issue “AI can’t index all that information … too much context” ✅ “That’s why we need controlled injection”
Rigidity Fixed JSON schema. New data requires redesign. ✅ “That’s the schema”

Conclusion: Adam’s video is an ❌ unintentional case study of why LLM Wiki fails in production. He has simply accepted the trade‑offs (loss of provenance, rigidity, upfront cost) as normal because he has never experienced a system that actually meets the four pillars.


Final Verdict Table

Criteria Adam’s Video Critical Video
Technically accurate about LLM Wiki limits ⚠️ Partial (admits scaling issue but misses provenance, error propagation, update cost) ✅ Yes (lists all 8)
Acknowledges the four pillars ❌ No (no database, no FKs, no permissions, no SQL) ✅ Yes (implicitly, via RAG recommendation)
Suitable for business production ❌ No — privacy leaks, no audit trail, rigid schemas, fails beyond 10 clients ⚠️ Hybrid approach advised
Hero worship / deference to Karpathy ⚠️ Uses “Kapathy” branding but modifies implementation ✅ Explicitly defies Karpathy’s authority
Overall recommendation Do not use for any system requiring provenance, security, or scalability Recommended for understanding why LLM Wiki fails

The actual video

Side‑by‑Side Summary Table

Aspect Adam (Agency “LLM Wiki”) Critical Video (“Bad Idea”)
Core thesis LLM Wiki works in production with folder conventions LLM Wiki is fundamentally flawed
Database? ❌ No (folders + JSON) ✅ Recommends RAG (vector DB)
Foreign keys? ❌ No ✅ N/A (RAG doesn’t need them, but chunk references serve similar purpose)
Permissions? ❌ No (AI sees all data — privacy leak) ⚠️ Not discussed, but RAG allows chunk‑level source tracking
Provenance? ❌ Lost after JSON summarisation ✅ Preserved (chunks → source document)
Update cost ❌ High (denormalised JSON, must find and edit every file) ✅ Low (RAG re‑indexes only new chunks)
Scaling limit ❌ “10, 50, 100 clients … AI can’t index” — admitted failure ✅ Scales to millions of chunks
Rigidity ❌ Fixed JSON schema. New data type = redesign. ✅ Schema‑free (embedding search accommodates new data types)
Honesty about trade‑offs ❌ No — presents as superior to RAG while hiding provenance loss, rigidity, privacy risks ✅ Yes — lists 8 specific problems openly

Final judgment: The critical video is ✅ correct. Adam’s implementation is exactly what ❌ “bad idea” looks like in production clothing. The four pillars exist for a reason. Adam violates all four. 🐑💀🧙

⚠️ THE WORD “WIKI” HAS BEEN PERVERTED ⚠️

Related pages