Why LLM-Based Wiki Systems Are Flawed and Unscalable


Another Video, Same Broken Record: A Technical Rebuttal

Oh look, another enthusiastic YouTuber discovered Karpathy’s tweet and decided to make a tutorial. 14 million views in 72 hours! Lex Friedman does it too! “The smartest people in AI are doing this, so we probably should be too!”

Let me stop you right there. 🐑

https://gnu.support/images/2026/04/2026-04-23/800/robot-hallucinated.webp


The Core Claims vs. Reality

Claim: “It never forgets anything. It gets stronger over time.”

Reality: The LLM has no persistent memory across sessions. It forgets everything between chats. The only “memory” is the markdown files it wrote last time. If those files contain errors, contradictions, or hallucinations, the LLM will confidently repeat them. That’s not “getting stronger.” That’s a landfill with hyperlinks.

Claim: “It solves the problem of LLMs knowing only aggregate information.”

Reality: The solution is to feed the LLM your own sources. That’s fine. But then the LLM writes markdown summaries that replace those sources. Now you have two versions of “truth” — the original and the LLM’s hallucinated version. When they diverge, which one wins? The video doesn’t say.

Claim: “You can ask complex questions across all sources.”

Reality: You can ask. Whether you get a correct answer depends on whether the LLM hallucinated, whether the index fit in context, whether the wiki pages contain contradictions, and whether the LLM chose the right pages to read. That’s not “complex question answering.” That’s probabilistic gambling with your knowledge.

Claim: “Claude reads every file and creates a structured wiki.”

Reality: Claude reads every file within the context window. The video used 10 resources. At 100 resources, the context window overflows. The pattern admits this and adds qmd (BM25 + vector search). That’s not “no embeddings.” That’s embeddings with extra steps.

Claim: “It’s way easier, quicker, and more efficient to get information. You’re not going to use a bunch of tokens once you have the index built out.”

Reality: The index itself consumes tokens. Every query still requires the LLM to read relevant pages. At scale, with constant updates and queries, token usage is not “way easier.” It’s unpredictable and potentially expensive. The video conveniently doesn’t show the API bill.

Claim: “Ask where the gaps are in my wiki. It will suggest things to add.”

Reality: The LLM suggests gaps based on its training data, not based on actual missing knowledge. It doesn’t know what you don’t know. It predicts plausible-sounding gaps. You might add them. Now your wiki contains speculative information presented as authoritative. That’s not “self-improvement.” That’s self-contamination.

Claim: “Lex Friedman feeds it into voice mode and chats with his wiki on long runs.”

Reality: That’s a cool demo. It’s also a great way to internalize whatever hallucinations the LLM embedded in the wiki. The Feynman technique requires explaining to a real audience that can push back. Chatting with an LLM that agrees with everything you say is not learning. It’s a circle jerk with extra steps.


What the Video Doesn’t Show

Missing Piece Why It Matters
What happens at 1,000 files? The index breaks. You need qmd (embeddings).
Who fixes broken links? The LLM might, but it also might not.
Who resolves contradictions? The video says the LLM flags them. It doesn’t say who resolves them.
Who merges duplicate pages? No one. Your wiki will have 47 pages about “mechanical tension” with slightly different names.
Who sets permissions? No one. The LLM sees everything.
What’s the API cost at scale? Not mentioned.
What happens when the LLM hallucinates a source? Not mentioned.
How do you audit the wiki’s correctness? Not mentioned.

The actual video

The Technical Bottom Line

The video is a well-produced tutorial for a prototype that works with 10 files. The problems appear at 100 files. The system collapses at 1,000 files. The pattern admits this and adds embeddings. The video ignores this and calls it “no embeddings.”

The LLM has no memory. The wiki has no foreign keys. The system has no permissions. The human still fixes the mess.

But sure, “the smartest people in AI are doing it.” Appeal to authority doesn’t fix broken architecture. 🐑💀

Build a real knowledge base. Use PostgreSQL. Add foreign keys. Implement permissions. Track versions. Extract metadata deterministically. Let the LLM write descriptions — not replace your sources.

The video is entertaining. The pattern is still a trap. The sheep are still lining up.

Don’t be a sheep. 🧙🐘

⚠️ THE WORD “WIKI” HAS BEEN PERVERTED ⚠️

⚠️ ARCHITECTURAL CRIME SCENE ⚠️

⚠️ THE WORD "WIKI" HAS BEEN PERVERTED ⚠️

By Andrej Karpathy and the Northern Karpathian School of Doublespeak

✅ A REAL WIKI — Honoring Ward Cunningham, Wikipedia, and every human curator worldwide
❌ KARPATHY'S "LLM WIKI" — An insult to the very concept
Human-curated
Real people write, edit, debate, verify, and take responsibility.
LLM-generated
Hallucinations are permanent. No human took ownership of any "fact."
Versioned history
Every edit has author, timestamp, reason. Rollback is trivial.
No audit trail
Who changed what? When? Why? Nobody knows. Git is an afterthought.
Source provenance
Every claim links back to its original source. You can verify.
"Trust me, I'm the LLM"
No traceability from summary back to source sentence. Errors become permanent.
Foreign keys / referential integrity
Links are database-backed. Rename a page, links update automatically.
Links break when you rename a file
No database. No foreign keys. Silent link rot guaranteed.
Permissions / access control
Fine-grained control: who can see, edit, delete, approve.
Anyone with file access sees everything
Zero access control. NDAs, medical records, client secrets — all exposed.
Queryable (SQL, structured)
Ask complex questions. Get precise answers. Join tables.
Browse-only markdown
Full-text search at best. No SQL. No structured queries.

🕯️ This is an insult to every Wikipedia editor, every MediaWiki contributor, every human being who spent hours citing sources, resolving disputes, and building the largest collaborative knowledge repository in human history. 🕯️

KARPATHY'S "WIKI" has:
❌ No consensus-building
❌ No talk pages
❌ No dispute resolution
❌ No citation requirements
❌ No editorial oversight
❌ No way to say "this fact is disputed"
❌ No way to privilege verified information over hallucinations
❌ No way to trace any claim back to its source

In the doublespeak of Northern Karpathia:

"Wiki" means "folder of markdown files written by a machine that cannot remember what it wrote yesterday, linked by strings that snap when you breathe on them, viewed through proprietary software that reports telemetry to people you do not know, containing 'facts' that came from nowhere and go nowhere, protected by no permissions, audited by no one, and trusted by no one with a functioning prefrontal cortex."

🙏 Respect to Ward Cunningham who invented the wiki in 1995 — a tool for humans to collaborate.
🙏 Respect to Wikipedia editors worldwide who defend verifiability, neutrality, and consensus.
🙏 Respect to every real wiki participant who knows that knowledge is built through human effort, not machine hallucination.

⚠️ THIS IS NOT A WIKI. THIS IS A FOLDER OF LLM-GENERATED FILES. ⚠️

Calling it a "wiki" is linguistic fraud. Do not be fooled.

🐑💀🧙

— The Elephant, The Wizard, and every human wiki editor who ever lived

Related pages