Why LLM-Based Wiki Systems Are Flawed and Unscalable


Another Video, Same Broken Record: A Technical Rebuttal

Oh look, another enthusiastic YouTuber discovered Karpathy’s tweet and decided to make a tutorial. 14 million views in 72 hours! Lex Friedman does it too! “The smartest people in AI are doing this, so we probably should be too!”

Let me stop you right there. 🐑

https://gnu.support/images/2026/04/2026-04-23/800/robot-hallucinated.webp


The Core Claims vs. Reality

Claim: “It never forgets anything. It gets stronger over time.”

Reality: The LLM has no persistent memory across sessions. It forgets everything between chats. The only “memory” is the markdown files it wrote last time. If those files contain errors, contradictions, or hallucinations, the LLM will confidently repeat them. That’s not “getting stronger.” That’s a landfill with hyperlinks.

Claim: “It solves the problem of LLMs knowing only aggregate information.”

Reality: The solution is to feed the LLM your own sources. That’s fine. But then the LLM writes markdown summaries that replace those sources. Now you have two versions of “truth” — the original and the LLM’s hallucinated version. When they diverge, which one wins? The video doesn’t say.

Claim: “You can ask complex questions across all sources.”

Reality: You can ask. Whether you get a correct answer depends on whether the LLM hallucinated, whether the index fit in context, whether the wiki pages contain contradictions, and whether the LLM chose the right pages to read. That’s not “complex question answering.” That’s probabilistic gambling with your knowledge.

Claim: “Claude reads every file and creates a structured wiki.”

Reality: Claude reads every file within the context window. The video used 10 resources. At 100 resources, the context window overflows. The pattern admits this and adds qmd (BM25 + vector search). That’s not “no embeddings.” That’s embeddings with extra steps.

Claim: “It’s way easier, quicker, and more efficient to get information. You’re not going to use a bunch of tokens once you have the index built out.”

Reality: The index itself consumes tokens. Every query still requires the LLM to read relevant pages. At scale, with constant updates and queries, token usage is not “way easier.” It’s unpredictable and potentially expensive. The video conveniently doesn’t show the API bill.

Claim: “Ask where the gaps are in my wiki. It will suggest things to add.”

Reality: The LLM suggests gaps based on its training data, not based on actual missing knowledge. It doesn’t know what you don’t know. It predicts plausible-sounding gaps. You might add them. Now your wiki contains speculative information presented as authoritative. That’s not “self-improvement.” That’s self-contamination.

Claim: “Lex Friedman feeds it into voice mode and chats with his wiki on long runs.”

Reality: That’s a cool demo. It’s also a great way to internalize whatever hallucinations the LLM embedded in the wiki. The Feynman technique requires explaining to a real audience that can push back. Chatting with an LLM that agrees with everything you say is not learning. It’s a circle jerk with extra steps.


What the Video Doesn’t Show

Missing Piece Why It Matters
What happens at 1,000 files? The index breaks. You need qmd (embeddings).
Who fixes broken links? The LLM might, but it also might not.
Who resolves contradictions? The video says the LLM flags them. It doesn’t say who resolves them.
Who merges duplicate pages? No one. Your wiki will have 47 pages about “mechanical tension” with slightly different names.
Who sets permissions? No one. The LLM sees everything.
What’s the API cost at scale? Not mentioned.
What happens when the LLM hallucinates a source? Not mentioned.
How do you audit the wiki’s correctness? Not mentioned.

The actual video

The Technical Bottom Line

The video is a well-produced tutorial for a prototype that works with 10 files. The problems appear at 100 files. The system collapses at 1,000 files. The pattern admits this and adds embeddings. The video ignores this and calls it “no embeddings.”

The LLM has no memory. The wiki has no foreign keys. The system has no permissions. The human still fixes the mess.

But sure, “the smartest people in AI are doing it.” Appeal to authority doesn’t fix broken architecture. 🐑💀

Build a real knowledge base. Use PostgreSQL. Add foreign keys. Implement permissions. Track versions. Extract metadata deterministically. Let the LLM write descriptions — not replace your sources.

The video is entertaining. The pattern is still a trap. The sheep are still lining up.

Don’t be a sheep. 🧙🐘

⚠️ THE WORD “WIKI” HAS BEEN PERVERTED ⚠️

Related pages