The LLM-Wiki Pattern: A Flawed and Misleading Alternative to RAG


jjjjjj# The LLM-Wiki Comedy Hour: A Technical Takedown of the Markdown Graveyard

Or: How I Learned to Stop Worrying and Love Foreign Keys


Introduction: The Setup

Another day, another enthusiastic YouTuber explaining how Andrej Karpathy’s LLM-Wiki pattern is going to revolutionize knowledge management. No vector databases! No embeddings! No complicated RAG! Just markdown files, Obsidian, and the power of vibe-coding!

I watched the video so you don’t have to. And then I laughed. A lot.

Let me walk you through the technical claims — and why each one deserves a chuckle.


Claim 1: “No Vector Database. No Embeddings.”

What they said: “This Obsidian-powered knowledge base has no vector database, no embeddings, and no complicated retrieval process.”

What they didn’t say: The pattern itself admits that the index.md approach only works at “small enough” scale. When the wiki grows beyond a few hundred pages, the LLM cannot read the entire index. The solution? Add qmd — a local search engine with BM25 and vector search.

That’s embeddings, folks. 🐑

You can’t claim “no embeddings” and then quietly add them back as a “scaling tool.” That’s like saying “I’m on a diet” while eating a cake and calling it a “nutritional scaling tool.”

Technical reality: At scale, you need vector search. The pattern knows this. The video ignores it. The sheep don’t notice.

https://gnu.support/images/2026/04/2026-04-23/800/sheep-embeddings.webp


Claim 2: “It Solves the Same Problem as RAG”

What they said: “It solves the exact same problem that these more complicated RAG structures claim to do.”

What it actually does: RAG retrieves chunks from your original source documents. The LLM reads the source and answers.

LLM-Wiki retrieves from LLM-generated markdown pages. The LLM reads its own summaries, which may contain hallucinations, contradictions, or outright fabrications.

These are not the same. 🐑💀

RAG is like asking a librarian to find a book and read you the relevant page. LLM-Wiki is like asking a librarian to write a new book based on what they vaguely remember, then read you that book, then pretend it’s the original.

One preserves source integrity. The other launders mistakes into truth.

Technical reality: Retrieval from generated content is not retrieval. It’s a game of telephone played with an LLM.


Claim 3: “Essentially Free”

What they said: “It’s essentially free.”

What they didn’t mention: API calls cost money. Every ingest consumes tokens. Every query consumes tokens. Every lint pass consumes tokens.

At small scale, sure, it’s cheap. At 1,000 documents with constant updates? That “free” system starts looking like a line item on your AWS bill.

Technical reality: The cost doesn’t disappear. It shifts from your labor to your API provider. The video conveniently forgets to mention pricing.

And let’s not forget: The human still pays with time. Fixing broken links. Resolving contradictions. Merging duplicate pages. Auditing hallucinations. That’s not “free.” That’s a second job.


Claim 4: “The LLM Automaintains Index Files Pretty Well”

What they said: “The large language model has been pretty good about automaintaining index files and brief summaries.”

What they demonstrated: Eight transcript files. A handful of trading concepts.

What they didn’t test: 100 files. 500 files. 1,000 files. The moment the index exceeds the LLM’s context window, the whole system collapses.

Technical reality: The LLM has no memory across sessions. It doesn’t “maintain” anything between chats. It reads the index.md file fresh each time. When that file is too large to fit in context, the LLM cannot navigate the wiki. You then add qmd — which is just RAG with extra steps.

The demo is a prototype. A toddler can organize eight toys. Try organizing 8,000.


Claim 5: “If It Doesn’t Work, Just Move to RAG”

What they said: “If it’s clear your scale goes well beyond the bounds of what this thing can handle, then just move into RAG.”

This is an admission. 🐑

The video is literally telling you: “Try this broken system. When it fails, use a real one.”

Imagine a car salesman saying: “This car has no engine, but if it doesn’t work, just buy a real car.” That’s not a solution. That’s a waste of time.

Technical reality: LLM-Wiki is not a stepping stone to RAG. It’s a detour. A distraction. A markdown graveyard that you will eventually abandon when you realize you need foreign keys, permissions, version control, and actual referential integrity.


Claim 6: “Most People Don’t Need a Real RAG System”

What they said: “Most people don’t need a real RAG system. They just don’t, right?”

What they ignored: Most people also don’t need a broken system that collapses at scale. What they need is a proper knowledge base with:

LLM-Wiki has none of these. But sure, tell me I don’t need RAG. I need a database. 🧙


Claim 7: “Just Try It. Experiment.”

What they said: “Just try it out. Just experiment. It’s not costing you anything.”

What it actually costs: Your time. Your API credits. Your sanity when the wiki becomes an unmanageable mess of contradictions, broken links, and hallucinated facts.

Technical reality: Experimentation is fine. But don’t confuse a prototype with a production system. Don’t build your knowledge base on sand and then wonder why it collapses.

The video ends with a shrug: “If it doesn’t work, it doesn’t work. Fine.”

That’s not engineering. That’s a vibe check. 🐑💀


The Bottom Line: What the Video Won’t Tell You

Claim Reality
No embeddings Adds qmd (BM25 + vector search) at scale
Solves same problem as RAG Retrieves from LLM-generated pages, not sources
Essentially free API calls cost money; human labor shifts
LLM maintains indexes Works at 8 files; fails at 1,000
If it fails, use RAG Admits the pattern is disposable
Most people don’t need RAG Most people need a database
Just try it Costs time, money, and sanity

The actual video

My Advice

Build a real knowledge base. Use PostgreSQL. Add foreign keys. Implement permissions. Track versions. Extract metadata deterministically. Let the LLM write descriptions — not replace your sources.

The video is entertaining. The pattern is a trap. The sheep are still lining up.

Don’t be a sheep. 🐑💀

“Not today.” — Engelbart’s ghost, probably. 🧙🐘

⚠️ THE WORD “WIKI” HAS BEEN PERVERTED ⚠️

⚠️ ARCHITECTURAL CRIME SCENE ⚠️

⚠️ THE WORD "WIKI" HAS BEEN PERVERTED ⚠️

By Andrej Karpathy and the Northern Karpathian School of Doublespeak

✅ A REAL WIKI — Honoring Ward Cunningham, Wikipedia, and every human curator worldwide
❌ KARPATHY'S "LLM WIKI" — An insult to the very concept
Human-curated
Real people write, edit, debate, verify, and take responsibility.
LLM-generated
Hallucinations are permanent. No human took ownership of any "fact."
Versioned history
Every edit has author, timestamp, reason. Rollback is trivial.
No audit trail
Who changed what? When? Why? Nobody knows. Git is an afterthought.
Source provenance
Every claim links back to its original source. You can verify.
"Trust me, I'm the LLM"
No traceability from summary back to source sentence. Errors become permanent.
Foreign keys / referential integrity
Links are database-backed. Rename a page, links update automatically.
Links break when you rename a file
No database. No foreign keys. Silent link rot guaranteed.
Permissions / access control
Fine-grained control: who can see, edit, delete, approve.
Anyone with file access sees everything
Zero access control. NDAs, medical records, client secrets — all exposed.
Queryable (SQL, structured)
Ask complex questions. Get precise answers. Join tables.
Browse-only markdown
Full-text search at best. No SQL. No structured queries.

🕯️ This is an insult to every Wikipedia editor, every MediaWiki contributor, every human being who spent hours citing sources, resolving disputes, and building the largest collaborative knowledge repository in human history. 🕯️

KARPATHY'S "WIKI" has:
❌ No consensus-building
❌ No talk pages
❌ No dispute resolution
❌ No citation requirements
❌ No editorial oversight
❌ No way to say "this fact is disputed"
❌ No way to privilege verified information over hallucinations
❌ No way to trace any claim back to its source

In the doublespeak of Northern Karpathia:

"Wiki" means "folder of markdown files written by a machine that cannot remember what it wrote yesterday, linked by strings that snap when you breathe on them, viewed through proprietary software that reports telemetry to people you do not know, containing 'facts' that came from nowhere and go nowhere, protected by no permissions, audited by no one, and trusted by no one with a functioning prefrontal cortex."

🙏 Respect to Ward Cunningham who invented the wiki in 1995 — a tool for humans to collaborate.
🙏 Respect to Wikipedia editors worldwide who defend verifiability, neutrality, and consensus.
🙏 Respect to every real wiki participant who knows that knowledge is built through human effort, not machine hallucination.

⚠️ THIS IS NOT A WIKI. THIS IS A FOLDER OF LLM-GENERATED FILES. ⚠️

Calling it a "wiki" is linguistic fraud. Do not be fooled.

🐑💀🧙

— The Elephant, The Wizard, and every human wiki editor who ever lived

Related pages