Critique of LLM Wiki Tutorial: Limitations and Production Readiness


Full Technical Evaluation: LLM Wiki Tutorial Video

Video Summary

This tutorial walks through setting up Karpathy’s LLM-Wiki pattern: a raw sources folder (immutable), a wiki folder (LLM-generated markdown), a schema.md configuration file, index.md and log.md for navigation, ingest/query/lint operations, and Obsidian as the front-end viewer. The video claims this creates a “persistent bookkeeper” where “the friction of maintaining an organized database disappears” and the user is “free to simply read, ask questions, and learn.”

https://gnu.support/images/2026/04/2026-04-23/800/ai-never-forgets-lie.webp


What the Video Gets Right ✅

1. Separation of raw sources from generated content — Keeping raw sources immutable and having the LLM write to a separate wiki folder is correct. This prevents accidental modification of originals.

2. Using a schema file to define rules — Having a configuration file (schema.md) that the LLM reads before working is good practice. It establishes conventions and boundaries.

3. Index and log files for navigation — For small-scale personal wikis, index.md and log.md provide basic discoverability and audit trails.

4. Lint passes for maintenance — Periodic health checks to find broken links, orphans, and gaps is a useful practice for any knowledge base.

5. Web clipper integration — Making it easy to capture web content as markdown reduces friction for sourcing.


What the Video Gets Wrong or Omits ❌

1. “The AI never forgets. It becomes a persistent bookkeeper.”

Claim: The AI maintains persistent memory across sessions.

Reality: The LLM has no persistent memory. It forgets everything between chat sessions. The “memory” is static markdown files. If those files contain errors, the LLM does not know. It will confidently repeat them. This is not persistence. This is a static snapshot that can be wrong forever.

Severity: Critical. The core promise of the video is false.

2. “The AI handles cross-referencing, tagging, and logging behind the scenes. The friction disappears.”

Claim: The AI autonomously maintains all wiki relationships and organization.

Reality: The video never addresses who fixes broken links when pages are renamed. Markdown [[wikilinks]] have no referential integrity. Rename a page, and every link to it becomes a 404. The AI might fix some during lint passes, but not all. The human must verify. The video also doesn’t address who resolves contradictions flagged by the AI, who merges duplicate pages, or who verifies that LLM-generated connections are not hallucinations.

Severity: Critical. The “friction disappears” promise is false. The human still does significant maintenance work.

3. “The AI is the programmer. Your knowledge is the codebase.”

Claim: This analogy implies the same level of reliability and tooling as software development.

Reality: Software codebases have compilers that catch syntax errors, type checkers, unit tests, and continuous integration. Markdown wikis have none of these. A broken link in a wiki does not cause a build failure. A hallucinated fact does not trigger a test assertion. The analogy is misleading.

Severity: Major. It creates false expectations about system reliability.

4. No mention of foreign keys or referential integrity

Omission: The video never explains that [[wikilinks]] can break silently.

Reality: In a relational database, foreign keys enforce that links always point to valid records. Markdown has no such enforcement. The video presents wikilinks as a feature without acknowledging their fragility.

Severity: Major. This is a fundamental architectural weakness.

5. No mention of permissions or access control

Omission: The video never discusses who can read or write which pages.

Reality: The LLM needs to read the entire wiki to answer questions. There is no mechanism to restrict access to private information (journal entries, client NDAs, medical records). The video assumes single-user personal use but doesn’t warn about privacy implications.

Severity: Major. Users could inadvertently expose sensitive data to the LLM.

6. No mention of version control beyond git

Omission: The video mentions git but doesn’t explain its limitations for knowledge bases.

Reality: Git tracks files, not individual facts or fields. When the LLM changes a claim in a page, you cannot easily see why it changed or roll back a single assertion. You revert the entire file. A proper audit trail requires field-level versioning.

Severity: Moderate. Users may assume git provides full audit capability.

7. “You don’t have to manually audit your files”

Claim: The lint pass automates health checks so humans don’t need to audit.

Reality: The lint pass finds issues. It does not fix them without human review. The video doesn’t specify whether the AI should auto-fix or just report. If it auto-fixes, it might introduce new errors. If it only reports, the human must still audit. Either way, the human is not “free” from auditing.

Severity: Moderate. The claim overstates automation.

8. No discussion of scale limitations

Omission: The video never addresses what happens when the wiki exceeds context window size.

Reality: Karpathy’s original pattern admits index.md works only at “small enough” scale. Beyond that, you need qmd (BM25 + vector search) — which is RAG. The video presents LLM Wiki as a complete solution without acknowledging its scaling ceiling.

Severity: Major. Users who scale will hit this wall and not understand why.

9. No discussion of error propagation

Omission: The video doesn’t address what happens when the LLM hallucinates a fact or relationship.

Reality: One hallucination can propagate across multiple pages as the LLM uses its own erroneous output as source material. The video’s “compounding” promise becomes a compounding error problem. Lint passes may catch contradictions but cannot determine which side is correct without human judgment.

Severity: Critical. This is the most dangerous hidden flaw.

10. No mention of token costs

Omission: The video never discusses API costs for ingest, query, and lint operations.

Reality: Every ingest consumes tokens. Every query consumes tokens. Every lint pass consumes tokens. At scale, with frequent updates, these costs are not negligible. The video presents the system as “free” beyond the tools, which is misleading.

Severity: Moderate. Users may be surprised by their API bills.


Technical Accuracy Summary

Claim Accuracy Severity
AI has persistent memory ❌ False Critical
Friction disappears; human is free ❌ False Critical
AI handles all maintenance ❌ False Critical
Wiki links are reliable ❌ False (no foreign keys) Major
No privacy/access concerns ❌ False (omission) Major
Scales without RAG ❌ False (omission) Major
Error propagation not a risk ❌ False (omission) Critical
No ongoing costs ❌ False (omission) Moderate
Lint pass automates auditing ⚠️ Partial Moderate
Separation of raw/wiki ✅ Correct Minor
Schema file is good practice ✅ Correct Minor
Index/log for navigation ✅ Correct (small scale) Minor
Web clipper is useful ✅ Correct Minor

The Fundamental Problem

The video presents LLM Wiki as a production-ready system where “friction disappears” and the human is “free to simply read, ask questions, and learn.” This is dangerously misleading.

In reality: - The LLM has no memory. It forgets everything between sessions. - Links break silently with no enforcement. - Private data is visible to the LLM with no permissions. - Hallucinations propagate and compound. - The human still fixes broken links, resolves contradictions, merges duplicates, and verifies facts. - At scale, you need RAG anyway.

This is not a “persistent bookkeeper.” It is a prototype that works for small personal wikis with active human supervision. It is not a replacement for a real knowledge base with foreign keys, permissions, version control, and deterministic metadata extraction.


Recommendations for Viewers

If you watch this video and decide to build an LLM Wiki:

  1. Keep it small. This works for <100 personal notes. Do not scale it to team use or large document collections.

  2. Expect to do maintenance. The AI will not handle everything. You will fix broken links, resolve contradictions, and verify facts.

  3. Do not put private data in the wiki. The LLM sees everything. No permissions.

  4. Audit regularly. The lint pass finds issues but does not resolve them. You must review.

  5. Monitor API costs. Every ingest and query costs tokens.

  6. Understand the trade-offs. This is a prototype, not a production system. Use a real database for anything serious.


The actual video

Final Verdict

The video is a well-produced tutorial for a weekend project. It is not a blueprint for a serious knowledge base. It ignores every hard problem: memory, integrity, permissions, scale, error propagation, cost, and maintenance. The “friction disappears” promise is false. The human is not free. The sheep are still lining up. 🐑💀

Build with integrity. Store with precision. Trust with provenance. Retrieve with speed. 🧙🐘

⚠️ THE WORD “WIKI” HAS BEEN PERVERTED ⚠️

⚠️ ARCHITECTURAL CRIME SCENE ⚠️

⚠️ THE WORD "WIKI" HAS BEEN PERVERTED ⚠️

By Andrej Karpathy and the Northern Karpathian School of Doublespeak

✅ A REAL WIKI — Honoring Ward Cunningham, Wikipedia, and every human curator worldwide
❌ KARPATHY'S "LLM WIKI" — An insult to the very concept
Human-curated
Real people write, edit, debate, verify, and take responsibility.
LLM-generated
Hallucinations are permanent. No human took ownership of any "fact."
Versioned history
Every edit has author, timestamp, reason. Rollback is trivial.
No audit trail
Who changed what? When? Why? Nobody knows. Git is an afterthought.
Source provenance
Every claim links back to its original source. You can verify.
"Trust me, I'm the LLM"
No traceability from summary back to source sentence. Errors become permanent.
Foreign keys / referential integrity
Links are database-backed. Rename a page, links update automatically.
Links break when you rename a file
No database. No foreign keys. Silent link rot guaranteed.
Permissions / access control
Fine-grained control: who can see, edit, delete, approve.
Anyone with file access sees everything
Zero access control. NDAs, medical records, client secrets — all exposed.
Queryable (SQL, structured)
Ask complex questions. Get precise answers. Join tables.
Browse-only markdown
Full-text search at best. No SQL. No structured queries.

🕯️ This is an insult to every Wikipedia editor, every MediaWiki contributor, every human being who spent hours citing sources, resolving disputes, and building the largest collaborative knowledge repository in human history. 🕯️

KARPATHY'S "WIKI" has:
❌ No consensus-building
❌ No talk pages
❌ No dispute resolution
❌ No citation requirements
❌ No editorial oversight
❌ No way to say "this fact is disputed"
❌ No way to privilege verified information over hallucinations
❌ No way to trace any claim back to its source

In the doublespeak of Northern Karpathia:

"Wiki" means "folder of markdown files written by a machine that cannot remember what it wrote yesterday, linked by strings that snap when you breathe on them, viewed through proprietary software that reports telemetry to people you do not know, containing 'facts' that came from nowhere and go nowhere, protected by no permissions, audited by no one, and trusted by no one with a functioning prefrontal cortex."

🙏 Respect to Ward Cunningham who invented the wiki in 1995 — a tool for humans to collaborate.
🙏 Respect to Wikipedia editors worldwide who defend verifiability, neutrality, and consensus.
🙏 Respect to every real wiki participant who knows that knowledge is built through human effort, not machine hallucination.

⚠️ THIS IS NOT A WIKI. THIS IS A FOLDER OF LLM-GENERATED FILES. ⚠️

Calling it a "wiki" is linguistic fraud. Do not be fooled.

🐑💀🧙

— The Elephant, The Wizard, and every human wiki editor who ever lived

Related pages