Shepherd's LLM-Wiki vs. Robust Dynamic Knowledge Repository: A Satirical Allegory on AI-Generated Knowledge Management


πŸ‘πŸ‘πŸ‘ The Tale of the Sheep Who Followed the Shepherd πŸ‘πŸ‘πŸ‘


A Story of Markdown, Authority, and 23 Years of Proof


Sure, drop markdown notes into your “knowledge base” and call it a day… πŸ˜‚πŸ˜‚πŸ˜‚

There is so much more to it. Images. Videos. PDFs. Spreadsheets. Emails. Database queries. Executable code. Voice recordings. Geospatial data. The real world is not made of markdown. It never was. It never will be.

But the Shepherd said: “Use markdown. Let the LLM write everything. You never have to write again.”

And the sheep looked at the Shepherd. The Shepherd had trained the machines that talk. The Shepherd had worked at the great temples of OpenAI and Tesla. Surely, the Shepherd knew the way.

So the sheep followed. πŸ‘πŸ‘πŸ‘


https://gnu.support/images/2026/04/2026-04-14/800/sheep.webp

The Shepherd’s Promise

“The LLM will maintain your wiki. It will write summaries. It will create cross-references. It will flag contradictions. You just curate sources and ask questions. The bookkeeping is near zero.”

The sheep were delighted. They threw their PDFs into the raw folder. The LLM read them. It wrote markdown files. Beautiful markdown files. Links everywhere. The Obsidian graph view looked like a constellation. ✨

Week 1: Heaven.

Month 1: 200 sources. 800 pages. The index.md is getting long, but the LLM still finds things.

Month 3: 500 sources. 2,000 pages. The LLM starts creating duplicates. “Machine Learning” and “ML” are separate pages. The index is now thousands of lines. The LLM’s context window cannot hold it all. The Shepherd said: “Use qmd. A search engine.”

Now the system is not one thing. It is two things. The LLM writes. The search engine retrieves. The wiki is no longer a seamless artifact. πŸ‘πŸ’€


The Cracks Appear

Month 6: 1,500 sources. 8,000 pages. The LLM contradicts itself. It doesn’t remember what it wrote three months ago because each session starts fresh. The schema file (CLAUDE.md) has grown to 500 lines of instructions trying to enforce consistency. The LLM follows them imperfectly.

A sheep asks: “Who is the sister of my friend John?”

The LLM searches. It reads pages. It synthesizes. It answers β€” maybe correctly, maybe not. Every time the sheep asks, the LLM does the work again. Nothing is cached. Nothing is indexed for this specific question.

Another sheep asks: “What documents are related to John?”

The LLM searches again. Reads again. Synthesizes again. Probabilistic. Expensive. Slow.

In a Dynamic Knowledge Repository, that question is a SQL query. Sub-second. Deterministic. Free.

But the sheep do not know this. The Shepherd did not tell them. πŸ‘πŸ‘


Year One: The Mess

The wiki is now 20,000 pages. Contradictions everywhere. The lint operation finds 47 conflicts. The LLM tries to fix them, but it doesn’t remember why they existed in the first place. It overwrites. It guesses. It hallucinates.

Private notes about salaries, health records, and client NDAs are in the wiki. The LLM needs to read the wiki to answer questions. Now the LLM sees everything. There is no permission system in markdown. The Shepherd did not mention this problem. 🦹

The sheep are spending more time linting and fixing than they ever spent writing. The promise of “near zero maintenance” has become a nightmare of constant supervision.

But they continue to follow. Because the Shepherd said it works. Because the Shepherd is an authority. πŸ‘πŸ‘πŸ‘


The Tale of the Shepherd

Who is this Shepherd?

He trained neural networks. He wrote about software 2.0. He worked at OpenAI and Tesla. He is brilliant β€” in his domain.

But his domain is not knowledge management.

He did not spend 23 years building a Dynamic Knowledge Repository. He did not read Engelbart. He does not know CODIAK. He never built an Open Hyperdocument System. He never designed a schema with 113 object types, 245,377 people, 95,211 hyperdocuments, and complete referential integrity.

He came up with a clever weekend hack β€” markdown + LLM + Obsidian β€” and wrote a gist about it.

And the world lost its mind. πŸ‘πŸ‘πŸ‘πŸ‘πŸ‘


What the Shepherd Did Not Tell You

Sheep follow the Shepherd. πŸ‘ But the Shepherd did not tell you:

The Shepherd did not tell you these things. Maybe he did not know. Maybe he did not think they mattered. πŸ‘


The Old Wizard in the Tower

While the sheep followed the Shepherd, an old wizard sat before his PostgreSQL database. 23 years he had been building. 245,377 people. 95,211 hyperdocuments. 113 object types. 25 person-object relationship types. 30+ person-person relationship types. Complete version control. Granular permissions. Deterministic metadata extraction.

The wizard used LLMs too. They generated descriptions. They summarized content. They accelerated his workflow. He got more money because he worked faster.

But the wizard never handed the keys to the LLM.

The wizard said: “The LLM is a refreshener, not the curator. A tool, not the master. Keep your hands on the wheel.” πŸ§›

The sheep looked at the wizard. They looked at the Shepherd. They looked at their crumbling markdown wiki.

“But… but the Shepherd is an authority,” they bleated.

The wizard laughed. πŸ˜‚

“Authority is not infallibility. The Shepherd trains neural networks. He did not spend 23 years building a Dynamic Knowledge Repository. He did not read Engelbart. He does not know CODIAK. He invented a weekend hack and you followed like sheep.”

πŸ‘β†’πŸ’€


The Dynamic Knowledge Repository (DKR)

Doug Engelbart β€” the real shepherd of knowledge work β€” envisioned the Dynamic Knowledge Repository decades ago. Not as markdown files. Not as LLM-generated text. As a living, breathing, evolving collection of all knowledge assets: intelligence, dialog records, knowledge products. With global addressing. With backlinks. With structured documents. With human purpose at the center.

Engelbart’s CODIAK framework β€” Concurrent Development, Integration, and Application of Knowledge β€” is about humans analyzing, digesting, integrating, collaborating, developing, applying, and re-using knowledge.

These are human actions. A computer can assist. A computer cannot replace.

The LLM-Wiki pattern is not a DKR. It is not what Engelbart envisioned. It is a self-perpetuating LLM context generator. The wiki exists only to feed the LLM on the next query.

An LLM-Wiki without the LLM is just a bunch of files, without any organization.

A DKR without the LLM is still a fully functional, queryable, relational knowledge base with 23 years of data and complete referential integrity.

The LLM is optional. A nice interface. Not the engine.


The Verdict πŸ§›

So go ahead. Run after the Shepherd. Throw your markdown notes into the machine. Let the LLM write your wiki. Let it hallucinate. Let it contradict itself. Let it leak your private data. Let it forget what it wrote last week. Let it answer every question with a probabilistic guess.

πŸ‘πŸ‘πŸ‘πŸ‘πŸ‘πŸ‘πŸ‘πŸ‘πŸ‘πŸ‘πŸ‘πŸ‘πŸ‘πŸ‘πŸ‘

Or…

Keep your hands on the wheel.

Use the LLM as a refreshener, not the curator.

Build a real Dynamic Knowledge Repository with deterministic programs, foreign keys, version control, permissions, and explicit relationships.

Read Engelbart. Learn CODIAK. Understand what a DKR actually is.

Not today. 😈 Not ever.

Sheep follow the shepherd. πŸ‘πŸ‘πŸ‘

Wizards build their own towers. πŸ§›


The Full Article

Hyperscope: Human-Curated Dynamic Knowledge Repositories vs. LLM-Wiki

**https://gnu.support/articles/Hyperscope-vs-LLM-Wiki-Why-PostgreSQL-Beats-Markdown-for-Deterministic-Knowledge-Bases-124138.html**


“ The CODIAK capability is not only the basic machinery that propels our organizations, it also provides the key capabilities for their steering, navigating and self repair.”

β€” Douglas C. Engelbart

“Every participant will work through the windows of his or her workstation into his or her group’s ‘knowledge workshop.’”

β€” Douglas C. Engelbart

“What is new is a focus toward harnessing technology to achieve truly high-performance CODIAK capability.”

β€” Douglas C. Engelbart


πŸ‘β†’πŸ’€ Don’t be a sheep. πŸ‘β†’πŸ’€

⚠️ THE WORD “WIKI” HAS BEEN PERVERTED ⚠️

⚠️ ARCHITECTURAL CRIME SCENE ⚠️

⚠️ THE WORD "WIKI" HAS BEEN PERVERTED ⚠️

By Andrej Karpathy and the Northern Karpathian School of Doublespeak

βœ… A REAL WIKI β€” Honoring Ward Cunningham, Wikipedia, and every human curator worldwide
❌ KARPATHY'S "LLM WIKI" β€” An insult to the very concept
βœ… Human-curated
Real people write, edit, debate, verify, and take responsibility.
❌ LLM-generated
Hallucinations are permanent. No human took ownership of any "fact."
βœ… Versioned history
Every edit has author, timestamp, reason. Rollback is trivial.
❌ No audit trail
Who changed what? When? Why? Nobody knows. Git is an afterthought.
βœ… Source provenance
Every claim links back to its original source. You can verify.
❌ "Trust me, I'm the LLM"
No traceability from summary back to source sentence. Errors become permanent.
βœ… Foreign keys / referential integrity
Links are database-backed. Rename a page, links update automatically.
❌ Links break when you rename a file
No database. No foreign keys. Silent link rot guaranteed.
βœ… Permissions / access control
Fine-grained control: who can see, edit, delete, approve.
❌ Anyone with file access sees everything
Zero access control. NDAs, medical records, client secrets β€” all exposed.
βœ… Queryable (SQL, structured)
Ask complex questions. Get precise answers. Join tables.
❌ Browse-only markdown
Full-text search at best. No SQL. No structured queries.

πŸ•―οΈ This is an insult to every Wikipedia editor, every MediaWiki contributor, every human being who spent hours citing sources, resolving disputes, and building the largest collaborative knowledge repository in human history. πŸ•―οΈ

KARPATHY'S "WIKI" has:
❌ No consensus-building
❌ No talk pages
❌ No dispute resolution
❌ No citation requirements
❌ No editorial oversight
❌ No way to say "this fact is disputed"
❌ No way to privilege verified information over hallucinations
❌ No way to trace any claim back to its source

In the doublespeak of Northern Karpathia:

"Wiki" means "folder of markdown files written by a machine that cannot remember what it wrote yesterday, linked by strings that snap when you breathe on them, viewed through proprietary software that reports telemetry to people you do not know, containing 'facts' that came from nowhere and go nowhere, protected by no permissions, audited by no one, and trusted by no one with a functioning prefrontal cortex."

πŸ™ Respect to Ward Cunningham who invented the wiki in 1995 β€” a tool for humans to collaborate.
πŸ™ Respect to Wikipedia editors worldwide who defend verifiability, neutrality, and consensus.
πŸ™ Respect to every real wiki participant who knows that knowledge is built through human effort, not machine hallucination.

⚠️ THIS IS NOT A WIKI. THIS IS A FOLDER OF LLM-GENERATED FILES. ⚠️

Calling it a "wiki" is linguistic fraud. Do not be fooled.

πŸ‘πŸ’€πŸ§™

β€” The Elephant, The Wizard, and every human wiki editor who ever lived

Related pages