← Docs hub

Phase 1.25 — Research Report

Date: 2026-04-08 Method: Cloned every referenced implementation from Karpathy's gist and the top related GitHub searches into .temp/ (gitignored) for side-by-side comparison.

This document is the deliverable for Phase 1.25 Research, a new phase added to the llmwiki framework (see docs/framework.md). It is the source of truth for prior-art analysis and the 10x gap that llmwiki targets.

Summary

15 reference implementations were cloned and analysed. They fall into four clusters:

Cluster What they do Examples llmwiki differentiation
Pure-markdown skills A Claude Code skill/plugin + CLAUDE.md schema; rely on the agent to do all writes kfchou/wiki-skills, Astro-Han/karpathy-llm-wiki, bashiraziz/llm-wiki-template ✅ Same base + native .jsonl → markdown + static HTML + multi-agent
Markdown-first + light Python Schema + a few Python scripts for ingest/query/lint SamurAIGPT/llm-wiki-agent, Ss1024sS/LLM-wiki, hsuanguo/llm-wiki ✅ Same shape + session-transcript adapter + beautiful static site
Obsidian-coupled Wiki lives inside an Obsidian vault; user views via Obsidian AgriciDaniel/claude-obsidian, louiswang524/llm-knowledge-base, kytmanov/obsidian-llm-wiki-local, remember-md/remember 🔀 Obsidian as one of many connectors (input adapter) — not the only view
Heavy Python / hosted Backend services, databases, hosted demos lucasastorian/llmwiki (Apache, Supabase + MCP, hosted at llmwiki.app), bitsofchris/openaugi ❌ Too heavy — violates llmwiki's stdlib-first rule
Session browsers (not wikis) Search/TUI over raw .jsonl; no wiki compilation raine/claude-history, sinzin91/search-sessions 🔀 Complementary — they search raw; llmwiki builds the wiki on top

Per-repo analysis

Pure-markdown skills

kfchou/wiki-skills

Astro-Han/karpathy-llm-wiki

bashiraziz/llm-wiki-template

Markdown-first + light Python

SamurAIGPT/llm-wiki-agent

Ss1024sS/LLM-wiki

hsuanguo/llm-wiki

Obsidian-coupled

AgriciDaniel/claude-obsidian

kytmanov/obsidian-llm-wiki-local

louiswang524/llm-knowledge-base

remember-md/remember

Heavy Python / hosted

lucasastorian/llmwiki

bitsofchris/openaugi

Session browsers (complementary, not competitors)

raine/claude-history

sinzin91/search-sessions

Not cloned (mentioned but out of scope)

The 10x gap (feature matrix)

Feature Most existing llmwiki
Ingests .jsonl session transcripts ❌ (generic markdown only)
Claude Code adapter Some
Codex CLI adapter None ✅ stub (v0.2)
Multi-agent adapter pattern None
Pure stdlib + markdown (no DB, no MCP, no Node) ~50%
Beautiful static HTML viewer ✅ god-level UI
Global search (Cmd+K) ✅ client-side index
Syntax highlighting Rarely ✅ highlight.js (CDN)
Redaction by default ✅ username + API keys + tokens + emails
Live-session detection ✅ skips <60min old
Idempotent incremental sync Some ✅ mtime state file
Windows .bat scripts Rarely
Obsidian connector Some (only input) ✅ input and output
No cloud, no telemetry, no auth Some ✅ hard rule
Build time <15s for 300 sessions Varies ✅ 9s measured

Borrowed ideas (with attribution)

Decisions informed by this research

  1. Keep llmwiki stdlib-first. lucasastorian/llmwiki shows the "full-stack" approach exists; llmwiki is the local alternative.
  2. Ship an Obsidian adapter in v0.1. Four of 15 reference implementations use Obsidian — clearly important to users. Make it an optional input adapter, not the only path.
  3. Ship the HTML viewer as the hero feature. None of the reference implementations have a beautiful static HTML output. This is llmwiki's most visible 10x.
  4. Keep the slash commands compatible with SamurAIGPT/llm-wiki-agent and kfchou/wiki-skills. Users can switch between implementations.
  5. Use Karpathy's three-layer structure exactly (raw/ immutable, wiki/ LLM-maintained, schema in CLAUDE.md/AGENTS.md). No deviations.
  6. Build-time redaction is non-negotiable — none of the reference implementations do this, and session transcripts leak PII by default.

References