If you're able to learn how to create an LLM knowledge base then you have essentially created your own "external brain" that you can utilise, on top of that, this could change how you run your business, your content, your network, your life, it can all be run through your own personalised external brains that you have access to. In fact, there's probably a market for creating this for local businesses in your community if you're able to connect to them and showcase them the value in creating an external brain for their entire business.
What happened?
Andrej Karpathy happened and showcased how he is using LLM knowledge bases which essentially work as external brains...
So, I wanted to showcase this system in this complete guide, step by step, so I had some back and fourths with my mate claude and we decided that we'd chat through 3 versions to create LLM knowledge bases, here they are...
Pick your version:
1. Complete beginner > Click CTRL F and type: "YOUR FIRST KNOWLEDGE BASE"
1. Comfortable with AI tools > Click CTRL F and type: "THE FULL SYSTEM" and read it up to automation.
1. Builder / developer > Click CTRL F and type "THE FULL SYSTEM" and read it all the way through.
Before you do that ask yourself if you want to do this and read this:
Most people use AI like a search engine with amnesia.
You ask a question, get an answer, close the tab. Tomorrow you start from scratch. Nothing accumulates. Nothing compounds. You're burning tokens to rediscover the same context over and over again.
Karpathy's system flips this completely:
1. You collect raw material. Articles, papers, YouTube transcripts, PDFs, anything related to a topic you care about
1. The AI reads everything and writes a structured wiki. Summaries, concept explanations, connections between ideas, a master index
1. You ask questions against the wiki. It researches across its own compiled knowledge and gives you cited, synthesised answers
1. Every answer gets filed back into the wiki. So the next question benefits from all previous work
1. The AI periodically health-checks the wiki. Finding contradictions, gaps, outdated information, and fixing them
The result? A personal knowledge base that gets smarter every time you touch it.
After a month of feeding it, you have a deeply interlinked resource that no Google search could replicate. Because it's been *synthesised*, not just indexed.
This works for literally any topic. Crypto markets. Medical research. Legal case law. Competitive intelligence. Academic study. Philosophy. If you want to accumulate and connect knowledge over time, this is the system.
1: YOUR FIRST KNOWLEDGE BASE:
Zero tekkerzzz are required here (tekkerzzz is like British slang for a baller, which is British slang for someone who has skill with a football, which actually just means you got skills, I could have just said zero skills but that's not me, so zero tekkerzzz it is init)...
If you can install an app and copy-paste text, you can do this right now.
What you need
- Obsidian (free). A note-taking app that works with plain text files. Download from obsidian.md. Mac, Windows, Linux, iOS, Android.
- A Claude subscription ($20/month Pro at claude.ai). Or any AI chatbot you prefer: ChatGPT, Gemini, whatever.
That's it.
Step 1: Create your vault (2 minutes)
Open Obsidian. It'll ask you to create or open a "vault." A vault is just a folder on your computer where your notes live.
1. Click "Create new vault"
1. Name it something descriptive (e.g. "crypto-research" or "health-knowledge")
1. Choose where to save it (Documents folder is fine)
1. Click "Create"
You now have an empty vault. Obsidian watches this folder. Any markdown file you put inside will appear as a note automatically.
Step 2: Create two folders (1 minute)
In Obsidian's left sidebar, right-click and select "New folder". Create these two:
- raw for your source material (articles, notes, anything you collect)
- wiki for where the AI will build your compiled knowledge base
That's your entire starting structure.
Step 3: Add your first raw sources (5 minutes)
Pick a topic you're genuinely interested in. Find 3 to 5 good articles about it. For each one:
1. In Obsidian, right-click the raw folder > "New note"
1. Give it a descriptive name (e.g. "bitcoin-halving-2024-explainer")
1. Copy-paste the article text into the note
1. At the very top, add a line like: Source: [paste the URL here]
Don't overthink the formatting. Don't stress about structure. Just get the raw text in there.
Quick win: Install the Obsidian Web Clipper browser extension (free, Chrome/Firefox/Safari/Edge). It saves web pages directly into your vault as formatted markdown notes with one click. But for your first go, copy-paste is perfectly fine.
Step 4: Ask the AI to compile your wiki (5 minutes)
Open Claude (or your preferred AI). Copy and paste this prompt, replacing the bracketed sections:
Claude will produce structured output. Copy each section into a new note inside your wiki folder:
- Save the summaries as individual notes (e.g. wiki/summary-bitcoin-halving.md)
- Save the master index as wiki/index.md
- Save the concept article in wiki/ with a descriptive name
Step 5: See the magic
Open Obsidian's Graph View (click the graph icon in the left sidebar, or press Ctrl/Cmd+G).
You'll see your notes as dots, connected by the [[wikilinks]] the AI created. This is your knowledge base, visualised as a network of connected ideas.
Click on any [[linked concept]] in a note. If a page exists for it, Obsidian opens it. If it doesn't exist yet, Obsidian offers to create it. This is how the wiki grows organically.
You now have a working knowledge base. Everything from here is about making it bigger, faster, and more powerful.
Growing Your Knowledge Base (The Daily Habit)
Once you have the basic structure, the workflow becomes dead simple:
Adding new sources
Whenever you read something worth keeping:
1. Clip or paste it into your raw folder
1. Open Claude and paste this:
1. Save the outputs into your wiki folder, replacing the old index with the updated one
Asking questions
This is where it gets genuinely powerful. Once you have 10+ compiled articles:
The critical habit: always file the answer back into the wiki. Save it in your wiki folder. This is the compounding loop. Every question enriches the base for future questions.
Weekly health check
Once a week, paste your full index to Claude with this:
Three interactions per week. One or two source additions, an occasional question, and a health check. That's all it takes for the knowledge base to grow steadily.
2: THE FULL SYSTEM:
The Full System: Architecture and Setup
Everything above works with just Obsidian and Claude Chat. But if you want the system Karpathy actually described, where the AI handles file creation, maintenance, and indexing automatically, here's the complete architecture.
The three-layer design
Layer 1: Raw sources (raw/ folder). Your single source of truth. The AI reads from here but never modifies it. Articles, papers, repos, datasets, images all go here. Think of it as your library's intake shelf.
Layer 2: The compiled wiki (wiki/ folder). AI-generated and AI-maintained. Summaries, concept articles, entity pages (people, organisations, tools), cross-links, indexes, and query outputs all live here. You rarely edit this directly. The AI does the writing.
Layer 3: The schema (CLAUDE.md file). A configuration document that tells the AI how the wiki is structured, what naming conventions to follow, and what operations are available. This is the programme that controls the AI's behaviour. Lives in your vault's root folder.
The four operational cycles
These repeat continuously, compounding the wiki's value:
Ingest. You add raw sources. The AI reads them and creates summaries, concept pages, and connections.
Compile. The AI builds and updates wiki pages, maintains the index, and weaves new information into the existing structure.
Query. You ask questions. The AI researches across the wiki and produces cited answers, which get filed back.
Lint. The AI health-checks for contradictions, gaps, broken links, stale content, and missing pages. It fixes what it can and flags what it can't.
The complete folder structure
Create this inside your Obsidian vault:
You don't need every subfolder on day one. Start with raw/, wiki/, and CLAUDE.md. Add subfolders as the wiki grows.
Naming your files
Use kebab-case (lowercase, words separated by hyphens) for all filenames:
- active-inference.md ✓
- Active Inference.md ✗
- active_inference.md ✗
For source summaries, use author-year-short-title.md (e.g. friston-2010-free-energy.md). For concepts, use the concept name directly (e.g. transformer-architecture.md).
Why this matters: kebab-case works across every operating system, is URL-friendly, and AI references it consistently. Trust me on this one. You'll thank yourself later when you have 200+ files.
Setting Up Your AI Environment
The fundamental workflow is identical regardless of which tool you pick. The AI reads your markdown files, processes them, and produces markdown output. What changes is *how* the AI accesses your files.
Option A: Claude Chat (Easiest, No Technical Skill Required)
This is what most people should start with.
Claude Chat at claude.ai can't write files to your computer directly. You work in a copy-paste loop: upload or paste your files, ask Claude to process them, copy the output back into your vault.
Setup:
1. Go to claude.ai and sign in (Pro subscription at $20/month recommended)
1. Create a Project (left sidebar > Projects > New Project)
1. In the Project instructions, paste the contents of your CLAUDE.md file
1. Upload your raw sources and existing wiki files to the Project
Every conversation in that Project now has persistent context. Claude remembers your wiki structure and conventions across conversations.
The workflow:
- Paste new sources > Claude generates wiki pages > you copy them into Obsidian
- Ask questions > Claude produces cited answers > you save them as notes
- Request a health check > Claude identifies issues > you apply fixes
Yes, it's manual. But it works reliably and requires zero terminal knowledge.
Option B: Claude Code (Most Powerful, Requires Terminal Comfort)
Claude Code is Anthropic's command-line tool with full filesystem access. It reads, writes, creates, and modifies files directly. Ideal for automated wiki maintenance.
Before choosing this option:
- You need basic terminal comfort (navigating folders, running commands)
- Minimum Pro subscription at $20/month, Max plans at $100 or $200/month for heavier usage
- No free tier for Claude Code
Okay if you want to continue then go to Claude Code and start a session.
Navigate to your vault folder in the terminal and type claude to start a session. Claude Code automatically reads your CLAUDE.md file from the vault root.
Why this is powerful: Claude Code can read all your files, create new wiki pages, edit existing ones, run search tools, and execute scripts. All without you copying and pasting anything. A single prompt like "process all new files in raw/" triggers the entire ingest-compile cycle automatically.
Option C: Other AI Tools
Several alternatives support this workflow as well but I'm using Claude Code but Codex would also work because the core requirement for any tool is the same: read markdown files in, produce markdown files out.
Writing CLAUDE.md: The Schema That Controls Everything
The CLAUDE.md file is the single most important file in your system. It tells the AI exactly how your wiki is structured, what rules to follow, and what operations it can perform.
Think of it as a job description for your AI research librarian.
If you're using Claude Code, this file loads automatically at the start of every session. If you're using Claude Chat, paste it into your Project instructions or at the start of each conversation.
The template (copy this, customise the first line)
Keep this file concise. Every line eats into your AI context window budget. Tell the AI *how to find* information rather than including everything inline. The template above is deliberately under 80 lines. Resist the urge to over-specify.
Supercharging Data Collection with the Web Clipper
Manually copying articles works, but the Obsidian Web Clipper browser extension makes collection dramatically faster. One click turns a web page into a clean markdown note inside your vault.
Installation
Install the free, open-source extension from your browser's extension store (Chrome, Firefox, Safari, Edge). Click the Obsidian icon in your browser toolbar, then the gear icon to configure.
Configuration
Set these:
- General > Vault: Type your vault name exactly as it appears in Obsidian
- Templates > Default template > Note location: Set to raw/articles/
- Templates > Default template > Properties: Add these metadata fields:
- Templates > Default template > Note content: Set to {{content}}
What this gives you
Reading an article worth saving? Click the Obsidian icon in your browser toolbar, confirm the template, click "Add to Obsidian". Done. The article appears as a formatted markdown note in your raw/articles/ folder, ready for the AI to process.
Making images work offline
By default, the Web Clipper saves images as web links, which break when pages go offline. Fix this with the Local Images Plus plugin:
1. In Obsidian: Settings > Community Plugins > Turn off Restricted Mode > Browse
1. Search for "Local images plus" and install it
1. In the plugin settings, set the download folder to raw/assets/
1. Run the "Localise attachments" command after clipping sessions
This downloads all referenced images locally. Everything works offline and the AI can reference images if you're using a vision-capable model.
Converting non-web sources
For PDFs, Word docs, and other files, use MarkItDown (a free tool by Microsoft):
Not comfortable with the command line? Open the PDF, select all, copy, paste into a new note in raw/. Formatting won't be perfect, but the AI only needs the text content.
Wiki Compilation: Turning Raw Sources Into Structured Knowledge
This is the core operation. You've collected sources. Now the AI compiles them into a living wiki.
First-time compilation
If using Claude Code, navigate to your vault in the terminal and run: claude, and then enter:
If using Claude Chat, upload your raw source files (or paste them) along with your CLAUDE.md, and use the same prompt. You'll need to copy each generated page into the correct folder manually.
For ten articles, expect roughly 10 source summaries, 15 to 30 concept/entity pages, and a comprehensive index. Typically takes one conversation with a few follow-ups.
Incremental compilation (adding new sources)
After the initial build, each new source triggers an incremental update rather than reprocessing everything:
This is the key efficiency principle: the AI integrates new sources into the existing structure rather than rebuilding from scratch. Fast and token-efficient.
When things go wrong (and they will)
The AI will occasionally produce imperfect output. Completely normal. Here's what you'll run into:
Missing or malformed frontmatter. The AI sometimes forgets the YAML block at the top of a page, or gets the formatting wrong. Fix: remind it in your next prompt ("please ensure all pages have complete YAML frontmatter as specified in CLAUDE.md").
Broken wikilinks. The AI creates a [[link]] to a page that doesn't exist yet. Fix: this is actually fine. In Obsidian, clicking a broken link offers to create the page. You can also ask the AI to create stub pages for all unlinked concepts during a lint pass.
Hallucinated connections. The AI claims two concepts are related when the source material doesn't support it. Fix: this is exactly why you keep raw sources immutable. You can always check claims against the originals. Flag suspicious connections and ask the AI to verify with specific citations.
Context window overflow. If your wiki gets large, the AI can't read everything at once. Fix: always start with the index file, then load only the specific pages needed.
The Master Index: How the AI Navigates Your Wiki
The index file is what makes this entire system work without complex database infrastructure. It's a table of contents the AI reads first, allowing it to decide which specific pages to load for any given task.
What a good index looks like (example):
Each entry has a [[wikilink]], a roughly 50-word summary, key tags, and source count. This format lets the AI scan your entire wiki's contents in about 6,500 tokens for 100 articles. Trivially small compared to modern context windows.
The activity log
wiki/log.md is an append-only record tracking every operation:
Full audit trail. Helps the AI understand what's been done recently.
Asking Questions: The Compounding Loop
Once your wiki reaches even 10 to 20 compiled articles, you can start asking complex questions that synthesise across multiple sources.
The query prompt:
The AI reads the index, identifies relevant pages, reads them, and produces a cited answer. If using Claude Chat, you'll paste the index and relevant pages manually.
The crucial step is filing the answer back. Save it in wiki/outputs/. This is the compounding loop. Every question enriches the knowledge base for future queries.
Different output formats
Your wiki doesn't have to produce only text:
Slide presentations (using Marp format):
Data visualisations (if using Claude Code or another tool with code execution):
Comparison tables, decision frameworks, reading lists. The wiki is your canvas.
Obsidian Plugins
These free plugins transform Obsidian from a simple note viewer into a serious knowledge base interface (please DYOR on plugins you utilise, my LLM has suggested the following):
How to install plugins
Settings > Community Plugins > Turn off Restricted Mode > Browse. Search for the plugin name, click Install, then Enable.
The essentials
Dataview is the single most powerful plugin for this workflow. It treats your vault as a queryable database, reading all your YAML frontmatter automatically. Embed live queries in any note:
This creates a live table of all your concept articles, sorted by most recently updated. Use it to build dashboards, find unprocessed sources, and identify gaps.
Templater auto-populates dates, filenames, and other variables when you create notes. Saves time when manually creating wiki pages.
Obsidian Git gives you automatic version control. Configure it to auto-commit every 30 minutes and push to a remote repo. Every change is tracked and reversible. Your safety net when the AI makes a mistake.
Tag Wrangler for bulk renaming and merging tags across your entire vault. Essential once your taxonomy grows.
Linter auto-formats notes on save. Enforces consistent YAML frontmatter, heading levels, spacing. Critical when the AI writes many files since formatting drifts.
Marp Slides renders markdown files as presentation slides. Any note with marp: true in its frontmatter becomes a slideshow. Export to PDF, HTML, or PowerPoint.
Homepage designates a note as your vault's landing page. Build a dashboard with Dataview queries showing recent activity and statistics.
Graph View (built in, no plugin needed)
Obsidian's Graph View visualises your wiki as an interactive network. Notes are dots. [[wikilinks]] are connecting lines. Highly connected concepts appear as larger nodes.
Use it to spot clusters of related knowledge, find orphan notes that need integration, and discover unexpected connections between topics.
Health Checks and Quality Maintenance
Periodic health checks prevent your wiki from decaying into a collection of disconnected notes. Run these weekly, or after every major batch of new sources.
The lint prompt:
Supplementary automated checks
If you're comfortable with the command line, these free tools catch structural issues the AI might miss:
- markdownlint (npm install -g markdownlint-cli) enforces consistent markdown formatting
- markdown-link-check (npm install -g markdown-link-check) validates all hyperlinks
Optional. The AI's semantic lint catches the important stuff. These catch formatting nits.
The two-model validation pattern
For high-stakes knowledge bases (medical research, legal analysis, investment theses), use two different AI models: one writes the wiki pages, a second independently validates them before they enter the "live" wiki.
This prevents compounding hallucinations. A real concern flagged by many commentators on Karpathy's original post.
Compile with Claude, validate with GPT-4o (or vice versa). If both agree, the content is likely sound. If they disagree, investigate.
Scaling: When Your Wiki Grows Large
How much can you fit?
At small scale (under roughly 100 articles), the index-file approach works brilliantly. The AI reads the index, picks relevant pages, loads only what it needs.
Adding search with QMD
QMD is a free, open-source local search engine built by Tobi Lütke (CEO of Shopify). Purpose-built for markdown knowledge bases. Combines keyword search, semantic (meaning-based) search, and AI reranking. All running locally on your machine with no cloud dependencies.
Automating the Entire Workflow
The manual copy-paste workflow works. But it's slow. The real power of this system unlocks when the wiki builds itself.
The Best and Easiest Way: Install a Ready-Made Plugin (2 Minutes)
Since Karpathy shared his pattern, the community has built ready-made Claude Code plugins that turn the entire wiki workflow into simple slash commands. You don't need to write any configuration, create any templates, or craft any prompts. Install, type two commands, done.
The fastest path is the wiki-skills plugin:
You now have these commands:
- /wiki-init scaffolds the entire folder structure in seconds
- /wiki-ingest processes a raw source into the wiki (summary, concepts, entities, wikilinks, index update)
- /wiki-query researches a question across your wiki and files the answer back
- /wiki-lint runs a health check and fixes what it can
Your workflow becomes:
1. Drop articles into raw/ (via Web Clipper or copy-paste)
1. Type /wiki-ingest in Claude Code
1. Done. Open Obsidian and browse your wiki.
IMPORTANT:
If you don't want a plugin, or you want to understand what's happening under the hood, the rest of this section walks through five levels of DIY automation.
Level 1: One-Command Compilation (Claude Code CLI)
If you have Claude Code installed ($20/month Pro minimum), you can process every new source in your raw/ folder with a single command:
That's it. Claude Code reads your raw sources, writes every markdown file, places them in the correct folders, creates all the cross-links, updates your index. Open Obsidian and everything is there.
The -p flag means "prompt." It runs non-interactively and exits when done.
Level 2: Custom Slash Commands (Type /compile and Walk Away)
Claude Code supports custom slash commands. Reusable workflows saved as markdown files, invoked with /command-name.
Create this file at .claude/commands/wiki-compile.md:
Now in any Claude Code session inside your vault, type:
Claude reads the command file, follows every step, builds your wiki.
Create similar commands for other operations:
- .claude/commands/wiki-lint.md > /wiki-lint for health checks
- .claude/commands/wiki-query.md > /wiki-query How do AMMs work? for research
Level 3: Scheduled Tasks (The Wiki Builds Itself Every Day)
This is where it gets properly autonomous. Claude Code supports scheduled tasks that run automatically without you typing anything.
Using Claude Desktop (Mac/Windows):
Open a task, type /schedule, configure:
- What: "Read CLAUDE.md. Process all new files in raw/. Compile into wiki."
- When: Daily at 9am (or whatever suits you)
- Repeat: Daily / weekdays / weekly
Each run starts a fresh session, processes your raw sources, writes the wiki pages, exits. You clip articles throughout the day; the wiki compiles itself overnight.
Using the CLI:
Using cron (Mac/Linux):
This runs in the background. Your wiki grows while you sleep.
Important notes:
- Each scheduled run consumes your Claude usage quota
- Desktop tasks only run while your computer is awake and Claude Desktop is open
- Cron tasks require Claude Code to be installed and authenticated on the machine
Level 4: GitHub Actions (Cloud Automation, Computer Off)
The most robust setup. Your computer can be off. Compilation happens on GitHub's servers.
1. Store your vault in a GitHub repository
1. When you push new files to raw/, a GitHub Action triggers
1. Claude Code compiles the wiki
1. Updated wiki files commit back to the repo
1. Pull changes into your local Obsidian vault (or use Obsidian Git to auto-sync)
The workflow file (.github/workflows/compile-wiki.yml):
Cost note: This uses the Claude API (pay-per-token), not your subscription. You'll need an API key from console.anthropic.com. For 5 to 10 new sources, expect under $0.50 per run on Sonnet 4.6.
Level 5: Full Agent Skills (Self-Maintaining Wiki)
Claude Code supports Agent Skills that trigger automatically when it detects the right context. Unlike slash commands (which you invoke), skills activate on their own.
Create .claude/skills/wiki-maintainer/SKILL.md:
With this skill in place, just say "I added three new articles to raw/" and Claude knows exactly what to do. No commands needed.
Start with Level 1. Layer on more automation as you get comfortable. Each level builds on the previous. Nothing breaks if you jump ahead.
Synthetic Data Generation (Advanced)
Once your wiki is mature, you can use it to create training data for a fine-tuned model that "knows" your domain in its weights.
Feed each wiki article to Claude:
From 100 wiki articles, expect 300 to 500 QA pairs. Enough to begin fine-tuning a smaller model via OpenAI's API (roughly $8 per million training tokens) or locally with QLoRA on a consumer GPU.
Fine-tuning teaches the model your domain's vocabulary and reasoning patterns. It's poor at memorising specific facts, so those are better served by the wiki itself. The ideal setup? A fine-tuned model for behaviour combined with the wiki for specifics.
Existing Tools and Community Implementations
You're not building from zero. Several open-source projects already implement this pattern:
- llm-wiki-compiler with /wiki-init and /wiki-compile commands
- sage-wiki as a full CLI with compile, search, query, and serve commands (exposes wiki as an MCP server)
- CRATE as a Python CLI implementing the three-layer pattern, OpenAI-compatible, Obsidian-friendly
- QMD as a local search engine for markdown with keyword, vector, and hybrid search plus MCP server support
- Fabric with 140+ curated prompt patterns for standardised operations
YAML Frontmatter Quick Reference
Every wiki page should have metadata at the top. This enables Dataview queries, automated maintenance, and consistent AI processing.
Troubleshooting Common Issues
"The AI keeps rewriting pages instead of updating them." Add explicit instructions: "Append new information to existing sections. Do not rewrite content that hasn't changed. Preserve all existing wikilinks."
"My index is getting too long and unwieldy." Split into category indexes. Keep wiki/index.md as a lightweight overview (page title + one-line summary only). Put detailed summaries in each _index.md file within subfolders.
"The AI hallucinates connections between unrelated topics." The most common quality issue. Mitigate by always tracing claims to specific source pages, running the two-model validation pattern for critical content, and including "only make connections explicitly supported by source material" in your CLAUDE.md.
"I hit Claude's usage limits before finishing compilation." Process sources in smaller batches (3 to 5 at a time). Use incremental compilation rather than reprocessing everything. If you regularly hit limits, consider Max ($100/month) or API pay-as-you-go.
"My wiki has duplicate concepts under different names." Happens naturally as the AI encounters the same idea phrased differently across sources. Add a lint rule to your CLAUDE.md: "Check for concept pages describing the same thing under different names. Merge them, keeping the most common name and adding redirects."
"I want to share my wiki with someone else." Everything is plain markdown files. Push the vault to a GitHub repo, share via cloud storage (Dropbox, Google Drive, iCloud), or deploy as a website using MkDocs Material (pip install mkdocs-material && mkdocs serve).
The Bottom Line
The shift Karpathy identified, from "AI as answer machine" to "AI as knowledge infrastructure," changes how individuals manage expertise. The system requires no programming background, no database administration, and no infrastructure beyond a free note-taking app and an AI subscription.
Three things make this different from normal AI usage:
The filing loop is the superpower. Every query saved back to the wiki enriches it for future queries. Compound returns on every interaction. No chat interface gives you this.
The AI is a compiler, not a search engine. It synthesises and connects ideas in ways that chunked retrieval never can. It understands the *relationships* between documents, not just their similarity.
Plain text is forever. Your knowledge base is a folder of markdown files, readable by any tool, on any operating system, for as long as computers exist. No vendor lock-in. No proprietary format. No opaque database.
Start small. Pick one topic. Clip five articles. Run the compilation. Ask a question. See what the wiki gives you back. Then keep feeding it.
The only question is what you build your first wiki about.
Drop a comment with your topic. I'll tell you if this system fits it.
and finally, as I have taken like a good section of my day to put this post together for free for you I believe I'm allowed to shill my own newsletter.
please go check out yesterday's free newsletter covering a ton of useful AI alpha you can utilise:
