Karpathy's Wiki vs. Open Brain. One Fails When You Need It Most.
By AI News & Strategy Daily | Nate B Jones
Summary
Topics Covered
- Knowledge Compiles Once, It Doesn't Re-derive
- Study Guide vs. Filing Cabinet: Two Knowledge Philosophies
- The Quiet Risk of Trusting AI Summaries
- Wiki Staleness Looks Like Active Misinformation
- From Oracle to Maintainer: AI's Evolving Role
Full Transcript
Everybody is going nuts about the wiki idea that Andre Carpathy posted a couple of days ago. 41,000 people have bookmarked it. And on the surface, it
bookmarked it. And on the surface, it sounds ridiculously simple. You use your AI to build and maintain a personal wiki. As someone who launched something
wiki. As someone who launched something somewhat similar with OpenBrain, I have gotten hundreds of DMs and emails saying, "Nate, is this the same thing?
Nate, is this different? Nate, does this make OpenBrain obsolete? Nate, is the wiki better?" I want to actually step
wiki better?" I want to actually step back and give you a different answer than you expected. We are going to go through, we are going to talk about what
the wiki approach gets right, why Andre put it together, the principles underlying this wiki approach versus the open brain approach. And I want to be honest about why we're doing that. We're
not doing that just for giggles. We're
doing that because deciding how you organize your context layer is one of the single most important things you can do in 2026. It is a really big deal. And
Karpathy's approach is really different from OpenBrain. It's not the same thing
from OpenBrain. It's not the same thing at all. And there are specific reasons
at all. And there are specific reasons why he went with his decision and why I went with the decisions that I made to construct open brand. And I'm going to be really open about both. I'm not going to give you a steelman. I'm not going to
say one is better than the other. So
just get that out of your head. We're
going to talk about where each one breaks, the scale issues that each have, and where each one is strong. And then I promise you, I am going to help you solve the problem because I know that
what most of you are going to say is, "I want the best of both worlds." We are going to get there. I put a plugin into OpenBrain that will help you have the best of both worlds. So you can have the
wiki approach Carpathy takes with the structured data that OpenBrain brings.
And by the end of this video, you'll understand why it matters to be able to do both and you'll feel equipped to make that decision. So let me explain a
that decision. So let me explain a little bit about what's under the surface that is a little bit more complex than it might appear at first glance. Because on the surface,
glance. Because on the surface, Carpathy's implementation is very simple. It's just folders and it's text
simple. It's just folders and it's text files. That is the big idea for him. And
files. That is the big idea for him. And
you might wonder why did he go so simple? Well, his insight is the idea
simple? Well, his insight is the idea that everybody uses AI with their documents today all day, right? You
upload files to chat GPT, you use notebook LM, you use cloud, and you always end up in a place where you have bunches of documents everywhere and then you ask a question and the AI has to go
and find relevant chunks of documents, previous chats, reads them, gives you an answer. It's all hyper fragmented all
answer. It's all hyper fragmented all over. This works okay, but it's actually
over. This works okay, but it's actually not ideal because what's happening under the hood is that AI is effectively rediscovering your knowledge from scratch every single time you ask a question. So, if you ask a question that
question. So, if you ask a question that requires connecting five different documents across six different chats, the AI has to find all five, read the six chats, figure out how they relate, and produce some kind of a synthesis. If
you ask a similar question tomorrow, it does the whole thing over again. Nothing
was saved about that synthesis. It
doesn't inherently preserve connections between documents, even if you're in the same chat application, let alone if you're in two or three different AIs like so many of us are. In other words,
the AI did real cognitive work and then threw it all away. And that is what bothered Carpathy enough to put together a different solution. What he asked is, what if instead of just finding relevant
chunks and answering questions, the AI actually wrote down what it learned?
What if every time you added a new source, the AI read that source, figured out what mattered and updated a set of organized notes that already contained everything it learned from the previous
source? In other words, what if it
source? In other words, what if it started to auto update based on its own synthesis? What if those notes already
synthesis? What if those notes already included cross references, flagged contradictions, tracked how your understanding evolved? If you're trying
understanding evolved? If you're trying to learn how you think and evolve over time, that's what this wiki idea is for.
It essentially becomes not just a clever file organization system, but actually a persistent artifact as a whole where
you're capturing the AI's evolving understanding of your thinking over time. So the AI might read a paper you
time. So the AI might read a paper you give it on Monday, write up what it learned, and link back to it what it learned the previous week from some other thread that you were working on.
The next Friday when you ask a question, the AI doesn't have to reread everything from scratch. Instead, it looks at the
from scratch. Instead, it looks at the synthesis that is already sitting there, the cross references that are already built, the contradictions that have been flagged. And in Carpathy's own words,
flagged. And in Carpathy's own words, the knowledge is compiled once and then kept current. It's not rederived on
kept current. It's not rederived on every query. And that is really the key
every query. And that is really the key here. Most AI knowledge tools spend
here. Most AI knowledge tools spend compute and tokens to rederive. whereas
his wiki compiles. That's a
fundamentally different relationship between AI and your information and it has different strengths and weaknesses than a lot of other predominant memory paradigms including OpenBrain. He
described his working setup as having the AI agent on one side and Obsidian which is just a note viewing app on the other side. And so that allows the AI to
other side. And so that allows the AI to make edits based on their conversation and he can browse the results in real time and follow links and check the graph view and read the updated pages and kind of have an evolving
conversation with the agent. And of
course being a programmer the way he talks about it is the LLM is sort of the programmer for the codebase of the wiki which means in plain English the note app is where he actually reads and the
AI is the one writing in the notes app based on a collected series of built documents. Some of them are raw sourced
documents. Some of them are raw sourced that he inputs from previous chats. Some
of them are docs. Some of them are synthesis the AI puts together. Think of
it like most AI tools are like having a brilliant research assistant who reads your entire filing cabinet every time you ask a question and then it gives you a really great answer and then immediately forgets everything that it
figured out. Andre Carpathy's wiki is
figured out. Andre Carpathy's wiki is like that same assistant keeping a running set of notes that are organized and cross-referenced and updated so that each question builds on the last one
instead of starting over. I think that idea of being able to build on your learnings is why 41,000 people jumped on it and bookmarked the post. Right? It's
not because the folders and the text files themselves are exciting. It's
because an AI that builds understanding over time instead of starting from zero is something we're all hungry for right now. It's the thing everyone's been
now. It's the thing everyone's been waiting for. It's what made people
waiting for. It's what made people really excited about OpenBrain. But
there's a catch that almost nobody in these 41,000 bookmarks is thinking about yet. Every time the AI turns a raw
yet. Every time the AI turns a raw source into a wiki page, it is making editorial decisions. It's making
editorial decisions. It's making decisions about how to frame the connections between ideas that may be right or wrong, but those are the AI's choices, right? They're not the human's
choices, right? They're not the human's choices. It's making synthesis choices
choices. It's making synthesis choices somewhat independently of what you may or may not think. And so important nuance could get dropped that m might matter a few months from now and you would literally never know. You wouldn't
know what's missing because the wiki reads so cleanly. This is sort of the same trap as dashboards and analytics. A
dashboard is so much easier to read than a spreadsheet, but it is a condensation of data, right? It can hide exactly the thing you need to see because it just shows only the thing that it thinks you
want to see in the moment. And to his credit, Andre Carpath has been very honest about this, right? His
architecture keeps all the raw sources untouched in their own folders. So he
can always go back to the originals, which is a really smart design. But to
be honest, most people building on his pattern are not going to maintain the discipline to go back to raw sources. In
practice, I think the wiki will become the thing that is trusted as this open source project rolls out. And the source of truth is quietly going to shift from
raw material to an AI summary of that material, which may be correct 80% of the time. maybe 90% of the time, but
the time. maybe 90% of the time, but where it misses something or where it frames something slightly wrong, there are going to be issues that arise and that errors will now be baked into our
understanding if that is the only way we're approaching memory. And you won't get in the habit of questioning it because the whole premise of this when I started this video is that we are lazy people. It's really nice to have a wiki
people. It's really nice to have a wiki where you can just chuck stuff in and it sort of automatically organizes, learns, and comes back with written artifacts.
And so this is where open brain enters the picture and things get really really interesting. Every knowledge system with
interesting. Every knowledge system with an AI at its core has to answer one question. When does the AI do the hard
question. When does the AI do the hard thinking? Is it when information comes
thinking? Is it when information comes in or is it when you ask about that information you got to pick that's the fork everything else follows from that.
Carpathy's wiki is a right time system.
So when a new source arrives like an article, a paper, a set of notes, the AI does not just store it. The AI actually actively works against it. It reads the
source. It extracts what matters and it
source. It extracts what matters and it writes that understanding into the wiki.
It will update topic pages for you. It
will write relevant summaries for you.
You get the idea, right? It's going to actively work to add links between related ideas, develop concepts, note where new data contradicts something
that was filed last week. It will do a lot of that thinking at input, right?
The hard work then happens one time at the beginning, the moment the information comes in the door. After
that, you can browse the wiki and get pre-built understanding without the AI doing virtually any work at all. It's
just retrieving. Open brain is different. It is a query time system.
different. It is a query time system.
When new information arrives, OpenBrain is designed to store it faithfully. It
tags it. It categorizes it. It makes it searchable. But we're not assuming that
searchable. But we're not assuming that you need to synthesize that information yet. Nobody's synthesizing. Nobody's
yet. Nobody's synthesizing. Nobody's
doing work. The data is sitting in structured tables waiting. When you or your AI agent asks a question, that's when the AI goes to work. It reads the
relevant entries. at that point at query
relevant entries. at that point at query time. It does the thinking fresh and it
time. It does the thinking fresh and it produces an answer on the fly. So the
hard work happens at the moment you need it, not before you need it. So think of it this way. Carpathy's wiki is like a study guide that a really good tutor writes for you as you learn the subject.
Every time you cover new material, the tutor will update the guide for you so you don't get lost along the way. The
tutor will add new sections, revise old sections, connect ideas across chapters, and really help you dig in so that when exam day comes, you just read the guide, and you're good to go. Which is exactly
kind of how that wiki is supposed to work. Ideally, the thinking has been
work. Ideally, the thinking has been done for you. The tutor has prepared everything so perfectly, you can't fail.
Open brain is like a perfectly organized filing cabinet with a brilliant librarian standing next to that filing cabinet for you. Every document is filed. It's indexed. It's searchable. So
filed. It's indexed. It's searchable. So
that when you need an answer, the librarian can very quickly pull the relevant file, read through that relevant file for you, and then pinpoint exactly what you need to find in that
file. It will tell you what you're
file. It will tell you what you're looking for. The filing is really clean
looking for. The filing is really clean and pristine, and that enables the metaphorical librarian to do the thinking fresh in a very efficient way every time you ask, so you get exactly
the synthesis you're looking for. To be
honest with you, I'm not here to compare and contrast and give you an easy winner. Study guides and filing cabinets
winner. Study guides and filing cabinets can both be useful, but they're good at different things, and I don't want them compared inaccurately. And that's really
compared inaccurately. And that's really important. So, why does this matter for
important. So, why does this matter for you? This is the way to think about it
you? This is the way to think about it that's not about architecture. If you
only store stuff, your AI has to figure out what it all means every time you ask. You've been feeding it articles,
ask. You've been feeding it articles, meeting notes, and research for months and months and months and months. You
ask a question that requires connecting a bunch of different sources together.
And the AI has to go and burn tokens. It
has to find those sources. It has to make sense of them. It has to read them.
It has to think through them. It has to understand what's going on, figure out how they relate together. And
ultimately, it has to produce a synthesis that actually works from scratch. And it has to do that every
scratch. And it has to do that every single time. Nothing has been pre-built.
single time. Nothing has been pre-built.
Now, here's the other side. If you only build a wiki, your AI can read the summary, but it cannot do anything precise with the raw data underneath.
You want to pull every deal over $50,000 from the last quarter. You want to filter all your meetings by client name.
You want to have three different AI tools that query your knowledge base at the same time. A folder of text files cannot answer complex questions like that. The understanding is there in
that. The understanding is there in synthesized form, but the detailed structured data to make meaningful decisions just isn't there. And it isn't
there by design. It's just not going to be there. In addition, if you have three
be there. In addition, if you have three or more agents, that's just going to break when they're all trying to write Markdown files at once. The wiki
structure presupposes a single agent working for you that just writes in one place. Whereas the open brain structure
place. Whereas the open brain structure assumes you may want to hook in multiple agents at multiple points to contribute to or pull from a structured database.
Let's move on from structured data to talk about a different kind of challenge with AI. It is difficult right now to
with AI. It is difficult right now to actually trace how an AI learns or improves over time when there is no memory architecture under the system.
And I want to talk about a distinction between remembering detailed facts which open brain is designed to do and remembering narrative or synthesis which
the wiki is designed to do. And I want to help you understand how that plays out for a team because it's really important to understand that our storage architectures shape the futures that we
are unlocking for teams because that we're effectively choosing a context layer that you need to make sense of, use, input information into, believe,
trust, and depend on for decisions. The
stakes could not be higher. Most
organizations are generating enormous volumes of AI touched knowledge right now. We're generating meeting summaries.
now. We're generating meeting summaries.
We're generating strategy documents touched by AI. We're generating research outputs. We're generating Slack threads.
outputs. We're generating Slack threads.
And almost all of it is write once, read never because nobody is maintaining any of it. Nobody is synthesizing across any
of it. Nobody is synthesizing across any of those documents. Nobody is flagging that the Q2 strategy deck contradicts what the CEO said in last week's all hands. Your company's AI generated
hands. Your company's AI generated knowledge right now is either a compounding asset or it's just a growing pile of noise. And so the choice between the two memory structures here is a lot
more than a design decision. It's
actually the thing that most teams are making by accident that determines how reliable their northstar compass in product decisioning is. And the subtlety that matters here is that sometimes
contradictions are the most valuable thing in your knowledge base. And one of the things that you worry about is that you're going to lose the distinctions that you need to make good decisions in a wiki format. So engineering might
think the timeline for the build is 12 weeks and sales promise the client 8.
And something like a smart wiki might resolve that contradiction into one coherent narrative rather than flagging that you have a fundamental misalignment. And that is a strategic
misalignment. And that is a strategic signal in the system that you would not want to synthesize across with an estimate of 10 weeks. The gap between what engineering knows and what sales
promised is exactly the problem your leadership would need to see in that situation. A database that stores both
situation. A database that stores both views without resolving them preserves that tension and a well-meaning wiki might smooth those all away. So those
are some of the structural differences.
But if we go past the structural differences in these two memory systems, the open brain system and the wiki system, I want to talk a little bit about the job that the AI does in each
system and why it's important to name the AI job description really clearly.
One of the sharpest practical differences between these approaches is what the AI will spend its time doing.
And you need to decide like where do you want to invest in your AI. In Karpathy's
system, the AI is primarily a writer.
The job is to maintain a document. And
when you add a new source, you have to write to that, right? You have to read the raw material, synthesize it, write what you think about it. Update wiki
pages, connect new links, make sense of it, add concept explanations, cross reference it, create an index. There's a
ton to do. It's effectively doing editorial work. It's making judgment
editorial work. It's making judgment calls about what's important, about what connects to what, and where those contradictions might lie. Whereas in
open brain, we think of the AI as primarily a reader. Its job is to answer questions by pulling from the structured data and when you or an agent will ask something, the AI will just search the
database that has been carefully read and carefully organized, read the relevant entries and come back with a precise, fresh synthesis based on all of
the available data. So effectively, it is doing the analytical work on the fly, but it's able to produce more detailed results because all of the detail is immediately available in the database.
And so those different job descriptions have real consequences. When the AI is a writer, you interact with it intensively when new information comes in. Is that a job that you want to do? Do you want to
interact with it a lot when the new information comes in? it does adding a single research paper trigger updates across a dozen wiki pages and is that something you're comfortable doing as you think through and and figure out the
connections? It's a somewhat heavy
connections? It's a somewhat heavy operation on the front end, but afterward you end up getting answers that are very cheap because all of your thinking is captured in that wiki. The
thinking has been done. When the AI is more of a reader, as in open brain, what you get is adding new information is lazy and cheap. That's sort of why I did
it because I'm a lazy person and I want my stuff autocatategorized as cheaply and easily as possible. We just write a row, we tag it and we're done. The heavy
operation is when you ask a question because the AI has to reconstruct understanding from the data each time.
So simple lookups can be fast and complex lookups will take time as the AI does deep synthesis because it's actually interrogating the raw data.
That cost is going to recur every time if you ask something similar. But on the other hand, you are not going to lose detail if you need to get into the grounds and really understand what is going on. The difference between these
going on. The difference between these approaches raises a question that that I think most of us aren't asking yet.
Whose understanding matters here? When
your AI maintains a wiki, what you are effectively saying is that when a colleague asks you about a topic, you are willing to check the wiki and trust what the AI says before answering. And
you are trusting that the AI's capture of your understanding or your thinking or the article you gave it is good enough to share with your colleagues as yours. Whereas if you have an open brain
yours. Whereas if you have an open brain style database, the providence is very clear. These are facts from identified
clear. These are facts from identified sources with timestamps. You can trace any claim back to where it came from.
What you know, you know, and you know why you know it. And you can come back with a fair bit of authority and say, "I'm not just trusting the AI's ability to synthesize information. I'm actually
saying this is the raw material I got.
This is the facts that I'm basing this on, and this is a considered opinion based on a query across all of the data that I've collected over the last few months or the last few weeks or whatever it is for you." That is a deeper and
more consequential kind of trust. It
also means the instructions you give the AI that tells it how to organize your wiki becomes the highest leverage document in the whole system. Because if
you're building a wiki, I want you to think about this for a second. If you're
building a wiki, you basically are telling in one markdown file the AI to organize and synthesize in a way that's profoundly useful to you and profoundly
accurate. and you're betting your career
accurate. and you're betting your career that it will get it right or you're going to invest time on every single ingest to make sure it's correct and to doublech checkck it. Most people will underinvest in that and the wiki will be
worse than it should be as a result. Not
because it can't be good, but because we're lazy. If we were to talk about
we're lazy. If we were to talk about what each approach is good at and where the advantages are, I would say that Carpathy's wiki wins when you're deep in research mode, when you're reading 10 papers on a topic over a couple of
weeks, which sounds a lot like what Andre does, right? like it's written for him. You could tell, right? And each one
him. You could tell, right? And each one might build on. It might contradict the last. It's a thinking person's tool. The
last. It's a thinking person's tool. The
wiki approach is going to be dramatically better in that situation because by paper five, you're continuing to wrestle with it. You're continuing to read. You're giving input. The wiki
read. You're giving input. The wiki
contains a synthesis of the first four.
You've read all of the primary sources.
You have them in your head as well. And
paper 5 can get integrated into that existing picture and help you evolve your thinking. contradictions get
your thinking. contradictions get flagged at the moment of ingest and you can see them really quickly. Cross
references get built automatically. It's
basically an academic researcher's dream. And so by paper 10, you have a
dream. And so by paper 10, you have a really rich navigable artifact that represents the current state of your understanding of a very difficult subject. It's sort of like notebook LM
subject. It's sort of like notebook LM on steroid. It's not just the current
on steroid. It's not just the current state of your files. It also wins because your personal knowledge evolves over months and you can actually see it grow. Right? So if you're thinking about
grow. Right? So if you're thinking about your health over months about self-improvement about competitive analysis for any domain where the value is in the connections between the
sources rather than in any single source alone then that's where Carpathy's approach is going to win right because you're really looking at how it can help
you understand a complex synthesis problem but open brain wins when you need precise structured operations across your knowledge base. If you want to ask, "Show me every meeting note from
Q1 where pricing was discussed," that's an open-brain type question. If you want to pull the three most recent competitor updates and compare them, that's an open-brain question. Or find all
open-brain question. Or find all actionable items assigned to me in the last two weeks, open brain. Again, these
are database queries, right? You are
digging in for specific facts. They
return exact filterable results. A
folder of text files can approximate this with some keyword search, but it's not going to be perfect, right? It's
going to miss stuff. it's going to break fast and it's not really what that whole wiki system was designed for, especially when you need to combine filters, sort by date, or work across hundreds of
entries. OpenBrain also wins for multi-
entries. OpenBrain also wins for multi- aent access when you have clouded code and chat GPT and cursor and a scheduled automation all working against the same data source at once, all needing to read from and write to the same knowledge
store at the same time. Well, you need a database that handles simultaneous access in that situation, not a directory of files where two agents editing the same page creates a complete mess. And OpenBrain wins on volume, too,
mess. And OpenBrain wins on volume, too, right? OpenBrain can handle thousands of
right? OpenBrain can handle thousands of entries across dozens of categories with search, with metadata, with relational queries. And and Carpathy absolutely
queries. And and Carpathy absolutely acknowledges this. It works best at
acknowledges this. It works best at roughly 100 to 10,000 high signal documents. It is not corporate level
documents. It is not corporate level memory. And I hear corporations saying
memory. And I hear corporations saying we should just use this for for our company level context layer. And that
will not work. And at the upper end, 10,000 documents, you already need extra search tooling just to stay manageable.
And so when you're dealing with thousands of contacts and transactions and events and tasks and documents on top of all of that, structured storage is the only sane option that scales. But
to be fair, we should look at where both systems break, right? Every system has a load where it starts to break. They just
tend to break in different ways. So as
I've called out, the wiki approach tends to break at scale. So, if you have a team that's using it where you are starting to hit that wiki structure from multiple directions, well, now the wiki doesn't know how to autooptimize, right?
If person A has an understanding that's evolving differently than person B or or agent A and agent B all have different approaches and they're trying to update the same wiki page. One, you have a conflict and that's going to be a
problem. But two, the wiki is going to
problem. But two, the wiki is going to look like a weird merge of these different approaches that doesn't reflect deep personal understanding.
Fundamentally, the semantic understanding that you're evolving with the wiki is designed for a world that's kind of like Andre's world where he's a researcher and he's thinking deeply about a problem and it's for him and
it's his evolving understanding with the agent. So for a solo practitioner, you
agent. So for a solo practitioner, you don't get issues here. But for a team, this becomes a really serious problem.
If your knowledge changes daily, if you are an operation where you have project status, you have competitive positioning, you have live deal flow, the cost of reynthesizing the wiki every time something comes in becomes really
punishing because every change potentially ripples across multiple pages in ways that you can't control.
And it should not, right? It should just be another data point in the row. And
so, think of the wiki system as being optimized for like papers and articles speed, not Slack message and ticket update speed. And that's the thing that
update speed. And that's the thing that worries me the most is that people don't recognize that a particular knowledge system is designed to work at a particular speed of business. And if you don't think about it that way, you might
implement the wrong one. A neglected
database has gaps, but the old facts are still true. as opposed to a wiki. A
still true. as opposed to a wiki. A
neglected wiki tends to drift because old syntheses become increasingly wrong as new information is not integrated, but they still read with the confidence that comes from well-written pros. And
so database staleness can look like ignorance. It can look like you're
ignorance. It can look like you're missing something. I forgot to put stuff
missing something. I forgot to put stuff into my open brain. But wiki staleness looks differently. It actually looks
looks differently. It actually looks like active misinformation because you don't know that you're wrong because the page reads like it knows what it's talking about because that is the entire purpose. It's supposed to synthesize
purpose. It's supposed to synthesize stuff and write confident pros that helps you understand a situation and you might not question the gap that you do not see. Now, let's get at some of the
not see. Now, let's get at some of the scale breakpoints for OpenBrain. And by
the way, yes, I am launching fixes for these because that's what we're all about with AI. We make things better over time. In the past, Open Brain has
over time. In the past, Open Brain has really cracked around deep synthesis quality. If you try to synthesize 15
quality. If you try to synthesize 15 different facts at once, the AI can do it, but it tends to do it in slightly unpredictable ways because it has no previous map of how that worked in the
past to do it well. It's essentially
searching the shelves of the database every single time from scratch. Now, the
answer is usually good because the AI is good, but it's rarely as good as a pre-built synthesis that had the time to integrate everything deliberately from the beginning. And that is something
the beginning. And that is something that we're addressing. Browsability is
another area that we can think about here. Open brain is deliberately
here. Open brain is deliberately headless. There's no artifact you open
headless. There's no artifact you open and wander through. And I built it that way because it gives you the flexibility to decide how you want to access it.
Now, the nice thing is it's very very easy to build the right head over the top. There are people who have added
top. There are people who have added Obsidian to OpenBrain. There's a plugin for that already. So, if that's something where you're like, I just I just can't browse the database, you're absolutely right. Just pick the plugin
absolutely right. Just pick the plugin of your choice and you can browse it.
whether that's Obsidian or something else. Here's another one we're building
else. Here's another one we're building to improve in the wiki. Contradictions
surface when new information comes in as long as your initial markdown file deliberately says look for contradiction because the AI is actively integrating against existing pages following your
prompt. But in a database environment,
prompt. But in a database environment, the contradiction might just sit silently in adjacent rows unless you specifically ask the right question to expose that contradiction. I'm building
a plugin that helps with that. If you
are interested in essentially running audits that check for contradictions in your data set, we're launching a plug-in that helps you use OpenBrain as a
contradiction surfacing tool. You can
actually build out and understand a map of the contradictions in your team or your org data sets really, really easily because you can look through the raw material and see it right away. Yes,
databases store facts. they're not
contradiction aware by default, but it's relatively easy in the age of AI to extend something like open brain and make it aware of those contradictions.
That's what I did. And I know I've spent a lot of time talking about differences, but one of the things I want to call out is that there are a lot of common principles that these systems share.
They might disagree on implementation details, but a lot of the underlying thesis or principles about AI and about data they agree on. They agree that you own the artifact, not the tool. So,
Carpathy's files are text in a folder you control. Open brains data is in a
you control. Open brains data is in a database you own. It's the same principle. Neither system hands your
principle. Neither system hands your knowledge to a platform that can repric or lock you in. Carpathy calls this file over app. I've called it building with
over app. I've called it building with no SAS middlemen. It's a very similar mindset. It's the same conviction at
mindset. It's the same conviction at root. In the age of AI, we should own
root. In the age of AI, we should own our own context layer. Right? There
should not be someone who is out there whom we are paying just to own our context layer. Also, in both systems,
context layer. Also, in both systems, the human's job is curation and questioning. We have to ask what sources
questioning. We have to ask what sources go in. We have to figure out what
go in. We have to figure out what questions to ask. We humans retain a big job in both cases. There's no substitute for thinking carefully about how to
organize your personal context layer.
And yes, the AI has lots of work to do.
It has to understand the facts that you put in an open brain. It has to be able to effectively synthesize on the wiki side. It's effectively a similar
side. It's effectively a similar division of labor. It's just timing that work differently because on the carpathy wiki approach, it's doing all of that up front and on the open brain approach, it's doing all of that at query time
when you ask. In both cases, memory compounds through intentional structure, not just through random accumulation.
The only difference is how that structure is positioned and where that structure lives. So it might live in a
structure lives. So it might live in a wiki in Karpathy's case and it lives in a SQL database in OpenBrain's case. But
in both cases, the structure is intentionally framed to enable a certain kind of connection to occur. And so for wiki work where you might want the
connections to be between documents, that makes a ton of sense, right? You
want all of the documents there. You
want the AI thinking it through. And
that's an intentional structure. Whereas
for OpenBrain, the intentional structure is a SQL database that you know can scale and it is designed to hold operational facts and make sure that they are in a neat place where you can reason against them and get audit ready
results from day one. Both systems
assume that the primary user of the knowledge base isn't you reading in a browser. It's an AI agent working on
browser. It's an AI agent working on your behalf. And I think increasingly
your behalf. And I think increasingly that's going to be the assumption of all of these memory systems. Human readability is a bonus. Asian
accessibility is actually the requirement. So now we come to what I've
requirement. So now we come to what I've built because let's be honest, we want a mature system that gives us the strength of both approaches. It's not either of those alone. And so the specific
those alone. And so the specific architecture that I'm putting together and proposing is the next major open brain extension.
You want to keep openrain as your permanent store. Don't change that. It's
permanent store. Don't change that. It's
a great spot for fax. Everything goes in there. That's fantastic. Every meeting
there. That's fantastic. Every meeting
note, every article clip, every research finding, every task, every contact, it's all tagged. It's all searchable. It's
all tagged. It's all searchable. It's
all queryable. That makes sense. That is
your durable memory layer right there.
And it can handle high volumes. It can
handle precise query. It can recall across multiple domains in your life. It
can be the source of truth. And a wiki layer can act as a compiled view on demand. And and so I'm launching a new
demand. And and so I'm launching a new process, a new plug-in where a compilation agent can run on a schedule daily, weekly, on demand. And the agent can read from open brain structured
data. Effectively, it becomes an open
data. Effectively, it becomes an open brain graph. It can synthesize across
brain graph. It can synthesize across entries. It can produce wiki pages on
entries. It can produce wiki pages on demand. It can produce topic summaries.
demand. It can produce topic summaries.
All driven by the idea that if you form a graph of your knowledge base, you can get the advantages of the wiki approach with the solidity and the factuality
that comes from an open-brain SQL database. And so these pages can be
database. And so these pages can be generated artifacts for you. Think of
them like a daily briefing that a really good chief of staff writes by reading everything in your files and distilling it into something you can browse. The
graph approach allows you to follow Karpathy's patterns for synthesis to cross reference to link related topics to flag contradictions to maintain an
evolving synthesis but it works from structured data not raw files and that means it can do things Carpathy's ingest can't like filter entries by date or category before synthesizing. It can
wait by confidence. It can exclude outdated items. In other words, the synthesis is richer because the underlying data is more detailed. The
wiki pages are an easy to read layer and you can browse them in Obsidian. You can
browse them in a note app, but they're all powered by a pre-built context graph that lives on your structured data that would not exist without your structured data. They end up being your hot
data. They end up being your hot reference for when you're actively working on a topic. And the structured data ends up being like the raw files that Carpathy uses when he wants to look
at the raw material in his wiki. But
unlike the raw files, these are easily queryable and organized in a SQL database. So you can scale them in a way
database. So you can scale them in a way that you can't with raw files. You do
not have a 10,000 file limit with OpenBrain in the same way. So the
database stays the single source of truth. New information always goes into
truth. New information always goes into the core SQL open brain first. The wiki
is never edited directly and this prevents the error compounding problem that several commenters on Carpathy's post flagged. If the AI writes something
post flagged. If the AI writes something slightly wrong into the wiki and it stays there, the next answer will build on that wrong thing and you start to get drift and errors start to accumulate.
Whereas in the hybrid model that I'm proposing with OpenBrain, the database is always authoritative. The wiki is generated from a graph built off of that
database. So if the wiki has an error,
database. So if the wiki has an error, well, you fix the source data and you regenerate. You're not dependent on the
regenerate. You're not dependent on the wiki as a source of truth. The wiki
never drifts from reality because it's always rebuilt from ground reality in the SQL database. In open brain terms, this is like a recipe. It's a composable workflow that reads from the database
and produces an output based on a graph.
A wiki compiler recipe can query relevant tables, synthesize pages through AI, and effectively develop a network of relationships and write that output to a wiki directory. And if
you're wondering, yes, it can run on an automated schedule. It can get better
automated schedule. It can get better every cycle because the underlying data hopefully, if you're committing to it, grew since last time. It becomes a compounding loop as long as you are good
at putting data in. And so what you end up with is OpenBrain for structured storage and Asian access and a Carpathy style wiki over the top for compiled understanding and human browsability.
The database ends up feeding the wiki and the wiki never contradicts the database. You can query either one
database. You can query either one depending on what you need, whether it's a precise fact or a synthesized narrative, and you can decide which you want to go for depending on the kind of
problem that you're solving. And so just to be really really blunt about which of these because I know I'm going to get asked which do I build? If you are going deep on a single research topic, if you're a solo user, if you don't need
precise queries, if you don't need multi- aent access, if you want to think by reading and by browsing, you want something running in 30 minutes with zero infrastructure. In those
zero infrastructure. In those situations, then it absolutely makes sense to use straight up Carpathy's wiki that he posted on the GitHub because the AI will build the whole system for you
and it's designed for exactly that kind of solo use case. But you should build with open brain if you need multiple AI tools accessing the same memory. If you
are assuming that you have a team working with this information, if you're capturing high volume information across many categories, if that information is not necessarily narrative based, if it's numbersbased,
if you need structured queries, if you're building automated agent workflows off of this, if you're thinking about this as infrastructure that lasts for a long time and needs to
scale and not just for a single project.
In a sense, a lot of what the wiki feels like is a better, cooler version of Notebook LM, which is an amazing tool, but not a tool that you can use for an
entire team. And so, right now, I tend
entire team. And so, right now, I tend to say have it both ways. Have your open brain running, and if you want a browsable presynthesized understanding layer, just grab the graph plugin and
add that over the top. And then neither replaces the other, and you get the benefits of both. None of this is to say that Andre Carpathy isn't right about what he built. He built a phenomenal tool for himself and for other researchers in a similar position. And
regardless of which system you end up going with, there are two ideas from Karpathy's post that are worth adopting right away. The idea file as a
right away. The idea file as a publishing format is one of those. And
one of those is really simple. It's the
way he shared it. The idea file is his publishing format. Carpathy didn't ship
publishing format. Carpathy didn't ship a tool. He published a high-level
a tool. He published a high-level description of an idea that was designed to be pasted into an AI agent that would build the specifics with you. This is
what I have been saying when I tell you to take my YouTube transcript and feed it to an AI. It is a genuinely new way to share technical knowledge. It is a great blueprint for an AI to execute.
And I think we're going to see more of it because it's much simpler than just having to give an exhaustive step by step that a human has to follow. It ends
up respecting the reader's agency because they can give their own commentary on the idea and then them and the agent can decide the details together while giving them a proven pattern to start from. And yes, if
you're wondering, you can absolutely take the transcript from this YouTube video and get started on your own memory project as we've been going through this video together. Just plug it into your
video together. Just plug it into your agent and go. But the deepest insight here is that Carpathy is moving the AI from Oracle to maintainer. The role AI plays is starting to change. Most of us
have treated AI as something you ask questions to. Whereas Karpathy correctly
questions to. Whereas Karpathy correctly treats it as something that has an ongoing job, maintaining a knowledge artifact that gets better over time. The
AI isn't here for magical pie in the sky one-off answers from the clouds. It's
here for building sustained work that compounds. And the question that we're
compounds. And the question that we're all facing is just is this the right interface for that maintenance role.
Right? I don't want to lose the fact that underneath that there is a profound insight here about moving from an answer engine mindset to moving to a mindset
where AI is a maintainer of thinking systems that allow you to think deliberately and do your work better. I
think that's a profound insight because it allows us to be the ones who curate, who think, who select, who explore, and it allows the AI to support us, right?
As we ask the right questions, the AI can help us by doing so much of the grunt work. And isn't that what we
grunt work. And isn't that what we wanted in the first place? Didn't we
want that division of labor in the AI dream world to be the AI doing more of the grunt work and human judgment being relevant? That's the dream. What Andre
relevant? That's the dream. What Andre
Karpathy is describing is one way to get there, especially if you are in a deep solo research project. And OpenBrain
describes another way to do the same thing. It's just focused on more
thing. It's just focused on more scalable structured data. And yes, you can have the best of both worlds because we can build a graph over the top of OpenBrain. This is exactly why I built
OpenBrain. This is exactly why I built it Extensible because I knew that we would have more stuff coming out around memory in 2026 and I wanted to build a foundation we could build on. So here we
are. It's our first major test and we
are. It's our first major test and we can build something over the top that allows us to have the best of this wiki approach as well as the best of the structure data that open brain gives us.
Ultimately, I think the lesson that we get from Karpathy's wiki is that we need to become thinkers about how we want our
memory and our context layers to work in order to be effective builders of agents and effective partners with AI over the second half of 2026 and into 2027. None
of what I am describing excuses us from doing that thinking. In fact, it's the opposite. What I've been spending time
opposite. What I've been spending time telling you in this video is that there is no substitute for making really clear distinctions and really clear decisions
about the way you want your knowledge structured. Whether that's just you in
structured. Whether that's just you in your room with a laptop and it's your personal knowledge base or whether it's for your team or whether it's for your org. It is up to you to say I want
org. It is up to you to say I want structured data because I know that I need to query against structured data and get reliable results above 10,000 artifacts. Or it's up to you to say you
artifacts. Or it's up to you to say you know what I want the best of both world.
There's going to be some stuff where I'm going to actually want to query with multiple agents and get structured results for three different reports at the same time. But over the top of that, I want a graph database that allows me
to think in connections between materials. That would be a little bit
materials. That would be a little bit more difficult to do if I was just querying structured data by itself. It's
up to you. It is not up to me. We all
have to wrestle with this. And if you are an engineer thinking about this or a product manager thinking about this in an org, you cannot substitute for that level of thoughtfulness. I'm sorry. You
got to do the thinking. And so I hope this video has helped give you the tools to make that decision clearly.
Loading video analysis...