Every Claude Code Memory System Compared (So You Don't Have To)
By Simon Scrapes
Summary
Topics Covered
- Context Roof: The Inevitable Limitation of AI Memory
- OpenClaw's Three-Layer Memory Architecture
- Mem Palace Uses an Ancient Memory Technique for Word-for-Word Recall
- One Brain to Rule All AI Tools
Full Transcript
Right now, if you search for the best claw code memory system, you get absolutely buried. Mem zero, Carpathy's
absolutely buried. Mem zero, Carpathy's Obsidian Wiki, Open Claw and Hermes, Mem Palace, Light Rag, Clawude Mem, Claw Mem. It's getting ridiculous to keep up
Mem. It's getting ridiculous to keep up with it. But after spending hours
with it. But after spending hours researching it, these aren't competing tools. They're different ways of
tools. They're different ways of handling memory depending on your use case. And the reason I went deep on this
case. And the reason I went deep on this is because I've been building out my own agentic operating system. basically a
business brain that runs across everything I do. So, I needed to figure out what actually works, what scales, and what just adds complexity to my existing setup. So, in this video, I'll
existing setup. So, in this video, I'll walk through six levels of memory that build on top of each other, all the way to systems that work across every single AI tool that you use. So, by the end, you'll know exactly which one makes
sense for your use case and which ones you can completely ignore. But before we go ahead through the levels, one thing I want you to know now because it will make the rest of the video easier to follow is that every memory system is
answering exactly the same question.
When you give claw code a task, how does it pull the right context at the right time? So the difference I'm going to
time? So the difference I'm going to show you between each level is just two things. First is where your memory
things. First is where your memory lives. So it's going to be the storage
lives. So it's going to be the storage mechanism for your data and what the file structure looks like, whether it's marked down or stored as vectors, for example. And secondly is how Claude
example. And secondly is how Claude actually gets hold of that data, which is just the retrieval stage. So some
stuff's going to get copied into Claude's context automatically. Some
stuff is going to sit in a database and get searched exactly when it's needed.
Some stuff's going to live on your laptop. Some's going to live on the
laptop. Some's going to live on the cloud, etc., etc. So let's jump now into level one. So you can think of level one
level one. So you can think of level one as the native stuff that ships inside claw code. So most of you have already
claw code. So most of you have already heard of the claw.md file. Some of you have heard of memory.mmd, but very few people are actually using both together properly. So, let's do claw.md first.
properly. So, let's do claw.md first.
And as most of you know this one already, I'll keep it quick, but it's effectively a plain markdown file inside your project folder where you store your rules, stuff about your brand, things about how you work with clients, coding
style, and it's always loaded into every single session you pull up inside the terminal. You can think of it like a
terminal. You can think of it like a note that is just at the start of the context, almost like a system prompt that you're giving to Claude. So you can actually have this at the Claude project folder level to pass to all of your
Claude instances. You can have it
Claude instances. You can have it slightly lower at the root folder level if you have multiple projects which sit within the same folder. Or you can even have it in the individual project level for specific project instructions. The
lower down the folder structure you go, it still inherits the parent claude.mmd,
but any conflicting rules, it will always take the local claw.md as gospel for specific project level instructions.
But here's the thing that most people actually get wrong with their claude.md.
And I've seen this a lot, which is stuffing too much into the claude.md. So
putting in your full brand guide, your tone of voice document, your complete client list. Think about when Claude
client list. Think about when Claude actually starts a session every single time. It's going to burn through that
time. It's going to burn through that context window reading all of it. So
yes, it's important context, but only when you load it in at the right time.
And this is exactly the problem we're trying to solve called context ro. So
this is the inability for AI and LLM models to actually recall 100% of the information that we've loaded in as we increase the amount of context that's loaded in in the first place. So as a general rule of thumb, keep your claud
under 200 lines. And then if you've got more context like a brand voice document then you can put that in a separate external file that's referenced from the claw.md and therefore it will only pull
claw.md and therefore it will only pull it in at the right time when it needs it. So the second built-in memory system
it. So the second built-in memory system inside claw code already running on your machine is called automemory and that forms a memory.mmd file. So you can jump into an existing project folder to see
this best and what you can do is type in /memory.
You'll hit enter and you'll see there are a few different things. There can be imported memory if you're using a separate memory system and we'll come to separate memory systems. You've got project and user memory which are effectively your claude.md files at
different levels in the root folder.
We're going to open up the automemory folder. You'll see that we have on a per
folder. You'll see that we have on a per project basis effectively a memory structure with an me memory file. So if
I open up all these files inside the memory MD file, we effectively have an index of lots of different memories that claw code can actually reference and open up other files of. So it's not just
dumped a a huge amount of memories in here. What it's done is said on a
here. What it's done is said on a project basis, we've got a specific project level feedback and on a more general scale, we've got a bunch of different feedback for specific things
that it can refer back to kept as separate documents. It's following
separate documents. It's following exactly what we've talked about, which is don't just dump more than 200 lines in a single file. Always reference a separate file and treat the memory MD or
the claw.md as an index and then point
the claw.md as an index and then point to another file. So for example, we've got here a file that's feedback on project structure. So level two or three
project structure. So level two or three projects go under project/briefs.
And you can see separately we've got feedback project structure.md which is a specific bit of feedback that only needs to be loaded in when Claude understands that we're looking at project structures
table of contents of different bits of feedback that you've given it over time or it's understood from the way that you work. Hence this file here on a specific
work. Hence this file here on a specific project basis. And to demonstrate this
project basis. And to demonstrate this further, I'll go to an empty folder. And
you'll see that there's no memory. MD
file or claude.md file inside that. So
if I run the /memory command and I hit autoopen memory folder, then you can see that because it's done on a project level basis. There's no memory
level basis. There's no memory consolidation there. So you don't
consolidation there. So you don't actually have to create that memory.md
file. It's just quietly taking notes on your tasks and on your feedback in the background. And then it's created all
background. And then it's created all those files as an index for me. So
claude.md and memory. IMD is what Claude has today. And it's also very clear that
has today. And it's also very clear that Anthropic is trying to tackle this problem themselves. And that's evident
problem themselves. And that's evident because when the Claude code source accidentally leaked a couple of weeks back, people found references to a system or framework called Chyros and had a look through the code. And it's an
unreleased always on demon that watches your project continuously, decides what's worth remembering, and then consolidates old notes while you sleep.
So it's not in any public build yet, but for sure Anthropic in the background is working on features to actually solve this context rock problem and improve the native memory system. So it's only
going to get better from this point. So
level one, the setup is already done for you. It's in the background. It's
you. It's in the background. It's
completely free to use. You just need to be able to actually use them properly.
So you might have noticed this if you keep repeating information to Claude that you think it should already know.
And that is where level two comes in. So
level two improves the reliability of feeding those memory files in at the right time as well as improves the in-built structure inside cla MD. I
found an article by a guy called John Connelly who's taken a concept by Pavl Hurin a huge name in tech who wrote this on the 18th of Feb. How to give claude code memory paste this into your
projects claude. MD and bearing in mind
projects claude. MD and bearing in mind this was before it was announced that Claude had this automemory feature but what it now does is effectively build on that structure of Claude and then when
we combine it with John's we've got the ability to actually inject that automatically at the start of every single session. So basically what we're
single session. So basically what we're doing is we're adding in to the claude.md a big prompt all about how to
claude.md a big prompt all about how to manage memory and the rules in managing memory too structured memory system rooted atclaude/memory.
We already saw our automatic memory system inside Claude, but instead we've got a structure here. We've got
memory.mmd general.mmd which stores all cross project facts preferences environment setups but then we've also got domainspecific knowledge where we effectively have domain/topic
and each topic can have its own markdown file with domain specific knowledge one file per topic and the same with tools we can have individual tools like slack.md with a specific tool config
slack.md with a specific tool config workaround edge cases for that specific tool and what this basically does is we're going to add this to our claw md and we're going to go into plan mode.
We're effectively going to get it to set up a memory structure for our project that we then can automatically inject into every single session. Now, this
works using a claude code session start hook to ensure that when you open up that new terminal or even spin-off sub aents, it automatically loads the memory or the memory MD index into session
files. And if you've not used a hook
files. And if you've not used a hook before, it does exactly that. It's just
a little script that basically says at specific moments like session start I'm going to inject that context into claude and we're not loading in the full context. We're just loading in the
context. We're just loading in the memory.mmd file which is that index if
memory.mmd file which is that index if you remember. And the second most
you remember. And the second most exciting thing about this is the potential for sharing this between teammates. So imagine if you could sync
teammates. So imagine if you could sync those domain files that we talked about.
So we have different products and different projects between your teammates. So if somebody's working on a
teammates. So if somebody's working on a different project, they can save memory files to do with that context in the shared folder that we all can access inside our team. And the idea is it's
specific granular detail that claw code actually needs to be useful for that domain or that tool. Now let's get it set up to show you exactly how it works.
So what we're going to do is go to the link that I'll leave down below in the description under level one and we're going to copy this whole markdown section for part one. So basically
adding this to claw.md. What it's going to do is create a global memory directory structure which you'll probably have already based on claude's auto memory. We're going to update the
auto memory. We're going to update the claude.md file. We're going to
claude.md file. We're going to initialize project memory.mmd files. And
what it's basically looking out for is when I say reorganize memory, it's going to read all the memory files. Remove the
duplicates and outdated entries. Merge
our entries that belong together. So
basically consolidating our memories, split files that cover too many topics, restore entries by date, and update the memory. MD index. So every so often
memory. MD index. So every so often we're effectively going to be reorganizing our memories. But first we just need to copy this into a prompt.
We're not actually putting this directly in our claude.md file. What we are doing is opening up a new terminal window inside an existing project or an empty project. I've got my Aenttos demo
project. I've got my Aenttos demo project here. I'm going to switch mode
project here. I'm going to switch mode using shift and tab twice to put it on plan mode. And I'm going to paste in
plan mode. And I'm going to paste in that entire prompt. What it's going to effectively do then is plan out a structure for our memory files and what to add to our Claude. MD file based on
what already exists and based on what already exists in our memory files. And
once Claude has built the plan for part one, which was the one we copied in above, we're going to switch into plan mode again and run reorganize memory.
So, we're planning out this first. We're
going to build it and then we're going to run reorganize memory with a plan to basically tidy up our memory files and see what it's built out of our previous project. So super simple. We're just
project. So super simple. We're just
going to wait for that to execute, check the plan, build it, and then execute the phrase reorganize memory to actually then push through all those changes. And
you can see right now it's exploring the memory system state before it comes back to us with a plan. And it's come back to us with a plan structured persistent memory management system, setting up a global structured memory system for claw
code so that knowledge persists across sessions and projects. So currently our root claw.md exists but is empty. Claude
root claw.md exists but is empty. Claude
memory does not exist and only one of our 28 projects has a memory/memory.mmd
file. So you can see this has checked all the way up the parent tree to understand what exists at the highest level that will be inherited downwards.
And the current project because it's a demo has an empty memory/ directory. No
memory md. So it's going to create a global memory directory structure with general tools domain. It's going to update our claw.md
with all five sections. Memory
management, global memory, repo memory, auto init, which basically autocreates memory.mmd on a session start. And it's
memory.mmd on a session start. And it's
effectively going to go through all of our files and create that memory system and structure that we can then use to actually reorganize memory and basically set up a bunch of different domain and
tool files. So, we're going to yes,
tool files. So, we're going to yes, clear context and bypass permissions and get it to execute on that. Then after
that we will set up the ability to actually auto inject using the hook so that every time we load up a session it auto injects that memory index. So it's
now built all of that. So it's created the global memory structure and you can actually commandclick into any of these files and we can see sync rules between
global memories and project memories the structure of the file and it's got the general andd the tools the domains. We
can also see that got a uh cross project conventions like a general MD file and it's also done project memory.mmd file.
So on a specific project basis for all the projects we've done 28 projects it's created a specific memory MD file or consolidated that for that specific project and you saw before that I didn't
have any for previous projects except for the Aentic OS in which there was one memory file. So what we need to do do
memory file. So what we need to do do now is just run reorganize memory. We'll
make sure it's on plan mode and not bypass permissions mode. And it's going to come up with a plan to reorganize the memory and therefore it will start filling these memory MD and the claude.mmd files with the relevant
claude.mmd files with the relevant folder structure outlined here. So we
should see the general MD start to fill out the tools and the specific domain structure and it will tell us what it's actually going to create and then it will actually start creating that.
Whilst that is happening, let's take a look at step two which is actually introducing the hook. So this hook basically injects your project memory and the global memory. MD index, if you
remember that just points to the right memory files before the first tool call of every session. So this means that when you spin up an agent or a sub agent, it's going to basically throw in
that index so it knows what memories it has access to. And we're just going to copy this file here. And in summary, tells it to create the settings.json
JSON hook that says when it's uh pre-tool use, make sure to run the pre-toolmemory.sh
pre-toolmemory.sh file and it's going to create the pre-toolmemory.sh file, which is just
pre-toolmemory.sh file, which is just the script to effectively load in the memory files that we've just created.
And then it's going to go and update our claude.md to say instead of reading the
claude.md to say instead of reading the claude/memory/memory md at session start, it's going to auto inject it this time. So it's not relying on Claude actually reading it and taking
action. It's going to auto inject it
action. It's going to auto inject it automatically because this is a hook.
And you can see after 5 days of usage for John specifically the number of memory files had increased significantly. So up 90% and
significantly. So up 90% and significantly more lines but it hasn't increased the memory. MD or general MD by a huge amount because those are just index files pointing to the correct
tools, the correct domains and the correct cycle files. So, I've asked it to reorganize the memory. And this is really brilliant. I'm just reading
really brilliant. I'm just reading through this now. It's basically looked through my Aenttos folder and said it's
found 11 or 12, sorry, empty memory files, like daily memory files. So,
there's an empty template, empty template, two empty sessions. So, it's
saying actually, we just need to delete those because what's the point in having them? It's a waste of loading them in.
them? It's a waste of loading them in.
Some identical first run onboarding empties. A bunch of empties there. And
empties. A bunch of empties there. And
then part of phase two, it's basically going to trim the empty sessions. Phase
three, resolve stale open threads. So
it's actually read through the specific threads in our daily memories and basically going to go and find exactly what happened after those actions were taken to make sure that they're updated.
And you can see it does this for a few different daily memory files. And then
it's going to go through and actually add some cross references to all of these different projects that we've done to make sure that they link together correctly. It's going to update our
correctly. It's going to update our general MD with all of these different projects that we've been working on, update the indexes inside the memory.mmd
file, and then go through a verification phase to make sure that that all happens. So, we're basically going to
happens. So, we're basically going to run yes, clear context, and bypass permissions. It's going to make all
permissions. It's going to make all those changes for us. And from this point onwards, we have a memory system that has a clear structure at our global directory. So, all verifications pass.
directory. So, all verifications pass.
It deleted the 13 files, trimmed six sessions, finished the nine open threads, added 10 cross references, three patterns were added to general.mmd.
general.mmd.
It has remaining daily files of 12 with meaningful content, so it's got rid of the rubbish, and it's cleared all the empty memory directories inside the client's folders. So now we're going to
client's folders. So now we're going to run the second prompt. And again, we're going to run it on plan mode, and then we're going to execute the plan to add to the settings.json. So that injects this every time we start a new session.
Okay. So it's come up with a plan. It's
going to basically register the hook in settings.json, update our global memory
settings.json, update our global memory and claw.md so that it effectively has
and claw.md so that it effectively has all of this knowledge that we're building into it. And that means everything will be executed automatically when a session starts. So
for a lot of you, this will be enough.
But the moment you start running this for more than a few weeks, you're going to hit a couple of problems and your files are starting to get really messy.
And the keyword search that works for a small amount of files using this system is going to stop working as we grow and start to scale. So this is where level three comes in. And this is where we
start to get a bit more serious about how memory is actually organized and how it is searchable. But before we move on to that, YouTube tells me that 97% of you are watching this video right now
haven't subscribed to the channel. So do
me a quick favor and hit the subscribe button below if you've made it this far.
So only add to level two if these two things are true. One, you've been using Claude code for more than a month and your memory files have grown past a handful of files. And two, if you've
ever typed a question to Claude where you know the answer is somewhere in your notes, but Claude couldn't actually find it, then it might be a sign to move on to level three. So at level two, we've got Claude loading memory more reliable.
But as you start to scale this, when you're running it for real for months across multiple clients or projects, the structure eventually stops scaling. So
you end up with a giant general MD file that can't be read efficiently. And
because it's summarizing key topics and not taking words verbatim, the keyword search starts falling apart as well. So
level 3 is designed to solve both. And
the template for level 3 comes from an AI agent that you're probably familiar with called openclaw. So the open claw file structure is effectively from that standalone open claw agent. It's not
from claw code but their memory design system is something that's really genuinely powerful and clean. So it's
got three different layers. First you
have a memory MD which is your long-term durable facts. So things that don't
durable facts. So things that don't change often like your preferences, your decisions, facts about your business, and these are loaded in at the start of a recession. Second, you've got your
a recession. Second, you've got your daily note files. So you've got one file per date and you can think of these as a running log. What happened today, what
running log. What happened today, what got decided, what got shipped, what didn't work, and it's basically today and yesterday's notes that get loaded automatically. And older ones are going
automatically. And older ones are going to stay on the disc, but they're not going to get loaded into context. And
then third, with open claw, there's an optional thing called dreaming. So it's
a background process that's going to read through your daily notes, score them, and promote the ones that keep coming up again and again into your long-term memories. Memory MD. and the
long-term memories. Memory MD. and the
stale stuff is going to get forgotten completely so it doesn't blow your context. So you basically then have a
context. So you basically then have a mechanism with a proper short-term memory deciding with the dreaming process what is actually embedded into our long-term memory and fed into
claude. So it brings us to a framework
claude. So it brings us to a framework called memsarch which is actually a plugin and it basically ports that framework from open core over to claw code. So it's built by a company called
code. So it's built by a company called zillus and that's a team behind one of the most popular open-source vector databases out there. So they know a thing or two about searching by semantics and searching by meaning and
that's exactly what this is designed to do. So they've extracted open claw's
do. So they've extracted open claw's memory architecture into a standalone library that plugs straight into claw code with a twoline install. So it's got all the benefits of the openclaw method which is the markdown first philosophy.
So everything we can go we can see we can read through and we can port it to different systems. It's got the same chunking strategy as open claw, the same file structure with the memory.mmd and
the daily notes, but now it works inside claw code too. And the way it works is it basically chunks your documents, everything you write into semantic vectors so that when we are actually
searching for information, it's understanding the meaning behind our questions and the meaning behind the stuff we've stored rather than just using keyword search. But it takes that one level further and applies a concept
we've already seen already, which is actually hooks. So it auto injects the
actually hooks. So it auto injects the top matches into every single prompt. So
as soon as we write a prompt, it's going to use a hook called user prompt submit and it's going to feed in our top three semantic matches from our memory files
directly into the context for claw code.
So you don't have to remember to ask cla to search it. It's going to inject the most relevant parts of your memory files into the context pretty quickly as soon as you're actually typing in a query. So
if you want to install this, we just need to take these two lines here. I'm
going to go and open up a new folder and open up cl code. And I'm just going to use the /plugin marketplace. Add
zillastte mem search. It's going to find the repository and clone the repository.
And we're going to choose to add this to our local folder so that we're not adding it above all the other memory systems that we're considering today.
And then I'll grab the second line, plugin install memarch, and I'm going to install for you in this repo only like we just mentioned. And you can see it's spun up a settings.local.json file
enabled plugins inside there and mem search plugins is true inside there. So
you need to run / reload plugins and that will help activate it. And we'll
also just c twice and then reopen claw code. And then to actually verify it's
code. And then to actually verify it's working there's some GitHub instructions which I'll leave down below. We
basically need to have a few conversations. So have a few
conversations. So have a few conversations with it and then check your memory files. So to check your memory files, you just need to basically use the ls command inside a terminal, which is going to list whatever we're
asking it to list. So mem search/memory files, and you should see today's date inside a memory file. So go ahead, have a conversation and then list and verify that that's working correctly. Or you
can actually use the built-in skill that installs and that is uh memory recall.
So you can see me search, search and recall relevant memories from past sessions via mem search. So you can run that even as a slash command if you wanted to, but obviously we've got no daily memory files yet. It will start from this point as we start building out
conversations. And you can see it's
conversations. And you can see it's using the hook user prompt submit and it's searching through the mem search files now. And this system is super
files now. And this system is super similar to what we're actually using in our own Aentic OS to keep those daily log files as well as that long-term memory file that we can actually semantically push into the context at
the right time. Now, there's another alternative worth knowing about, but in my opinion, not really as suitable for us called Claude MEEN. And it's a claude code plugin that automatically captures everything Claude does during your
coding session, compresses it with AI, injects relevant context back into future sessions. So, sounds almost
future sessions. So, sounds almost exactly the same, super popular, has a similar goal, but a different philosophy. So, this uses MCP tools and
philosophy. So, this uses MCP tools and a three tier storage model. So, it
stores summaries, timeline, and observations. And it has more features
observations. And it has more features than me search out the box. like it's
got a little dashboard to actually search through the memories if you want.
It's got team collaboration features, cost tracking, privacy labels, but in my opinion, it's a bit overkill for what we actually need. So, if you want that
actually need. So, if you want that management layer on top of your memory, then claude mem might be able to give you that. But for our use case, it's MCP
you that. But for our use case, it's MCP based, which means Claude has to actively decide to call the search tool rather than just injecting it from a local database like me search. And the
other major difference between claude mem and mems search is claude mem is actually going to store everything in the background whereas mems search is going to keep everything in plain markdown so you can actually go back and read it. So you've got auto injections
read it. So you've got auto injections everything in readable markdown and open clause great memory patterns without actually having to leave claw code. Now
level four is where we get serious with conversation recall. So word for word
conversation recall. So word for word recall using a service called me palace or a framework called me palace. So, you
only need to go to this level if you catch yourself thinking, I know we discussed this a few weeks ago, but I can't remember exactly what was decided.
So, where level three finds our notes and a summary of those notes, level four is designed to find exactly the words that were used when you made a decision.
And the best part about it is it's all stored locally on your system. So,
compared to the other levels so far, this is a proper RAG system. It's free
and right now it's got the highest benchmark score of any memory system ever published apparently. So because it stores words verbatim, nothing ever gets summarized. So nothing can actually
summarized. So nothing can actually theoretically get lost. So here are a few snippets from the website. You can
see the content stays verbatim always.
The index above it is written in a a k a dense symbolic dialect and LLM can scan at a glance. So it's written in effectively a different language. And
this is what it looks like. It's got a bunch of pointers that are pointing to locations of specific data. So you
effectively are searching an index which is pointing to a specific area or draw as they call it inside the second database and therefore it's able to go and retrieve that information straight
away. So they've got an example here. My
away. So they've got an example here. My
son's name is Noah. He turns six on September the 12th. So this is what you'd say into the LLM. He loves
dinosaurs, etc., etc. And the pointer is then indexed as okay, Noah, son, age six, date of birth. And it stores a bunch of information in that drawer that
can then basically point it directly to draw 007 inside wing 42 room 11. And
we'll come to the importance of that structure in a second. But we can basically pull information out directly as it is stored. So when we are semantically searching for information,
it basically uses this memory palace which is an ancient method of memory to store things in wings, rooms, closets and drawers. And we just saw exactly
and drawers. And we just saw exactly that. The pointer of a wing represents
that. The pointer of a wing represents people's projects and topics. The rooms
could be days, sessions or threads. The
closets could be topics, threads or bundles. And then inside that is the
bundles. And then inside that is the verbatim text. So it's got a series of
verbatim text. So it's got a series of nested bits of information that points it to the exact correct place. And
that's how it's able to actually retrieve information or store information. And because it's stored in
information. And because it's stored in that symbolic index, the AA language, it lets the model scan thousands of draws in a single pass. And it happens super quickly. So you can see an example up
quickly. So you can see an example up here without me palace and with meal. So
with me palace, it basically as soon as you say something, it's going to go and file it in a specific draw. So wing room draw. Two weeks later when you ask it a
draw. Two weeks later when you ask it a question, you're going to see it's going to retrieve in such a quick time 42 milliseconds from that exact draw and it can even pull the verbatim text to. Now
this effectively uses two separate databases. An SQL database that tracks
databases. An SQL database that tracks entities and relationships between them and a Chroma DB or a Chroma vector database that stores every single conversation as searchable chunks as
we've seen in the drawers. And similar
to some of the previous levels, it's also using background hooks to silently store and index information. So filing
and indexing happen silently through claw code hooks on session end. So we've
seen the session start hook on session end on pre-ompaction. So if you're about to compact, it will also fire the hook.
You write it and the palace is going to basically put that information somewhere behind the curtain. Now it's super easy to install. You're basically building
to install. You're basically building your palace. One command to install. So,
your palace. One command to install. So,
you're going to run the commands listed down here. I'll link down in the
down here. I'll link down in the description below to the me palaceofficial.com. But you're going to
palaceofficial.com. But you're going to basically install it, initialize it, and it's going to create a palace structure, wings, rooms, drawers in a me palace folder. It's going to register all the
folder. It's going to register all the hooks in your settings.json file, and then it's going to start working. And
you can even retrospectively mine your previous information, your previous sessions by using this mine function. So
at this level then with me palace Claude can literally search every single conversation you've ever had in plain English word for word because it's all going to be indexed in that memory palace file structure in the background.
Now one of the downsides here is we don't have that verbatim in markdown. So
it's not exactly readable directly but we can retrieve information easily and quickly in that way. But even now everything's still locked to your local machine. Everything we talked about so
machine. Everything we talked about so far most of it is locked on your local machine. But what about the research you
machine. But what about the research you did in chat GPT 3 or 4 months ago, the brainstorming you did on your phone? So
that's where level five and six come in.
And they're totally different because it's not about remembering conversations anymore. It's about how do we actually
anymore. It's about how do we actually have conversations across multiple tools and also reliably store research that connects all of our different information. So everything sits in an
information. So everything sits in an isolated drawer in this instance and actually none of the knowledge is connected. So let's move on to level
connected. So let's move on to level five. So at this level for the me
five. So at this level for the me palace, Claude can literally search every single conversation you've ever had in plain English word for word because it's all index in your memory
palace. But even now up to this point,
palace. But even now up to this point, everything's still locked into your machine. So what about the research you
machine. So what about the research you did in chatbt last year, the brainstorm you did on your phone the other day?
That's where level five and six come in.
And they're totally different. They're
not about remembering conversations anymore. It's about building up a
anymore. It's about building up a knowledge base across tools. So level
five is totally different actually. It's
not anymore about remembering conversations. The me palace is kind of
conversations. The me palace is kind of the best the pinnacle of remembering word for word conversations. So now it's about actually if you want to build a
knowledge base of interconnected information. So you want to add level
information. So you want to add level five if you regularly consume information that you want to keep and connect over time. So think about articles videos podcasts client
notes, anywhere where you're actually just reading a lot of information and losing track of what you actually learned. You can connect that all into a
learned. You can connect that all into a interconnected knowledge base. So it's
less about one-off projects and more about building that second brain on topics that you actually care about. But
to be honest, I'd skip this if you consume content casually and don't need to come back to it later. The most
popular library out there for this you might have seen already. It's Andre
Carpathy's GitHub repo called LLM wiki or his markdown file which talks about the idea of LLM wikis and it's received a huge amount of coverage. So what makes it a bit different is that it's a
pattern for using claude to build a living wiki about any topic that you care about. So you literally just create
care about. So you literally just create two folders inside your project. You
create a raw folder where you drop your source documents which your articles, your reports, your podcast transcripts, YouTube video transcripts, PDFs and Claude is going to read from this folder but never actually write to this folder.
And then you have a wiki folder which Claude owns completely. So it's going to write every file, maintain the structure, update cross references as you go and you never actually write yourself to the wiki. And the whole
thing again is in just plain markdown files. So we don't have an external
files. So we don't have an external database or a vector database. It's all
in readable markdown files again. And
all you effectively need are the files that we've spoken about Obsidian and an AI. And effectively inside Obsidian,
AI. And effectively inside Obsidian, which is a separate software that you can download for free. Once you compile all of your resources, you'll get this knowledge graph of how your different
resources interconnect. And you've
resources interconnect. And you've probably seen a ton of other people demonstrating this on videos. I'm not
going to do a video specifically dedicated to this method because actually I don't think it's super relevant for how we want to synthesize information and I think Donnelly in this article summarized it quite well. It's
best in use cases where you have a folder called save for later which is basically a pile of bookmarked Slack threads, YouTube videos, client notes that you never go back to that you want to see the interlinking relationships
with and as it sounds effectively create a Wikipedia on a specific subject topic.
So this is if you want to do deep research on interconnected topics. But
in my opinion, I can't see the directly applicable use cases for retrieving this information in this manner inside a system like the business OS or the Aentic OS. If you disagree with me, I'd
Aentic OS. If you disagree with me, I'd love to hear why in the comments below.
Maybe I've missed the point of this one.
I can see some use cases like actually opening up wikis for specific key topic areas you're researching in and actually having those visualized and interconnected. But in the way that
interconnected. But in the way that we're using it for business use cases to build out projects, I'm not sure it's the best way to utilize a memory system apart from just doing deep research on a
specific topic. I've also had a few
specific topic. I've also had a few people actually surface a similar alternative called recall because it's been blowing up online. So recall is a hosted service that's basically
Karpathy's wiki, the LLM wiki done for you. So you install a browser extension,
you. So you install a browser extension, you can save articles, you can save all your podcasts, PDFs, etc. and Recall summarizes, tags, and auto builds that knowledge graph for you without the the
setup headache, but to be honest, it's quite easy to set up the LLM wiki for yourself. There's plenty of tutorials
yourself. There's plenty of tutorials out there that can show you how. And
Recall seems to give you its own AI chat interface, so you can actually go and chat to your saved content, so it works on your phone. And they also have MCP access, so you can directly connect it to claw code. And this in some ways is
better than LLM Wiki because there's zero setup. You don't need to configure
zero setup. You don't need to configure or download Obsidian and it's basically just available through a login and all the work is obviously done for you too.
Why I think it's not as good as Carpathy's pattern for most people. The
first one is ownership. So you don't actually own the data. Recall lets you export everything to markdown if you want to. But while it's live, it's
want to. But while it's live, it's actually live on their servers. So
you're effectively renting and not owning your own information. And
secondly, Recall seems to be built for content consumption, not operational memory. So, it would be really good for
memory. So, it would be really good for a use case like I've watched 40 YouTube videos on blog code. Uh, which one talked about memory and what did they mention about memory? But it's not
necessarily designed for what did we decide about client X's landing page and the conversion rates back in March. And
then thirdly, because it is a off-the-shelf software, there is a pricing implication too with that. So,
my take is if you're non-technical and you mainly want to organize articles, videos, and podcasts, then recall is great. But if you want maximum control
great. But if you want maximum control and you still want to build out that Wikipedia, then the LLM wiki will definitely beat you there. There's also
a quick mention for a heavyweight alternative which is similar like knowledge graph style called light rag.
So light rag is an enterprisegrade knowledge graph and it's designed for heavier entity extraction dual level retrieval the full research grade thing basically but it's completely overkill
for 99% of business owners. You don't
need the setup headache. You don't need the massive database in the background for that and you're better off either going back to level four or if you're looking for a visual knowledge base then you can install carpathies or use recall
but to be clear I think both of these are for building knowledge bases not for operational memory and therefore it probably wouldn't be used in many of our own use cases apart from research so at this point we've covered a huge amount
and well done if you stuck with it so far but also up until this point everything lives inside core code and some of you may actually not just use core code you might use chatbt on your phone, you might use cursor. And when
you start using multiple AI tools, none of them know anything about each other.
So everything we've installed so far has been inside CL code. But level six is a method I've seen called open brain by Nate Jones, who has brilliant knowledge in this space. And the intent of this is
to actually allow you to switch between AI tools constantly like chat on your phone, claw code at your desk, and it's the only system today where all of your different AIs that you're using are able to see the same memory automatically in
real time. So arguably it's the most
real time. So arguably it's the most futurep proof and most portable because actually your memory for this is going to live in a Postgress database that you own. So when the next big AI tool drops
own. So when the next big AI tool drops in 6 months, say you just connect it to your existing Postgress brain and there's no actual migration needed. But
if you're still living mostly in Cycl code on one machine, then it's probably worth skipping this part. So it is what it says on the tin. It's an
infrastructure layer for your thinking.
So it's one database, one chat channel, and any AI you use can plug into it, which makes it super powerful if you work across apps. You're not connected to any SAS and it's super cheap to run.
I think it's less than a dollar a month to run this. So it runs on a Postgress database hosted on Superbase and inside that database there's basically one table called thoughts and every row is a
single memory. So it's a chunk of text
single memory. So it's a chunk of text an embedding vector to be able to search for it and some tags and an associated time stamp when it was said and then it uses in-built extensions that handle all
that semantic search for you. So we
basically have an MCP server running from claw code that connects up to superbas's edge functions that act as a front door for every ai tool that wants to read from the database. So if that's
sounding a little bit complicated it's because one of the downsides of this it does have an AI assisted setup but it is going to take significantly longer to set up than some of these others and significantly longer to understand
versus others here. So, you have a full setup guide, which I'll link down below, or a video walkthrough on YouTube from Nate himself, as well as some companion prompts that help you migrate your
memories from existing Claw Code memory systems over to this open brain. And the
idea is super simple and follows a lot of what we've seen so far. For
retrieval, Claude Code is going to directly query it and chat is also able to query it. Claude Desktop is able to query it. Cursor is able to query it.
query it. Cursor is able to query it.
They all use just the same brain effectively. And I'm not going to set
effectively. And I'm not going to set this one up live as it is probably overkill for most people, but I'll link to Nate's guide which is written for nontechnical people and you'll get through it if you want to do the setup
in about 30 to 45 minutes. Now, there is a live community about this. So, I
recommend having a look in the GitHub and actually joining that community if you're interested in hearing more or getting some help with the setup there.
But if you're somebody that wants to actually keep your memory portable and not associated with just claw code, then this is probably the right solution for you. Now, there is an alternative that I
you. Now, there is an alternative that I quickly wanted to mention here called Mem0ero. And it's another cross tool
Mem0ero. And it's another cross tool memory layer. It's wellunded. It's
memory layer. It's wellunded. It's
really well marketed and widely used already by developers. And you can see like used by 100,000 developers from all of these large companies. So, if you wanted something that's more production
ready strength, then definitely check out mem.ai. Especially because the setup
out mem.ai. Especially because the setup is designed to actually just get you started in less than a minute and you still have a universal self-improving AI memory layer for your LLM applications
and it's completely cross tool as well.
So if you're actually shipping an AI product for example, then this might be a sensible choice for you. But again,
one of the downsides here is that your data is going to live on their servers permanently. Whereas something like open
permanently. Whereas something like open brain, you actually own the superbase project and you can export the whole Postgress database whenever you want and control access to multiple different
tools. So in terms of level six versus
tools. So in terms of level six versus all the different levels, this is probably the most complex to set up. Not
mem, but open brain. And it's the only one really that's going to introduce significant costs per month. Although
Open Brain, as we mentioned, is going to be like 10 to 30 p per month on the free tier of Superbase. But one of the downsides compared to the other levels we've seen one we're still storing away
from our laptop. So it's not locally held data. It is secure in a database
held data. It is secure in a database but it effectively means we have to query an external facing database every single time which is going to add some latency to our process. But that might be the compromise you're willing to make
if you want every single AI tool you use to share that same memory for you. So I
know that has been a ton of different information. So I wanted to just touch
information. So I wanted to just touch on how do you just get started today and which one should you pick. So if you are literally just starting then don't worry about these memory systems. Turn on
level one by just utilizing your claude.md and your memory.mmd files in
claude.md and your memory.mmd files in the right way. It's going to take you 10 minutes and it's going to be a massive improvement on the inbuilt memory inside. If you've been using claude code
inside. If you've been using claude code for a little bit then you can go to level two install John's hook and then honestly most of you should stop right there. But if you started to build up
there. But if you started to build up significant amount of context and you're losing old decisions across months of work, then you can either go to level three and install mem search or level
four where we're actually verbatim word for word retrieving information from me palace. Level five and six kind of took
palace. Level five and six kind of took us into a completely different realm for specific use cases. Level five was Karpathy's LLM wiki, which is best if you want to do some deep research on a specific topic and reference the
relationship between the subtopics in that topic. And then level six was all
that topic. And then level six was all about actually if you want something that you own that's still relatively cheap but you can plug into any external AI tool or LLM in the future then you want to go with something like open
brain or mem. And the best part about all the things that we covered today is a lot of them do stack together. So you
can run levels 1 2 and three together with no issues and actually the folder structure is fairly sim similar and you can ask claw code to actually integrate those yourself. And in terms of what I
those yourself. And in terms of what I am personally implementing, I'm implementing up to level three inside the agentic operating system. So
following those open claw conventions, but also adding on the semantic search capability and the hooks that enable actually injection of certain context at a certain time and that just improves
the search and recall functionality inside the business operating system. So
if you want to see more about exactly how to build out your own agentic operating system, then watch the next video. Thanks for watching. didn't see
video. Thanks for watching. didn't see
you in the next one.
Loading video analysis...