LongCut logo

Why are top engineers DITCHING MCP Servers? (3 PROVEN Solutions)

By IndyDevDan

Summary

## Key takeaways - **MCP Servers Bleed Context**: MCP servers can consume a significant portion of an agent's context window, with a single server potentially using 10,000 tokens, which is 5% of the total context. Stacking multiple servers can lead to over 20% context loss. [00:05], [00:20] - **CLI Tools: Build Once, Use Everywhere**: Using CLI tools with agents offers greater control and reduces context consumption compared to MCP servers. Agents can be taught to use these CLIs by providing them with a concise prompt and access to the CLI's documentation and code. [03:59], [05:48] - **Scripts Save 90% Context**: A script-based approach with progressive disclosure can drastically reduce context window usage, potentially down to 10% of the original consumption. This is achieved by prompting the agent to only read specific script files when needed, rather than preloading all tools. [12:51], [13:34] - **Skills: Claude's Ecosystem Approach**: Skills offer a way to bundle scripts within a directory, providing a self-contained and isolated toolset for agents. This approach also leverages progressive disclosure, consuming minimal context by only exposing the necessary skill definition. [18:36], [19:53] - **80/15/5 Rule for Tooling**: The recommended approach for agent tooling is: 80% of the time use MCP servers for simplicity, 15% use CLI for control, and 5% use scripts or skills for context preservation when absolutely necessary. [25:44], [26:30]

Topics Covered

  • MCP Servers: The Hidden Cost of Agent Integration.
  • Progressive Disclosure: The Future of Agent Context.
  • CLI: The Optimal Approach for New Agent Tools.
  • CLI-First Strategy: Enabling Flexible Agent Architectures.

Full Transcript

What's up, engineers? Indie Dev Dan

here. Once again, my MCP server just ate

10,000 tokens before my agent even

started working. That's 5% of my agent's

context window gone, and my Kshi

prediction market MCP server isn't even

that large. Stack up two or three more

MCP servers, and I'll be bleeding 20%

plus context in no time. This isn't new

but what is new are these three proven

alternatives to MCP servers. You know

about skills. We're going to break down

two additional approaches you can use to

connect your agents to external tools

without torching your contact window

while maintaining control. These

approaches are being used by some of the

top leading companies in the agent space

like Anthropic and by some of the top

engineers who use agent coding tools

every single day. We'll break down when

to use each and their tradeoffs so your

agent can ship for hours not minutes

with focus contacts. Let's go beyond

MCP.

So the first version of this is of

course the classic MCP server. The great

part about this is that your MCP server

manages everything when it comes to the

connection to your external data source.

So here we have a Haiku model. We don't

need sonnet. That's going to be overkill

for this specific problem. What does

this MCP server do? This is a Kshi

prediction markets MCP server. This

gives us a gentic access to the Koshi

prediction markets betting platform. If

we type MCP, you can see we have that KI

server connected. Let's go ahead and

search some prediction markets. We'll

type KI search markets OpenAI achieves

AGI. So this is one market that we can

access. And you can see our agent diving

into the information with the search

markets tool. And we can of course open

up that market right here. And we can

see this when will open AAI achieve AGI.

You can see here probabilities are very

very low before 2030 mark. There's a lot

of information missing from this UI. We

can of course have our agent tap into

that for us. Right? One of the key value

propositions of agents is that they can

manipulate and crud information on your

behalf faster than ever. So here's a

great breakdown, right? Get recent

trades and get the order book for this

market. I want to understand this data

at a deeper level. Haiku is doing a

great job moving at light speed for us

breaking down the markets here. Here's

the order book, all the shares, total

volume on each side, recent trades

placed. Fantastic. These agents are

really powerful at understanding

information. So, we can say something

like this. Summarize bets and market

sentiment in a concise table. I just

want to showcase what this MCP server

can do. This is a readonly couchy

marketplace MCP server. Clean sentiment

analysis. Bearish by 2029. And then we

still only have a 43% chance. So the

market is telling us with their dollars

that by 2029 we expect only a 43% chance

that OpenAI achieves whatever AGI is.

The devil's always in the details here

right? OpenAI has to announce that

they've achieved AGI. Who knows what AGI

is? No one knows anymore. But this is

the power of this MCP server, right? So

we can understand the markets at light

speed. And here we're doing it with a

classic MCP server. you know exactly how

this works. It's a great tool. It's a

great resource for connecting your agent

to services. But the big problem here

is, of course, slashcontext right away.

10k tokens torched for my small, you

know, relatively well-built MCP server.

So, let's move beyond MCP server with

three new approaches where we can get

some of the value without torching our

agents most important resource. And

these approaches trade off complexity

for control. The theme though is the

same. Use raw code as tools.

For our next approach, we're using the

CLI. So here we prompt our agent and we

give our agent a specific prompt that

teaches them how to use a CLI, a set of

functions that they can then call to

access the thing we're trying to

interact with. So how does a CLI first

approach work? Let's open up a brand new

agent. And you can see here I've dropped

the MCP server. So now we're just

running our cheap fast IQ model. If we

drop MCP, there's nothing there. How

does this work? If we don't have an MCP

server, how can we enable and teach our

agents how to use a specific set of

tools? We can do this in many ways, but

it all boils down to just context model

prompt and tools. Prime Ky CLI tools

right? Our fast agent is just going to

read these two files, a readme and a

CLI. Now, our agent has summarized how

to use this tool exactly. It understands

the CLI. It understands its settings.

It's broken down the common workflows.

And if we scroll up, you can see it

understands exactly what's going on. So

what exactly does this file look like?

Inside of the Beyond MCP server codebase

where we're breaking down each approach.

If we look for this prime couch CLI

tools, you can see a concise 25line

prompt that tells our agent how to use

these tools. You can see the key line

here is in our workflow step read only

these two. The readme of our second app

and the CLI we're on the report section

as you work with the user call the right

tools to get the data you need. So let's

look at another market. We'll search

trillionaire and I'll prefix this with

market search. So check this out.

Instead of running an MCP tool, our

agent is running a CLI command. Let's

scroll up so we can see that we have

taught our agent how to use the CLI

instead of an MCP server. Now, there's a

little more upfront cost here, right? We

are prompt engineering how to use this

MCP server, but as a reward, we get to

fully control everything the agent can

and cannot do. MCP just says, "Here's a

bunch of tools. Here's all the

descriptions. Here's all the context

you're going to need to consume every

time you boot the agent up." Here we're

in full control over everything. Here's

the readme. Here's the CLI. This is what

you should do. And then we said do not

read, any, other, Python, files., All right.

So this is important. We're saying don't

consume any additional context. This is

all you need. We can do the exact same

work, right? Market search trillionaire.

We're pulling up some of the

trillionaire couch marketplaces, right?

Who's going to be the first

trillionaire? And then we have the Elon

Musk trillionaire market. We're betting

on when he'll be a trillionaire. Great

summary here. Highest conviction. Musk

by 2030. We have the most liquid, most

traded. Of course, we can pull up these

markets here and see them at a glance.

Betting on on Elon is pretty high here.

We have this market and then we also

have who else, you know, will be the

first trillionaire. You can see a lot of

sentiment around Elon, but not uh so

much anyone else, right? So, very

interesting stuff there. We can push

this further. Summarize bets and market

sentiment. This is something that I keep

prompting. I need to turn this into a

reasonable prompt. sentiment for Elon

and first trillionaire markets. So our

agent is going to do the, you know

exactly what you would expect to kind of

pull this information together and then

we can set up one more prompt. Quick

simple summary. These markets are giving

us information. This is a really

interesting thing about these betting

markets, these emerging poly market cali

you can use this in two ways, right? As

Vitalic said, the creator of Ethereum

you know, there's two ways to use these

betting markets. It's a betting site for

some people and you know for those

looking to make high return on

investment decisions and figure out

where to bet their time and understand

events happening in the world. This is

also a news site, right? It's a place of

finding valuable information and it

helps us understand the future before it

occurs because people are placing their

bets, right? Vitalic calls this info

finance. You can use these betting

platforms to understand incentives

before things happen. So very

interesting way to think about betting

markets and it's a great way to get an

edge with agents. These agents can

understand this information faster and

better and you can get multiple

perspectives on this data. So just by

looking at these betting markets, we can

see here that by 2030 the market becomes

bullish that Elon is going to be a

trillionaire. Some of these other

prediction markets are actually really

interesting, right? Like when will

OpenAI achieves AGI? This actually tells

you the sentiment of the market around

OpenAI achieving and pushing forward and

becoming a great valuable company.

That's the information underneath the

data, but only a 43% chance by 2030.

Very interesting. Cool stuff here

right? And you can continue to prompt

your agent however you want. Web search

2025, the net worth, uh, Jensen, Elon

and Sam, what market cap would their

companies need to make them

trillionaire? This is all thanks to

agents helping us move super super fast

and understand data. But focusing back

in on the CLI. If we open up the code

here while our agent works. So we have

the CLI and here's the important part.

Right? In the CLI, we just showcase all

of the functions, right? And notice how

via the CLI syntax, click, typer

whatever you want to use, our agent can

easily see how this works. As long as

you're not using a super new tool that

the agent has no idea about, it's pretty

straightforward. Right? Here's an

option. It's called D-limit default.

Here's how you can use it. We

effectively have all the capabilities of

MCP server with raw code. And this is

something that Mario mentions, pretty

top tier engineer covering some hot

topics. He argues, what if you don't

need MCP at all? And the way he does it

is he just sets up a prompt as we're

doing here through a readme file and he

basically just tells the agent, you

know, look at this. Here are all the

tools you can use. They're all in a

specific file. We're going to break down

this approach in just a second. And it's

very similar to the CLI, but you get

improved context management. The

benefits here are are really obvious

here, right? You can pull in the readme

whenever you need. So for this agent, we

ran our prime calcium CLI tools. We only

activate this. We only set our agent up

when we need it. Okay? When we need the

specific tool set. So this is really

powerful. It's a lot more dynamic than

MCP servers. Not a ton to go over here.

I think the most interesting thing here

though is that you know Sam Alman

currently reported worth2 billion while

Elon at 500 billion already and Jensen

he's about 175 billion and then we get a

breakdown of their their companies and

how much more revenue they would need

right to really hit these marks. So, you

know, Jensen would need 450% increase

not that much, but Elon only 100% away

between all of his assets. And Sam

actually a lot further away, not as as

wealthy as a lot of people think.

Important thing to mention here, I

missed this at the start. If we clear

this agent, restart it, and we run this

prime command, only reading these two

files. Okay. Uh, we're going to conserve

our context window. If we hit /context

you'll see here our tokens are down from

that 10% down to just, you know, 5.6%.

All right,, so, we've, saved, roughly, 4%

context window with our CLI approach.

Very powerful, but we can push this

further. Let me show you the scriptbased

approach.

So scripts look a lot like skills. In

particular here, we have that same setup

where you prompt your agent with a

specific prompt, right? A prime prompt.

You're preparing your agent for

something just like you would set

yourself up for a great day in the

morning. So, you're priming about

specific files. But the trick here is is

that you want to use progressive

disclosure. This is something that

Enthropic mentions inside of their blog

when they're talking about this using

direct tool calls. Scale better by

writing code to tool calls instead. Now

interestingly here, Enthropic actually

ends up calling the MCP server under the

hood. I think Mario and I would argue

that you don't actually need to go that

far. You can just hand them the script

or the tools and have them run that

directly. The only con there is of

course you have to build out that tool

and the interaction versus rely on the

existing MCP server that exists. So

what is this scripts approach? So, if we

boot up our agent here and we run clear

and let's just go ahead and start from

scratch so it's super clear. So we have

the ha coup model here and then we're

going to, run, slash, file, system, scripts.

So again we're priming our agent with a

specific set of information. Okay. So

same deal and we can of course look at

this prompt. There you go. Okay. So it's

already finished. This model is so fast.

This time we only read a read me. And

check this out. Our agent understands

when to use each script. Okay. So this

isn't preloaded in the context. It just

understands at a high level when to use

each script. We effectively have a

condition mapped to files. This is a

powerful agentic data structure that you

can use to activate or ignore context.

Okay, so you know, check this out. This

is crazy. /context less than 1% just

under 2K context. Here's the key part. I

will not read scripts themselves

unless-help doesn't provide information

needed. We are prompt engineering this

outcome. A lot of engineers are super

obsessed with context engineering.

Everyone's jumping on the context train.

Even before context comes prompt

engineering. This is still a critical

skill. In fact, it is the critical skill

for engineers in 2025 and beyond. The

prompt shows up before the context gets

in your context window. We have just

prompt engineered out 10,000 tokens that

don't show up via a default MCP server

or even our CLI script. Right? Our CLI

script got it down by say 50 60%. But

our scriptbased approach is taking it

all the way down to let's say 10% of the

original. All right. So, how does this

work? We have again another great prompt

with a typical structure that we use on

the channel all the time. Make sure you

subscribe so you don't miss out. We're

breaking our prompt down piece by piece.

Every section has a specific purpose.

And there are some other sections here

we just didn't need, so we didn't add

them. Here's the important part. Do not

read the scripts themselves. Then we

have d-help where we explain that as you

work through each script you use d-help

to understand how to use it. Now here's

a cool part right if we look in this

directory. So this is our apps 3 file

system. Every single script is of course

a self-contained script. If we open up

the read me you can see exactly why we

had our agent read this file. These are

file system scripts, right? Isolated

single file scripts that you can use uh

on their own or together, right? And we

have conditions on each when to use each

file. And that's it, right? A 58 line

readme file. Um we could have put this

in the prompt, but having it here in the

readme is fine as well. And now we can

just, prompt, as, usual., All right,, so

business as usual. Let's go ahead and

have our agent actually do some work for

us. Let's look at another prediction

market on Kelsey. Let's run something

like this. Koshi market search. I want

to understand the government shutdown.

Okay. What's going on around the

government shutdown? How long will this

last? So check this out, right? UV run

app 3 file system scripts search-json.

There's the search. There's the result.

This is effectively the same thing as

the MCP server as the CLI server. And

that's the big kicker here, right? All

these approaches help you solve the same

problem. Give your agent access to

tools. The question is how you do that

and what does it cost? In the

scriptbased approach, you know, we have

dedicated single file scripts, right? If

we open up our search, you can see it's

self-contained. We're using Astral UV.

Shout out Astral. They are the best

Python dependency manager. The industry

is really picking up on this finally.

We've been using it for over a year now

because we can do cracked stuff like

this. Python single file scripts all

over the place with dependencies

declared at the top, right? But each one

of these scripts is their own world of

code. Of course, the trade-off here is

we have code duplication all over the

place. But that's fine. We're willing to

pay that price because it makes our

agents more effective. When you have

less garbage context, your agent can

perform better. So, here's this market.

Basically, we're betting on how long

will the government shutdown last.

There's no value. There's no information

in a bet that's 99% certain, really even

90% certain. But there is more

information here in this, you know, 66

63% bet. And we can pull up the full bet

here. How long will this last? You can

see the interesting places are this 45

to 52 mark. Right? This is where the

interesting bets are actually happening.

This is where we can actually get real

information. If we want to predict how

long, this, will, actually, last,, all right,

we can say something like summarize

bets. When can we expect the government

shutdown to end? So based on the bets

again, based on the information

available through people placing real

bets on these markets, we can kind of

understand the future in a more uh

deeper way. We're pretty much guaranteed

that it'll last longer than 39 days. And

this is since October 1st, by the way.

Um so longer than 39 days, longer than

40, but then the probability really

drops off. Based on this information, we

can see that, you know, roughly 40 to 50

days is the consensus. expected end date

based on the probabilities 18th to 20th

based on again these market bets. So

really interesting way to actually use

these markets is information about a

future state in the world. This is the

huge advantage of using these scripts

and this is something that Mario points

out here. I'll of course add all these

links into the description. Benchmarks

have shown that there's no degradation

in quality by going right for scripts

going right for CLI, right? basically

handing your agent code versus giving

them an MCP server. Now, Anthropic has a

slightly interesting approach to this.

They have this uh call MCP tool approach

where they basically recommend that you

wrap and expose specific functionality

in a CLI or individual scripts, right?

Get document and then you call the MCP

server underneath. I think you can just

cut the MCP server out completely if

you're going to script it out yourself.

They mention here that you know the big

benefit is progressive disclosure. Very

interestingly here you can see we're

getting that progressive disclosure by

ourselves by prompt engineering and it's

not a complex prompt, right? It really

is not a complex prompt. We're saying

when to use every single file and then

we're saying don't read and here's a

couple tools you can use, a couple ways

you can understand every single script

without blowing up your context window.

I like to call this incremental context.

We'll talk about the trails of all this

in a second. There's one more way to

manage your context. It is of course

with skills.

Skills look like this. They look very

similar to actually scripting things

out. The big key difference between

skills and scripts is how you invoke

them. So with scripts, we have to have a

priming prompt to kind of fire things

off. And with the skill, the prime

prompt is the skill.md file, right? So

you still have to set up the prompt that

kicks things off and that lets your

agent understand the tools available

but how you do it is just a little bit

different, right? And that's a key thing

to mention here, right? I talk about

this a lot on the channel. Don't give

away your understanding of how to write

great prompts because at the end of the

day, everything is just the core for

context, model, prompt, and tools

right? every feature just builds on top

of your agents context model prompt and

tools., All right,, so, the, interesting

thing is where and how the tools are

discovered, right? And that's what we're

really focused on here. So if we close

all of our approaches and if we open up

apps skill, you can see all we have here

is acloud file. And so of course we can

have our agent look at that file. We can

close this. So we need to cd apps for

then we boot up haiku again /mcp. You

can see nothing there. We can of course

do slashcontext and here's the best part

about skills. They have progressive

disclosure. All our agent sees here is

the definition of our skill which is

just at the top. So this is all the

context being consumed right now by our

agent. So we can prompt something like

this. Khi market search top LLM. And so

this is going to kick off the skill.

There it is. Cali markets. We're using

skills and we're using the scripts

inside of our skill. So if you look at

the structure here, you can see same

thing as our third approach, right, with

the file system scripts. The only

difference is that we've embedded, we've

bundled all the scripts into the skills

directory. Ky markets only has skill.md

and then it has all the scripts it needs

to run, right? So it's self-contained

it's isolated. This is a powerful

approach to giving your agents specific

tools without building an MCP server

right? These are MCP server

alternatives. So, you can see a very

similar structure here to our scripts.

Calcy market instructions. There's that

d--help. All of our scripts are

self-contained and they're useful and

informative for agents. If we just open

up a random one here, you can see

exactly the detail we're putting into

this. The agent doesn't even need to

look at the top of the file. But we have

all this code self-contained in a single

file script. And our skill is telling

our agent when to use each. You can see

here that it has searched, right? Ran a

couple searches here. And now it has our

top language model by the end of the

year and we have early top ranking

there. And right so we can go ahead and

open this up. You can see Gemini is

hugely biased here. There's a lot of

information missing from this. Let's go

ahead and open this bet up. Best AI end

of year. Probably you're raising

eyebrows already. We we know generally

that Gemini is not the best model. So

why is this true? Again, the devil's

always in the details. It's looking at a

single benchmark here. It's not just

looking at this benchmark which tells a

very different story. Cloud models are

also tied for first place. But they have

this remove style control on. You really

have to understand the details in these

betting markets to understand what it's

really about. But you can see here

there's a lot of specific settings that

change the actual leaderboard. So

anyway, you know, we can see the bets

here and of course we can have our agent

summarize and market sentiment, right?

Same deal. And it's kind of already done

that for us here, right? You can see

here early top rankings. Not a lot of

volume here though. Overwhelming

consensus. Everyone's saying Gemini is

going to dominate this leaderboard. We

have a skill-based approach to accessing

and running tools through scripts. Now

to be super clear here, your skill could

also be a CLI. Your skill could call

right to the API endpoints. So, instead

of these skills here, we could have

additional markdown files that just

detail how to run these commands

directly via bash and curl. If our

endpoint doesn't require a lot of, you

know, security or authentication, there

are many approaches you can take when

you're building out your custom skills

and your file system scripts, right? You

can do anything under the sun, right?

But that leads us to the tradeoffs of

each one of these approaches to giving

your agents tools.

Everything has trade-offs, right? It's

not just that we want to go beyond MCP

and that MCP is bad and you should never

use MCP. That's almost never the case

with engineering. Everything has

trade-offs. There's no one winner takes

all approach. There are options and

tradeoffs., All right., So,, for, instance,

who's invoking this? Is this agent

invoked with MCP? Yes. CLI scripts? No.

You need to run a slash command to set

your agent up and then you can prompt

it. Right. With skills, this just

happens automatically. Context window

consumption. Of course, this is the big

winner here for CLI scripts and skills

and MCP is the big loser. Especially if

you're using external MCP servers. You

have no control over this.

Customizability, no for MCP server

unless you own it. And then CLI scripts

skills, of course, you have full control

over this because you own the CLI, the

scripts, and the skill. Portability

super low for MCP, higher for CLI, but

you get even more portability out of

your scripts and your skills. Why is

that? It's because your skill is just a

single directory. So, just copy this

entire directory. Your scripts have even

higher portability, right? Because it's

just a single file. I can literally just

copy all the content in this file, paste

it to you, and you could run the code.

Composability is interesting. All of

these are composable. The kind of key

thing to mention here is that you need

to build out local prompts, sub aents

system prompts for CLI scripts, and

skills that are always going to be kind

of separated from the actual core code.

And this is where MCP wins because with

your MCP servers, there are a bunch of

features that frankly engineers

completely gloss over. Everyone thinks

it's just for tools, which yes, tools

are the most powerful piece of the MCP

server. But if we hop down to the

feature set here, you can see tools

resources prompts elicitation

completion, sampling. There's tons of

features people miss in MCP server all

the time, right? Simplicity super high.

And then things get a little more

complex here for CLI, scripts, and

skills cuz you need to manage and roll

everything out yourself. Obviously, the

trade-off there is that you get more

customizability, and, control., All right,

engineering investment kind of following

that exact same vein with simplicity.

The great part about MCP servers is that

if it's an external MCP server, you're

just done. They have everything done for

you. Just use the tools and get running

right away. Right? This is why MCP is so

great. It's standard. It's open- source.

No one controls this. Right? Skills are

very different. This is clawed ecosystem

lockin., All right., Let's, be, super, clear

about that. It's great. Uh there's a lot

you can do with it, but it is clawed

ecosystem lockin. Okay. On the other

hand, CLI and scripts, you're in full

control of these. Okay? So, you can do

whatever you want with these. You can

share them however you like. And it's

relatively simple to set this up and

maintain this. So, these are the key

differences you'll want to know. So, how

am I using all of these different

approaches? There are trade-offs across

all these. Here's how I like to think

about this, right? So, we have the tool

belt, access layer, MCP, you can use

scripts, CLI, and you can use skills.

Here's my approach, and this is what I

recommend to engineers as well. If

you're using external tools, 80% of the

time, just use MCP servers. Don't think

about it. Don't waste your time trying

to reinvent the wheel. It's just simple

to get started and running with this.

All right, but 50% of the time or 15% of

the time, you should dial in to a CLI.

And you do this if you need to modify

extend, specifically control like tools

and you can control here as well. When

you're using external tools

specifically, so you don't own these.

The the problem here by jumping down to

the CLI level is that you need to either

interface into the MCP server via code

or you need to build out the MCP server

again as a CLI. So I don't do this often

but when I do I go for CLI and then 5%

remaining the time I go for scripts or

skills and I do this only if I need

context preservation because this is

really the value ad of scripts or skills

right it's the progressive disclosure

right as anthropic mentions you get this

progressive disclosure where you give

your agent just a little bit of

information because very rarely are you

actually using every single tool every

single time. Right? If we dial back up a

couple directories here and we hop into

our MCP server, right? MCP enter inter

enter interview, tools., There, are, 13

tools here. Are you really using all 13

tools every single time? The answer is

no. Very clearly the answer is no. So

you can use progressive disclosure with

scripts or skills to just see and use

the tools you need for that one

instance, right? And you can save a ton

of context with a single MCP server. No

one really cares, right? even at the

beginning, you know, chewing up

slashcontext, you know, chewing up this

5%, it's fine. The problem is when you

stack up two or three more larger ones

and 20% of my context window is gone.

That's a problem. And so when that

becomes a problem, you can push to CLI

and you can control tools and context.

And if that's still a problem, you can

go all the way to script or skills. Now

for new tools, uh the approach is quite

different and you can see this inside

this codebase. So let me just draw this

up first and then I'll show you exactly

what I mean. So 80% of the time I just

use CLI, right? I do CLI and then I give

a prompt and it's use a prime prompt to

set my agent up with how to use the

tools, right? And this is as simple as

this. This isn't complicated, right? I

know some engineers always think things

are overengineered and things are

complicated. This is not hard at all.

Okay, look at what we're doing here.

It's a basic prompt. We have some a few

instructions and then we have a dead

simple three-step workflow. read just

these files, run the report section, and

then as you work, call the right tools.

Step three here really isn't even

necessary., All right?, I'm, just, being

super clear with my agent. And then I'm

having it report it back to me to add

more weight to the tokens. Okay? And so

what's happening here, it's simple. I'm

telling my agent how to use these tools

by showing it the exact file. That's it.

80% of the time, I just use this. Now

the nice part is is that CLI works for

you, your team, and your agents. The

trifecta can be met here. You're not

just building for you. You're not

building for your team. You're building

for you, your team, and your agents. And

CLI gets you all three out of the box. I

will push, you know, 10% of the time I

will wrap MCP server. So why do I do

this? I do this when I need multiple

agents at scale and don't want to focus

on the context. Usually the MCP server

problem isn't a problem at all if you're

using dedicated focused onepurpose

agents., All right,, this, is, something

that we talk about in tactical agent

coding. You can sidestep every single

context engineering problem by just

focusing your agents on one problem and

then you delete them when they're done.

All right,, but, sometimes, you, do, need, to

stack MCP servers. you need to have

larger tool sets and when you need that

I go from CLI to MCP server and I do it

in a very specific way. I build CLI

first so that it is very simple to then

wrap an MCP server. If you go into the

one MCP server server py you'll notice

something really interesting. All of my

methods here all of the tools you can

just search this decorator right 13

tools. All this does guess what it does

it calls right into the CLI. So I get

interoperability with MCP because I have

a CLI server first. So I always build

CLI servers instead of MCP servers when

I'm building new tools. And then 10% of

the time if I need agents at scale and I

just want to pass it a MCP.json.star

file, I will then wrap it in an MCP

server. Right? And then the last 10% of

the time I'll use a script or skill

again for the same reason if I need

context, preservation., All right?, So, if

you really need to protect the context

and you have lots of MCP servers, then

you go all the way to scripts or skills.

But most of the time, I recommend for

your new skills, you just roll out a

CLI. Why? Because it works for you

works for your team, and your agents

understand it as well. And there's an

even simpler version of, you know, all

these prompts. A CLI prompt. You can get

rid of every other section here. If

you're really just being, you know

super lazy and you're rolling out that

first version of your CLI prompt, you

can just do this five line prompt now.

Just read these files, summarize these

tools, right? And then bam, you're off

to the races. No MCP server, no anything

else. You just actually build against

the use case you're focused on. So this

is what I recommend most of the time.

And then if you need to, again, you go

to scripts or you go to skills. Now, I

think this order changes. If you're deep

in the cloud code ecosystem, then you

can of course go right for skills. But I

like to keep a balanced approach. As

much as I love the cloud code ecosystem

and the clawed ecosystem, I am always

aware of lock in and skills is a claude

specific lockin mechanism. That's fine.

Again, these are just trade-offs that we

need, to, manage., All right., So,, this, is

how I think about alternatives to MCP

server. This codebase is going to be

available to you. Link in the

description. Everything's detailed out

here so you can understand each of these

approaches. You'll be able to get

quickly up and running with this. Read

through this code before you start it.

There are a couple caveats here in this

codebase. I highly recommend you check

out these three blogs from Vitalic, from

Mario, and from the Anthropic team.

There's a lot of rich information in

each one of these. You know where to

find me every single Monday. Stay

focused and keep building.

Loading...

Loading video analysis...