I Tried Every AI Coding Agent... Here's My 2026 Setup
By Your Average Tech Bro
Summary
## Key takeaways - **Cursor for Tab Completion Only**: I'm still using Cursor as my primary AI IDE, but I primarily use Cursor for the tab to complete, which I still find to be the best that I've tested throughout various other apps. [01:47], [02:03] - **Model-Specific Use Cases in Cursor**: I use GPT 5.2 for large refactors needing tons of context despite being slow, Composer 1 for fast responses to stay in coding groove, and Gemini 3 Pro for superior UI/UX even with vague instructions like 'make this prettier'. [03:27], [05:10] - **Always Plan Before Execute**: Pretty much every single time I always go with plan first and then execute, as it gives the model time to gather context and implement better quality solutions, supported by an Anthropic paper on improved performance from thinking time. [05:36], [06:06] - **Claude.md Files Auto-Updated**: I create claude.md files for every major feature as source of truth, and use a 'PR creator with docs' skill that analyzes file changes in PRs to update these files, reducing manual tagging and improving context. [08:44], [09:39] - **Parallel Agents via Git Worktrees**: I use Branchlit to manage git worktrees for isolated branches, enabling parallel AI agent work on multiple tasks without merge conflicts, including separate Supabase instances on different ports for database changes. [11:32], [13:20] - **Avoid One-Shot Features**: I try my best not to one-shot features because models overbuild like creating a flying car with machine guns instead of a simple car, wasting time peeling back; better to build piece by piece from the ground up. [14:37], [15:05]
Topics Covered
- Model Switching Beats Model Monoculture
- Always Plan Before Executing
- Parallel Agents Unlock True Velocity
- Iterate Incrementally Over One-Shots
- No-Code Accelerates Production Apps
Full Transcript
In this video, I'm going to break down my new AI coding workflow for the year of 2026. I've made this video in the
of 2026. I've made this video in the past on my channel before, but the state of AI and how quickly things change, I'm always testing out a bunch of different tools and honestly just testing out the different skill sets and features that
all the new AI coding agents are coming out with. And that's why I wanted to
out with. And that's why I wanted to create this brand new video to talk about what I'm doing these days in January of 2026 compared to 6 months ago when I last made this type of video.
Now, just to provide a little bit of context and background of myself, I'm a coding YouTuber, as you can see here.
So, I'm always keeping up with the latest tooling and I'm always testing tools out for my channel so I can give my opinions on the new tooling that are out there. But outside of YouTube, I
out there. But outside of YouTube, I also code a lot every single day because I'm also building my own startup called Yori. Yori is a social media marketing
Yori. Yori is a social media marketing tool geared towards businesses to help them more easily find content inspiration on how to make high-erforming marketing content as well as create that high performing marketing
social media content themselves. That is
my primary focus. I spend probably 80 to 90% of my week strictly building that.
And even before that, I spent the past 5 years building over 14 different apps.
So I code a lot every single day for the past 5 years. And on top of all that, I also worked a day job for the past 5 years in big tech. So I am coding a lot every single day and I haven't stopped
for many, many years. Now I'm not trying to preach that my coding workflow is the best way or the right way to do things.
I think a lot of you might be surprised with how relatively simple and not complex my workflow is. I'm just a random dude on the internet trying to share my learnings and what I do on a day-to-day basis and what has worked for me. Now, let's get into it. We're going
me. Now, let's get into it. We're going
to splitting up this video into two primary sections. Number one is a deep
primary sections. Number one is a deep dive into the actual tech stack and toolings that I use. And then
afterwards, we're going to go over some highlevel workflow of how I use the AI coding and toolings and just general strategies as well. All right. So, first
let's go over the text stack that I use.
So, I'm still using cursor as my primary AI IDE, but I primarily use cursor for the tab to complete. You know, like when you're typing around and you just get these like tab to complete stuff. I
still find that to be the best tab to complete that I've tested throughout various other apps and I like it the most. So, that is what I primarily use
most. So, that is what I primarily use cursor for. And I know that they've
cursor for. And I know that they've introduced a lot of other really fancy bells and whistles. Like for example, I know that cursor now has the ability for you to open up your app within a web browser directly within cursor and then
you can point at things with the cursor mouse and then you can modify everything through there. I personally haven't
through there. I personally haven't played around with that all too much. I
still just stick with using Google Chrome. Maybe I'm old school that way.
Chrome. Maybe I'm old school that way.
I'm sure it's a great feature. I just
haven't quite adopted that feature yet because I just really quite haven't found the need to do so quite yet. And I
know they've also introduced a lot of other things like debug mode, agent review, bug bot, all that stuff. But
once again, I keep things relatively simple with my cursor usage, largely relying on the tap to complete. And the
reason for that is because I do use cloud code as my primary AI coding agent. And we'll come back to that soon.
agent. And we'll come back to that soon.
But roughly 6 months ago, I was really really almost exclusively using Claude Code. But as of late, I actually have
Code. But as of late, I actually have found myself using cursor a little bit more and using Claude code less exclusively. Don't get me wrong, I still
exclusively. Don't get me wrong, I still do the vast majority of my AI agent coding with cloud code, but I have found some actual genuine use cases for using cursors AI aentic coding as well. And
I've primarily been using Cursor for the ability to use different models, primarily GPT 5.2, Gemini 3 Pro, and also their new or not that new at the time of recording this, but their own
in-house composer 1 model because each of these models in my opinion have a different use case for when I use them.
So once again, like I said, claude code, primarily cloud opus 4.5, that is my primary go-to coding agent for agent coding stuff, but every now and then when cla is not able to solve a certain problem, I just want to get different
opinions from different models. So I've
been testing out this GPT 5.2 codeex and I've been seeing pretty good performance. Now, at least from my
performance. Now, at least from my experience and just from what I read on the internet, GPT 5.2 codeex, it is slow. It can be a very very slow model.
slow. It can be a very very slow model.
But that's why I don't use it for my day-to-day coding and instead when I have something like a really large refactor that I need to do, I like to delegate this task for GPT 5.2 codecs because at least from what I've read and
a little bit of personal experience, I've heard that GPT 5.2 codec is better for those really long running tasks that require a ton of context and can run for like many many minutes if not a few
hours at a time. So I primarily use GPT 5.2 to when I'm doing something that I know is very time inensive, requires a ton of code changes, but it's a little bit too slow for me to use in my day-to-day coding changes just cuz, you
know, it kind of gets me out of my groove. Now, for when I do want a bit of
groove. Now, for when I do want a bit of quicker and faster response time to keep myself in like the coding groove a little bit better, that's when I actually use Composer 1's model. I
believe Composer 1 is based off of the openweight GLM model. Don't quote me on that, but I believe that is the case.
And what I really like about Composer 1 is the fact that it is really fast and you still get pretty good performance.
Now, obviously because it thinks and processes a lot more slowly compared to like the Opus 4.5s, the 5.2 codeexes or even the Gemini 3 Pros, I do think like the total raw intelligence is less, but
at the trade-off it comes with is the faster response time and the faster output tokens per second or time to first token. And I really like that as
first token. And I really like that as it helps me stay in the groove a lot better in my coding flow. So, I have been mixing that in a little bit as well. And then last but not least, I do
well. And then last but not least, I do love Gemini. I'm a Gemini die hard. I've
love Gemini. I'm a Gemini die hard. I've
been strictly Gemini 3 Pro. I think it is a step function above any other models out there in terms of UI and UX performance. Even if I give it really
performance. Even if I give it really vague instructions, I'm just like, make this look better, make this prettier.
Gemini 3 Pro's output, in my opinion, for my taste, is substantially better than Claude or GPT or Composer. So that
is what I primarily use Gemini 3 Pro for. So that's kind of how I use all of
for. So that's kind of how I use all of the different models on a day-to-day basis, specifically within Cursor. And
pretty much every single time I always go with plan first and then execute. I
pretty rarely ever just go raw dog execute. I'm like, whoa, whoa, whoa,
execute. I'm like, whoa, whoa, whoa, show me what you're going to do first.
Let me develop some understanding about it and then let's start moving forward with it. I also think it's a little bit
with it. I also think it's a little bit telling because I also remember reading a paper published by Anthropic a couple months ago, maybe a year ago at this point where they said that just by telling the model to wait and think just
for like an arbitrary amount of time, it significantly improved the output performance. Now, I don't know how true
performance. Now, I don't know how true that is. I'm just trusting anthropic as
that is. I'm just trusting anthropic as LLM experts and their opinion on that.
So, ever since I read that article, I almost always default to using any plan mode first just to make sure it has sufficient time to gather the context.
Think I often find that it will implement a better quality solution. So,
that's kind of how I use cursor. I know
that there's probably a lot of alpha in terms of using all the various debug modes. I've heard really good things
modes. I've heard really good things about it. Using the built-in browser
about it. Using the built-in browser within cursor, I think that could be really compelling as well. I just
haven't quite adopted it yet. I don't
think they're bad features. I think
they're good features. It's just, you know, I'm kind of a boomer at this point and it's a little bit harder to break out of my old existing workflow and my habits. But I do want to try out going
habits. But I do want to try out going allin on cursor again and exclusively using cursor and seeing how that goes.
Now, outside of cursor, I've already spoiled it a little bit earlier in this video, but I did talk about how cloud code to date is still my primary go-to way of doing the vast majority of my agent coding. I still go back and forth
agent coding. I still go back and forth a little bit between whether I like doing it within like the terminal view down here, but I've also been using a lot of the VS Code extension. I actually
like the, you know, the UI is pretty good. I flip-flop back and forth
good. I flip-flop back and forth depending how I'm feeling these days.
Right now, at the time of filming this video, I am back to the VS Code extension rather than the raw dog terminal. I do think the readability of
terminal. I do think the readability of the plan is easier. I don't really think there's a difference in performance per se, but I just think strictly from a UIUX perspective, I still do use cloud code. Now, in terms of how I use Claude
code. Now, in terms of how I use Claude Code though, obviously there are some crazy people out there using Claude Code to like the absolute bleeding edge. I
wouldn't say I'm quite there yet. So,
for example, I know Claude has a lot of skills or agents and stuff that you can build in and add into your Claude code.
Personally, I don't do a lot of that. I
often feel like it's kind of like that Midwit meme. Like, I'm talking about
Midwit meme. Like, I'm talking about this meme right here where on the left on like the the bumbling idiot and then the awakened Jedi over here is like just tell just tell AI to write code for you.
And then the middle is like orchestrator swarm uh agent skills parallel agents all coding and checking out at the same time. I feel like I don't know maybe
time. I feel like I don't know maybe I'll be wrong about this and I'll like bite my words for this but I don't want to be that guy stuck in the middle over complicating my AI workflow. I try to keep things really simple with my cloud code usage and I just tell it to build.
So, in general with Cloud Code, I always use Opus 4.5 and I am paying for the $200 a month 20x max plan, which is very expensive, but that gets me the absolute most usage, so I never have to worry about running out of usage and getting
rate limited. And once again, I almost
rate limited. And once again, I almost always default to using plan mode before implementing. Now, within Cloud Code, I
implementing. Now, within Cloud Code, I do have one skill in particular that I do use a lot. Obviously, you can see I have like a ton of these different like skills. I don't remember adding a lot of
skills. I don't remember adding a lot of these. I think a lot of them are kind of
these. I think a lot of them are kind of just like built in already. But the one that I do use a lot is this one, the PR creator with docs. So essentially, this is a skill that I also had AI help me
write that helps me keep the context up to date throughout my entire application. And the way that I do that,
application. And the way that I do that, essentially what the skill does is whenever I tell cloud code to make a PR, it'll execute the skill. And the skill will then analyze all of the file changes that I made in the active
branch. And then from there, it'll
branch. And then from there, it'll update any necessary cloud.md files to keep the context up to date. So as you can see, I have a ton of different cloud
files and essentially I create cloud.md
file for every major feature that the app has. So that claude.md file becomes
app has. So that claude.md file becomes the major source of truth for that particular feature. So then with that
particular feature. So then with that one skill that I have, the PR creator with docs, I think that's what it was called. It'll look for any claude.md
called. It'll look for any claude.md
files for the specific features that I modified and keep those up to date over and over again. And this has been a really big unlock for me because I like you all know for maximum and best LLM
usage and output context is key and I hate having to tag a bunch of files every single time I'm working on a repeat feature. And just creating these
repeat feature. And just creating these claw.md files and making sure they get
claw.md files and making sure they get updated with every single code change that I ever do, at least for me in my personal experience, has significantly decreased the amount of manual files that I have to tag within Claude. And it
has increased that almost magical experience that Claude code gives me in terms of making the changes and knowing where all the changes live. It's a very small skill, nothing much, but it has made a huge difference in my opinion.
And then last but not least, we'll talk about the terminal that I use. Once
again, I do the vast majority of my work within Cursor, but I do use Warp as my go-to terminal of choice. I've been
using them since like 2021. This video
is not sponsored by them. I'm just
sharing the tools that I'm using. But in
the recent years, they really dived really deep into a lot more like AI, agentic coding stuff. And they have a pretty good uh agent coding experience built into it. So with Warp, if I ever need to do some very complex terminal command, I'll often just tell it in
natural language and then Warp will then turn that natural language into actual terminal commands. And they also have
terminal commands. And they also have some pretty good like lightweight agentic coding stuff that you can do directly within the terminal that I do every now and then if I know the change is really small and I don't really want to open up cursor for example. So that's
kind of a deep dive into the actual tech stack that I use. Last 6 months that I made the video to this video talking about my AI coding workflow. The actual
tools that I'm using honestly haven't changed that much, but the way that I actually use the tools on a day-to-day basis have changed as I explained throughout the video. So now let's kind of transition over into like a more higher level overview about how exactly
I use AI and how how I use AI has changed. So the first big change that I
changed. So the first big change that I want to talk about is parallel agent. So
right now it's really invogue to run a lot of parallel agents at once, especially with models like the GPT 5.2 codeexes, Gemini 3 Pros, and Opus 4.5s that are pretty slow and you get really good performance, but often times you
can be stuck there waiting a couple minutes for a change to complete. So
rather than twiddling my thumbs and doing brain rod doom scrolling on my phone, I decided, hey, let's do parallel agents and do parallel work. And this is what a lot of other people have been doing as well. So here is exactly how I
structure my parallel agent workflow.
Now specifically, I use this tool. Once
again, I'm not sponsored by them. It's
called the branchlet. It is a open- source like DLI git workree manager. So
essentially what you can do is over here you can see I have my Yorby schema 1 and Yorbi schema 1.workree. So within Yorb schema 1, I will go do branch lit. And
for those of you that don't know, like GitHub work trees is a way to work in completely separate branches that have their own unique file set of changes and tracks their own set of changes so that you don't have to worry about any merge
conflicts if you're working on two or three different tasks at once all in one branch. All the changes are isolated
branch. All the changes are isolated into their own file directory. And this
has become really useful in my day-to-day work when especially when using parallel AI agent work, working across various different changes. So
within branch lit you can list all changes you can create new work trees directly from the parent directory that you're in. Really lightweight tool and
you're in. Really lightweight tool and it has really improved my day-to-day coding usage. So for example when you
coding usage. So for example when you run a branch lit and you create a new branch for example within Yorbi schema 1 it then creates this Yorby schema 1.workree directory and that's where all
1.workree directory and that's where all the subwork trees of the parent Yorby schema 1 are defined in there and then I just open up a separate cursor instance and just start going to town making
changes there. As you can see, I have
changes there. As you can see, I have this other directory called Yorby schema 2. Now, this is particularly like very
2. Now, this is particularly like very specific towards working on a superbase application, which is what my app is built on. I use Nex.js with superbase
built on. I use Nex.js with superbase running my background. I primarily build with Nex.js with superbase running the entire backend essentially. Now, with
superbase, the way that you run your local Superbase instance is you essentially create your own local superbase docker image and you run that docker container and then your local app connects to that docker container. The
reason why I have this Yorb schema 2 directory is because sometimes I want to work on two separate tasks that require their own database changes and I just want to have these database changes completely isolated from one another and
I don't want them to like overwrite or conflict with each other. That's why I created this Yorby schema 2 directory strictly designed for all work its own database changes outside of the Yorbi
schema 1. And then I modify my local
schema 1. And then I modify my local superbase instance for Yorbi schema 2 to have different ports assigned to it compared to Yorb schema 1 so that I can have different apps connecting to their
own individual instances of my local Superbase instance. So I can be free to
Superbase instance. So I can be free to make any database changes without having to worry about conflicts arising. So
that's how I've been tackling parallel agents and it has been phenomenal. It
has been so much more productive being able to work on multiple things at once.
10 out of 10. A little bit of a side note, but I have been testing out anti-gravity and I think anti-gravity has so much potential because I love the UI that it has for like that agent inbox
thing that you see right here. Like this
UI is great, but the big downside right now is the fact that they don't support GitHub work trees and then you can't really work on changes isolated in their own like space. All the changes are going to be merged and conflicting with
each other all in the same branch. This
makes it not as efficient and not as great potential. So, I do think once
great potential. So, I do think once anti-gravity kind of gets that set up, I think it'll unlock a lot of stuff, a lot of potential. So, I am looking out for
of potential. So, I am looking out for that. I also do know cursor does have
that. I also do know cursor does have their agent mode as well. Haven't quite
used it all that much, but I think it could be useful. I need to play around with that a little bit more. Next up, I really try my best not to oneshot anything anymore. Like, don't get me
anything anymore. Like, don't get me wrong, I think that the models are capable of oneshotting features to a certain extent, but I find myself like with when you're trying to oneshot a particular feature being built, it's
like I tell the model, "Hey, build me a car." And previously in the old school
car." And previously in the old school way of coding, it'd be like, "Okay, you want to build the car? Let's first
create the engine. Let's create the tires. Let's attach the tires to the
tires. Let's attach the tires to the engine. Let's create the frame." Piece
engine. Let's create the frame." Piece
by piece, building up to the final product of the car. But right now, with one shoting, I feel like I tell it to do something, and I say, "Hey, build me a car." And then it'll build me a flying
car." And then it'll build me a flying car with machine guns attached to it with 50 wheels and 50 different sound systems going on at once, and I have to peel back every single feature to get it as bare bones and as simple as possible.
So, don't get me wrong, it works. But
often times I do find the code sometimes not be implemented exactly how I wanted.
And I know some people in the comments are going to be like, "Bro, this is a skill issue. You just got to write
skill issue. You just got to write better spec driven development." Bro,
look, I've tried it all. Once again, it might actually be a skill issue. I could
be just like a trash developer. Don't
get me wrong. I'm not disagreeing with you there. But I also do think like this
you there. But I also do think like this wasn't a problem for me when my code base was small. But my as my code base has gotten larger and larger and I build more features and it's just gotten more complex. I found it much harder for the
complex. I found it much harder for the models to handle oneshotting anything these days. So I really really try my
these days. So I really really try my best not to oneshot anything. Obviously
I still find myself being a little crazy when I'm I'm like hey build out this entire feature on one shot. I still find myself doing that. Mixed results whether it really works or not. I really try my best not to just try to oneshot anything
and I try to just do building everything like one step at a time piece by piece from the ground up. really just trying to delegate the actual code writing to the LLMs while I still maintain and handle a lot of the architectural
decisions myself. I feel like
decisions myself. I feel like oneshotting and then peeling back all the features and trying to fix the oneshot results makes me waste more time than if I were to just try to build it up piece by piece correctly on the first
go. All right, now let's talk about how
go. All right, now let's talk about how I create better UI within my apps. Once
again, I'm not a UI expert. My apps
aren't actually gorgeous. Like they're
honestly like they're okay UI-wise. This
is really just talking about how to get better UI output from your models. And
the way that I've primarily been doing this is I find the various apps and websites that I do like in terms of their design language and I just create a bunch of screenshots of them. The
landing page, the actual app itself. And
then like I mentioned earlier, I do find Gemini 3 to be a the best UI model. I
then dump all of these models into Gemini 3 Pro and I say, "Hey, take all these images and create like a universal design language for myself that I can use within my app." And then from that
design language and design system output, I then import it into a separate cloud.md file within my app. And then
cloud.md file within my app. And then
whenever I need to make any UI changes, I just make sure to reference that cloud.md file for all the UI work that
cloud.md file for all the UI work that gets done. Once again, not prolevel
gets done. Once again, not prolevel design stuff, but it is powerful to let someone like me, a traditionally trained engineer, to have like a 6 out of 10 design output. Whereas previously before
design output. Whereas previously before all this, I was getting like two out of 10, three out of 10 design output. It's
kind of similar to how AI coding tools allows designers and literally anybody in the world build apps not at like a 10 out of 10 developer level but at like a five out of 10, six out of 10 developer level. So, and I still find that to be
level. So, and I still find that to be one of the biggest drawbacks of the whole AIdriven development just like trying to create good design. That is
something that I've really been trying to focus on making better. And Gemini 3 was a good step in the right direction.
And then creating this like design system documentation with a bunch of screenshots has been another step in the right direction. So, we're getting
right direction. So, we're getting better and better and I'm still trying to crack that part of my workflow better. So, we'll see how much more
better. So, we'll see how much more progress I can make there. That's been
something that I've really been trying to focus on as of late. And another part of my AI workflow that I've been trying to work on is actually incorporating more noode/ low code tools in our tech stack in general. And there's honestly a lot of reasons for this. Number one is
the fact that our team is expanding. We
have some technical and nontechnical people as well. And I think it's important to let nontechnical people also contribute to the product development. And while it may not be
development. And while it may not be with code, they definitely could help out with building no code automations.
And also in general, I do think that using no code tools to power certain parts of your company or your app is actually faster than building everything out from scratch. And velocity is something that I'm really focusing a lot
on this coming year. I used to think that velocity meant locking in grinding for hours and hours and hours building everything completely from scratch. But
honestly, that's really not the case.
And really what I found is that true velocity is just using whatever is the best and fastest solution out there and to not reinvent the wheel because that is just not necessary. And I've recently been using Mind Studio, who is also the
sponsor of today's video, to start powering huge portions of my app. For
example, within Yori, we have our viral content database, which is a handpicked, handcurated set of proven viral videos that other businesses have posted on social media to go viral and market their business and market their product.
And this is actually powered by Mind Studio. This entire workflow of pulling
Studio. This entire workflow of pulling all the data, categorizing it, and saving the data into our actual superbase database is actually all powered by Mind Studio. And don't get me wrong, we could build this out from
scratch ourselves, but that requires a lot of overhead of creating a really complex workflow management process. But
by using Mind Studio, we were able to create the first version of this feature way faster than if we had tried to build it out ourselves. And we liked it so much that we continue to use it to power this entire flow in our app. It's
powerful enough for me, a developer, to go in and do a lot of technical things with it, but it's also simple enough that non-technical people can jump in and create automations for themselves.
If you want to try it out yourself, you can go visit Mind Studio and they have a completely free plan that you can try out yourself. And once again, thank you
out yourself. And once again, thank you to Mind Studio for sponsoring this portion of the video. But that is a quick overview about how I use everything and how I code with AI. But
I'm also curious to hear how you code with AI. Leave me some comments down
with AI. Leave me some comments down below. Leave any questions that you may
below. Leave any questions that you may have. But I also want to hear about what
have. But I also want to hear about what your coding workflow is. Is there a certain tool that I'm not trying out that you'd find really useful? Is the
Ralph Wiggams method really that good?
Let me know in the comments. I want to hear some thoughts on that. But that's
all I got for today's video. Thanks so
much for watching and I'll see you in the next one. Peace.
Loading video analysis...