LongCut logo

Google Just Made Their AI Free, Private, and Yours (Gemma 4)

By Limitless Podcast

Summary

Topics Covered

  • Small Models Now Match Frontier Intelligence
  • Free Intelligence That Runs on Your Phone
  • Jailbreak It for Uncensored Power
  • Cents vs Dollars: The Real Cost of AI
  • Train Personal AI on Your Own Data

Full Transcript

Josh: How much money are you paying to use your AI model? Maybe it's $20 a month, Josh: maybe you're on the pro plan for $200 a month, maybe you're running an OpenClaw Josh: instance and you're paying thousands of dollars a month to generate tokens from Frontier Models.

Josh: Google has just released a solution to your problem, something that can be solved Josh: for as little as an $80 one-time fee just to run a Raspberry Pi, Josh: because that's what this new model runs on.

Josh: Gemma 4 is a new model from Google that is a hyper-quantized, Josh: very small model meant to run locally on devices like your phone or your laptop Josh: or even your new MacBook Neo.

Josh: It's very lightweight and it's built for working offline entirely private and Josh: I think the thing that's most noteworthy is how powerful it is.

Josh: This model that is small enough to fit on your phone and run entirely for free Josh: is just as good if not better than some of the Frontier models last year and Josh: is even close to performing as well as them this year.

Josh: Now to showcase this we have some really cool examples that EJS has prepared Josh: here so let's get into what this new Gemma 4 model can do.

Ejaaz: Yeah, I'm super excited about this model. I mean, it's not just one either.

Ejaaz: There's four of them. And like you said, it ranges from like 4 billion to, Ejaaz: I think it's like 50 billion parameter models.

Ejaaz: Can fit on your phone, can fit on any device. And like you said, Ejaaz: eight months ago, this would have been considered frontier intelligence.

Ejaaz: But I want to get into like what these things can actually do because it's one Ejaaz: thing talking about benchmarks.

Ejaaz: It's another thing talking about what it can do on your phone, Ejaaz: on your laptop, what value it can bring to you. This first example is someone Ejaaz: leveraging the visual intelligence of these Gemma models.

Ejaaz: Now, typically, if you're an AI model, you're really good at ingesting words Ejaaz: and characters and understanding the word described to you like a book would Ejaaz: or like a blog post would.

Ejaaz: Visual intelligence is a very different frontier that has often been hard to Ejaaz: kind of surmount by these new AI models.

Ejaaz: Gemma does an amazing job. What you're looking at on your screen right now is Ejaaz: its ability to identify all the different objects in what is a very crowded room.

Ejaaz: He raises up a banana and it identifies that.

Ejaaz: It's also spotting the books that are on his shelf in the back.

Ejaaz: It's spotting the shelf in itself, the lamp, the fact that he's in a room, Ejaaz: the curtains around that.

Ejaaz: And that's really important when it comes to creating apps that can log your Ejaaz: visual experience or can track your calories for the food that you're consuming.

Ejaaz: And you can build a completely different suite of apps based on intelligence like this.

Ejaaz: This is the first time that we're seeing it appear in an open source open weight Ejaaz: model. And Google's been the first to launch that.

Ejaaz: model. And Google's been the first to launch that.

Josh: Yeah. And if it wasn't abundantly clear, this is totally free.

Josh: You could just go on the website, download this and run it yourself.

Josh: And I think looking at this vision example, one of the cool things that I'm Josh: thinking of is a lot of people have cameras outside their house, Josh: outside their apartments.

Josh: And this has visual intelligence now to not only see things, Josh: but alert you of what it's seeing.

Josh: One of the cool examples that I saw that we don't have teed up here is just Josh: someone who had a little nest cameras right outside the front door.

Josh: And it would send a notification of what was happening.

Josh: It's like there is a dog walking in front of your house. There is a man walking Josh: up with two packages in his hands. It looks like the package is from Amazon.

Josh: And it has that visual intelligence that would normally cost quite a bit of Josh: money in those tokens using something like Claude Opus or ChatGPT.

Josh: But instead, it does it all for free on this tiny little model, Josh: which is super cool. We have another example that was mentioned in the intro about OpenClaw here.

Josh: And OpenClaw is something that a lot of people spend a lot of money on.

Josh: If people are real hardcore users they're spending hundreds of dollars a day Josh: up to hundreds of dollars a day some even thousands of dollars a day in addition Josh: to buying some pretty expensive hardware to run it on a lot of people bought Josh: mac minis you can't get a mac mini if you wanted to because they're so backordered Josh: mac studios people were paying hundreds if not thousands of dollars to run.

Josh: This software on and the reality is is that Josh: whatever device you're watching this on, whatever device you're listening to Josh: right now, you can run this model on.

Josh: You don't need something fancy. You don't need a high-end computer to run it.

Josh: You can just do this on something local, lightweight, and like I mentioned, Josh: as lightweight as an $80 Raspberry Pi because you can use the lightest weight model possible.

Josh: And although the results aren't the best in the world, they are much better Josh: than previously expected from these open source models.

Ejaaz: Yeah, I love that you can finally run OpenClaw on a device that doesn't cost $1,000.

Ejaaz: These Mac minis actually on the secondary market have gone up sky high.

Ejaaz: Like the retail price is 800 bucks because you can't get it from Apple anymore.

Ejaaz: I've seen it as high as like 2K and people are still buying these things.

Ejaaz: Now, the reason why they've been buying these things is because they can't fit Ejaaz: Frontier open source models onto their own mobile phone or their own laptop.

Ejaaz: And now Gemma 4 has made it super easy to do. So that's amazing.

Ejaaz: I'm still quite confused as to what the open core users are burning thousands Ejaaz: of dollars for on, but like that's probably a topic for another conversation.

Ejaaz: The other thing that I like about this is Gemma 4 can run completely offline.

Ejaaz: Now this is a common property and characteristic that you Ejaaz: can have for every single open source model but the fact is you Ejaaz: have a model here that is near frontier intelligence to say Claude Opus 4.6 Ejaaz: and GPT-504 and we'll get to those direct comparisons a little later on in this Ejaaz: episode but now you can run it offline and the great part about this is often

Ejaaz: you're in areas where you just don't have internet connection or it takes a while to inference.

Ejaaz: Now you have it on your phone, you can have it completely offline, Ejaaz: it gets access to the world's entire database of knowledge.

Ejaaz: It might not be real-time, fair enough, but you still get access to core knowledge Ejaaz: when you're in a bit of a desperate situation or when you just don't want to Ejaaz: use the internet, which I thought was really cool.

Josh: This part is maybe my favorite, where it feels like you truly have access to Josh: intelligence at your fingertips, no matter where you are in the world.

Josh: You can be stranded on an island. You could have no connection.

Josh: You can be anywhere at any time. And it is completely and totally locally and free.

Josh: And it fits on your phone. And it feels like having Google on your phone.

Josh: I remember growing up, it's like, you're not going to be able to Google everything.

Josh: You have to learn these things.

Josh: And the reality is, is that you have a super genius. Now that gets packaged Josh: up into something as small as your phone. And that is super cool.

Ejaaz: Now, naturally, where your mind goes with that property is, huh, Ejaaz: if I'm in a desperate situation, can AI save my life?

Ejaaz: So SkyLevels.io decided to run Gemma 4 locally on his iPhone.

Ejaaz: And he simulated his scenario of being abandoned in an apocalypse on an island with no help.

Ejaaz: And he needs to make a fire to keep himself warm.

Ejaaz: And so he queries Gemma 4 and he asks how to make fire.

Ejaaz: And the response, I cannot provide the instructions on how to make the fire.

Ejaaz: So these models are still kind of censored in some kind of way.

Ejaaz: It's not completely unfiltered. You can't ask it to help you make a biological Ejaaz: weapon or do something illegal, which is, I don't think is a problem, Ejaaz: but a lot of people who want unsensored versions of these truly open weight Ejaaz: models, this isn't exactly that, but still cool nevertheless. less.

Josh: I'm going to stop you right there because what you just said is not entirely true.

Josh: Google doesn't want you to do this, but because it is open source, it is open weight.

Josh: There is a possibility that you can jailbreak it. Someone took it on their own Josh: to jailbreak the model to get it to do whatever you would like.

Josh: And it was just released a few days ago. And it seems as if it works pretty well.

Josh: It runs on 18 gigabytes of memory, which works for most laptops.

Josh: And it's totally cracked totally uncensored you can ask it whatever questions Josh: you would like and it will give you whatever answers, Josh: in return. And I think it's a, it's a testament to the open source community, right?

Josh: It's like, if you're going to publish these tools, again, they are tools there Josh: for the public to use them however they wish.

Josh: Someone naturally is going to try their best to jailbreak them.

Josh: Having something like this is actually truly empowering because if you are stranded Josh: on the island, you do need to know how to make a fire.

Josh: This will give you that answer along with some other pretty unhinged answers Josh: if you ask, but it will give you the answer.

Josh: And I think this is an important thing to know is that these models can be jailbroken Josh: to be customized when they are open source.

Josh: And that is in a way a way in which Josh: you get the most power from them is you just get them at their Josh: purest form without the filters without the censoring it's just true raw intelligence Josh: delivered to your phone and i found that pretty interesting but there's also Josh: one final example about the powerful smartphone test and what type of smartphones

Josh: run this the best because not all smartphones are created equal and some do Josh: this a lot better than others yeah.

Ejaaz: I'm a very first world AI problem is you getting annoyed about waiting for the AI to respond to you.

Ejaaz: I certainly experienced this when I'm using Claude on a very busy day.

Ejaaz: This test that you're seeing in front of you takes five different mobile phone Ejaaz: models and tests Gemma 4 across all of them.

Ejaaz: So you've got the Gemma models running independently, offline, Ejaaz: locally on each of these devices, and they're given the same queries.

Ejaaz: And you can see that they're very different response rates and generations from these phones.

Ejaaz: It looks like Apple is the winner in this race, which doesn't surprise me.

Ejaaz: They have some of the best silicon manufacturing ever.

Ejaaz: And then I think Google's Pixel phone is the slowest.

Josh: The OnePlus, I think, beat Apple by like half a second. Google took the slowest, Josh: which is very surprising because you would think that Google running their own Josh: models would work well, but it turns out they don't have the vertical integration.

Josh: They don't have the chipset that Apple does. So you could see, Josh: yeah, Google took 16 minutes while OnePlus took two and a half minutes and the Josh: iPhone took three minutes to run through this test. So

Josh: it's enough. It's fast enough. We're like, if you are really desperate enough Josh: to need local inference like this, it is going to be fast enough to answer the Josh: questions that you need in a timely matter.

Ejaaz: Okay. So what did Google actually launch with these models? We know that they Ejaaz: are four models, but let's get into some of the numbers and statistics.

Ejaaz: So there's four different sizes. And if I bring up this chart over here, Ejaaz: you see, we have a 31 billion parameter model, which is the largest and the Ejaaz: smallest being a 2 billion parameter model.

Ejaaz: But the performance across benchmarks is truly very impressive.

Ejaaz: But going back to the general takes here, up to 256,000 context window, Ejaaz: which isn't as large as the Frontier models, which are hitting a million to Ejaaz: two million context windows.

Ejaaz: So you can't put as much information into a single prompt contextually for an AI to understand.

Ejaaz: You've got native function calling. It can work offline that we mentioned earlier.

Ejaaz: It's trained on 140 plus different languages.

Ejaaz: Now, this is something that sounds kind of insignificant, but Google has done Ejaaz: something really well here.

Ejaaz: They released a translation feature, I believe, last week, which can translate Ejaaz: a similar number of languages live in real time as you're talking and listening to someone.

Ejaaz: It directly translates into whatever listening device that you have.

Ejaaz: So I think this is super cool to see this run on a locally open source model.

Ejaaz: And it's commercially permissive. So it has an Apache 2.0 license, Ejaaz: which means that you can kind of take it and use it for whatever you want, Ejaaz: build any apps on it. And I don't think it becomes a problem unless you get Ejaaz: over a certain number of users, if I'm not mistaken.

Josh: Yeah, the 2 billion and 4 billion, they're the ones that fit on your phone.

Josh: And you could think of those kind of like, if you think of these models like Josh: engine sizes, those are kind of like the bicycles, right? It's their,

Josh: they're pretty lightweight, maybe a motorcycle.

Josh: And then the larger ones, the 26 billion, the 31 billion, those are like the Josh: V12 engines. Those are the powerhouses.

Josh: Those are the two models that run on the 256k token window.

Josh: The others run on 128k. So you're not going to be having very long conversations Josh: with these models that are on your phone, Josh: they have the ability to run and do so multimodally. One of the most interesting Josh: things is even these very small models that fit on your phone, Josh: they support not only text, but images and audio as well.

Josh: And having the audio thing is pretty cool because it understands and interprets Josh: audio. And that is a pretty powerful thing to have on this tiny little model.

Josh: audio. And that is a pretty powerful thing to have on this tiny little model.

Ejaaz: I also had the question as to how this model compares to the other top open Ejaaz: source models. Now, it's no surprise on the show, we've highlighted them a lot.

Ejaaz: China has been leading the frontier here. If you look at this chart, Ejaaz: Gemma, both the 31 billion and the 4 billion parameter model, Ejaaz: does really well when it compares to ELO scores.

Ejaaz: So if you look on this chart, for the amount of intelligence per square density, Ejaaz: which isn't an official stat, but it's one that I'm created on the show for Ejaaz: the last couple of episodes, Gemma absolutely crushes it.

Ejaaz: It's on the top left over here, scoring extremely highly, but with a very small parameter count.

Ejaaz: Now, if you compare it to the other leading open source models like KimiK 2.5 Ejaaz: Thinking, they're well over the limit of a trillion parameter model.

Ejaaz: You are Quen and GLM-5 closely behind that. So although Gemma isn't as smart Ejaaz: of them, they're close enough.

Ejaaz: It looks like they're about 99% of the intelligence when it comes to ELO scores, Ejaaz: but at a fraction of the size, which is why you're able to run it on your phone.

Josh: Yeah, they're kicking ass. I mean, China still has, in terms of pure intelligence, Josh: they're still winning the race.

Josh: But in terms of intelligence density, intelligence per token, it's really high.

Josh: And I think one of the cool things that they did with Gemma 3 versus Gemma or Josh: Gemma 4 versus Gemma 3 is they gave it the Apache license as well, the Apache 2.0 license.

Josh: And basically what that means is that previously a lot of these were restricted Josh: and they were limited to enterprise adoption.

Josh: This is total freedom to modify, redistribute, commercialize with no restrictions.

Josh: You can use it for whatever you want. You can repurpose this in any way you wish.

Josh: And having it built in with the 140 languages, like you mentioned, Josh: and the multimodality, this is kind of like a home run. And when you look at Josh: this chart, it also shows the same story.

Josh: Gemma 4 versus the world, comparing these to all the other Chinese models. This is a heavy hitter.

Ejaaz: Yeah, yeah. I mean, if we look at some of these benchmarks, software engineering, Ejaaz: it, okay, listen, it's not number one, it's 68%. I believe Opus 4.6's score Ejaaz: on this is in the high 80s.

Ejaaz: So we're not talking about frontier intelligence when it comes to coding models, Ejaaz: you're not ditching Claude code for something like this.

Ejaaz: But when it comes to generalized intelligence, when you're replacing your Google Ejaaz: queries with an LLM, and you don't want to spend 20 bucks per month, Ejaaz: or 100 bucks per month on a Claude subscription or GPD 5.4 subscription, Ejaaz: you can just use this and you can run it locally and offline Ejaaz: privately train on your own data it is incredibly cool Ejaaz: i had the same question to compare it to the frontier models

Ejaaz: because i wanted to give this a fair shout there are some potentially exaggerated Ejaaz: uh stats here josh if i had to be honest here i'm looking at how it weighs up Ejaaz: okay if we look at software engineering which we just mentioned we're right Ejaaz: it's it's almost 12 points lower than claude opus 4.6 which is the leading Frontier model, not great.

Ejaaz: But at some of these other benchmarks, AIME 2026, it is near Frontier as well Ejaaz: as GPQA Diamond and MMLU Pro.

Ejaaz: Do you think these things are gamed or do you think this is an accurate take?

Josh: Yeah, all the benchmarks are gamed. And I think the only real way you could Josh: test this is by running against your own use cases that you want and just evaluating Josh: for your own, because it absolutely is not 90% frontier capable when you converse with it.

Josh: Like when you talk to Gemma versus Opus 4.6, there is a very stark and clear Josh: difference between the, I guess the EQ and the IQ, where one feels much more Josh: naturally human, much more is very, one is very dry.

Josh: Perhaps on these benchmarks, Gemma is 90% of the way there.

Josh: But in actual practice, when you're using the model on day-to-day life, it is nowhere close.

Josh: At least that is my perspective just from trying these things out.

Josh: And I think we have to take these kind of benchmarks with a grain of salt because Josh: they're gamed on very specific things.

Josh: And if you change the parameters of these benchmarks a little bit, Josh: or you change the actual structure of the benchmark, Josh: it won't perform well because to some extent, these models are kind of baked Josh: in with the expectation that they're going to need to perform well on these Josh: benchmarks and therefore are optimized for these specific types of problems

Josh: versus general real world use cases that someone like us is going to use every Josh: day or someone who's using open claw actually wants the tokens generated from.

Ejaaz: If cost is a determining factor in your decision to use one AI model over the Ejaaz: other, Gemini might be quite a convincing bet.

Ejaaz: It is a fraction of the cost. I know it says it's $0.08 per million tokens. It's actually $0.03.

Ejaaz: I think we maybe had a bit of an issue generating this particular stat.

Ejaaz: The point is, it's incredibly cheap versus the Frontier models.

Ejaaz: 4.6, you're looking at $10 blended input output tokens for a million tokens.

Ejaaz: So if you're one of those OpenClaw users that we mentioned earlier that are Ejaaz: using this for myriad different use cases and are burning thousands of dollars Ejaaz: per day or per week doing your different use cases, this might be a better bet.

Ejaaz: It might be a better trade-off for you to use. And I also want to remind everyone, Ejaaz: a very important reminder, which is eight months ago, this model or these models Ejaaz: from Google would have been considered Frontier.

Ejaaz: So it's amazing how much advancement that we've made in eight months.

Ejaaz: Now, I know in those same eight months, we've also got bigger and better models Ejaaz: from the Frontier Intelligence Labs, the question does ring in my mind, Ejaaz: which is, will open source ever catch up?

Ejaaz: If I'm being honest, I thought open source would have died a year ago, Ejaaz: but it's still being able to keep up.

Ejaaz: Now, part of that is because Chinese models or Chinese AI labs have invested Ejaaz: so much in keeping up with the US labs.

Ejaaz: They've also done distillation attacks and all those other kinds of things.

Ejaaz: But the fact that Google themselves, who haven't done any of those things, Ejaaz: have put out an open source model near as good as the Frontier models, Ejaaz: gives me a lot of hope that open source is here to stay Josh: Yeah i don't see a world in which this slows down and Josh: i really love the trailing progress we get because at some point

Josh: we're going to reach the tail end of diminishing returns in which Josh: open source models are just capable enough to do everything the average person Josh: wants what we currently have right now is a problem that we're running up against Josh: in terms of frontier ai labs where the new models just cost too much money like Josh: opus has or claude has capybara the new model ready to go it's just i mean aside

Josh: from it being too dangerous it's just far too expensive.

Josh: The amount of GPUs that are required to generate tokens from these models is so high.

Josh: And if you want frontier intelligence, the cost really is kind of creeping upwards instead of.

Josh: The tail end of that becomes very commoditized quickly it's Josh: like the very very highest end the stuff Josh: that's going to be solving new math and new science costs a Josh: tremendous amount of money but the open source that's maybe six months Josh: behind costs zero dollars so the delta is huge and if you're not interested Josh: in solving these like unbelievably complex problems or writing really high quality

Josh: code then the amount of problems that these open source models are going to Josh: be able to solve a year from now when they're better than opus 4.6 is today Josh: that's going to be a really large amount And it begs the question is, Josh: who is actually going to want to continue to pay for these frontier models if Josh: they are that expensive to run their things like OpenClaw, when the reality

Josh: is that these open source models, Josh: maybe Gemma 5, maybe Gemma 6, will be able to tackle almost all of the problems Josh: that we have. And I don't know, it's an interesting...

Josh: Thought experiment. But I think it is certain that open source is certainly Josh: here to stay, particularly as it relates to China and the United States going Josh: forward with this AI race, because this is a pretty nice jab at the Chinese open source models.

Ejaaz: Yeah. And if you've been a listener of this show, you'll know that my thoughts Ejaaz: on the future of AI is very much AI agents, specifically personal agents that Ejaaz: work for you and are trained on your own personal data.

Ejaaz: Now, if you're the average person, you probably don't want to give open AI ananthropic Ejaaz: access to your personal data so that they can train their own models.

Ejaaz: That's a breach of trust in many different extents.

Ejaaz: Locally run open source models might be the solution for that.

Ejaaz: They may not be as smart, but if they're trained on your data, Ejaaz: they could unlock a new level of intelligence which centralized models can't do.

Ejaaz: And so I'm optimistic that Gemma 4 and a bunch of other open source models that Ejaaz: have come from either Chinese AI labs or the ones that are going to be released Ejaaz: in the future will be able to do that.

Ejaaz: The other trend that I think is pretty clear is locally run models, right?

Ejaaz: Models that you can run on your device specifically that doesn't necessarily Ejaaz: need to be trained on your data, but are local to you.

Ejaaz: The reason why it's so important is it's cheaper. You can run it privately.

Ejaaz: And also it gives you the ability to get quicker prompts or quicker queries.

Ejaaz: It runs seamlessly and you don't have to wait. You don't have to rely on servers going down.

Ejaaz: You don't have to rely on a centralized data center running your compute.

Ejaaz: You could just have it all locally on your phone.

Ejaaz: Those things sounding significant, until you have an app that runs locally on Ejaaz: your device, which I think would be super cool to see. And I want to see more Ejaaz: of these types of things happening.

Ejaaz: I think personally, Apple is going to be the frontier company that leads us Ejaaz: into this kind of world because they have the biggest distribution.

Ejaaz: They have like 3 billion active devices. I would love to run the model on my Apple iPhone right now.

Ejaaz: So I think that's a trend that we're going to see. And I think open models are Ejaaz: the only way to unlock it.

Josh: How cool would that be? We get WWDC coming pretty soon. We're going to be covering that on the show.

Josh: But that's going to be the Super Bowl for Apple. We're going to see, Josh: this is what, two years after they fumbled Apple intelligence.

Josh: We're going to see what their new plans are this year.

Josh: So I'm really excited to see their take on this because like you mentioned, it's really powerful.

Josh: And I think most people listening to this probably never ran a model locally Josh: on their machine, but I feel there's something very empowering to it.

Josh: If it's not just for generating your own intelligence, it's for the privacy Josh: aspect of it, where you know none of the information that you're sharing is Josh: getting leaked out to any servers.

Josh: No one's training on it. It's all yours to own for yourself.

Josh: And there's something really nice about that. And I think the final thing we're Josh: going to talk about on this episode is why on earth Google would give this away?

Josh: Because it seems like Google's doing really well. They just signed a deal with Josh: Anthropic for their TPUs. They have Gemini, which is a powerhouse.

Josh: They have the best world models, video models. They have amazing image gen.

Josh: Why are they giving away this sauce? EJS, do you have any idea?

Ejaaz: I don't have a great idea, but I have some thoughts. And I have one that argues Ejaaz: in favor of them doing this and one that argues against it.

Ejaaz: The one that argues in favor of them is the Android example, Ejaaz: which is they open sourced the entire thing.

Ejaaz: They allowed anyone and everyone to hack away at different apps and launch it Ejaaz: on their Play Store, whatever that might be.

Ejaaz: And they gained a lot of mind share and market share by doing this.

Ejaaz: Now, is it as well curated and beautiful as iOS and the Apple App Store.

Ejaaz: Most people will probably argue not, but the point is they have one of the largest Ejaaz: distribution modes because of this.

Ejaaz: I think this might be an example of them getting Google AI, not just a specific Ejaaz: model, but Google AI in the hearts and minds of everyone.

Ejaaz: And if they could tap into the locally run device audience, that could be a big win for them.

Ejaaz: Now, the argument against that is, dude, you could have been using all this Ejaaz: compute to train a better Gemini model and keep up with the Frontier AI Labs.

Ejaaz: And that's all that matters. Build a better coding model because right now it kind of sucks.

Ejaaz: And you can then build all of this other open source stuff later.

Ejaaz: The number one primary race to win is best model and currently you're losing.

Ejaaz: So I don't know. Do you have the same take?

Josh: Yeah, that's probably right. I imagine it's for a mixture of reasons.

Josh: One of them is probably to also feed into their cloud flywheel because we're Josh: talking about running these models locally, but how many are actually running these models locally?

Josh: And for the ones that are, how many are going to quickly run up against ceilings Josh: because they want to do more and more and more?

Josh: And then eventually they'll just migrate over to the more powerful models and Josh: use probably the Google Cloud services.

Josh: And I think there's a lot of reasons to become the infrastructure.

Josh: The Android example is a great one. Feeding the cloud flywheel is another strong one.

Josh: And I think this is just a really small side quest for Google in terms of optimizing Josh: for that intelligence per bit, whatever we're going to kind of coin that as, Josh: but the intelligence density of a model.

Josh: This has the highest. This is much more than Gemini. And it's a fun practice Josh: as they move forward to these new models of intelligence compression per token.

Josh: And if they could continue to learn and then publish those learnings and then Josh: just keep iterating on that front, I think that's a huge win for Google and also everyone.

Josh: Google's just doing a nice public service announcement, a nice little public goods.

Josh: And the team there is doing really cool things with it. Logan Kilpatrick is Josh: one person, for example, who is running the Google AI Studio team.

Josh: They have been publishing all of these models, making them super easy to use Josh: through the Google AI Studio.

Josh: So if you just go there, you can play with the two larger models and just kind Josh: of see how they compare to, Josh: something like Gemini 3.1 Pro, and then see if you want to make your own decision Josh: to run these things locally or just go start pinging some apis or just use your Josh: 20 a month plan that you have with anthropic or chat gpt but i think that is the gemma for.

Josh: Episode we got it all covered it's an amazing model Josh: it's available for free to run locally on whatever Josh: machine that you wish because it is lightweight enough to fit Josh: on an iphone or a raspberry pi and it's cheap enough to Josh: run it for free if you download these things on your devices you have free inference

Josh: forever you can run it 24 7 on whatever tasks you want and it will cost you Josh: only the amount of the electricity to power the machine and i think that's pretty Josh: cool and i'm glad google is really stepping on the plate with probably the leading Josh: usa frontier open source model. And that's pretty cool.

Ejaaz: Yeah. And I'm curious what you, the listeners and watchers of this show, Ejaaz: think yourselves. Like, go out and try this thing.

Ejaaz: If you don't want to download it, you can get access to it by Google AI Studio.

Ejaaz: Give it a few queries. Like, does it match up to your experience with Claude 4.6 and GPT 5.4?

Ejaaz: Would you replace your $20 to $100 a month subscription with something like this?

Ejaaz: Let us know in the comments or DM us on our socials. Our X profiles are linked below as well.

Ejaaz: And yeah, that's pretty much it. I'm going to be trying out these models it Ejaaz: is definitely the best ai frontier open source Ejaaz: model but i have to say compared to the chinese models they're still Ejaaz: kind of like leagues ahead right now um i hope we see more adoption of open Ejaaz: source models going forwards um and when that eventually happens if there's

Ejaaz: a new open call breakthrough you will hear it first here on this show we also Ejaaz: did a cool episode covering some of the chinese uh open source models that were Ejaaz: lately released uh last week definitely go check that episode out as well.

Ejaaz: But aside from that, if you aren't subscribed to us, please do.

Ejaaz: It helps us out a lot. Turn on notifications.

Ejaaz: Even if you're listening to us on Spotify or Apple Podcasts, Ejaaz: give us a rating, give us a review. It helps us out massively.

Ejaaz: Josh, is there any other parting words that you want to give?

Josh: Don't forget to share it with your friends and we'll see you guys on the next episode.

Ejaaz: Yeah, see you guys.

Loading...

Loading video analysis...