transcript
Speaker 1:
[00:00] GPT 5.5 has arrived. OpenAI's new flagship model has officially entered the chat.
Speaker 2:
[00:06] Smarter, faster, cheaper, thinkier. GPT 5 is better on long-term tasks, begins a new stage of iterative learning, which means much faster rollouts and might just make us lazy as hell.
Speaker 1:
[00:18] Yeah, the updates are causing that.
Speaker 3:
[00:20] Before previously, a lot of my prompts had to be very detailed, they're very instruction-y kind of, whereas with GPT 5.5, sometimes I become lazy and I kind of give it a very ambiguous task, but then it will figure it out.
Speaker 1:
[00:32] We will show you some incredible projects that people have already built with 5.5 and unveiled Gavin's latest Animal Deathmatch Arena thing, which we haven't even looked at yet.
Speaker 2:
[00:42] That's right, there are new crazy ways that you can integrate GPT Image 2, the image model, into Codex, into the new model, and I did it and I'm going to show you all right here. It's a ton of fun and it's GPT 5.5 day and this is AI For Humans. That's my brain dying at this point. I got a little Superman curl.
Speaker 1:
[01:03] Oh, that looks good. Yeah, I know, it doesn't look like a tapeworm at all.
Speaker 2:
[01:12] Welcome everybody to AI For Humans, your twice a week guide to the world of AI news. And Kevin, what a week, what a crazy month we have had. The AI world continues to kind of get nuts again and again, and it's not gonna slow down anytime soon. We have a new flagship model. Finally, Spud, the giant potato has landed.
Speaker 1:
[01:31] Okay, Gavin, hold on a second where I go to my Codex and I click a check for updates and it's okay. Sorry, hold on, it's not there. Let me go to my chat. Hold on, let me go to my ChatGPT app. Let me go to my ChatGPT app, a new file. Check for updates and it's not there. I have an enterprise account and a pro account, Gavin.
Speaker 2:
[01:49] It will come, some of us can't be as early as others, Kevin. I did get access, I do have access, Kevin.
Speaker 1:
[01:55] You found a magazine on your brother's floor under his bed?
Speaker 2:
[02:00] Yes, I did, I did. It's rolling out to everybody today. I got it, Kevin does not have it yet. We are recording this on Thursday.
Speaker 1:
[02:07] I'm dying.
Speaker 2:
[02:08] But it is a very cool new model, we need to dive in and really talk through some of this stuff. Kevin, I think the big thing that I was expecting from the get go with this was like, there's a lot of hype around this, there's a lot of vague posting, as they say in the AI world. We saw the mythos, like kind of mythical benchmarks, the model that Anthropix says is too dangerous for everybody to use. And then we saw Opus 4.7 come out last week. So this is kind of OpenAI's answer to that. Now, just the basics right now, some very important things to know, we're gonna dive into the specifics. The number one thing that they are touting here is that this is much more reliable on long running tasks. And later on the show, I'm gonna talk about a thing that I just did an hour ago that I literally only had to kind of input a couple things and I got something useful out of it and it ran for about an hour. The other thing that I have seen a lot of people talk about is that it thinks more for cheaper and better. And I know that's kind of a lot to unpack there, but one of the things that was going on with both the move from 4.6 to 4.7 on Opus was this idea that they were going to try to find a way to control the token costs and that it was going to get better thinking. I'm curious to know what you think about that kind of idea now that we're at this place where things are getting better, but also these companies are a little bit trying to maybe control their costs on the other side.
Speaker 1:
[03:21] Well, they need to control costs for the users. Obviously, they need to do it for themselves, but the meme was the GPT 5.5 nail in Anthropix coffin and people were posting and reposting that and sharing that because with the latest 4.7 Opus, there was a whole bunch of users felt regressions, right? It was costing more, their limits were being eaten up and 4.7 was supposed to help with that. So when you're running these long-term agents that need to spawn sub-agents and go out and read documentation and explore the web and write things and test things, all of that takes compute, it takes tokens. And so it behooves the companies that serve this, in some ways. It behooves them. If they are minotaurs, if they are horse-like, this is, yeah.
Speaker 2:
[04:10] They behooved. They behooved.
Speaker 1:
[04:11] It's in their best interest. It's in some ways in their best interest to make these models more efficient, right? Cause they can serve more, they can serve faster. On the other hand, in some ways they're not incentivized because the more tokens these models take, the more the end user has to pay. So it's this delicate balance of trying to extract as much from the end users and their corporate bank accounts as they can, while not extracting too much that people say, hey, Anthropic, we're done with you. We're now leaping to open AI. So this has been a huge issue. In fact, by the way, like not to get too in the weeds, but about 30 minutes after 5.5 was announced, Anthropic posted a big, hey, are bad. Y'all said, oh, they did.
Speaker 2:
[04:50] I didn't even see this.
Speaker 1:
[04:51] Yeah, they said Claude code was getting rough. A bunch of engineers that I chat with were like, hey, look at this. They basically found three major issues. So instead of gaslighting users, they said, actually, yeah, you're right. We had some issues on our end. We're going to reset all of your limits, even though some of you might have already paid out the nose because of these errors. I digress. Let's talk. It's 5.5 today. Let's give OpenAI some flowers. Because yes, this model should be more optimized.
Speaker 2:
[05:16] Yeah. I mean, I think this is all part of the big conversation right now that we have to talk about as we talk about these new models, is you would really have these two companies kind of neck and neck and getting into this. And I think, Kevin, it might as well, we might as well jump into it now. It's time for some Benchmark Boys conversation. Benchmark Boys! There was a lot of people last time who were confused that is benchmark boys and not bros. And we said benchmark bros. So just to clear that up from everybody's perspective.
Speaker 1:
[05:44] I think they're two separate war on factions, Kevin. Yeah. Yeah. But let's talk. Let's get the boys off the bleachers. Let's get a man here. Good game. Let's talk benchmarks.
Speaker 2:
[05:53] So I want everybody to know first and foremost, benchmarks are a weird thing in that these are, as many people in this audience probably already know, in case you don't, benchmarks are these numbers that are released that are testing these AI models on various specific tests and what they're good at. And every time they come out, they release a series of these numbers. And Kevin, the GPT 5.5 benchmarks are good. They are not as good as Mythos, right? And I think just to talk about this as a whole, Mythos had some higher numbers. The one thing I think that I was expecting a little bit from this kind of how people were talking about this was a larger jump. Because Mythos, when you saw the Mythos benchmark numbers, and again, all of this is about like what it really feels like to use the model, which we'll talk about in a bit. These numbers are still very strong. Like we were talking about the GPT 5.5 thinking number is at an 82.7 and Opus 4.7 is at a 69.4. So that is a significant jump over that particular agentic terminal use number. But in some of the other benchmarks that have come out, like even on the one that opening I released, the CS World Verified, the number is almost the same, right? So anyway, this is a long way of saying it's another step. It is not the kind of thing where you're like, it's going to do everything for me. But I do think it's important. And again, I'll get to this later on to kind of talk about what I did with the model already today, that the idea here is that you can give the model more stuff to do that's harder and it can go away and do it on its own. That is the life change that we're all looking at now.
Speaker 1:
[07:19] So a couple of things, like on the benchmark front, we've talked about this before, there's bench maxing, which is where companies overfit their model to crush the benchmarks, and it typically takes a couple of days or a few weeks for the vibes to come through, and people say, oh, this is what it excels at and here's where it falls short. Looking at the benchmark numbers, as you said, there's a couple of places where even the mythos is whatever, it's not out.
Speaker 2:
[07:42] Yeah, it's not out, so who knows?
Speaker 1:
[07:43] So comparing against Opus 4.7, which I have open right now in a terminal window, Opus bests this new model in some benchmarks. Yes, the early vibes coming out from like Dan Shipper and Evry and whatnot, like the early vibes are that this thing is is is the best model in certain use cases. Yes, that for creative writing, it got a lot better for longer term horizon tasks, which are more specific, it got better. But that for for being a generalist, some people still prefer Opus. And so what I think is is is happening here is that, you know, companies have their own philosophies with how engineering should be done in general, forget the way these models should work, right? And they they tune the models to their preferences, to their tastes. And so we're getting like a Pepsi Cola or an Android iPhone sort of existence where it's like, look, iPhones are amazing. Android phones are amazing. Some people absolutely hate Android.
Speaker 2:
[08:41] Ooh, I hate Android. I hate Android. I don't want to ever see Android in my face ever. So there you go, Kevin.
Speaker 1:
[08:45] Oh, wow. That's right, Gavin will kick a clanker. If you're getting tacos delivered in a little rolly-bot, he will kick it. He doesn't want an Android in his face. He said it. He's a clanker kicker.
Speaker 2:
[08:57] You're right, though, Kevin.
Speaker 1:
[08:57] Hashtag clanker kicker in the chat.
Speaker 2:
[08:59] No hashtag clanker kicker.
Speaker 1:
[09:01] I love clankers. Yeah, put it in the comments.
Speaker 2:
[09:02] OpenAI CTO Jakub, I think his name is Jakub. Let me make sure I understand. Jakub Pachocki. Jakub Pachocki had something really interesting to say about this. And Sam kind of reiterated this in a couple of tweets. Basically, they are saying we see pretty significant improvements in the short term, but extremely significant improvements in the medium term. I would say the last few years have been surprisingly slow. So everybody at OpenAI is kind of saying this is a new way that they are developing. They're going to be much more iterative with rolling this stuff out, which we've also seen from Opus. And Kevin, there's a little piece of this in the blog post, but like this is another model that did a lot of work on itself. And I think this is just the speeding up of stuff. And as we've seen, Opus ship all those features for Cloud Code and other stuff, I suspect we're about to see the same thing with OpenAI as well.
Speaker 1:
[09:52] Please, let's go. Let's take off, friends. Let's do it. I mean, look, we even see it with like in the open source model community, right? A new Quen model will drop and then you wait 30 minutes and then there's a distilled or fine tuned. And then a couple of minutes later, there's another one that's optimized for a different operating system or a different processor entirely. Like the pace of the evolution here is getting faster and faster. And it would make sense that if as their foundational models get better, they're better at improving themselves as well.
Speaker 2:
[10:25] Yeah. And I want to call out a couple other tweets that are really interesting. Prince has had it for a little bit and said that the GPT 5.5 Thinking Heavy, there's a different version of this, delivers better answers in two minutes than GPT 5.4 Heavy delivered in 10. So that's a little bit of what's going on here. The other thing I do want to shout out is Sam wrote a longer tweet, which was this idea about iterative development. But then he also then said, we believe in democratization. We want people to be able to use lots of AI. We aim to have the most efficient models, the most efficient inference stack, and the most compute, blah, blah, blah. So this is definitely a shot that feels like being taken at Anthropic. And then I love at the end of this, he says, we love you and we want to win. We want to be a platform for every company, scientist or entrepreneur in person. My whole career has been largely about magic of startups. And I think we're about to see that magic in hyperscale, but we love you and we want to win. So we have a combination of things going on here. This is a little bit of interesting stuff that's happening over a lot. The other thing we should talk about is Codex, right? So not only is this new model out, but Codex actually dropped a bunch of new features, which is really cool and I just used Codex with these features. I'm sorry, Kevin, I know you don't have it yet. Keep updating and see if it arrives. Better browser use, better docs. One of the experiences I had with this, Kev, was in Codex in the past, I don't know if you've had this experience, but I'm trying to build something. The browser is funky and the in-browser, which just came out a week ago or a week and a half ago, sometimes pops up, sometimes it doesn't. This time, it was really solid, like it popped up, it showed me as it was working. I saw the little arrow moving around within the Codex window, all very clean. So to me, that's a pretty big deal. And this also follows up on the announcement that kind of didn't get enough hype earlier this week, which was about the shared agents in ChatGPT. Did you see this?
Speaker 1:
[12:12] Yes, yeah, yeah.
Speaker 2:
[12:13] Yeah, so that's another way that you can open the door to specific agents that have use cases within either Codex or ChatGPT. This whole world of having things that can be spun up, it feels like to me there's a little bit like a setting of the table for things like an open-claw-like world where you can go out and get all these agents that can do stuff for you, but maybe living within the OpenAI world itself.
Speaker 1:
[12:36] Well, that's exactly what it was. That's the open-clawification of the Codex app was adding these agents. So if you want an agent that just does email triage for you, now you can easily set that up. If you're running a small business and you need a dedicated agent to look at your CRM and check the status of your A-B testing of your ads and your marketplace, now you can have all of these dedicated agents that can talk to each other and be shared in the ecosystem. The browser and computer use, specifically the computer use on the Mac version of Codex is incredible. I think it bests the Anthropic Cloud plugin.
Speaker 2:
[13:11] It definitely does. I think it 100 percent does. Yeah.
Speaker 1:
[13:14] It seems way faster, seems way more capable. Odd to me that Sam Altman got on a live stream this week and it wasn't for GPT 5.5, it was for image generation. So that just goes to show you how powerful Tuesday's announcement was, how powerful the new Image 2 model is. Every day I'm seeing people generating wild stuff with Image 2, like generate a birthday cake that has code on it that when rendered, actually makes an image of a birthday cake, was one of the ones that I saw that kind of blew my mind, or complex mathematical functions integrated into like children's rugs, like they would play on like weird, weird stuff and when you start pairing that with a model like 5.5, now you start unlocking some really incredible capabilities.
Speaker 2:
[13:58] I'm very excited to talk about that and show off some really cool examples of what's been made with 5.5 with the Image model. But first, a message from a new sponsor. I'm going to do something I never thought I'd be able to do with a laptop, and that's because I have this HP Zbook Fury workstation to work with. There are powerful computers and then there is this. We are very thankful to HP and Intel for sponsoring AI For Humans this week and sending us this absolute beast of a PC. This thing is powered by an Intel Core Ultra V9 Pro processor, and it came ready to go right out of the box. I've been using it for everything, local AI, AI video, running cloud code, and even spinning up local LLMs for my own private research. It's that powerful. I'm going to spin up Humpy UI for local AI image gen right now. So I've installed a bunch of local models like Quenon Flex, which are free to download and free to generate, and I'm going to start making something really important. Images for my new AI series, The Raccoon Bachelor. Here's why this matters. Because I'm doing this locally and the models are open source, I'm not paying per generation. I'm not waiting in a cloud queue and I'm not sending anyone to anyone else's server. That's at least a subscription or two I'm saving per month, and I can just make a lot more. And because this bad boy has an NVIDIA RTX Pro 5000 Blackwell GPU, you can see just the size of it, it's crazy. It can handle the bigger models, and it has 256 gigabytes of RAM, a crazy powerful Intel CPU. I am running stuff that used to require a dedicated desktop computer on my laptop, which is pretty incredible. And now thanks to this computer, I've got all the images I need to make that little raccoon bachelor break the raccoon lady's hearts. Check out the link in our description if you want to spec out the Zbook Fury, and thanks again to HP and Intel for sponsoring AI For Humans.
Speaker 1:
[15:36] Well, as much as I love words from sponsors, Gavin, I love words from our dedicated followers, and you can leave them as a comment below. And if you don't want to say anything, I guess that's chill too. Just like and subscribe, leave a five-star review. And if you want to back us on Patreon or buy us a coffee, you can do all that too. AIForHumans.show, that's our site. But sincerely, thank you to our sponsor, and thank you to everybody who helps grow this operation each and every week. We appreciate your time.
Speaker 2:
[16:00] That's right, and last week, thank you to everybody who said Kevin is beautiful at the end of the show. I see you YouTube commenters. There were a lot of them, Kevin. You're very happy. Okay, let's talk more about 5.5 because there are some really cool examples I've seen already, and I'm gonna show off my 64 animal tournament game. First and foremost, Kevin, there was a really interesting demo from Peter Gostev, which he made, he asked 5.5 to make a toy train set in GPT 5.5 Heavy, kind of crushed it. What was really interesting here is seeing he compared it to what 5.4 did, and you can really get a sense of like, okay, these are the kind of different quality sets of the model. Like, if you're not watching it, it's just very, very detailed. It's all being done like in a browser, he can kind of spin around it. And it's just a much less detailed version than the 5.4 version. And I don't know, it's one of those cool things that lets you see what the differences are a little bit.
Speaker 1:
[16:50] Yeah, I love these same prompt tests. And for those that are just getting the audio version, the 5.4 is cool, right? It's like a table with a model train set, literally chugging along, and then you could jump into like the conductor seat and look first person through it. But it looks a little primitive. It looks like an old Roblox type game. When you jump to the new 5.5 high, the town that the toy track is going around is fully flushed out. There's buildings, there's trees, there's a little river with a boat going through it or whatever. And when you jump to the first person mode, you have controls that make sense and they're labeled appropriately. And it's like just staring at it and going like, oh, that's a cool prompt. I like that comparison. It makes my head spin about what this test is going to look like in a year from now, Gavin.
Speaker 2:
[17:34] Or six months from now, right?
Speaker 1:
[17:36] Yeah, sure. But like the whole room is going to be modeled and you'll be able to go in and take full control and it will be multiplayer and it will run in browser. And it's just like, I'm so excited for this near future.
Speaker 2:
[17:48] I know I had a moment of that this morning thinking about like a year or a year and a half ago when you and I would be excited about what these new models would look like and the fact that we can just spin up these things so much faster is crazy to me right now. Another cool thing from Sebastien Bubeck, who actually works at OpenAI, put a unicorn together with an SVG and he said basically, he says, GPT 5.5 not fully saturating the TICZ unicorn test yet but getting awfully close. He says, this is actual TICZ code. I find it so unbelievable that I'm putting the code below for anyone to verify for themselves. So what you're seeing here is a code generated unicorn that kind of looks like a My Little Pony, but it's definitely a few far steps from what we used to see with code generated graphics. Like even the unicorn looks a little demure, like it's kind of like sadly winking at us or maybe not winking, maybe it's closing its eyes, it could be sleeping. I don't know what you think, Kevin. Is it winking? We don't see the other eyes, so who knows?
Speaker 1:
[18:41] Gavin, I actually don't want to explore this. This is a weird unicorn Vorschach test for you.
Speaker 2:
[18:46] Move on to your thing.
Speaker 1:
[18:48] I actually love the way the unicorn is playing coy and it's suddenly kind of just, it's a little wink towards me, letting me know, Gavin, everything you're doing is working these days in the gym, and you're really looking good. I came across a UFO tank game by In The World Of AI. And this was supposedly like a one-shot and there is a 3D tank that you can drive around a map as little UFOs whiz about and shoot at you and you can shoot at them. And when you make a collision with a bullet, pew pew, UFO go bye bye. This is just like, again, the new grounds of gaming. I'm sure there's a thousand startups that are going after it, but the games are going to start being good enough that you're going to actually want to participate in them and create them and remix them.
Speaker 2:
[19:33] Yeah, so let's talk about that. The project I gave 5.5 this morning was a classic project that I have given lots of times to AI models. Kevin will remember this well. You in the audience may be new, you may not. I had an idea forever ago. I think it was two and a half years ago, which was I wanted to make a March Madness tournament of the world's most dangerous animals. You take 64 of the world's most dangerous animals and you fight them one by one until there's a champion. The goal here is you as the player play one animal, and then you go through this. This morning, literally this is 45 minutes ago, I gave it two additional prompts for this. I said, go make this as a card battler. I gave it a pretty complicated prompt to start just so it had the information on it. But Kevin, the big difference here is I gave it the ImageGen tool in Codex. So what I said to it was like, hey, don't just give me, because often what happens with this when you try to get to make a game, it'll give you some sort of almost looks like a website. I said, don't do that. Pull up images. So you're going to pull it up for the first time right now. I've pulled it up earlier. It's not great, but it's also amazing that I made this in 45 minutes.
Speaker 1:
[20:35] Okay. So I'm at the dangerous animal madness site. I love that there's some particle effects going on in the background or whatever, right off the rip gap. Nice.
Speaker 2:
[20:44] Okay.
Speaker 1:
[20:45] When six ridiculous fights with the animal, the wheel gives you. I'm going to spin for my animal here and I got chaos in turn, which is a, which is a chimpanzee.
Speaker 2:
[20:57] Parking lot menace is a goose. So again, play with yours and we'll keep your record.
Speaker 1:
[21:00] I'm going to enter the bracket here. So I see the dangerous animal madness bracket. I'm using chaos in turn versus the buzzkill committee, which is a titsy fly swarm. So let's see if I can win. I'm going to zoom to the match here. Chaos in turn versus buzzkill committee. I'm entering the match. Opponent intent clamp down, attack eight, block nine. Let's go. Come on. Oh, I have to choose my hand, right?
Speaker 2:
[21:24] You have to choose your hand. Yes, you have to choose your hand. It kind of plays out like Slay the Spire or another game like that.
Speaker 1:
[21:30] Well, I guess I'll brace for weirdness, which is a defense move. And then I'm going to do a wild swing. Go for it. Okay. Yeah. Yeah. Take that titsy fly swarm. Okay. I guess I got to end my turn now. All right. This is actually too complex for me to just shoot from the hip and start clicking, Gavin. Yeah. Like, dude, I don't want to actually lose here.
Speaker 2:
[21:52] Well, so here's an interesting thing about this. So basically, again, this is the first time I'm testing it or seeing it. What's very cool about this is it's the speed to demo, right? Like that's what we've been talking about here before. The idea that you can get from zero to like, this is probably, I'd say maybe 25 to 50 percent through a game. But the idea that you can play it right away makes a huge difference. And-
Speaker 1:
[22:14] Oh, dude, I'm OP. Yeah. I'm OP. Sorry. Yeah. No, no, you go ahead. You go ahead. I'm just OP. I am crushing this titsy fly swarm.
Speaker 2:
[22:21] Get the sense of what it means to be able to demo something quickly in your brain and just drop it out. This was about a one paragraph prompt and I sent it away. It worked for about a half an hour for the first time. It came back and I said, do it a little bit better, make sure you're using the Imagen tool. It worked then for 45 minutes and came back with this. Now, it's not perfect yet clearly, but the speed to demo idea is pretty phenomenal.
Speaker 1:
[22:44] Chaos Intern survives, choose one card. I can choose an evasive flop, panic geometry or double tap dance. Gavin.
Speaker 2:
[22:53] All of that was stuff that was just prompted in. Now again, there's going to be a lot of balancing in the game like this. I'm playing a lot of Slay the Spire 2 right now. That's part of where this inspiration came from. But you get the sense that like you, the person at home, I am dummy. I do not have coding abilities. But the fact that you could spin up a demo like this very quickly and actually get it playable and get it so, I mean, it's not pretty yet, but like, it's not ugly, right? Like this idea that like, it's not just like a prototype that looks like, you know, boxes knocking to each other, that sort of thing.
Speaker 1:
[23:23] The fact that there's any graphics on screen this early in what would be a development cycle is wild. The fact that that's deployed and playable and you can share it is also wild. And I'm assuming you just told it, hey, go put this website up on Versaille or whatever and it deployed it for you.
Speaker 2:
[23:40] Yep, that's exactly right. So even while it was working, I said, hey, I steered it, I said like, hey, throw this up on Versaille so I can share it with Kevin in the middle of this conversation. So again, speed to demo, capability, long form agents, like all of this stuff is finally coming together.
Speaker 1:
[23:58] Let's focus in on GPT Image too as well because it has only been out for a few days. I am amazed at how good it is at certain tasks to the point where it has disrupted my usual workflow which I'm working on a feature for Tele right now. I typically make a PRD, I talk with our designer, I make some mockups, whatever. But now the speed with which I am iterating is it's almost quicker and easier for me to make the full thing. Have the designer anointed, like make their adjustments because they're better at design than me. But then I go and implement it as well like that. And that just changed this week.
Speaker 2:
[24:35] I had a crazy moment. I'm consulting with a friend of mine on some stuff for him and he had an idea. So I spun up, me not Coder, I spun up the demo, I spun up the design. And one of the things you can do with Image 2 is so fascinating. It's like, hey, give me a website what this might look like, right? So you get a file back. But the thing that I did, Kevin, which I was kind of blew me away because when it tries to implement that file, sometimes it's better or worse at it, knowing all the different elements on the screen. You can ask GPT to image to send you just the elements on the screen so that like in my thing it had a really good logo and it had a couple other things that were cool. He said, give me all that stuff as individual elements and then you put that in your file and you let it build, you can do it all. It's like a one person shop. It really is shocking.
Speaker 1:
[25:20] So I had it do the mock up of this product that I'm building basically. And then I said, oh, go ahead and install hyperframes or use remotion. In fact, use both and then make the mock up move like this. I want the icons to come in. I want things to highlight, animated, ba ba ba. And then give me like a 15 second video. It went off. This was 5.4, but it went off and did all of that using the GPT, the Image 2 image. And it looks great. It just it looks like a fantastic little mock up. And I mean, that's like, OK, whatever. That's me being actually productive. Let's get to the Where's Waldo games.
Speaker 2:
[25:58] Yeah. Well, that's there's a bunch of people making Where's Waldo versions with this because one of the things they can do is very detailed, very specific, larger prompts. There's a good example from Jeff Lattish, who made a University of Berkeley anti AI, Where's Waldo sort of thing where there's a bunch of jokes and then I stole his prompt and used it to make a thing about the NFL draft. Today, if you're a football fan, you know the NFL draft happened. So I had it make one of these things. And what was interesting for me was going like we said last time on the show, the little jokes and little things that ads are so interesting. And this image, this NFL draft image I made is so complicated. There's so much stuff going on in there. And now not all of it's perfect. There's a few things that are wrong, but it's making jokes like a mad magazine sort of thing, right? Like it almost feels like it's like this giant thing that somebody drew and wrote a bunch of stuff on. So it is a shocking moment when it comes to what's possible with that. And then when you compare it and contrast it with what you can do with the code, those two things together just like overpower a person I feel like.
Speaker 1:
[26:59] Yeah, I saw your draft image and I zoomed around and was like looking at things. I don't understand, like this is me looking at like actual code. I don't understand half the references, but I can tell that every little frame is packed, like every little pixel is playing some sort of joke or being part of like referential humor. I don't even know what some of these things are, Gavin.
Speaker 2:
[27:19] Well, the funny thing about it is like there's a couple of things that gets wrong. Like one of the teams that gets the wrong team, stuff like that, but it goes through, there's 10 draft picks in the middle, and it's the actual people, I asked it when I created the image, I said like go find who these draft picks are and put little jokes about each of them in. And some of them have very specific jokes, but then all around the edges, there are other jokes about what happens during the draft or things like that. So anyway, this is a very fun prompt to try for yourself for whatever world that you live in. Like it's probably a good thing if you're a corporate person, like you could do a thing where it's like make it about my company, like it probably knows a fair amount of stuff. And you can make these little jokes. It's a very cool thing to show off. I do want to say one more thing, Kevin. I sent this image to my daughter last night because my daughter was like, oh, OpenAI's new image model is interesting. She had made a picture of herself and did some stuff with it. When my daughter was a kid, hopefully they don't kill me for telling this story, there was a character that she created called Mr. Brewster where she wore this kind of white wig and she went around. It was like an old man character that she made. She was very embarrassed of that character. We loved it. My wife and I thought it was one of the funniest things in the world at the time. She was probably eight or nine. She's always had this thing of like, oh, you guys thought Mr. Brewster was so funny. It was stupid, but I think it's funny. Anyway, I sent her back this image and I said, hey, you wouldn't believe what I saw at Whole Foods. I made an image that was Mr. Brewster's wonderful concoction. It was kombucha. She's like, wait, what is that? I was like, did somebody take our name?
Speaker 1:
[28:40] It looks like a real end cap with all the different kombuchas available in a Whole Foods branded appropriately. That's amazing. She actually thought that someone made Mr. Brewster's for a moment.
Speaker 2:
[28:54] My daughter said, I thought you saw this in the store. My other daughter said, is this AI? It's just an interesting thing at large. This is where we're at right now, folks. All right, everybody, that is it for now. We will see you all next week. Thank you for joining us in Playground with 5.5.
Speaker 1:
[29:07] Oh, I still don't have 5.5.
Speaker 2:
[29:10] Kevin still doesn't have it. He'll have it soon. All right.
Speaker 3:
[29:12] Bye, y'all. We'll see you next week.