transcript
Speaker 1:
[00:00] OpenAI's new ImageGen2 model is officially here, and it is very, very, very good.
Speaker 2:
[00:07] Nano Banana who? This model is awesome. It can handle images up to 2K resolution, multiple languages in and out, and it is very good at image-to-image editing.
Speaker 1:
[00:18] And as OpenAI pointed out in their livestream, it can actually write on individual grains of fake rice.
Speaker 2:
[00:25] Thanks, Kevin. Now I'm very hungry. We'll walk you through what it does well, what it doesn't do as well, and where we go from here.
Speaker 1:
[00:31] Oh, also, Elon is buying Cursor to be the new XAI.
Speaker 2:
[00:35] Ooh, and a cool new way to use Claude Code to make motion graphics.
Speaker 1:
[00:38] We got all of that and...
Speaker 2:
[00:40] Hey, look, Kevin, I won the Super Bowl. That's me, I'm the MVP quarterback.
Speaker 1:
[00:43] Great job, buddy. It's really great.
Speaker 2:
[00:46] This is AI For Humans, baby! Welcome everybody to AI For Humans, your twice a week guide to the wonderful world of AI. And we have a banger of an episode today. This is a big one, a big one. It doesn't seem it's that big, but it is a big one, Kevin.
Speaker 1:
[01:09] I warn you now. This one's good. No matter who you are or what you do, the announcement today, the availability of this new image model from OpenAI fundamentally changes the way you can do anything or everything, not an understatement, friends.
Speaker 2:
[01:26] Yes, this is a real human use case. This is a thing that will be good for humans, might be bad for humans, but first and foremost, we have to say GPT ImageGen 2, AKA ChatGPT Images Version 2 by any other name. This is OpenAI's new image model. We have been talking about it being teased for a while here, and it is out. Kevin, it is very good. We'll be showing some examples of things that they made for their promo videos, but there was a live stream today where Sam Altman and four gentlemen all sat on the couch and showed off the ability to make themselves into fashion models and do all sorts of other stuff. We should run through some of the basics of what this does. And then I spent quite a bit of time with this. We also want to highlight some of the cool things we saw other people do, but first and foremost, let's tell everybody what it is and what it does.
Speaker 1:
[02:14] Yeah, so the broad strokes is image model get better, right? The quality of the images you can generate, the format and resolutions of any image that you can generate. We're talking like if you want a crazy long, like vertical style billboard, this thing can do it. If you want a one by one square, it can do it. And when Sora was released, and for those who don't know, that's their video product. This was OpenAI, it's like Mimi. Yep, I know RIP Sora, gone too soon, too long. We never had the chance to dance. But Sora, what made it so special was that a basic prompt could be enhanced into this really funny or interesting diverse set of video clips because there was some thinking, some prompt magic going on in between, you give it a query and outcomes your video. And I gotta say like, that's the secret sauce that's happening here, right? If you will, there's a harness, invisible though it may be, built around this image model so that it can go out, it can think, it can do research, it can reason about what you're asking for and then generate an image for you.
Speaker 2:
[03:19] Yeah, well, I don't, but let's try to make one now while we're waiting. We wanna come up with a, let's think of something to kind of come up with as an image model so that we can kind of show people what comes out right in real time. What do you think?
Speaker 1:
[03:30] Okay, yeah, let's do it. How about like a butcher's diagram for like a muppet?
Speaker 2:
[03:36] Okay, fair enough. Let's try it. Let's see what happens. Put that in. We'll come back and check it a little bit. So there's a couple of other big things to know about this. One of the big things is, Kevin mentioned that it's a thinking model, so it is using different thinking settings. You cannot use GPT Pro right now. I've tried doing Pro a couple of times and it gave me like weird SVGs trying to interpret it, but you can use it on the thinking mode for paid users. If you are a free user, you can still use it, but you are not going to get as good a result. It's a different model. In fact, Kev, one of the things we've talked about was like the three different tiers of testing models they did before. The different tapes, if you remember, packing tape, masking tape. Right, duct tape, right? Yeah, so the thinking model is a different model than the free model, and it is very good. I think we're going to go through some really interesting examples here. But first, we should show off this thing we teased at the top. There's a sound up of the live stream that I want to show about how they wrote words on an individual grain of rice. Let's play that and just let everybody listen to this.
Speaker 3:
[04:33] Everyone, how far we can go with our image generation model. So this is an image I generated with our experimental 4K API. This is just a pile of rice, but this is also not just one pile of rice. What if I tell you there's one single grain yet with the text GPT image on it? Can you find it?
Speaker 1:
[04:54] Here we go.
Speaker 2:
[04:55] Can you find it?
Speaker 1:
[04:58] Yes. By the way, Sam Altman found it very quickly. Very smart gentleman. For the audio only folks, there is an image that was generated of thousands of grains of rice in a little pile on top of a wooden table and sure enough, smack in the middle on one singular grain of rice, they've got some text written. It says GPT image too, but you have to zoom genuinely very far in like you are looking at these individual atoms or getting into the microscope and looking at a single cell organism. It is there and that's impressive.
Speaker 2:
[05:31] Yeah, I mean, I think the biggest thing here is text is so much better and so much more pronounced, right? So you can have images, we tested this before, but you can have images with just a ton of text on them and it will still show up. In fact, I tried doing a thing a couple of weeks ago when they were leaking out these models where I could have like the periodic table be created with an image model and see what would happen and I did it today. And Kev, if you go and look at my test, my periodic table test, like it actually did a good job. My prompt here is really like make the periodic table, but then behind each element, make sure you get the symbol right and have a representation of what that element is. And some of the ones that's cheating on, but you have to know from 1.5 to this is a massive leap because this is asking the model to think a lot. It's got to put all the stuff in the right place. It's got to put the right image behind the right symbol. What's going on?
Speaker 1:
[06:22] What are you looking at there? Americ, Americium is a smoke detector?
Speaker 2:
[06:28] Well, I see some of this stuff, I'm not an expert enough scientist to know, but my gut is telling me that Americium probably is in a smoke detector.
Speaker 1:
[06:36] Is used inside of a smoke detector?
Speaker 2:
[06:38] Let's find out. Yeah, whatever, nerd.
Speaker 1:
[06:40] I don't believe it. That's fake. Also earth flat, prove it.
Speaker 2:
[06:44] So I want to also go through a couple other things. In the Grain of Rice mold, Simon Williamson, who if you're not reading his blog, it's a great blog, had a great test that he's been doing with Where's Waldo? And we know that Where's Waldo is a very complicated image with lots of people in it. And one of the things he showed off was like how good it was at burying a specific image within a group of people. Kevin, I think you remember you and I tried doing Where's Waldo tests like years ago at this point, and generally it was pretty bad at it. If you look at Simon's image, like you can just see that like zoom in ability, which is something that really didn't exist in the same way with AI Images for a long time. Being able to do that without having to go out and up-res it and lose some of your backgrounds and things like that is I think a pretty big deal.
Speaker 1:
[07:28] Wow. Okay. Oh, we got Gavin.
Speaker 2:
[07:31] We got our Muppets. Let's see, can we see how they turn out?
Speaker 1:
[07:34] We got some puppets. Standby here, I'm gonna send you. I got two of them coming in. I did ask for a modification on the second one to make it more kind of cartoon-like, and maybe that's the one we wanna focus on.
Speaker 2:
[07:45] Oh, wow.
Speaker 1:
[07:45] Because the Snorf butcher diagram. So the prompt, for those that are seeing the results on the screen, the prompt was for a detailed butcher's diagram that cuts up a quote puppet-like cartoon character with recommendations on how to prepare and tasting notes for each thing. And Snorf head, the roasted whole head is great because Snorf doesn't lose his smile or his charismatic wide eyes when his head is severed and served on a plate.
Speaker 2:
[08:17] That's a great centerpiece for my event, Kevin. I'm so happy I finally know that I can have Snorf as part of my world.
Speaker 1:
[08:24] Did you see the Snorf facts at the bottom as well as the sustainability note?
Speaker 2:
[08:29] So this is what's so interesting. So the Snorf facts for everybody are he is an herbivore, he loves tickles behind the ears, he speaks and snoofles and Snorbs and is an excellent hugger. And tonight we are going to be eating him. I'm sure he is delicious.
Speaker 1:
[08:43] But don't worry, don't worry, Gavin. Snorfs are farm raised with love and care. Use every part, waste not, be kind and keep the Snorf spirit alive. Oh my God. Gosh, Gavin, I hope you like my feet, they're chewy and gelatinous.
Speaker 2:
[09:01] We're gonna have people that are against that. So they're gonna be protesting the Snorf devastation that we're perpetuating here. But I want to go to the next thing that might be another example of mine. This kind of speaks directly to one of the other things that this can do is really good images of websites or screenshots or things that exist in the real world. And one of the reasons people are assuming that's the case is because a giant AI model needs to train and all this stuff. So it's probably, you know, seen a ton of the Internet at this point and all these different sort of things. And one of the use cases of this is that they're talking about in codex, the codex model. We do still expect a newer AI model to come from OpenAI soon, like an actual state of the art LLM model. With codex though, they're saying that in the codex app, you can use this to generate screenshots that can then build websites based on those screenshots. But Kevin, I want you to go look at my absolute worst meal thing that's in the rundown here. I basically made, I wanted to see what it could do of making a screenshot of a, like a cooking site, right?
Speaker 1:
[10:03] There's no snorf on the menu, but some of this looks delicious.
Speaker 2:
[10:06] Yeah, so I asked to create me a screenshot of a website that gives me a step by step recipe to making the worst possible meal, but take it clearly serious like a chef would come up with whatever the most horrifying but still real meal would be. And what's great about this is it created an entire website that's called the Culinary Institute of Disgusting Cuisine, and it looks like a very fancy place. And this is the recipe for the absolute worst meal. And basically, this is a combination of ramen, tuna, Coke, American cheese, chocolate chips, marshmallows, Cheetos, cocktail juice from a fruit cocktail, garlic, soy sauce, and blue cheese.
Speaker 1:
[10:41] So it's a big- The first ingredient is one can, five ounce, tuna, and water. Undrained. Which is a great note.
Speaker 2:
[10:48] So it gives you a lot of great step by steps. But Kevin, in the same way that we got a little bit extra from Snorf, one of my favorite things here is that it said, this recipe was developed in our test kitchen as a thought experiment. How bad can food possibly be while still technically being food? This result is nothing short of dreadful. Now, that is like a version of a surprising result that we used to talk about with Sora when you would not exactly say what you wanted, but it would bring something kind of created to the table. And Kevin, I did do one thing. I took this image, I handed it over to Claude, had Claude write a prompt for a TikTok video about a woman making this exact dish, and then I had Seedance generate it, so let's watch that.
Speaker 4:
[11:27] Okay, today we're making the absolute worst meal, and I promise you it hits. Drain most of the water, then stir in a half cup of cola. Trust me, whole can of tuna, do not drain, we want the juice. Mayo, chocolate chips, marshmallows, Cheetos, fruit cocktail, juice and all, blue cheese, soy sauce, stir it low and slow till the cheese melts in, top with American. And that's it, absolute worst meal. Don't forget to like and follow for more.
Speaker 2:
[11:43] First of all, it brings the grossness of it to life, right?
Speaker 1:
[11:46] It really comes to life, yeah.
Speaker 2:
[11:47] See how disgusting it is, but you just get a sense of like, okay, you have this crazy image, then you take it, break it down and you can make something out of it. This is the weird new creative pathway that we're going on through this AI stuff, right? So very cool.
Speaker 1:
[12:00] But, and I could see there being a culinary Institute of disgusting cuisine, an actual website we could go through and then submit videos of you preparing and eating the food and actually like reviewing it and becoming a meme. And then you got to get the shirt and then you got to have the pop-up restaurant at a South by like you could easily extrapolate like the, from idea to meme is just getting more and more condensed. That timeline. What I really love about this, Gavin, not only is it's like a great, it's a funny idea. The result is great. The coherence in this image from the ingredient list to the directions, which would actually make sense chronologically to of course the hero image of the completed meal where you see all of the individual ingredients incorporated into the meal. The fact that the cheese is slightly melted because it was on the hot marshmallow and tuna and ramen. All of that adds up to being just again, yeah, pure delightful as an output. And to be able to just take that and feed that into a video generator, you could imagine the TikTok account coming to life overnight.
Speaker 2:
[13:05] Yeah, exactly. And I think that that's kind of what this does. I do want to shout out a couple other examples. Ethan Mollick, who we've talked about in the show from the very beginning, did his famous Otter test. And what I love about this is it's so complicated that the Otter test is now giving a presentation about the Otter test. And you see him going through slides. And then the slides are all really good.
Speaker 1:
[13:25] What is the Otter test for those who don't know exactly?
Speaker 2:
[13:28] The Otter test was Ethan's original AI video test, where he would try to generate an Otter on an airplane. I think it did reach back to images originally. And so it kept getting more and more realistic. And at one point, kind of Ethan said, well, it's solved as a video test now because it looks like there's actually an Otter on an airplane doing this thing. And he would be on his phone. And now you can just see it opens the door even more creatively to think about different ways that Otters can present the information about the Otter test. But again, this is the thing I keep thinking about when new tools like this come out is, I love the day or two that they drop because A, as to your point last time, they're not in anywhere hobbled. It's probably the best version of them you're gonna get. So you can do a lot of cool stuff. But you also see human creativity on display, right? You really start to see why and how people can use these things to do things that they couldn't do in real life before. That's just a very cool thing.
Speaker 1:
[14:19] You also wanted to shout out Nathaniel Whitmore's magazine test because it was something that they did on the live stream as well. They sort of did like a photo of the boys from the stream on the cover of magazines. But now you can imagine these, I used to, I mean, I loved magazines growing up. I loved the crazy covers. I love this take on it. This is a magazine cover from the 90s about a time traveler from 2026 discussing GPT Image 2.0. And it's amazing. It's a leather jacket, white shirted gentleman holding up a disquette that says GPT Image 2.0 demo build 71825, which is interesting that that's the date. There's so much text on there. Yes.
Speaker 2:
[15:00] I mean, one of the things they showed off in the, in the live stream too, was this idea and maybe magazines will have somewhat of a comeback because like when you look at how these image models can actually stack text and what they can do, maybe it's like magazine, like websites that will kind of be a thing. Everything is going to be a lot more cool looking, I think, as long as we don't like kind of shave the edges off of everything. But like the ability for individuals to design, quote unquote, and I know we didn't even talk about Claude design drop this week, but that was another thing that's been going on is like Claude dropped a whole new system.
Speaker 1:
[15:29] I used it extensively. Yeah.
Speaker 2:
[15:31] I mean, it's pretty interesting. I used it a little bit too. I don't know what your thoughts were. Did you like it?
Speaker 1:
[15:35] Yeah. I mean, quick, quick departure. If you go to claude.ai/design, you can get a beta access to the new feature and it really comes alive if you already have existing assets or existing brand. You can use it to design a brand, sure. But I was able to, like for my day job, the company Telly, they've got massive figmas of all the component libraries and brand IDs and yada yada and the way things interact. And so I was able to cherry pick and put that into the system and then ask it, design me like a smart screen element for a widget that would do X, Y, Z. In fact, the jam that I did was design me a widget that displays the current status of the Strait of Hormuz.
Speaker 2:
[16:16] Oh, that's fun. Why not, right?
Speaker 1:
[16:18] And it used all of the design language and material science and thinking that we have. And it gave me like, here's what it would look like if it were open, contested or closed. Here's how it would look in like a miniature widget form or an expanded widget form. And it just fully, because we have really good documentation for all of those layers, it killed it. So if you've got your idea for a video game or a mobile app, or you've got your mom and pop restaurant, or you've got your consulting business, whatever it is, if you're able to communicate to Claude Design, these are the fonts, these are the colors, these are the principles, these are the ways we want our brand expressed or not expressed. The system is really good and it can generate motion graphics too. So I was very impressed.
Speaker 2:
[16:59] It's really interesting. And one of the things that's really interesting about this new image model that OpenAI came out with is that you can generate images in any ratio of aspect ratio. So when you think about a Claude design program, like you could say, like, hey, give me GPT-2 image, give me like 15 images, but I need them in these ratios and then plug those directly into Claude. I mean, it's a lot of things you can play back and forth with. I do want to say two other quick things. Kevin, I made a very quick YouTube screenshot of us in the future because that was something that people were doing. There's a version of us from the year 2045. But again, it gets all the copyright. You can see some interesting stuff. More importantly, I want to talk about the image-to-image capabilities of this because one of the things we know, for a long time with Nano Banana, it was one of the only models that could do this, was like image editing is really important. I just want this little section to be changed. Or I want to take the format that this thing is and put it into this other thing. So I used an image of an old Life magazine cover of Robert Oppenheimer. So there's a lot of people out there who kind of think of Sam Altman, he might think of himself this way, as like a kind of connector to Robert Oppenheimer because Oppenheimer is the person who created the atomic bomb. Sam Altman did not create AI, there's other people who are much more responsible for it, but there's this connection. So I took this image from Life magazine, and I said, make this image of Sam Altman, but keep it all in the same look. And it's pretty incredible when you see what it is. It's Sam in the same positioning, it's the same kind of like setup of the magazine, there's a slight difference in the date. And the only thing is like the Oppenheimer photo looks slightly worse. And I think that might be one of those weird things where like you can't get it exactly to match this sort of thing. But still, it looks pretty crazy. And then I used the FAL API, which is now live, you can go use the FAL API. And I actually used it, an example they had and there was like wrap a bus, a London double decker bus and something. And I wrapped it in, my wife has a new book coming out, a writing book called How to Write a Book in 100 Days, it's coming out in September, go pre-order it now. And I wrapped it in her book cover. And it fricking looks pretty good, right?
Speaker 1:
[18:55] And yeah, and just so Kim Purcell, who is magical, gets the proper shout out, is write a novel in 100 days, not write a book. And that's something that maybe her husband should know a little bit more intimately, but for the busy writers who need a guide, it just be very clear, not in her book, no. But in the promotion of it, just want to be clear. If you're a busy writer and you need a guide to finishing your book fast, how to write a novel in 100 days is the one you should go pre-order from Kim Purcell. We love Kim Purcell, shout out to Kim.
Speaker 2:
[19:21] She's yelling at me from the background here now.
Speaker 1:
[19:22] Well, she doesn't have to yell to thank us, Gavin. She should do what everybody should do. Then we'll round the discussion. She should like and or subscribe to this very podcast on YouTube. Click the thumbs up, click the bell, subscribe, it costs you nothing. If you want to give us money and dominate us financially, bring it on. We got a Patreon. You can buy us a coffee. You want to leave a comment below, do it. A five-star review, mwah. Leave a five-star review for Kim Purcell's book. The point is we need your time and attention.
Speaker 2:
[19:53] That's right. Fendom is Fendom a thing? Is that a term people say? Fendom?
Speaker 1:
[19:58] Oh, okay, good.
Speaker 2:
[19:59] Well, Fendom us then. Fendom us all.
Speaker 1:
[20:01] All of those communities. Oh, by the way, by the way, just a quick roundout. So the image model, again, we said like, oh, this is for everybody. Literally it's for anyone now. Like you can make incredible infographics with it. It works across multiple languages. It's very good. And there's a free version. At Asian languages now, there is a free version. You can do 360 imagery with it, like panoramic photos. I think we're only scratching at the surface of what this model can do. And it's early days. It was just released at the time of this recording. So excited to see what you guys create as well. So please hop in the Discord and share some of those creations.
Speaker 2:
[20:33] Oh, please.
Speaker 1:
[20:34] Or link us your best prompts below.
Speaker 2:
[20:35] Yeah. There are people already sharing the Discord, but jump in there. Kevin, there's another big piece of news that came out, which is a kind of not surprising one, but in some ways I was still kind of taken back. SpaceX is going to purchase or do a $10 billion deal with Cursor. And if you know what Cursor is, Cursor is the company that we've been following for a while. SpaceX is an agentic coding company mostly. They, in the past, were kind of a wrapper for other models where they were providing models that would be really good at coding and then they had a kind of a harness around it. They are now kind of entering a world where they are taking open source models and making their own models. But SpaceX, XAI now with all one company, is going to bring them in house. And Kevin, the interesting thing about this story to me was we haven't talked a ton about in this show, but the XAI founders have all left XAI. Did you know this, everybody that was there in the beginning stages, I think they've tracked them all now, have all left. So there might have been a thing, and there's rumors going around that Elon was pretty unhappy about the latest version of Grok. And what's interesting here to me is that this is kind of the consolidation of the AI space. Now it's a huge, huge windfall for the Cursor team still. I mean, it's a company that didn't exist whatever like three and a half, four years ago. To have a $60 billion potential acquisition is a big deal. But this is that kind of like we're all kind of seeing the gathering of the powers, right? There's the Google power, there's the OpenAI power, there's the Anthropic power, and there's the meta and all these other areas. It looks like Elon is trying to kind of bring forth to make sure he doesn't get left behind in this way.
Speaker 1:
[22:07] Well, how is he going to compete with all birds? Because that's the story we haven't really. I mean, that thing popped and dumped so quick. What happened, Gavin?
Speaker 2:
[22:17] Well, you want to tell people what that is because people listening might be like, you're talking about those terrible shoes?
Speaker 1:
[22:21] What was that? I'm talking about the shoe company. I had not worn them, so I cannot confirm or deny if they're terrible. But this would literally be like if Crocs announced they were getting into machine learning or maybe...
Speaker 2:
[22:32] Are they? I would try Crocs AI in a second. I would probably crock it up. Crock me, baby.
Speaker 1:
[22:37] If there was a crock averse, like if there was an augmented reality, like if Crocs bought Horizon Worlds and it was just Horizon Worlds, but everybody wears Crocs, first of all, they'd have to have legs, which would be amazing, and feet.
Speaker 2:
[22:49] I was going to say, yeah, where would they put them?
Speaker 1:
[22:50] But that's maybe they're just floaty Crocs, like Rayman Crocs.
Speaker 2:
[22:53] Why do Crocs not have hats? That's what I'd like to know. I'd wear a Croc hat, right? It would have little holes in it with rubber.
Speaker 1:
[22:58] Wait, is a sock a Croc hat, if you think about it? Cause it comes out the top of the Croc, it looks like a top hat, a sock would. A Croc sock? Okay, let's keep going. Suno, by the way, is responsible for the number one music download app in the world. What? Huh?
Speaker 2:
[23:16] Yeah. Isn't that crazy? So this is a small piece of news, but this Suno app right now is the number one music app in the world on top of Spotify, on top of Apple Music. So there's been a lot of stories lately about how people have been seeing AI music at the top of charts and all sorts of stuff. But just to know that the Mikey Shulman group, those guys over there, are doing something that people really love doing. And I know that AI music might be one of those things that you may not have a lot of friends that do it, or maybe you do, but it is a thing that has penetrated in a way that isn't a small thing. It has become a much bigger thing. I have this weird anecdote that I played a song from Suno V5, the most recent version, the other day for my family. And they didn't know it was AI. And I played it, did that thing, which we've all done, which is like there was an argument happening between my wife and my daughter, and I made a song, but I played it on the car radio. And everybody's like, wait a second, what is this? And they were just like, what? So some people, when you hear the newest version of them, just kind of blows them away. Anyway, big thing for Suno. Congrats to those guys.
Speaker 1:
[24:17] And that, by the way, is one of my favorite use cases. If you're on like a road trip with the friends or the family or whatever, like make little moments, like just instant generate dumb songs about the things that happen because the time that Jeremy shot milk out his nose will be a banger that like might be like, actually it might be a totally mid song. It doesn't matter when you hear it, it's going to bring you right back to that moment in a way that a photo may not. So, highly recommend.
Speaker 2:
[24:42] That's fantastic. Also, Kevin, this week, I spent some time this weekend with a new thing that came out of HeyGen and it is not digital avatars, it is a new thing called the HyperFrames. What this is, it's an AI editing harness that allows you to do motion graphics design with things like Claude Code or Codex. Kev, the one thing that I have wanted forever and no AI tool was ever able to deliver was, I don't know if you watch a lot of explainer videos on YouTube or know that world where you see the quote pop up, the quote kind of highlights, and then you see like a yellow highlighter go over it. When you're making a YouTube video, it's like a thing you need all the time because you'll say a quote and then it'll appear and you'd rather have a graphic somewhere to show it. Sure. I got within about probably like three or four prompts. It wasn't like an instant prompt, but I got a version of this to work. I have to tell you it was the beginning stages of an opening of my brain because you and I have both spent so much time in the TV production space. I was like, After Effects always felt like something that was just gobbledygook to me. I would go in, I would try to learn a couple of things, but it is so many weird things to learn. And this was a great example of like, oh, I can just go in and I can tell cloud code, hey, this is the kind of thing I want. Didn't get it right away, but when we worked together and figured out it was able to do it. And now I'm going to create this list of things that will basically, whenever I go to edit my own YouTube video, I'll be able to drop in and use. It's just a really useful case of AI.
Speaker 1:
[26:07] Yeah, first of all, if you can make a list of skills that achieve the visual effects that you're talking about here, immense value there that I'm sure everybody else would love as well. When you say it took a few prompts to get to, do you mean you had to build off of the progress from each prompt or you had to modify your prompts that fully understood the effect?
Speaker 2:
[26:25] No, it was all conversational back and forth. So the first time I think it came out, basically it got the line too big and it cut out the wrong thing. I also then gave it an example. I said this is the Vox style thing, like vox.com was the people that originated this. So I took it and gave it that. Yeah, exactly. So once you give it a couple of examples, it starts to understand. It is a little bit of back and forth with it still. To get it right really at the end, I needed to make sure I understood it. The coolest thing though is what HyperFrames does is it pops open essentially a window where you can see a bunch of stuff. It is not like re-motion in that it's popping up a whole editing soft suite because I don't know if you need it because you're doing this back and forth. Anyway, very cool. I had one other example this weekend where I was installing Comfy on another computer. One of the coolest things that I got Claude Code to do was Claude Code basically wrote an entire Comfy workflow and worked perfectly great. That's the sort of thing that you can now do with these coding services. Very cool. Go check out HyperFrames. It's fantastic. The new image model is out now. Play with it. Send it to us in the Discord and share everything.
Speaker 1:
[27:32] HyperFrames, super amazing, but you know what? Not as amazing as you, Gavin.
Speaker 2:
[27:37] Oh, thanks, Kevin. And not as amazing as you, the viewer, because if you made it this long, you are our true audience. I really thought that was going to be about... Just say the word something.
Speaker 1:
[27:48] I thought that was going to come back to me.
Speaker 2:
[27:49] No, it wasn't coming back to you. Tell the word Kevin's beautiful. Bye, everyone. Goodbye. Please write it. Write it a thousand times.