title The Agent Era: Building Software Beyond Chat with Box CEO Aaron Levie

description Erik Torenberg, Steve Sinofsky, and Martin Casado speak to Aaron Levie, CEO at Box, about what happens to enterprise software when agents become the primary users. They discuss why coding agents succeed where other knowledge work agents struggle, what abstraction layers mean for the workforce, and how data access and systems of record must change in an agent-first world.

Follow Aaron Levie on X: https://twitter.com/levie

Follow Steve Sinofsky on X: https://twitter.com/stevesi

Follow Martin Casado on X: https://twitter.com/martin_casado

Follow Erik Torenberg on X: https://twitter.com/eriktorenberg

Check out everything a16z is doing with artificial intelligence here, including articles, projects, and more podcasts.
 
Please note that the content here is for informational purposes only; should NOT be taken as legal, business, tax, or investment advice or be used to evaluate any investment or security; and is not directed at any investors or potential investors in any a16z fund. a16z and its affiliates may maintain investments in the companies discussed. For more details please see a16z.com/disclosures.


Hosted by Simplecast, an AdsWizz company. See pcm.adswizz.com for information about our collection and use of personal data for advertising.

pubDate Tue, 21 Apr 2026 17:15:00 GMT

author Aaron Levie, Erik Torenberg, Steve Sinofsky, Martin Casado

duration 3577000

transcript

Speaker 1:
[00:00] The diffusion of AI capability is going to take longer than people in Silicon Valley realize.

Speaker 2:
[00:04] It's just absurd to think you're going to vibe code your way to like SAP. All of that domain knowledge, it's not just represented in some well-orchestrated data layer.

Speaker 1:
[00:13] The engineering compute budget conversation is going to be the most wild one in the next couple of years.

Speaker 2:
[00:19] The biggest problem right now is everybody is trying to figure out the economics of all of this. When they're off by at least an order of magnitude on how big the opportunity is.

Speaker 1:
[00:29] If you have 100 or 1000 times more agents than people, then your software has to be built for agents.

Speaker 3:
[00:33] People in the abstract say things like, now you're marketing to agents, you're like an API, you've got a good idea. I actually think that's almost exactly wrong.

Speaker 1:
[00:41] Wow, this is breaking podcast news.

Speaker 4:
[00:44] Every major technology wave promised to eliminate the middleman, marketplaces would dismantle hotels, SaaS would replace on-premise, but the taxi medallion was the only real casualty. The layers persisted because they encoded organizational logic, not just software logic. Now, agents are arriving and the assumption is the same. They will flatten everything. But the first enterprise teams deploying agents at scale are discovering something different. Agents do not want simpler systems, they want better ones. They choose backends based on durability, cost parameters and reliability, not interface polish. The question for every software company is no longer whether to support agents, but what it means when agents outnumber employees a thousand to one. I speak with Aaron Levie, CEO at Box, alongside a16z board partner, Steve Sinofsky, and a16z general partner, Martin Casado.

Speaker 1:
[01:48] Do you start to imagine that we all have to build software for agents? I think we're all clear on that, right? So that trend is happening, which is we spend as much time now thinking about the agent interface to our tool as we do the human interface. Sure. Okay. And the reason we're doing that is because our hypothesis would be that if you have a hundred or a thousand times more agents than people, then your software has to be built for agents. And then what is the way that those agents are going to interact with your system? It's going to be through an API or a CLI or MCP or whatever. And the paradigm that appears to be taking off and is quite successful so far in terms of efficacy is what if you give a coding agent access to your SaaS tools and a coding agent access to your knowledge, work, sort of workflows and context. And that kind of becomes the superpower, which is the agent is not only capable of reading some data, understanding some information, it can actually code its way or use APIs through whatever task it's trying to achieve. That appears to be like a paradigm that is starting to compound. And that's the Claude Cowork phenomenon, that's the whatever OpenAI is kind of cooking up with the super app, perplexity computer, etc. And I actually think it kind of makes sense as like the ultimate manifestation of this stuff.

Speaker 2:
[03:04] I mean, I think you're right. It makes sense in a theoretical way. But in a practical way, we have to be really careful in that the way to say it is algorithmic thinking is really, really, really hard for the vast majority of people who have jobs. And so the easiest way to think about it is if you were to go into any person and ask them to create a flowchart for a particular thing that they have to go do, they would probably fail at producing that flowchart. So within any organization, say, doing a marketing plan and there's 50 marketing people working on a giant product line, one person probably understands and could document the flowchart.

Speaker 1:
[03:44] 100 percent.

Speaker 2:
[03:45] So if you put one of these agents or you put this tool, this co-working tool in front of people to create these things, their ability to explain to it what to do is really, really limited.

Speaker 1:
[03:58] 100 percent. But what if that becomes the new, this is the new way you have to interface with computers, and you just have to cycle that through?

Speaker 2:
[04:05] Well, then you're basically just developing the next abstraction layer for how people interact. And developing an abstraction layer has historically at each level of the abstraction layer been a highly skilled, very specific individual within an organization developing that. And then the little parts that they build just become little toolets in the world of people doing particular tasks. And some people are able to stitch them together and some can't. But that happened with paper clips and thumbtacks before. And I think it's going to happen with whatever we do next.

Speaker 1:
[04:38] I think basically the timeless part is the job just moves up a rung and you learn a new set of skills. And that's why I actually don't think anything about this is any different. It's just now the leverage you get is obviously fantastic. There was this viral kind of tweet that went around, which was the Anthropic Growth Marketer. Did you guys see this? Basically one person and he was using cod code at the time to basically more or less automate what maybe five or ten people would have done in various kind of silo jobs. And I think the reason why it's interesting is you had to have been a systems thinker to be able to accomplish that. So clearly he already was technical enough to be able to pull it off, but it did kind of represent what would each of these jobs look like if you had like, imagine you had X job in the economy and right next to that person was an infinite pool of engineers that could automate whatever that person wanted. And what would that job look like in the future as a result of that automation that now is possible? Yes, I agree that you'd have to find a way to think through your job as a system to be able to pull that off. Maybe the agent gets better and better over time at being able to nudge you in that direction. But it does sort of stand to reason that you will start to try and automate a lot of that kind of work. Well, why don't I take the keywords that are working in Google AdWords and then port them over to Facebook and make sure that those are replicated and then take in the new signal from what's happening in the market.

Speaker 2:
[05:56] That's a big leap. One thing first.

Speaker 1:
[05:58] I almost had you. You were nodding a little bit. I said something that went too far.

Speaker 2:
[06:02] The anthropic growth person, as an example, that's a job I could do. I could do that job.

Speaker 3:
[06:09] The demand is infinite. You've got the best thing.

Speaker 2:
[06:11] When demand is infinite and, frankly, supply is infinite, this is not a difficult job.

Speaker 3:
[06:16] And so, let's... Let me think either. The petrol pump in Australia right now is amazing. Right, right.

Speaker 2:
[06:22] So, instead, be the $600 PC marketing person and see how you can do against the NEO. That's a real job.

Speaker 1:
[06:28] All right, fine. We need a better example.

Speaker 2:
[06:30] But there is... I mean, it is really interesting. Like, let me do an old example, an old-person example. Like, my cousin, MBA, elite school, joined her first job, she's a little older than me, joined right on the cusp of computing. Like, she actually didn't use a spreadsheet in grad school. And then, the spreadsheet showed up, but she wasn't a spreadsheet person. So, instead, they told her hire as many interns as you want. And so, her first year on the job, she, like, supervised, like, essentially, a whole room of agents. And the kids, who was me, not literally, but they were in college, came and just did all the spreadsheeting. But then what happened, sort of magically, over the next couple years was, she and her cohort all became the spreadsheet people. And then this idea that you, being a manager in a bank, or just two years in, meant you had a cadre of people doing the spreadsheet. The whole abstraction layer moved up. And the old job before those interns was, you just sat there with basically calculators and an HP calculator, figuring out the model for some M&A deal or whatever, and you only got to do like two iterations before you had to put out the pitch deck or just go to the customer or the client or whatever. And then all of a sudden, they're doing 30 iterations themselves. But they see, and so I think where we are with agents is just at this step where you think you need 50 and the abstraction layer is such that we're dividing up in these really small pieces with one super smart person coordinating them all. And pretty soon, that whole thing is just gonna, they're all gonna collapse on each other. And there is just gonna be like a skillset amount of code, call it an agent, that is like marketing-ish. And you'll be able to ask it marketing stuff. And then the next step will be and have it go do things. I'm a little skeptical of the, until the whole like non-reproducible, non-random element of this AI stuff goes away, the doing stuff is gonna get very costly. And so then you get into the human in a loop discussion and all of that. But I feel like when I talk to people trying to do stuff, that we're right, I feel like I'm at Thanksgiving dinner talking to my cousin six months in her job, when I'm using a spreadsheet already, and I'm like, I don't know why this is so hard. You should just use one. And then two years later, she's doing it. And I think this right now, you have to be an absolute, you have to be a rocket scientist and the growth marketing person to create 42 agents and spin them all up and do all of this stuff. But the rocket science part of it just is going to evaporate in a very short order. And then you're talking about, wow, there's a giant chunk of domain expertise.

Speaker 1:
[09:06] Yeah, it goes back to the domain expert.

Speaker 3:
[09:07] So I actually think something that you said, I'll take the other side of, which is, I think it's very tempting to be like, these agents are going to code and do X. But I think we're going the opposite way. So I think actually where we started was, we'd take a piece of SaaS software and we'd add AI. Then that's the new AI enabled. So that's the extreme version of using code for these types of things. But now what are we actually doing? We're like, okay, the SaaS software is still SaaS software, and the agent uses it as a computer because it's actually very good at that. So I'd say we started with code, then we went to the terminal, which is actually less code, and now this year is going to be the year of computer use. So it's almost like they're much more like humans using computers, than them generating code, and that feels like very much like this mezzanine step. I actually come from the generating code type of the world. I would argue that that's happening less, not more.

Speaker 1:
[09:58] Yeah. So to me, whether it's computer use, API use, or writing code on the fly, I maybe erroneously put that all in one category.

Speaker 3:
[10:08] Well, they're very different.

Speaker 1:
[10:09] They're very different, but we have an agent that we're working on, where it just makes a determination whether it should use an existing skill, is it using an existing tool from box, or it should write code to solve that problem. And its ability to do any one of those three at any moment ends up being incredibly useful because sometimes there's just some specific operation you want to be able to do. We're writing code to be able to do that operation is just faster. And we can't possibly pre-plan for everything that anybody would ever want to do on their documents. And so the fact that the model is good enough to also write code on the fly for that use case, ends up just being an amazing property. Even though maybe 90% of the things that it's going to do should just be using an existing API.

Speaker 3:
[10:51] And over time, Preto takes over. And over time, there's literally like seven apps on her iPhone. There's seven SaaS apps we end up. Like over time, these things tend to consolidate.

Speaker 1:
[11:00] But the seven apps on the iPhone is an issue of humans don't want to learn these things over and over again. And so I, as a human, I don't have the mental bandwidth to learn that many apps. But an agent that is going to use tools and APIs and be able to code things doesn't have any of the same constraints that we have. So I don't know, like I don't mind.

Speaker 3:
[11:20] You could argue that there's just so many things to do and you can make interfaces sufficiently general.

Speaker 1:
[11:25] Yes, fine.

Speaker 2:
[11:25] Well, fair. Let me say, I think I like what you said then because...

Speaker 1:
[11:29] Oh, I'm back.

Speaker 3:
[11:30] We're back.

Speaker 2:
[11:32] We're aligned. So open. No, but I think there's something super interesting here, which I do really, really like, which is that where software has evolved, you know, like I use SAP all day. I work in finance. I have to go and generate all these reports. And then somebody shows up and says, I want a report that does this view sliced this way. And I'm like, oh God, I don't know how to make that. And like, now let me go wade through the SAP help system and try to find it. One thing that, let's just say AI could be very good at is it actually can navigate that surface area much, much better. You know, the help is all there. And so it's a matter of finding it, mapping language. And humans have been a bottleneck in tapping the past 25 years of software capabilities. I mean, like, I spent my life, my life with sitting next to people on airplanes saying, how can I make PowerPoint do X? Just go to the ribbon. And, you know, it was because it hurt, physically hurt to watch somebody suffering with bullets and numbering in Word or trying to figure out, you know, like, oh, let me just make a two-sided, a two-axis graph in Excel, which like is rocket science. Like, almost no one can do that, but yet it's super common. And so people are like, have not, and so that impedance mismatch was a human user interface design.

Speaker 3:
[12:50] On the consumption layer, I totally buy it, which is like the perfectly fluid, like UI or consumption layer. I just feel the back end, like the systems of record, it'll probably converge into like some database, like some generic set of APIs, like that they'll connect to. And like that seems to be the direction it's going.

Speaker 2:
[13:09] I agree, you go ahead, sorry, sorry.

Speaker 3:
[13:11] So I spent all weekend implementing my NanoClaw bot. And when you first start out, it's like you're building an integration for everything. NanoClaw is very, like OpenClaw has all of the integrations, NanoClaw has a few of them. And so you have it build all of its own tools. But after two or three days of these, like you kind of have the tool integrations that you need.

Speaker 1:
[13:31] Yeah, but back to the essay, we're talking about personal productivity probably, like you're organizing your life or something.

Speaker 3:
[13:37] Well, it's work productivity.

Speaker 1:
[13:38] Okay, fine, work productivity, and then an SAP system. And so there's an infinite amount of complexity when you get to, okay, some company that has a global supply chain and they're dealing with 75 pieces of information across 30 different systems, that does require a certain amount of horsepower from the agent that is just, I mean, we just haven't been able to get from any architecture up until now.

Speaker 2:
[14:04] But what you just described is literally what IT has been doing for 50 years and will continue to do, which is, yeah, I have a friend who was the CIO of the VA, and all he spent his time on was gluing the 75 VA systems together, and it's all just integration redundant.

Speaker 3:
[14:24] Perfect. For this, I told you.

Speaker 1:
[14:27] Okay, great.

Speaker 3:
[14:28] For integration, these things are the best, but it's integration. It's literally, how do I stitch these two systems together?

Speaker 1:
[14:33] But now the thing that I think is happening is, it's kind of like integration on demand. Yeah, totally. It's my new query in the system that the IT team didn't pre-wire. Now I need it to happen at runtime.

Speaker 2:
[14:46] Let me get off my lawn.

Speaker 1:
[14:48] Okay.

Speaker 2:
[14:49] Okay. So the reason, I just was in a room filled with a bunch of CFOs and CIOs, and they all looked at me when I said something along these lines, although not as optimistic as you can imagine. But they just, they like, no, it caused like six of them to come running up afterwards and say, you're insane. You've lost all credibility with me. Because it's back to-

Speaker 3:
[15:11] What specifically that the agents are going to do integration?

Speaker 2:
[15:14] That the integration is a problem that will get a lot easier.

Speaker 3:
[15:18] Yes.

Speaker 1:
[15:19] They were against that?

Speaker 2:
[15:20] No, they're no one's against it.

Speaker 1:
[15:21] I know.

Speaker 3:
[15:21] They think it's practical.

Speaker 2:
[15:23] But their fear is like unleashing, not just the agents themselves, but humans to do integration. Because you put people creating new integrations, and you just say, please break my system of record.

Speaker 1:
[15:36] Oh, yeah.

Speaker 2:
[15:36] So this idea that you just create a new API between system 27 and system 38, and then that might be fine for a report, because if that person wants to be wrong, that's their business.

Speaker 1:
[15:48] But you're not going to- I think we have a read-only version of this for a number of years before.

Speaker 2:
[15:53] Where n is very large.

Speaker 3:
[15:56] And a lot of it is just the consumption layer where the consumer is a human being. It really feels right now, a lot of the AI stuff is consumption layer.

Speaker 1:
[16:02] But yeah, I mean, we actually have, so we just rolled out the official box CLI. Thank you for liking the tweet on that.

Speaker 2:
[16:12] I used it. I have some feedback.

Speaker 1:
[16:13] We'll talk about it. I'll take all the feedback. But it's a really interesting thing. So we had all these debates internally of like, okay, you give Cloud Code the box CLI, and you can now interact with your entire box system via natural language, and you get the horsepower of Opus 4.6 being the orchestrator of doing a bunch of operations. And it blows your mind. I guess I'll get some feedback, but it blows your mind in some ways because you could just be like, upload this entire folder from my desktop in the box and it'll work, or process all these documents in this folder and it'll work. And it's amazing. And then we started thinking through like, well, let's say you were a company with 5,000 employees, and everybody had access to some shared repository, like engineering documentation and marketing assets or whatever, and everybody had Coddex or Codex running with the CLI. Wow, we now have some really interesting new challenges, which is like, how do you coordinate possibly the fact that you might be hitting this system 10,000 times an hour or something? Not from a performance standpoint, but just like, how do you make sure that people didn't move like a file from one thing accidentally from one folder to another folder while the other person is trying to do a write operation and somebody else was trying to delete something because you should have these agents running wild. This is going to be like the new big question that every CFO, CIO, etc. is running around trying to, with their hair on fire, trying to figure out.

Speaker 2:
[17:37] That's exactly what I ran into, which is I played around with your example, which is create, the video example, which is create a marketing plan directory or something. And all of a sudden, I'm in some loop creating directories. Yes.

Speaker 1:
[17:48] And it's going to go on as long as it can.

Speaker 2:
[17:51] Right. And I was like, I wonder what the limit is on Vox for nested directories because I'm about to hit it.

Speaker 1:
[17:56] Actually, we're going to find out too.

Speaker 3:
[17:59] But it does feel to me that a lot of the intuition is to build a new layer of controls and whatever, but what's actually happening on the ground is the opposite. I'll give you an example. When we all picked up a lot of these personal agents, we would give them our API keys, we would give them our email addresses, and then they would access those things. They're like, oh, but how can I stop it from whatever? So what everybody's doing now is you give it its own phone number. I actually gave my NanoClaw its own credit card.

Speaker 2:
[18:30] Hopefully, it's just a Visa debit card that you bought at CVS.

Speaker 3:
[18:35] All the money. But then I gave it its own Gmail account, which you can log into. And then Gmail actually has all of these RBAC permissions. So you could make an argument that we've actually built in a lot of these permission systems. You have to treat it like a human as a separate human. And then instead of building another auth layer, bring another...

Speaker 1:
[18:54] Okay, now can I instantly do a take down of this element that we're gonna run into? Okay, so that is fantastic for personal productivity. And the question that we're gonna run into is in an enterprise. Let's say I have... Let's just make a simple example. I have a 50-person team of something. Should everybody also... Basically, will we have 100 people now collaborate... I mean, basically 50 humans...

Speaker 2:
[19:20] And then 50 credit cards.

Speaker 1:
[19:21] And then 50 agents in that same shared space. And do I have... I obviously have complete oversight over my agent. But what if my agent collaborates with somebody else and then accidentally gets access to some resource because they were sharing with the other person and I'm not supposed to have access to that resource and now this autonomous sort of stateful agent is running around working on somebody else's information.

Speaker 3:
[19:45] The default end-to-end argument is you treat them like human beings.

Speaker 1:
[19:49] It doesn't work. So you can't fully treat them like humans because here's the thing. And with regular humans, you don't get to look at the Slack channel of the person that is working with you or working for you. You don't get to log in as them. You don't get to oversee them. They are accountable for their own set of execution in the real world. You don't get penalized for how they screw up. The agent, you have all the liability of whatever they're doing. You do have complete oversight, and you're probably going to need to have that complete oversight. They have no right to privacy. So there's going to be some of these breakdowns that aren't as clean as just treat them like a person because I need to be able to give access to something to them, but I also need to be able to log in as them at some point and be like, no, no, you fucked up the whole thing, and I need to undo it all. But if I can log in as them, how could they have operated in the real world working with other people and keeping anything confidential or secure or whatever? So it really is still an extension of you. It's almost impossible to get around them being an extension of you. So now the thing that we're thinking through, that we're not going to be able to do anytime soon.

Speaker 3:
[20:52] It just doesn't logically follow. Maybe, but for example, for my employees, I can log in as them.

Speaker 1:
[20:58] You don't though.

Speaker 3:
[20:59] I can get access to their email.

Speaker 1:
[21:01] Yeah, if you get sued, you're not logging in as them on a regular basis because they sent one email.

Speaker 3:
[21:08] But isn't the right operating model with an agent? The same thing.

Speaker 1:
[21:11] The risk is a thousand times greater. These people, they will just leak your information whenever they want. They will happily just go and send some email to somebody because they got prompts injected.

Speaker 3:
[21:20] I think the terminal state is that these things are still these sloppy computers and therefore they will always...

Speaker 1:
[21:24] I don't like the word sloppy, unless we're saying it very in a colloquial sense, but like...

Speaker 3:
[21:29] They will never be able to contain information. They'll never...

Speaker 1:
[21:33] So, like, I think the ability for you to keep something in the context window a secret, like as in like you tell it, do not reveal X thing in the context window, I think that's a very hard problem to solve. And so then, so then thus, if anything can ever enter that context window, because they have access to a resource, then in theory, you should assume it can be, you know, prompt ejected out of the context window. And I don't know that we know of a way to solve that at the moment. Like, that's... Like, and so, so if I know your new agent's, you know, email address and I email it like it's an assistant, but like I can social engineer it 10 times easier than a human, like it'll be hard for you to pull off that that agent is now also has access to your like M&A documents and stuff.

Speaker 3:
[22:17] But isn't this like literally all of AI right now?

Speaker 1:
[22:20] Which part?

Speaker 3:
[22:21] I mean, the fact that we've got these shared systems that we use the intelligence for that have shared context.

Speaker 1:
[22:27] But what do you mean by it's all of AI?

Speaker 3:
[22:28] Well, I'm just saying like right now, when we use AI internally and agents internally, this is exactly how we use them.

Speaker 1:
[22:33] But this is why you were there working as you effectively right now and we don't yet know how to make them not work as you.

Speaker 2:
[22:40] Let me offer an example.

Speaker 1:
[22:42] Let me offer an example and then solving this problem though. The issue will be like you will just be able to trick the agent to reveal information. So then that's why having them have access to their own resources where they can fully make their own decisions is not yet something that we've been able to pull off.

Speaker 2:
[23:03] There's a perfect example for solving your problem, which is we already lived through this with open source. The model for open source was it's all there, you just use it and you pick and choose. Then nobody debated it because the world was much smaller then, and we weren't all on X doing podcasts when this was all happening. But then quickly everybody realized all the problems you were just talking about. If you're running a big company, you can't have some person just go copy in a bunch of source code from open source into your commercial product like that. There was a whole licensing problem, a whole bunch of stuff. So all these norms got developed. The debate that's happening right now is just this really interesting modern artifact of how new technologies develop, which is this is all happening in real time. During open source, we met in a conference room this big and debated how much open source we could use in Windows or Office. And nobody on the Internet knew we were having this debate. And I think it's just so interesting that not just the debate about specifics, but this whole notion of where is this heading is happening in writ large. And everybody is just trying to get to the end state, like way, way more, like in a sense more quickly than we can actually reach the end state. And so what really needs to happen is people just need to go build...

Speaker 1:
[24:26] We need standards.

Speaker 2:
[24:27] What?

Speaker 1:
[24:27] We just need some standards.

Speaker 3:
[24:29] I think you've got different intuitions on the end state.

Speaker 2:
[24:31] No, no. You don't want my intuition, but like...

Speaker 3:
[24:34] One could make an end-to-end argument that these things actually converge on the same type of reliability as a human being, which is exactly how we view like self-driving. And in that case, you use the exact same mechanisms that we use to protect with human beings. You consider it an insider threat. You consider the fact that people can be bought off. You consider the fact that people make mistakes.

Speaker 1:
[24:53] And that's just a risk.

Speaker 3:
[24:55] And that's operational processes. So one intuition is like, that will be the end state.

Speaker 1:
[25:00] There's another intuition. I'm talking about where we're at now. I don't know that we disagree on the end state. And by the way, strategically, we're hedging, because we're going to build agent users and regular users. I love the idea of OpenClaw having a box account, and it operates and you share with it.

Speaker 2:
[25:18] You just like twice as many accounts.

Speaker 1:
[25:20] Double the seats? No, I love it. I'm just saying, on the ground right now, we don't yet know how to give it an M&A data room to fully, securely be able to...

Speaker 2:
[25:30] But it is harder than that though, because the threat...

Speaker 1:
[25:34] He's the skeptic, I'm the skeptic.

Speaker 2:
[25:35] The threat vectors are going to be way more sophisticated. So we do have a cat and a mouse game going on, where you can't just assume that the agent acts like a human does today, because it's going to be the fastest, most thoughtful, craziest-ass human that ever existed trying to actually leak the information because it got injected in some way. And so part of what's going to happen is, we're going to go through this phase where the enterprise customers are just going to close everything off until there's some sense of sanity in all of this. But in the meantime, the individual and specifically the developers-

Speaker 1:
[26:14] We're going to have such a big gap.

Speaker 2:
[26:17] That, I think, is the most exciting tension that's going to happen, is that the enterprises are going to get left behind by these sort of advanced individuals, which will then start to look like the start-ups. And the start-ups will start to move much, much faster than enterprises because they just don't have any of these problems. You could end up with the agent going rogue in a start-up and doing that.

Speaker 1:
[26:41] And it's fine because you had no asset to begin with.

Speaker 3:
[26:43] But you have employees that go rogue routinely in the start-up.

Speaker 2:
[26:45] Yeah, well, it'll just be an episode of Silicon Valley. And so, you know, big deal.

Speaker 1:
[26:49] I agree with you on like the, okay, it's people, etc., the same risk. I think there's a couple differences though, in the sense that I can't really threaten the like, cloud code that it's just, I'm going to pull the plug on it. In the same way that you do have that threat as a regular employee is like, you at least like 95% of people are not, you know, trying to do bad stuff, you know, within an organization.

Speaker 2:
[27:13] They're not trying, but the ability to inadvertently do bad stuff. To your point about still not having that stuff fixed is real.

Speaker 1:
[27:21] I would argue that it's a lot easier to have people not share, let's say, files with somebody outside the company in a wrong way more than it is for an agent right now to have the same set of instructions.

Speaker 2:
[27:33] And also you have the tools so that you can basically stop that at a whole different level of abstraction.

Speaker 1:
[27:37] Which is why you have to build this into software in a lot of cases. But I do think actually if you were to like put a bow around your last point, a lot of this is actually why the diffusion of AI capability is going to take longer than people in Silicon Valley realize. Because what's happening is like we see startups that can start from the ground up without any of the risks that we're talking about, because they have nothing to blow up. And so we look at that as the trajectory that we're on. And then you go to like JP Morgan and you're like, how are you going to set up NanoClaw to be able to actually like, automate your business anytime soon? And it's like, oh, OK, there's going to be like a little bit of a gap there.

Speaker 2:
[28:15] Well, what do you guys think? I think that that opens up a pretty interesting problem, which is this split between big and small startup and enterprise, which is just that the current SaaS vendors who are all struggling in this SaaS apocalypse weirdness that I don't really agree with, but they are struggling with this problem that they don't really sell the line of business data. They actually sell this intelligence and domain expertise in this whole system. And the agent side of things wants to only buy the data now. And they only want to license the data, and they want to have unlimited access to the data. But they've actually never really enabled that. Like that's never been their business. And it's been a long standing tension point with the likes of Workday and SAP and stuff, how much API access to have. I mean, Salesforce went through three different massive platform redesigns. I think that that's a particularly interesting problem. Not for the same reason that Wall Street does. Wall Street's all wrong about the economics and the problem and all that stuff. But from a technology perspective, what does system of record mean in the face of people wanting to access the data? When the data...

Speaker 3:
[29:27] For training or for...

Speaker 2:
[29:29] Well, they are...

Speaker 3:
[29:30] What you're talking about from like actual day-to-day operations.

Speaker 2:
[29:32] I think of it as executing the day-to-day operations. Their concern is that somebody... That they want to do the training layer on your data. Like, I'm a big customer. They want to do the... My vendor wants to build a training...

Speaker 1:
[29:44] Actually, even if you don't even get into training, they're concerned because...

Speaker 2:
[29:47] Oh, yeah.

Speaker 1:
[29:48] Because like monetizing, you know, sending a little bit over the internet versus like you're in my UI. It's a very different level of monetization initially that you receive.

Speaker 2:
[29:59] But that's sort of that monetization part is the Wall Street point. Because I think like, look, there is so much domain stuff in an SAP, just to pick an example, not to pick on them or anything. But like, they're not going anywhere. Like, it's ridiculous. It's just absurd to think you're going to vibe code your way to like SAP. But also, all of that domain knowledge, it's not just represented in some well orchestrated data layer, much as they tried. There's like a whole bunch in the UI, there's a whole bunch in middle tiers, there's a whole bunch in just how you use it. And so I'm really unsure how this thing evolves, because SAP isn't going anywhere. So then that's going to slow the diffusion of AI on that particular data source, independent of whether or not it's a gentified AI that's doing stuff or just read-only reporting on stuff. So where do you come down on it? Where do you think that's going to go?

Speaker 1:
[30:52] I'm afraid of saying something that... Well, I want you to say something.

Speaker 2:
[30:57] Otherwise, you're not going to get invited back. So say something good.

Speaker 1:
[31:02] I think I've drunk the Kool-Aid on build something agents want. So this kind of the Paul Graham term kind of like emerged on, you know, the past year on this topic, which is just like, like event. I think we would actually then I fully agree on this, which is at some point you do enough sort of iterations of this. And at some point, the agent is largely in charge of what tools it wants to implement and use and whatnot. And yes, it can't, the agent is not going to be able to change out an enterprise system. But like, again, enough generations later, the agent might might just run into so many walls with your software that it's just going to say, you need to finally rip out your legacy HR system, or I'm not going to be able to automate this workflow for you. So I do think you have this really interesting dynamic, which is back to this whole point of, imagine that there's a hundred or a thousand times more agent volume on software than people. You do that enough times and eventually the software stack that agents talk to has to be built for them. And maybe there'll be a couple holdouts, maybe a couple ERP systems are like the final holdouts that don't do that. Everything else, your business performance will correlate to how well your agents can get access to the information they need to do their work. And so thus, your enterprise IT stack has to be set up in such a way to support that. And so agents are kind of in charge because basically your software has to support those agents being effective. And then that's going to mean everybody that built a SaaS business or a software business is like, the game is can you build really, really high quality APIs? Can you have a way of monetizing that? Do you have a way of handling the identities and all of the access controls for agents? And that becomes the new problem you have to solve if you're building a software company. And so yeah, and then how you monetize it? Do you monetize it? Does Workday charge a penny for every HR record? A poll is like, we'll figure that out. I do think that in some businesses it could mean less revenue and then in other businesses it can mean a lot more revenue. The thing we get excited by is every agent really loves working with files, so there will probably be more files in the future than there was going to be before. And so can we build a platform that makes it really easy for agents to work with that data? We're betting that that's actually a really optimistic outcome for our kind of business model. There might be some business models that are more constrained because the agent is doing more of the value than the software is in that kind of future scenario and then there will be everything in between.

Speaker 3:
[33:31] Can I quibble with one thing?

Speaker 1:
[33:32] You're going to quibble with that? I thought that was so not controversial.

Speaker 2:
[33:37] We're here to quibble with it.

Speaker 3:
[33:39] There's one thing I think like Paul Graham and many actually gloss over, which is they focus on the interface. They'll say things like you build something for the agent. I actually think that's exactly wrong. In the sense that...

Speaker 1:
[33:49] To be fair to Paul Graham, he had been extrapolated. I brought Paul Graham into this.

Speaker 3:
[34:00] People in the abstract say things like, now you're marketing to agents. The most important thing is to be like, whatever, you're like an API, you've got a good idea. I actually think that's almost exactly wrong.

Speaker 1:
[34:10] Wow. This is breaking podcast news.

Speaker 3:
[34:13] That's the one thing agents are really good at.

Speaker 1:
[34:15] Oh, okay. It's finding their way through.

Speaker 3:
[34:17] At the end of the day, it's the semantics that end up mattering a lot more. And so the agents, in my recollection, or in my experience, are very, very good at picking the right backend for whatever they're doing. So they're not like, oh, the interface for this is very good. It's none of that. They're like, the cost parameters of this, the durability of that. And so they actually have the collective wisdom of our experience using these platforms. Let's take cloud platforms. There's a bunch of cloud platforms out there. And whenever I ask an agent to choose a platform, it's actually using meaningful stuff, not interface stuff. So I think as an industry, we're so focused on these interfaces. Like, oh, you need to market to agents this and that. But really, I think that we're going to be pushed to actually build better systems. And that's what's going to be chosen.

Speaker 1:
[35:00] Okay, actually, so then there's probably no quibbling. I think we're actually a folio line. I'm sorry to ruin the quibble thing. I don't treat this as kind of a marketing-esque thing. I more mean like, if your tool is closed off to the agent, the agent eventually will find a better tool for that company to go use. And so what will happen is, is it used to be that you would go to Gartner to be like, tell me what system to use or whatnot. At some point with enough iterations, the agent is going to say, you should probably use this kind of database for this type of operation. And if you're not in there, then you're DOA.

Speaker 3:
[35:36] And I think we should actually be celebrating this, because agents are actually pretty smart at choosing the right technology. In the past, I really think it was a lot of the other things that caused people to buy it.

Speaker 1:
[35:45] But don't worry, we will, in Silicon Valley, we will ruin the meritocracy of this very quickly. Because you'll just be like, I'm going to outspend.

Speaker 2:
[35:52] Well, the agent, they'll be like an API to incent the agent to get them. But the marketing agent at Workday, well, the marketing agent at Workday will have the ability to purchase the recommendation.

Speaker 1:
[36:04] We will find a way to replicate steak dinners for agents.

Speaker 2:
[36:08] But here's a thing that, again, that happened with the web internally. Internal, just pick internal sites. Every company had file shares with the best documentation, the best slideshows, the best financial models for any department or working area. And people sort of got familiar with that. And then when they didn't find the one they wanted, they created a new one. And many organizations sort of operated like that was essentially a free market. In fact, because before the world of Box, like IT didn't, if it was in a file, they just didn't care. They only cared about if it was in SQL. And so one of the risks with the model you're describing is that the agents themselves will spin up but becomes like a de facto new system of record.

Speaker 1:
[36:52] Oh, they're going to fragment the heck out of it.

Speaker 2:
[36:54] In what the IT people think of as some middleware end-user BS area. And I think that that is a real risk. Is that like, in a sense, like the macros end up running the corporation.

Speaker 1:
[37:10] Yes.

Speaker 2:
[37:10] And so I think that they've seen this movie and they've seen what happens when you let marketing just go buy a website on the internet to do an event. And then it's like a huge security vulnerability and the mailing list is leaked and the whole company gets sued. Totally. And so I think there's a lot more real world tension in this dynamic than we just let on. Yeah. But I also think it's one of these ones where organizations are going to run at different paces. And JP Morgan is going to be the slowest at doing this, and the startups are going to be the fastest. But the Delta is huge, but even the startup one is a little far off, because even startups do need some systems of record. Yeah.

Speaker 1:
[37:51] Oh, 100%.

Speaker 2:
[37:52] And they are going to all start with some SaaS, and they're not going to replace it very quickly. So I think it's a little bit trickier.

Speaker 3:
[37:58] So it feels like there's two very competing viewpoints on this one. And like Elon said, it was like, okay, we're going to issue a prompt and it's going to spit out machine code. And that's basically the collapsing of layer view. Like whatever existing interfaces and layers that we've created in the past are all going to go away. And it's literally like prompt and machine code. The other argument, like the history of systems is layers never go away. They just get layered, right? Because a lot of the layers are actually more of like organizational boundaries or like state boundaries.

Speaker 2:
[38:27] Or compatibility. They stay for compatibility.

Speaker 3:
[38:30] So the other argument is, is like we've actually evolved these layers very specifically because of like more human and organizational needs. They're not going to change and the agents are going to go ahead and map to those. And I tend to be in that latter camp. Like I don't think that we're, I think like systems are going to continue to be used in fairly similar ways. Maybe there's more agents using them. I don't think they're going to evolve as much.

Speaker 1:
[38:50] Elon might be back in the like entropic category of the entropic growth marketer, which is like, he like, you know, over the years when you kind of like study the various IT departments of his companies, like, they are the most, he could do that.

Speaker 3:
[39:08] He can do it.

Speaker 1:
[39:08] He's the most homegrown, like everything is.

Speaker 3:
[39:11] Elon AI would do that.

Speaker 1:
[39:13] But Elon AI would like start his podcast. And then from your mortals, you're like, yeah, we kind of just want a CRM system. That kind of works the same way every time.

Speaker 2:
[39:20] I mean, this is not, it also hasn't been not tried before. Like if you were to look at an ERP system from first principles, you know, well, in 1970, whatever, when SAP started, there were a bunch of different assumptions. And today, you would start from a different set of assumptions about what's important, and you would architect a thing completely differently, but then it would still only last like 10 years until you thought, wow, that was a broken decision. And so I think that there's intentionality in layers, but there's also this first principles thing. And, you know, that always will exist because the decisions you can make at first principles at any given time mandate a whole bunch of different stuff. And so even if you don't go with LIDAR, which made total sense 10 years ago, you still need 10 or 15 years to get to where LIDAR, not having LIDAR worked. And then now there's going to be a whole bunch of other things that you're like, wow, we could have done that completely different. And so I feel like this is again like this discussion about trying to race to an endpoint. But let's see a first example of what you described happening. And I think that that's going to be the real tell because I think that there were just, companies will figure all this out. And I think that they will just fall back on layers and architectural models because it's the only way.

Speaker 3:
[40:37] We know how to think about it for policy. We know how to think about it for security.

Speaker 2:
[40:40] But it's also the only way to build a system. Otherwise, you're just building an app. And if you're building an app to do one thing, we don't need all of this. Like there's a whole different way to do it.

Speaker 1:
[40:52] The thing that I'm pretty fascinated by is, and I don't even have any amazing data points or anecdotes, but at least the notion of these sort of companies that are merging in these kind of services categories from the ground up, from the Pure First Principles approach, which is like, okay, well, if I could start a marketing agency or an engineering consulting company, or I don't know, maybe somebody's doing this for law firms.

Speaker 2:
[41:16] Construction work or anything, yeah.

Speaker 1:
[41:17] Well, maybe construction design, architecture design, anything that would be like a knowledge worker kind of services company, because you could kind of build your company pretty differently if you had no constraints of, I have no information barriers and boundaries of what people should have access to. I can give the agent just all the context it needs to do its work. I can write software on the fly for particular things. I do think that will be relatively disruptive for some time until the bigger incumbents can kind of get out of the way on this. And that will at least create some precedent or case studies of what this new sort of corporation could look like. But over time, they'll still run into the same exact problems of every other corporation.

Speaker 2:
[41:59] Well, they'll run into the geography or market segments, or distribution challenges. Anything outside your little walls, you will run into the physical world.

Speaker 1:
[42:10] Right. I do kind of like the idea that there are some new business models that open up now.

Speaker 2:
[42:17] Oh, of course. Yeah.

Speaker 1:
[42:18] Because there's so much either information or software that basically goes underutilized by 100x relative to what its economic value is. Simply because nobody wants to pay five cents for accessing a piece of data or use a tool for $1 once. But you do give these agents a budget and a protocol to work with, and all of a sudden, you're like, oh, on the fly, they can go get medical research for some deep research tasks they're doing, and I'll pay $3 for that, and the agent is able to go and transact. It opens up a whole new world of business models for the Internet.

Speaker 2:
[42:53] Oh, that was too nice.

Speaker 1:
[42:55] Oh, okay. You're going to go farther.

Speaker 2:
[42:57] No, that one is one where that's actually the biggest, I think that the biggest in the air problem right now, is everybody is trying to figure out the economics of all of this, when they're off by at least an order of magnitude on how big the opportunity is. Because the new models that people will come up with, nobody knows what they are right now, but they will absolutely come out with new models, because that's what happens with every new technology. The thing that holds back the discussion now, is you basically have a bunch of finance and Wall Street people trying to justify GPUs and tokens and things like, as if we're in some old world. They're viewing the world of revenue as this linear growth curve, and trying to justify all the expense. When people are going to create, like this was the problem with PCs. People viewed PCs as a finite market, because they just viewed the consumption of MIPS as some finite thing, and they didn't think what would happen if we put all those MIPS on every desktop. In particular, people thought software just came with the MIPS. Nobody thought, well, they'll just sell the software. One guy did. It turns out that was a really good idea. Was it Bill or somebody? Bill and Paul. The same thing happened with the Cloud, which was people looked at the Cloud and they said, oh, we're going to take all of the server business, which was literally like 60,000 units a year, and we're just going to move it to someone else's data center.

Speaker 3:
[44:28] That's the bit.

Speaker 2:
[44:29] That would be the business and then we'll divide up the price. Nobody went, oh, people are going to use a thousand times as much of the resource leveling if we move it there. And that's exactly, I mean, that's the thing that I, it just drives me absolutely bonkers that the Wall Street models have this fixed revenue right pie.

Speaker 3:
[44:47] Zero-sum thinking.

Speaker 2:
[44:48] And it's this weird zero-sum where they just think that the amount of money that a company is going to spend. And like this was the problem with Salesforce that they faced when you were starting too. But like Mark was just blazing the trail, which was like the CRM business was 2 billion a year. And it was 2 billion in like, you had to go buy all these servers and these Oracle licenses and this huge headache and years of deployment and consulting, when if you could just get salespeople to sign up individually, they all will sign up with no friction. And that is exactly, there is no, no doubt that that is what's going to happen with AI.

Speaker 3:
[45:22] Let me give you an example of this. So I've been in for investing for 10 years now. I probably have a portfolio of 240 companies at work. Let's have visibility, let's say in 50 of them. These are all infrastructure companies. Some historically have done well, some not so well. Every single one of them has gone asymptotic in the last six months. You're like, okay, why is this? It just turns out there's so much more software being written now than ever has been before.

Speaker 1:
[45:44] Of course.

Speaker 3:
[45:48] It's not because they've got enterprise customers, it's just because there's just so much consumption of the infrastructure layer right now. So with more software, with more agents, there's going to be a lot more consumption of computer resources.

Speaker 2:
[45:59] So certainly in the case of the computer side of things, well, we haven't even gotten to the point yet where everyone's phone is a huge consumer of AI. So once everybody's phone and on device, like once your phone on device is consuming AI, the amount of it is going to go up by a billion.

Speaker 1:
[46:17] So do you like the micropayment piece?

Speaker 2:
[46:19] All of them. The micropayments, there's a little bit of micropayments that has come with every technology where they always think that you'll be able to get a fraction of a penny. But in the end, especially in the enterprise, people are just going to consume things. It's just cheaper and easier to buy a bulk license for a bunch of stuff.

Speaker 1:
[46:36] You want some predictability in that.

Speaker 2:
[46:37] Well, you want predictability and you just want to not have to think about it.

Speaker 1:
[46:41] I like the idea that it is the first time that you could... The agent doesn't care about the friction of a small transaction. So it's the first time that you can have resources behind a paywall, that something will actually be willing to pay for that resource.

Speaker 2:
[46:54] The world has built up the infrastructure to aggregate those payments into something efficient for a customer.

Speaker 3:
[47:01] And because tokens are such a significant part of Cogs right now, it is pushing the industry to do usage base in a way that we have. I remember when we went from perpetual to recurring, and that required a bunch of huge changes. We're going through the exact same change right now towards usage base, and usage base is pretty granular, and it actually allows... I mean, again, you will have a contract with AWS or Google.

Speaker 2:
[47:23] We went through this with AWS. People learned to do the usage credit. We went through the phase where people were so terrified of cloud compute that they were like, we need companies in the middle to help us find the cheapest and to arbitrate it all.

Speaker 1:
[47:38] Okay, well, now you write tokens into this, and I don't see how we possibly have time in this conversation.

Speaker 4:
[47:43] I'm here as long as you guys can stay.

Speaker 1:
[47:44] But the engineering compute budget conversation, to me, is gonna be just the most wild one in the next couple of years. It's just like, how much should you allocate of your engineering expense to token? And it's like, depending on who you read on Twitter, it could be 1% and the other side could be 100%. And it's like, yeah.

Speaker 2:
[48:05] Yeah, but this stuff.

Speaker 1:
[48:06] Well, no, no, no, CFOs have to, literally, they actually have to know the answer to this.

Speaker 2:
[48:10] I understand they have to know. CFOs always wanna know the answers to things that don't have the answers.

Speaker 1:
[48:15] No, Wall Street is gonna make them know the answer.

Speaker 2:
[48:17] No, Wall Street is gonna make them come up with some number and hold them to it, then they'll get fired and then it'll, but it- Okay, okay. I hear you.

Speaker 1:
[48:24] R&D is somewhere between 14% to 30% of revenue of any public technology company, let's just say, okay? The difference between compute being 2X the cost of your engineering team or 3% more is like, that's all your EPS.

Speaker 2:
[48:41] I get it.

Speaker 1:
[48:42] So we will have to know the answer.

Speaker 2:
[48:44] I'm strictly willing to sacrifice a few CFOs at the altar of this. I want that. That's a good clip, by the way. But the reason is because again, this is trying to know what we just don't know right now. And this has happened with internet bandwidth, this has happened with...

Speaker 1:
[49:00] No, this is not even close to internet bandwidth.

Speaker 2:
[49:02] Oh? No, no, no. I beg to differ. Like people were free. It happened with vacuum tubes. It happened with transistors. It has happened with every technology. There was this, oh my god. It happened with programmers. There was a time when programmers were going to swallow every company.

Speaker 1:
[49:18] Yeah.

Speaker 2:
[49:19] And that's not... It was in my lifetime, not some made up...

Speaker 1:
[49:21] Yeah, but I don't think we've ever had a point where every end user in an organization has sort of a completely elastic ability to spin up a resource on their behalf.

Speaker 3:
[49:32] Well, it's certainly...

Speaker 1:
[49:33] That actually is in many cases very valid for them to go spin it up.

Speaker 3:
[49:36] But it certainly rhymes with what happened in the early 2000s with cloud. I remember very similar discussions when we went from CapEx to OpEx and then Unlimited Spend.

Speaker 2:
[49:45] Oh, no, and remember there were companies who the CFOs would sit in our briefing center here and say, you don't understand, we are an agriculture company.

Speaker 1:
[49:52] I can see the rhyming.

Speaker 2:
[49:53] We are an agriculture company. We only know CapEx. We have no... I sold through this shit. Right, right.

Speaker 3:
[50:01] We both did.

Speaker 2:
[50:02] Or like, oh no, we are an OpEx based company. So if you tell us... We love the cloud because we just shifted everything to OpEx. And so all of the stuff, like the rules of accounting work out, also don't, I keep thinking, do not discount the local compute engine as being a release valve for all of this.

Speaker 1:
[50:20] When's that gonna happen?

Speaker 2:
[50:21] Well, the question is, it's not when does it just happen with today's view of the technology, but how all of a sudden, wow, there's a whole...

Speaker 3:
[50:28] Has that historically ever gone that direction?

Speaker 2:
[50:29] Yeah, exactly.

Speaker 3:
[50:30] It goes the opposite, right?

Speaker 2:
[50:31] No, it went all to the client.

Speaker 3:
[50:33] Well, okay, back to the 80s, yes.

Speaker 1:
[50:36] No, that's most of the examples that we're hearing so far. Whoa.

Speaker 2:
[50:39] That was uncalled for.

Speaker 3:
[50:40] Okay. Since then, it's all...

Speaker 1:
[50:43] Vacuum tubes. He's talking about vacuum tubes.

Speaker 2:
[50:45] But I do those examples because you can't argue with them, and it's much easier that way.

Speaker 1:
[50:49] You're right. I can't prosecute for it.

Speaker 3:
[50:50] Since then, it's all kind of celebrated back.

Speaker 2:
[50:52] But it's only been 10 or 15 years that it moved back to all cloud, and then what has happened recently with that? A lot of people wake up in the morning and they say, oh, we're moving back to doing some critical but stationary workflows on on-prem.

Speaker 3:
[51:08] With AI, that's true.

Speaker 1:
[51:09] Dude, you wrote the blog post, man. Don't make me go through the archive. I had to deal with so many Wall Street questions on that one, by the way.

Speaker 2:
[51:18] Also, because your competitor went back to repeat stuff.

Speaker 3:
[51:22] We're talking about two very different... I agree with building your own data center. I'm talking about this notion of edge computing where things go to devices. That seems to be...

Speaker 1:
[51:34] I'm more in the cloud maximalist camp. But sorry, so you just don't think, you don't even think for one second that it matters whether, how you're supposed to be an engineering leader right now managing the compute budget of the engineering team?

Speaker 3:
[51:47] No, of course it matters. I just think in the long term this thing will get figured out.

Speaker 1:
[51:51] Oh, sure.

Speaker 2:
[51:53] Who cares? We don't need a podcast in the long term.

Speaker 3:
[51:55] Here's what I think.

Speaker 2:
[51:57] But here's a rule of thumb. First, the startups are going to burn through available capital pretending like it's not a problem. They are going to do that.

Speaker 3:
[52:06] Yeah, but they do that anyway.

Speaker 2:
[52:08] A lot of big companies are going to be so terrified, they're just going to freeze and not do anything. Then people are going to actually start buying it on their own, and they're going to do all the things that companies do when they're big, have a lot of money but don't want to spend it. In the middle, we are going to see if you pick a category of product or go to market or something, there are going to be people who are willing to make the bet for whatever reasons that they can because of their financials, and they are going to go ahead, and they are going to become the people who lead in the space, so long as they can maintain the financials. Now, they might do it in, they might say, we're going to just do it here in this particular application space or here in this particular usage space. But this idea that nobody is going to go in and because they're so terrified that the CFO is going to get fired or something. No, no, no. It's just crazy.

Speaker 1:
[52:57] Yeah.

Speaker 2:
[52:58] And but then there are going to be CFOs who make a mistake and like, okay, everybody gets a little.

Speaker 1:
[53:01] Yes.

Speaker 2:
[53:01] Well, if they do that, that's a complete fail.

Speaker 1:
[53:04] Yes. But also or like you get, there is a really interesting finesse here, which is like, you don't really want your engineers right now having to think about compute budget because we're still developing the, oh, okay, so that set you over in that.

Speaker 3:
[53:19] But I feel like we've been having the discussion for 15 years when it comes to Clag.

Speaker 1:
[53:24] This is totally new. Only like 10 percent of your engineering had to think about cloud infrastructure spend.

Speaker 3:
[53:32] In 2016 and 2018 timeframe, there's a whole set of companies that was basically like the dashboard for, what was it called, FinOps?

Speaker 1:
[53:42] FinOps is very cool right now because of the AI house.

Speaker 3:
[53:45] Developers have access because Cloud spends are getting out of control, and API spend are getting out of control, and so it was like, here's your Twilio spend, here's-

Speaker 1:
[53:52] But it's pretty different, and I'm going to wait for all the comments to come in on YouTube to call you out on this. It's like you can get into a conference room and just be like, hey, can you make that one algorithm a little bit more efficient so you don't use as much of our cluster at this time of night or whatever? Then you get out of the meeting, somebody improves it, and you're good. This is like every single prompt that every engineer is doing. You have to decide. Do you want it to be a long running prompt? Do you want it to be a long running agent? Do you want to parallelize that? What is your comfort level of wasted tokens? For me right now, I'm like, yeah, we should probably waste a lot of tokens because that means that we're trying new things. Should your head of engineering be happy if you run 10 experiments in parallel and thus you're obviously going to be wasting 90% of the tokens, but you're going to choose one of the successful paths? Or do you want to tell the team, no, before you go do that, make sure to really go and design the perfect system? We actually have a whole bunch of open questions that are going to start to happen. Literally, as of this recording time, people are freaking out right now on the new Cloud Code Max plan because they're getting blocked after three prompts. Well, this is going to be a very real topic until we can actually find a way to build data center capacity.

Speaker 2:
[55:09] Oh, that's a different problem.

Speaker 1:
[55:10] Okay.

Speaker 2:
[55:11] No, because, no, one is, well, assuming, well, wait. No, you can assume that if we build more capacity, the price will drop because there is more capacity and we're priced now based on limited capacity, whatever. But like, this is just going to get worked out and I feel bad for those that have to make a decision immediately about which 17 people get no more tokens this week or whatever, and that the whole company is walking around with like a token card. And the person in the lunch line is punching their card every time they do something. But like, I don't know, like somebody we were talking about today about performance and how like, we used to write command line tools that spit out the time it took after you ran a command line, just so you knew and if you knew you were getting better or worse. And you know, but the thing is this is all going to go away. There's absolutely no doubt that this just goes away.

Speaker 1:
[56:06] I think on the 10 year time frame, 100%.

Speaker 2:
[56:08] And the biggest reason it does is because you have to do the Benioff kind of math, which is if you're paying an enterprise salesperson, you know, a million dollars a year, you have to ask how much is their tool worth?

Speaker 1:
[56:20] Yeah.

Speaker 2:
[56:21] And if you're paying an engineer X dollars a year, well, at some point, their tooling is worth.

Speaker 1:
[56:28] It's absolutely worth it.

Speaker 2:
[56:29] And it's not going to even be an issue. And...

Speaker 1:
[56:32] Yeah, yeah. I don't think it's... I think it's...

Speaker 2:
[56:34] And so if there's a capacity thing in the short term, that is a different problem driving the price than this just we're going to forever have to be in some budgeting exercise.

Speaker 1:
[56:44] I think law of large number solves this because eventually you have enough engineers using this much compute, but like we're in a transition phase where like most people thought, you know, the two year ago level of spend on AI, which is like, ah, it's a chat bot.

Speaker 4:
[56:57] Yeah, yeah, but they were wrong.

Speaker 1:
[56:58] Yeah, right.

Speaker 4:
[56:59] Okay.

Speaker 2:
[56:59] But they were wrong. But they were wrong.

Speaker 1:
[57:01] We tried to warn them.

Speaker 2:
[57:02] No, but they were wrong because they saw it as this particular use case.

Speaker 1:
[57:07] Yes.

Speaker 2:
[57:07] And, but again, like, you know, like the vacuum tube thing you made fun of. Yeah. But like there was a time when they thought that like, like all of the Dakotas would be covered in vacuum tube warehouses and people on roller skates would be running up and down the aisles, placing vacuum tubes just so we could fight World War II. I mean, like that was how that was the, and they thought that, and then someone said, hey, how about a transistor?

Speaker 1:
[57:34] Right.

Speaker 2:
[57:35] And like we are going to have a transistor moment with all of this. It might just be more supply the way we think of it, but it also might be an actual algorithmic fundamental change. It could be a change in the hardware. There's a lot of stuff that can happen that changes this particular moment in time. It's just this, I think it's particularly weird that everybody has just gotten to token, which is the same thing that happened with IBM and Mainframes. People were on MIPS and then one day the reality was IBM was selling more MIPS for fewer dollars every year and didn't even realize it. And they were still pricing their Mainframes by MIPS until it got pointed out to them that they were on a decreasing curve because they were making MIPS faster than they can charge for. And that's what's going to happen. Guaranteed. I just said that in a hardcore way.

Speaker 3:
[58:22] I think that was great.

Speaker 2:
[58:24] It sounds really great to sound like I know what I'm making. Guaranteed.

Speaker 1:
[58:28] I actually probably believe it.

Speaker 4:
[58:32] Thanks for listening to this episode of the a16z podcast. If you liked this episode, be sure to like, comment, subscribe, leave us a rating or review, and share it with your friends and family. For more episodes, go to YouTube, Apple Podcasts, and Spotify, follow us on X at a16z, and subscribe to our sub stack at a16z.substack.com. Thanks again for listening, and I'll see you in the next episode. As a reminder, the content here is for informational purposes only. It should not be taken as legal business tax or investment advice, or be used to evaluate any investment or security, and is not directed at any investors or potential investors in any a16z fund. Please note that a16z and its affiliates may also maintain investments in the companies discussed in this podcast. For more details, including a link to our investments, please see a16z.com/disclosures.