transcript
Speaker 1:
[00:00] So the board goes to the CEO, what does the board say? We need more AI. And what does the CEO say? Oh, okay, I'll get like a consultant to do more AI. And then they have some centralized project that nobody knows how it works. They haven't aligned their operations and those things will fail.
Speaker 2:
[00:13] The funniest concept that the more code we write, the less we would need engineers would be the opposite because now your systems are even more complex than before, which means that you're gonna be running into even more challenges of when you need to do a system upgrade or when there's downtime and you have to figure out like, well, how do I fix that problem? Or when there's a security incident? So yeah, we're just getting started with the jobs on this front.
Speaker 3:
[00:33] They're gonna hit a wall at integration. The thing that's not different about AI and that agents don't fix, that nothing fix is that any enterprise of a thousand people or more, or that's older than 10 years, is just a massive stuff that's sitting there waiting to be integrated. And you can't just say it's gonna integrate. AI actually doesn't help to integrate anything.
Speaker 4:
[00:58] AI feels like it's moving fast, and for many companies, the real transformation is just getting started. There's a growing gap between what's possible in Silicon Valley and what's being deployed inside large organizations. Engineers are already shipping with agents and new workflows, while enterprises are beginning to adapt those capabilities to more complex systems and real world use cases. That creates a moment of opportunity. The tools are getting more powerful, and companies are learning how to integrate them into existing workflows, data systems and decision-making processes. At the same time, there's a deeper shift underway. AI isn't just another layer of software. It's starting to act more like a new kind of user, one that pushes companies to rethink how systems, permissions and workflows are designed. In this episode, Steven Sinofsky, board partner at a16z, Aaron Levie, CEO of Box, and Martin Casado, general partner at a16z, discuss what's working in enterprise today and where the transformation is heading.
Speaker 3:
[02:05] Hey, we are here monitoring the situation live, and we're very excited to talk about a bunch of AI stuff. And we have three of us are here today. There's me, Steven Sinofsky, and Martin Casado, who will wave and say hi, I'm Martin. And Aaron Levie, who is working on the elevation of his hair today. So we're excited about that.
Speaker 2:
[02:28] It just keeps getting more vertical. And I thought I could kind of tame it, but it didn't work.
Speaker 3:
[02:33] And is that just a token issue or a parameter, number of parameters issue?
Speaker 2:
[02:37] It's too many parameters. Too many parameters.
Speaker 3:
[02:39] OK, I have the same thing but in reverse. OK, so.
Speaker 2:
[02:44] You listen, you have a distilled model.
Speaker 3:
[02:47] I run local. So we had a lot of, there's been a busy week of things, but we want to bubble it up a little bit and just start talking about where things are heading. But I'll just kick it to you, Aaron, and you start where you are the most excited this moment, because you have visited a ton of customers this week and have learned a lot. You've shared a lot on X, but I think you're the most in-the-trenches CEO who is really talking to customers every single day in the enterprise, which is what the three of us tend to look at the most.
Speaker 2:
[03:24] Yeah, I think my, it feels like my job these days is just bring reality to the valley and then bring the valley to reality as much as possible. And it is a kind of a crazy divide that exists at the moment.
Speaker 3:
[03:38] Well, let's take a step back. I actually think it's a super interesting, what is it, what's the gap caused by?
Speaker 2:
[03:45] The gap is caused by, yeah, well, I think the gap is, and Martina, I'm sure you see this, but I think the gap is caused by the styles of work that exists in Silicon Valley and in engineering roles versus sort of the rest of the world. So, and we've talked about this a couple of times in different forms, but you know, the technical aptitude of an engineer is just like insanely high, the level of wired in this to what's going on in the Internet is insanely high. The ability to use your own tools and make your own choices is insanely high. And when things go wrong with the systems that you choose, you can just like quickly debug them and then make them sort of work for you. And then obviously you have all the benefits of just the models are really good at code and the work is verifiable. So you have like, you know, five or 10 things that make agents work in an enterprise context for engineering, or at least even a startup context for engineering, that tend to be a gulf between the way you work that way in engineering and the rest of sort of knowledge work. And so a lot of what I see is trying to figure out how do we kind of, you know, bottle up all of the greatness that is, you know, what we are seeing from coding agents, what we're seeing from agents that can use computers, to how do you bring that into the enterprise where the workflows are quite different, the users are less technical, the data is much more fragmented, the systems are much more legacy. And so, that tends to be the divide. So, it's not even that we're talking past each other in one of those classic government versus industry. It's just literally like there is just a pure workflow and technology set divide. And that's why it's gonna be a number of years for this sort of diffusion to roll from what we're seeing in Silicon Valley, what we're seeing as tech startups all around the world into the rest of knowledge work.
Speaker 3:
[05:29] Martin, just to build on that, you have a ton of experience in big companies. One of the other issues, though, is scale. And the difference in scale that Silicon Valley operates at the startup level versus everyone else.
Speaker 1:
[05:43] Yeah, I also think that, I mean, these secular trends, like the internet was like this, actually start with individuals. And big companies tend to make decisions centrally. And this is one of the fastest growing secular trends. Like, there's probably a lot of individuals in big companies that are doing it, where like, the big companies themselves don't know even how to think about it. And so when you hear stats like, oh, like MIT had this stat, like 95% of AI efforts in big companies fail. Like, that's clearly silly, because I am sure everybody is using ChatGPT very effectively. What they really should be saying is, you know, whatever. Like, I sit in these boards too. So the board goes to the CEO, what does the board say? We need more AI. And what does the CEO say? Okay, I will get like a consultant to do more AI. And then they have some centralized project that nobody knows how it works. They haven't aligned their operations and those things will fail. And so I don't, you know, when we say scale, often we think about things like system scale or number of peoples. I think the secular trend is scaling wonderfully, which is being reflected in the numbers of these companies. But organizations don't know how to adjust these kind of, you know, age-old processes that have been worked on for a decade around, you know, data and governance and operations and compliance, et cetera. That's kind of right now where I think, like Aaron is right between the secular trend and the organizational decision body. And this is something that we actually track very closely because we're starting to see now, I would say in the last few months, finally some real kind of inroads into the enterprise. But it's, it's, it's tepid because, and the last thing I'll say in this, one of the reasons is there's a lot of skepticism because the board wants AI, CEO, AI failures have created some amount of bruising, which is, you know, requiring these companies to get past it in order to do kind of the second go at it. And so I think this is exactly where we are.
Speaker 2:
[07:39] Yeah, I 100% agree with that, which is that it's good to start with agreements because we know how quickly that fades.
Speaker 1:
[07:46] We'll disagree the rest of the show.
Speaker 2:
[07:47] Exactly, exactly. That's the only time we're going to agree. I think maybe one more point on the board for agreements. Maybe you guys would agree. There's also this very interesting dynamic. I would say this is a minor one relative to everything else. It's probably 5% of the problem. I think it would be more fun to talk about the real problem. But there's a fun kind of, as an aside, there's a fun dynamic where you go to an engineering team classically for the past. Steven, you can take us back in history on this one. And one of the easiest ways to stall a project was just getting the architecture, the fights on what language to use, what architecture path to go down. That could take months and months to work through as your teams work through that. Because of the pace of change in AI, you actually have this incredible dynamic where the labs are obviously leapfrogging each other so frequently, but with not the exact same paradigm of how you should deploy agents and how they will work. And is the agent harness in the computer? Is it outside the computer? Do you run it in your cloud? Is it hosted? What tools does it have access to? This is not a point where these are completely fungible technologies. And so that actually creates a bit of paralysis because now as an enterprise architecture team in the real world, you're like, man, what horse do I want to get behind? And which architecture path do I want to get behind? Because I've been burned by doing the wrong thing in AI maybe three or four years ago, and I went down in some path that now is deprecated or not the right strategy anymore. So to some extent, the speed of our change in tech actually reduces the ability for the tech to get diffused into the really, really important workflows. Because now you have a lot of paralysis in just making decisions. So I actually think it's kind of fine because there's still so much upgrade work people need to do in their infrastructure and their systems and their data. But this is kind of an interesting dynamic where I'll go have conversations with CIOs and their AI teams and I'll say, hey, what are you using for your chat system or your core agent orchestration? And they'll say, yeah, we're in the middle of a debate between these two or three paradigms. And you hear that across almost every single customer because there is a little bit of a nervousness of like who do you get in bed with and how much do you sort of fully lock yourself into one particular path. And we also know that if you don't lock yourself into a path, it's always then you're building for this sort of duality, which also takes a lot of work architecturally.
Speaker 1:
[10:15] Sorry, I hate to jump in, Steve. So Aaron is totally correct. And there's a very specific instance of this playing out in product companies right now. And I'll tell you what it is. So software product companies, circa six months ago, they viewed integrating AI was like you're actually integrating it into the product. So everybody was like adding whatever this chat feature or like, and so it's kind of like this fusion or this hybrid model. What we're seeing instead is, instead of viewing AI as software, like just view it as a user. And so instead, like take your product, make it a CLI tool, and then have the AI be an agent that actually uses this. You're not fusing the two, you're just making it more useful for AI. This is a very, very significant architectural and mental shift, right? And so we started as pure product, and then we didn't quite know what the end thing looked like. So we created this AI software hybrid that hasn't worked, and now we're kind of going to the agentic model, which basically means the agent is going to be whatever. It's going to be cloud code or whatever, and then my product now just should be something that can be consumed by that, and that's the actual modality. But within a year now, you've had to re-architect your software twice. And so I think no matter what, many places that you look in the industry is having this dilemma of actually trying to figure out what the final form looks like. And Steven, you will remember, remember all the hybrid versions of cloud? Remember like, you know, like remote desktop and all these things? I think we're kind of like speed running that evolution to the final form.
Speaker 3:
[11:51] Right, and I think that people in Silicon Valley don't quite appreciate when a big company says, well, we have to map out our bet that we're going to make. Because like that just seems stupid. And you know, if you have, if your job history is, you know, five two-year stints at startups that went from seed to series A to aqua hire or something.
Speaker 1:
[12:16] Yeah, you didn't learn anything.
Speaker 3:
[12:17] Well, you, you never, you, you don't, your frame of reference is not, you know, picking an accounts payable system that's going to last 40 years. Yeah, I actually, I have like all these visual aids today. So here's like the ultimate, the ultimate engineer, if you're in Silicon Valley, is...
Speaker 2:
[12:36] Lower, lower, lower, lower. Yeah, exactly.
Speaker 3:
[12:39] Is Guilfoyle. And Guilfoyle is like, I don't want to talk to anyone.
Speaker 2:
[12:47] Yes.
Speaker 3:
[12:47] And I will just write the code and you go do your thing. And the thing is, is that you have people in enterprises that are saying, I'm going to use the model and do my thing. But they're only, they're going to hit a wall at integration. And this, the thing that's not different about AI, and that agents don't fix, that nothing fix, is that any enterprise of a thousand people or more, or that's older than 10 years, is just a mass of stuff that's sitting there waiting to be integrated. And you can't just say it's going to integrate. AI actually doesn't help to integrate anything. And so even if you change everything, the people say, oh no, if you make it an agent, then it can just go ahead and be a user. But if you're a user, like if you've ever called customer service for something, like literally you get bounced to another human if the system that you're talking to doesn't work. And they're like, well, that's a manager. Or no, you're talking about payment, not reservations. And so like, what I think is so exciting is that now we have proof of this technology that everybody likes it. I mean, you see all the people who don't like AI are saying, look at what's happening in law firms because people are seeing hallucinations and it's ruining legal cases and all this. And the reason that's happening is because the 25-year-old associate is the one using AI successfully already and had been using it for a year.
Speaker 1:
[14:19] Well, Steve, it's actually a little worse than that, where it is right now, many companies are incentivizing people to use AI by counting tokens.
Speaker 3:
[14:28] Oh, yeah, yeah, yeah.
Speaker 1:
[14:29] And so I'm not going to say the name of the guy. I spoke to someone yesterday who worked for one of these large companies that famously does this and he's like, me and my coworkers have agents do useless tasks just so that we can... No joke.
Speaker 2:
[14:42] No, no, no, totally. Well, you get whatever you measure, so.
Speaker 3:
[14:44] Yeah, yeah.
Speaker 1:
[14:45] That's right. So like, it's like the extreme form of what you're saying, Steve.
Speaker 3:
[14:48] Oh, yeah, yeah, yeah.
Speaker 1:
[14:49] People that are like being fake, productive and producing a lot of, you know, yeah, you could say perhaps problematic artifacts just because they're they're using these models.
Speaker 3:
[15:03] The when the Internet happened, all of a sudden, every company needed websites. And so like a very famous moment in time was not too long ago when every internal team had like a team website and they went out and they got like a vendor to write HTML and to create their site. And then there was a team, but like there's nothing dumber than having a team website at a large company because the team gets reorganized like six months later. And so companies were just filled with like with thousands of these dead web is what the expression was. But I think, go ahead, go ahead.
Speaker 2:
[15:37] No, no, but we should we should drill into your integration point because I do think this is something for, you know, sort of some reality to settle in, in the valley on, on the real world's sort of journey to fully being agentified and what that's going to take and what that's going to look like and your point about being passed to the different human, you know, based on the role that you needed to interact with, you know, agents basically don't have any, there's no real exception yet for the agent having the same problem because you basically, you know, as you pass through a different human, it's a different set of access controls that that human has. And if an agent can sort of bypass any of those steps, then that's how you instantly get the security risks that, like, you need to kind of pass through those steps, so that way you don't accidentally, you know, get to the wrong piece of information and those verification. And so there's a lot that you need to kind of build out for agents to be able to go and work with all these systems. And we've talked about this, but like most legacy environments don't have the most authoritative, you know, access controls. You're always as a human going and saying, hey, Sally, can you share that thing with me that I don't have access to? Or, hey, Bob, what's the number inside your data system for this question? And so if agents just get the exact same permissions that you had, then they'll just run into these walls everywhere, and they won't be able to complete the process. And unlike a human, they're not going to know to go talk to Sally or ask the question of Bob. So they're going to just be kind of, you know, stuck. So what's going to happen is you're going to have a lot of agents that don't have access to the right data. They're kind of working through systems that are, you know, not the real sources of truth for the information. They're getting the wrong number. They're getting the wrong document. So this is the real work that enterprises have to go through right now. The good news is that it's actually a great time, again, if you're a startup, because you can just, you get to know all the problems right out of the gate. So you can design your organizations to try and avoid this. But for big companies, there's real work that goes into how do I upgrade my systems? How do I modernize my technology environment? How do I make sure that agents will have access to the right data, the right documents, the right context to be able to do their work? And that's the work ahead. And there was this headline of OpenAI working with, in Codex, working with Accenture, Deloitte, all the major system integrators. And there were some snarky comments online around it that I was fascinated by, because it sort of showed how maybe great that divide is from the rest of the world versus those in tech. Because to me, it was like the most obvious announcement of all time, which is a large enterprise is going to have to go through the change management, the systems implementation, the integration of technology for these agents to be able to go and work. And so there was this sort of like, people thought it was somewhat ironic that, oh, we need people to implement the agents that are going to go automate the people. And it's like, no, that's exactly how it works. You actually do need to do lots and lots of work to be able to be in a position where agents can actually go and help you do any of the automation. So that is, and that's going to be, there's going to be businesses that are doing that for decades. Like, it's going to be an incredible opportunity for this kind of next generation set of firms, as well as existing ones that lean into that.
Speaker 3:
[18:52] Let me throw this out there. Well, first, I think the other thing that people shouldn't celebrate when those fail, because they will fail because they're, as Martin was describing, a lot of them are going to be these sort of top-down mandates where they pick the most acute problem in the company and think, oh, AI is going to go solve that. And the IT people are going to be like, oh, God, that's the worst system to try to do that. But the CEO or CFO or whatever is going to be obsessed with solving, or most likely, the customer service person will be obsessed. But I do think if I were advising a startup specifically in order to enter the enterprise space in that way, I definitely would be thinking about not just building a company that's step one, I only work with all the headless SaaS software that's out there, because there just won't be any. But the thing you can do is structure the value that you offer, and also this applies to what you go do in a company. It's really a fork. And the fork is, is this an agent that is seeking information and presenting it to some human, or is this an agent that's supposed to go act and do something? Like, is it acquiring or is it doing? Because if it turns out, that's what happened with the internet. The internet got very, very valuable when the first step was just providing access to things to people. And like all of a sudden, all the sites that were like, that literally did integration, like, hey, I need expense reports, but viewed by department, or I need to see our current inventory status across like the two companies we've acquired. All of a sudden, the web became the integration point. And so I do think that if you just show up first and just say, hey, we can actually use agents to learn stuff about what's going on in a company. And in particular, because you're here, Aaron, like learning across the files becomes way more possible than it ever was before. In fact, AI might be the first time that inside a company search can actually provide immediate value because the web just wasn't structured to deliver those results. And then you start to think, once you could bring them all together, then you can add like an agent that has an approve button or a reject button or something like that.
Speaker 1:
[21:12] Let me just try and provide, finally the point where I get to disagree.
Speaker 3:
[21:16] Uh-oh, we're in trouble now. No, no, no, I mean, you're invited back, so good.
Speaker 1:
[21:21] No, no, I think this is a very legit view, but it's not the only view. And in light of AI, I think it's not the only kind of compelling view. So here's the, so let me just try and rephrase. So the current view is we've got like AI as software. It works in a different way. We have a current set of systems and we have to integrate this new type of software with our existing system so that it can get access to data. It can do things, but in a safe way, right? So here's to the kind of end-to-end argument of why this isn't about evolving software systems. The end-to-end argument is these LLMs are non-deterministic. They are smart. They deal with a long tail of complexity. And it turns out those are all things humans do too. And we've spent 40 years building interfaces, processes and design to deal with messy humans. And we know who to access and we have access control. And so, if you have the mindset that an agent is more like a human and you hire the agent, you give it its own email address, it can access documents like humans can, it can log in, it can request the things that it needs, then it will be drafting on all of the process that we've put in place for humans, not for software. And so, I would just encourage us, as we have this discussion, like, listen, I grew up like you guys in software. I always think of every system like software, but these models don't integrate well with software, actually. I think it turns out, and what we're learning as an industry is, if you view them more like humans, and you draft on the mechanisms we put in place for humans, they're much easier to integrate. Well, and I think we're seeing-
Speaker 3:
[22:58] I love that point.
Speaker 2:
[22:59] I think we agree with that for sure. I think the issue is humans have a bunch of extra benefits that the agent doesn't have. The human has a lot of context that we get for free by virtue of, we can keep track of the myriad relationships that we've built in our organization and the person to tap on the shoulder when we need something done or we need to get information. That's not documented in a company yet, in a way that the agent can just draft on. And so I think we all would agree that you can't treat this like software. You treat these as people accessing systems and tools, but they're both at a massive advantage that they can work in parallel at infinite scale. And they're at a disadvantage in that they don't know who to tap on the shoulder.
Speaker 1:
[23:39] Hey, listen, Aaron, I am all for agent onboarding. Like, you know, the agent comes and it goes to orientation. And then the CEO gives it the culture discussion and then every... I'm not kidding.
Speaker 2:
[23:52] No, you're probably right.
Speaker 1:
[23:54] Every department does their pitch, like, this is what we do. I mean, I think, I actually, honestly, think given the technical nature of these agents and how much entropy they have and kind of how unruly they are, we're going to have to go through the processes that we've refined around humans because humans have all of those things. And so I just, you know, it's more about providing schools for them than somehow building some, you know, fancy index database.
Speaker 3:
[24:24] I mean, of course, what I love about that is you just keep going with the analogy because what that is, is it's the same argument that humanoid robots will be the best kind of robot, which is we have a whole world designed for humans. And like I saw at the Consumer Electronics Show, I saw this robot go into an elevator and then there was a button pushing robot on the elevator. So because the robot was a tiny little thing, like a Roomba on the floor, it couldn't push the button. So the same company that invented that robot invented a device you buy for the elevator that pushes the button. But then I asked, why did you need a device to push the button? And it was very interesting. They said, because the elevators don't have systems that they can hook into as a robot. Like there's no Wi-Fi, press the button in the elevator capability.
Speaker 2:
[25:17] There's no API for that.
Speaker 3:
[25:19] There is no headless version of the elevator.
Speaker 1:
[25:23] That's actually a great metaphor for like the problem. But I think that we're actually solving in the enterprise with these agents, which is we just, you know, we have two types of systems, those for humans and those for software. And these tend to be more like humans. So we should draft on those as much as possible, rather than try to retrofit them.
Speaker 2:
[25:39] Well, and so the big news last week was I think, Salesforce, I don't know if they surprised people or not, but I mean, based on the reaction, it seemed like it was maybe a surprise. They went full headless and they basically said, you know, like we want to be used everywhere across all of the different agents. And I see that as a little bit of a bellwether because I think as Salesforce goes, so does a lot of enterprise software. And I think a lot of people are going to try and, you know, have to figure out what is the new business model in this headless world, you know, do you charge a little bit of a small just API tax? Is there a seat for the agent? So there's obviously some work to do with that. And Steven, I saw one of your tweets on, you know, some of the complications there. But I think as a moment, it's a big deal because I think it's a recognition that, you know, software will be running in the background. It always has for machine users and applications. And now it is for these sort of probabilistic machine users or non-deterministic machine users. And what's cool and where I think this gets pretty exciting is, you know, as soon as I saw that announcement, like I had like five to ten personal use cases where I would need, you know, the headless version of Salesforce because I'm always doing just a tremendous amount of customer related intelligence work. I'm going into a meeting, I need some information. I need to do, I'm going into a city, who should I be meeting with? And so if you imagine, you know, being able to run compute in the form of agents across all of your data systems, like the use cases become pretty wild around what that opens up. So I think this gives a lot of software platforms, all new use cases that they can tap into, where again, you were normally constrained by the number of people on these platforms, but now the headless user can be 100 or 1000x the scale of those human users. So this is, I think, an exciting moment because as you have more of these agents running around and the headless software modes, you just have way more use cases for these tools.
Speaker 3:
[27:34] I also think, I think on this one, what's so super cool is that, of course, the first step is doing exactly what you described, which is just looking stuff up. And so the most interesting thing is using this notion that an agent is just an entity, it's incredibly obvious to me that it's another license. Now it might have a different license model, but it has to have an identity. Like when you go look something up in the box CRM system, I don't know if it's Salesforce or not, when you use the box CRM system, it has to be a person, like with a certain amount of access rights. And you presumably as CEO, you might have access to a bunch of stuff, but also there's a lot of ways that they actually don't want you to have the rights at the right time. Like you might be able to look and see who is on the account, but you don't need the up-to-date quota of those salespeople and stuff. And that might be HR sensitive, and you should probably have some other level to go see that. But as you go down the org, the agent is never going to have more permissions than the person who's getting it to go do something. And in fact, it's just going to be like a peer to somebody else in an organization. Because otherwise, you have all of these issues where the peer, where a human can just say, oh, get me the super smart agent that knows everything that I'm not allowed to know. And in the IT architecture sense, what's so fascinating about that is you have to build, you can't let the agent get the results and then try to figure out what works or not. But all the points that Martin made about the LLM stochastic model, which is you're not going to be able to figure out. It's not like a record in a SQL table that you could just apply ACLs to. It's actually like, it could be words in a sentence or just the number that shows up. And so I actually think that whole discussion about Headless for me made the SASpocalypse seem even dumber than it was already, and it was already dumb. So like, it was like, at first, it was dumb. And then I'm like, oh my god, it's actually much dumber than I thought it was in the first place. Because you're just going to have this explosion. Now, someone might come up with a very clever pricing scheme and that agents somehow cost less, because maybe for the first five years, they're read-only or they're always tied to a person or something. But it is another seat. There is no way around it. And if you're a SaaS company, you're crazy to try to say, oh, just use the credentials of another human. That would be bad security practice from the get-go.
Speaker 1:
[30:07] Exactly. So actually, in fact, this is playing out in many domains. You can even make the argument that a headless SaaS doesn't make sense. And here's the argument. The argument is, let me give you an example. So if you use OpenClaw, do you know why you use a Mac Mini with OpenClaw? Well, for two reasons.
Speaker 3:
[30:31] Is that what you're getting at?
Speaker 1:
[30:32] It's not. It's two things. It's number one for iMessage.
Speaker 3:
[30:37] So it's for the integration. Yeah.
Speaker 1:
[30:39] Because there is no headless version, so you're just going to use it. And then the second one is very interesting, which is if you've tried to use headless browsers with agents, the problem is, is all of the websites have anti-scraping measures, so they don't work. And so the reason you use a back mini is so it can actually use Safari proper. So to do anything headless kind of assumes that like the entire internet is going to go headless, when I think all of these models, like all of the data is humans working on the actual apps that are not headless. Like that's all of the data anyway. So I think these models are going to be very good at just using apps like they are today. And we're already seeing this happen. And rather than the headless versions, the non-headless versions are what's actually being used. So you could argue that it's just Salesforce, not a headless.
Speaker 2:
[31:28] It will go to a browser. Actually, just to clarify, do you literally mean the agent goes to the browser?
Speaker 1:
[31:34] Yes.
Speaker 2:
[31:36] I'm taking the other side on that one big time.
Speaker 1:
[31:38] Yeah. Let me simplify the argument so we can actually have it. So today, if you use an agent like NanoClaw or OpenClaw, you could use a headless browser. Let's say I wanted to look up the value of my house on Zillow. The headless browser simply doesn't work because Zillow is so tired of people scraping it, so we will detect headless browsers.
Speaker 2:
[31:59] Totally.
Speaker 1:
[31:59] So the thing that works is, is you pop up Safari and it uses a proper Safari directly. So then all of a sudden it works.
Speaker 2:
[32:08] No, but I would just say that any software that has a good API, the agent would absolutely prefer to use the API, and then you pop into the browser the moment that you run into some execution problem with the system.
Speaker 1:
[32:27] Seth is a fantastic long-term computer science software guy. However, these models are trained on data and RL environments from existing software that didn't have those APIs.
Speaker 2:
[32:38] Yeah.
Speaker 1:
[32:39] Right now, if you actually look at the adoption and the use of these agents, they look far more like what a human would do than what a program would do. So maybe you're right, but A, that's not what we're seeing. You can honestly make the end-to-end argument when it comes to data and all of the controls in the Internet. To Steven Smith, all the existing controls, do you just be like, these are going to actually have the same actions as humans?
Speaker 2:
[33:04] Well, the APIs, I think the APIs of any software provider will follow the same access controls of whatever the user is that is-
Speaker 1:
[33:14] Right, but they have to rebuild it. I mean, it's like you've got this existing app and all the models trained on all of the people using the app.
Speaker 2:
[33:24] Well, on that point, and it's a terribly fair point, but I would guess over time, you're going to have very accurate, rigorous data sets for models to be trained against the MCPs of every SaaS platform, the APIs of every SaaS platform. They're already training against all of our documentation on our products and our APIs. But I just think, to me, it's more just an inefficiency of navigating through pixels versus just you can just do it quick.
Speaker 1:
[34:01] An adage in systems, Aaron, is that layers never go away. They just get layered.
Speaker 2:
[34:07] Well, so, but I'll support your point 50%.
Speaker 3:
[34:11] Oh, well, there you go.
Speaker 2:
[34:12] Yeah, yeah, yeah. I just think if you need to do a search for a document, our search API is going to be a faster way to do it than clicking through an interface. Right, but to half support the point, like the new codecs, the computer use on the desktop is just insane. Steven obviously knows everything about how it would work. When I saw my ability to move a mouse and then this other mouse moving and clicking things, I was like, I don't understand computers anymore.
Speaker 3:
[34:43] Right, right.
Speaker 2:
[34:44] So there is a pretty, and to your point, Martin, my first instinct was to use it for something where I know there's no available API. So I did actually use it right away. For a thing that I don't have access to the API. And an agent over time is going to probably have to figure out, is there an easy MCP or CLI for this action? And if not, then I'm going to pop into some kind of cloud browser or cloud computer or maybe local thing that I can sort of parallel track and then go and execute that. So that does seem like a reasonable architecture. But I still think that I'm going to pound the Salesforce API massively in the headless mode just because that would be an efficient way to go look up records.
Speaker 3:
[35:25] Yeah, I mean, I think that you're sort of... I think you're both saying the same thing, but there's just a time, but no, but there's a time dimension. And I think like there was a moment at the Internet that I really was thinking about when you guys were, when I was seeing the time scale difference, which was suddenly the 8 million, 8 trillion pages of How to Use Word in Excel that we had written over the years, that we posted on the Internet. We had used to ship them with the product, and people would have them on their hard drive, not connected to anything, and they would say like, how do I make a nice chart or whatever? And it never worked. They could never find the thing that they wanted. But what happened with the Internet was the net result of everybody finding it caused us to make better documentation, but it also caused Google search to be better at finding the information that it needed, which then completely changed the way that we thought about doing documentation. And I think that with Headless, especially for the kind that's just finding things, it's going to really change the way that information is exposed. And so the way that Salesforce sees today of exposing a Headless API, is I'm almost certain if I were to go look at it, it's going to look like the developer API in front of it, behind a CLI. And it's going to look a lot like that. But in fact, that's not at all how humans using Salesforce interact as a human trying to solve, like I'm standing in the elevator, waiting to go see a customer, what is the stuff I need to know? Like that mapping is completely different. And so that API is going to really change as a result over time.
Speaker 2:
[37:03] And yes, I think the API changes for sure. I agree that, but I do think that unlike the humanoid kind of comparison, where sort of the physical world has interesting physics issues that you eventually run into, the digital world doesn't. And so at some point, your agent can run in parallel 500 times. And like I'm going into, I want to do a market map of customers across the Fortune 500. That agent can fan out and do that work in a way that I can't as a person in a browser. So to some extent, agents get to let you sort of bend the laws of normal human-based workflows. And so then, that's why, and I think that means the APIs maybe eventually evolve, but not obviously in the direction of the end user product, but maybe more toward an agentic sort of set of workflows of what is that agent looking to do.
Speaker 3:
[38:02] Well, but Martin, I think, would jump in and just say, wait, you didn't describe anything new. You described an architectural, no, but you described an architectural problem with today's software, which is it's API and performance gate was based on how much I can type, which is sort of the point I was making, which was our help system was designed on how much we could ship on one CD and had no data about what it is that people were trying to do and no context, but it didn't change like the problem, which is I needed to make a chart.
Speaker 2:
[38:33] Yes, exactly. Well, one real example of this, we've launched a box agent that has a bunch of our capabilities built into it. One of the capabilities is that it searches across your whole box environment, but it doesn't have the same limitations of a human-based search where you type in one query, you get back a set of results, you look through them. It fans out, does multiple queries, it can look through hundreds of results instantly and do its own re-ranking of that. That's just like, again, you wouldn't want to be rate-limited by the same process that a human went through, which is where the humanoid robot is, you're willing to be like, okay, the humanoid is still going to walk into the elevator, it's still going to press the button. When actually, in an agent world, you're like, no, I just want you to go and instantly press the floor that I'm going to.
Speaker 1:
[39:21] Yeah, but we should just be very clear. By the way, I very much agree. But we need to make a distinction between, would you ever build an indexing document that's only for AI, not for a human? I think that's less obvious. So clearly, there's performance gains based on automation. We've got to evolve our architectures for those. But if you find a great way to index documents and you don't expose it to a human, I think-
Speaker 2:
[39:44] Yeah, you got it. Yeah, exactly. 100%. Well, I think this probably reinforces some of Steven's internet analogy on documentation. There is this really interesting thing where it started out, where as we've been building our next set of agents, we first gave it the current set of tools, we saw how it used those, and then eventually we realized, oh, there's actually an even better way that the agent could do it, so we improved the underlying scaffolding, and then oh, by the way, that will actually help the end user also. So it does let you contribute back into the mothership of technology improvement that does lift all the boats of your users.
Speaker 3:
[40:25] Let me ask this, it occurred to me as you were saying it, like I sort of got all tense when the idea became, no, like, oh, we have 10,000 people hitting our SaaS system today, and we've got it all working, and it's all great, but now we're going to have 10,000 new people, which are the agents for each of those 10,000 employees, and they're actually hitting it 500 times as much. Okay, so that SaaS product will collapse. So, like, that's the first order, because it wasn't architected for that volume. Like, we saw this with all the BI tools. Like, when all the BI tools came out, all of a sudden, they were looking at the SAP data and trying to snapshot it and absorb the whole thing every night for a new kind of set of slices and dices. Like, your view across all 500. Yes. And, like, all the people making ERP were like, well, we don't do that. And so they had to go build all of this themselves because they had the knowledge of the data. Their API just couldn't, was not designed for that kind of lookup. So my sort of thing to throw out there and fight about is, what does the change management look like in a company? Because you can't let loose an agent that hits the system with 500x, the humans, and it's not a token thing. It's an actual like, wow, we don't have the network bandwidth and the throughput to handle 500x for any one of our customers. So what happens?
Speaker 1:
[41:57] So I've got a provocative adjacency, which you guys can tell me if I'm doing too much on a tangent here, but here's my provocative adjacency, which is, I don't know if having more agents is that big of an architectural shift. I just feel like we understand, like, whatever. If it's read-only data, you cache it, you know, like all the state issues are around mutable globally shared state. We understand the limits of those. We know how to architect around those. We had to tackle all of those things when we went to the Internet. And so if you built your system out to handle it, like you suck at building the system and you deserve to go down and just go build a system that doesn't suck. And like, I just feel like this is kind of standard computer science. However, I do think agents do introduce something that organizations and it technically have to deal with. And let me just give the analogy in code, which is...
Speaker 2:
[42:44] I think, Steven, this is what we call mogging on it.
Speaker 1:
[42:47] I don't know. You've been questioned.
Speaker 2:
[42:50] You've been questioned.
Speaker 3:
[42:51] I have no idea what he just did, but I'm just looking forward to how he magically made the problem go away. But go ahead.
Speaker 4:
[42:58] No, no, no, no, no, no, no.
Speaker 1:
[42:59] No, the problem is there. I just think like we know how to go from 10 users to a thousand users.
Speaker 3:
[43:03] It's there for stupid people. We just got rid of the stupid people. So now everybody is smart.
Speaker 1:
[43:09] No, no. Okay, so let me give you an example for coding. So this is where I actually think there's a shift in how work gets done. So when you code with AI, your code gets worse over time pretty materially. So it's almost like you're introducing as many problems as you are solutions. I don't think we've actually figured out how to manage that. Does this make sense?
Speaker 3:
[43:33] It's the whole world right now, yeah.
Speaker 1:
[43:36] I mean, this is kind of reasonable question is, if you're using AI, yes, you're productive, but are you creating more problems than you've actually solved for solutions? And I do think that there's this actually open question when it comes to using agents on existing systems for creating things, which is like, do we know how to wrap the growing set of entropy around that? And I would say anecdotally, watching companies struggle with AI coding, which of course I'm, you know, listen, I'm very close to many AI coding companies. I'm clearly very bullish on it. I don't think we know how to do that yet. And so the agents on a system, I think we can tackle those with known techniques. Using agents for long running things organizationally, where like, you know, the universe is kind of as clean as it was three days after then you started. I'm not actually quite sure we know how to do that at all.
Speaker 3:
[44:31] Well, I love that point, because that gets back to where we started, which is the difference between scale and not scale. And why it's perfectly rational for big company people to be like, no freaking way is this coming into our company. Because a big company is about to, the wheels are going to come off a big company or a division in a big company or a product in a big company at any minute. Like if you're, Martin, we were both giant company executives. Like literally we woke up every morning thinking, oh, the wheels are coming off today. This is the end of it. I'm getting fired by the five o'clock. And whatever started, what I left yesterday, thinking we were three months late and it's, we're now nine months late. And that's a typical day. And so, but the reason that that doesn't happen is because you put constraints all over the place. Which is exactly why Guilfoyle can't work at a big company. Because he thinks he knows, and it's also why all the one-shotting, vibe-coding kind of people have no problem saying, it's fine, because they've never had to live in an environment where the constraint was to prevent the whole thing from imploding.
Speaker 1:
[45:41] And I feel this is so critical, Steve, that you're like, so, again, this is going to sound like a little tangential, but it feeds into this, which is, I feel like core technology is kind of catered to some human need, like the internet catered to connectivity, and social networking kind of catered to vanity. And I feel like AI caters to our need to be productive. So I feel like we feel like we're being very productive when we do all of these things, but we may actually be creating mounds of extra work to do.
Speaker 3:
[46:12] Well, Aaron, you're deploying AI right now, like boxes all in. So tell us, share a story of the wheels coming off or not coming off.
Speaker 2:
[46:23] Well, I think we're probably in the more pragmatic part of the continuum, which is why we don't claim that it's a 10x productivity gain to our engineering team. It's like, no, because we have a lot of guardrails in place that create these constraints automatically in our system. We still rely heavily on code reviews. We still rely heavily on security reviews.
Speaker 3:
[46:45] So are you guys coding with like a rock and a chisel and stuff?
Speaker 2:
[46:48] It just sounds like that sometimes. We have chalkboards and like, but no, but like we had this new feature that we launched and I was like, go, go, go, go, go and AI built probably 80 to 90 percent of the feature. The thing that slowed down the release of it was we have to do a full security review because we can't let there be any accidental code injection into the thing that we created. So there's a lot of stuff where you go super fast, but then you get still rate limited or constrained by some other part of the process. I think that's relatively natural until we figure out then that other part of the process, security reviewing one or the actual code review being one or just even your pipeline for getting things into production being another one. So we're doing quite a bit of retooling of the whole product development life cycle, but I don't think that it's a 5-10x gain. I do think it's a 2-3x gain maybe across the board. You are still rate limited by how quickly can you review this stuff and check on the work. I do think that Martin is sort of pointing at though a thing that is the big open topic across enterprises, and to some extent, engineers will face it first and we'll find the right equilibrium. The harder part still remains in the rest of knowledge work. This is why if you're in accounting, we don't quite yet know when you can take your hands off the wheel, doing a full accounting audit because of AI. What you can do is have the AI go and comb through unlimited amounts of data to find anomalies that maybe would alert your accounting team to, oh, we actually have to go dig into this. That's awesome because that's only net new level of visibility. Versus the part of the accounting process where you're doing a fine-tooth comb on making sure every single number is accurate, that's probably still humans right now. I think the key is where do you find the productivity gains? I do think that if you are a CEO or a board of directors or a management team, you're trying to figure out, and you're also getting confused because Silicon Valley is telling you all the things. So you have to figure out where is the productivity most potent, where I actually can get the gain, I can get the success with less of the downside. I think as an industry, we're all figuring it out. By the way, this is actually why I remain unbelievably optimistic on jobs. Because I just think we've gotten it wrong on thinking all the places where you're going to remove humans from this. Because you still need a human somewhere in the loop. Maybe the abstraction is a little bit higher and you don't need the human in the loop at every single stage that you needed a year ago. But you do need a human kicking off the process, reviewing the process and incorporating whatever the work was. And so that creates just still a tremendous amount of opportunity in jobs across these organizations.
Speaker 3:
[49:36] Oh, let me, I have to jump in because I have a whole bunch of like visual aids I brought today to make it exciting. We got, you got a bunch of comments on the MTS live thing about people agreeing with you. So I don't want to let that slide because, you know, we complain about not agreeing with you. But like here, to your point, to your point, this was a book in the 80s called The End of Work.
Speaker 2:
[49:56] Yeah.
Speaker 3:
[49:57] And I, so actually, sorry, it was in the 90s. It came out like six months before the internet hit. And the whole thesis was that technology revolution was a complete bust, and we got no gains in productivity. But now there's going to be no more jobs because the economy is stagnant. And this was a guy, he called himself a futurist. And like, so the whole notion that it, that's like one of the neat things about this whole AI moment is like the number of things that when you hear them the first time you think they're stupid and then you go back and think about it, and you're like, oh my god, it's way stupider. And this idea that like AI just gets rid of jobs, it's as ancient as, like you talked about the accountant. Like one of the things people thought was that computers would get rid of accountants. And that was like IBM's pitch in like 1965. But what it actually did was like, oh my god, we could do so much more with accounting now that they're not like literally just adding numbers all day. And I think when you look at like, just the notion of like creating information, synthesizing and all that, like AI is, is an accelerant for that, for a person who knows what they're doing. And companies are suddenly going to want more of those people creating more of that information. Not to mention the fact that if AI is creating valuable information and there's more of it, then more people will need to consume it to do something. And the idea, the essence of a company is acting on information. And so this idea that information is just going to get produced easily and be in surplus and not used makes no sense at all. Because as you know, like in the unstructured information world, the problem is that you can make it, but the consumption of it effectively is the gaining factor. And that's the gaining factor now.
Speaker 2:
[51:46] I think we had a conversation with one of our board members who's chair of our audit committee and so he's a CPA and he was telling us early in his career, I can't even retell it because it felt so manual. But I don't even know how the world worked before all of the modern technology, but he was explaining the CPA's process and I was like, it seemed like the most manual thing of all time. But, and Steven, I think this is right out of your book is like, it was actually quite simple in the amount of things you could do because of how undigitized and relatively manual the whole thing was. Computers actually only made it more complicated, more comprehensive, and thus created even more jobs because of that complexity that we introduced. You can just see how easy this is to show up in so many areas of work. We can afford to make things more complex. So if you make things more complex, then actually you eventually still run into now new constraints of who can understand that complexity. To me, it's the funniest concept that the more code we write, the less we would need engineers would be the opposite because now your systems are even more complex than before, which means that you're going to be running into even more challenges of when you need to do a system upgrade or when there's downtime and you have to figure out, well, how do I fix that problem or when there's a security incident? So yeah, I mean, this is like we're just getting started with the jobs on this front.
Speaker 1:
[53:15] Right. Listen, we're a few years actually into this, and you can actually look a bit at the data too, right? Like what are the companies that are hiring the fastest? Like the AI data companies. They're hiring like crazy. But not only that, like I remember there was this early prognostication, which is AI writing code will get rid of infrastructure. Like it's going to commoditize infrastructure and like, which is this kind of very strange prediction, given the fact that there's more software than ever before been written. And sitting on the board of a bunch of infrastructure companies, some that have been flat for a while, they're all doing fantastic because there's so much software, there's so much more software out there now. And so listen, if you look at the data on the ground from the companies, it's more software, the AI native companies are hiring the most. And so it's very clear to me that we're in an expansion phase.
Speaker 2:
[54:02] And maybe just my only final point on this one at least is I think people, we have a little bit of a myopic view in Silicon Valley on thinking that engineering jobs are you go to work at Google or name your tech company and start up and that's an engineering job. And we're so wired into that because that's obviously the ecosystem that we're all part of and then you sort of forget, well, like John Deere is trying to make automated tractors and Caterpillar is trying to have AI systems and Eli Lilly is trying to design even more pharmaceutical therapeutics and just you can go through 5,000 other companies. They're going to now have the next set of engineers that are going to use Cloud Code and Codex and Cursor to be able to automate even more of their businesses and be able to design and develop even more software for their work flows and their systems. So it just might be that you don't go and work on a social network and improve the social network algorithm. You go work at John Deere and you improve the intelligent farming algorithm and we just have to, I mean, this is completely like Mark Andreessen predicted this 15 years ago of like software is going to eat the world. What that means though is that everybody's going to have lots of software and this gives everybody the ability to finally have lots of software, but you still need then an expert or a semi-expert to be actually going and prompting the agent on what to do, reviewing its work and managing the system that it builds. So all of the predictions on don't go into coding and don't go into software engineering, I think will be proven quite wrong.
Speaker 3:
[55:39] I think, I mean, look, that was super good, Aaron. And I think that the base case of all of this is just that there's too many people out there right now that don't like technology and have a static view of the world. So when they look to whatever it is that they think AI is going to do and people hear automation, they just assume it's going to take things away.
Speaker 2:
[56:01] Like here's a lot of people who like technology though, that are also creating that.
Speaker 3:
[56:05] Right, right. So here's like this is article fighting the paper chase.
Speaker 2:
[56:09] Lower, lower, lower, lower.
Speaker 3:
[56:10] Well, I'm looking at a feed.
Speaker 2:
[56:12] Even lower.
Speaker 3:
[56:12] I'm looking at a feed. What are they looking at a feed?
Speaker 2:
[56:14] Okay, oh, sorry. Okay.
Speaker 3:
[56:15] Okay. Oh, I see. Oh, you're looking at my Mac camera. And yeah, oh, that's why.
Speaker 2:
[56:20] Oh, you're fancy. You're fancy.
Speaker 3:
[56:22] Right. So this is like Time magazine. Every kid in high school read it. 1981. And but the whole view of what computers would do would be they would automate the paper in a company. And so the idea, like the whole first generation of computing was literally taking paper forms and turning them into something on a screen, then printing them out and then making it all easier. And you fast forward and it's all of these things that you just said, Aaron, like, you know, there was an era when lawyers didn't type. And so what happened was they just, they had people who were legal assistants, they called them paralegals, and they did all the typing. And then like some students at Harvard, they brought a computer into the classroom. And so this is, I'm lowering it so you guys can see. So they brought, that's an original laptop in there in the early 1980s. And they brought this computer into the classroom, and then they got thrown out for using it. But they were literally, they used to do law school and you'd write the essays in law hand in a book, and then the professor would have to read them. And now of course you just type them, and you have access to the database of all the citations. But that's exactly, like nobody deals with a lawyer who isn't in track changes with your contract. And last I checked, there are way more lawyers today than there were 30 years ago. And they all are, every human lawyer you talk to is a computerized lawyer. Their citations come from the internet, their information in the brief comes, and they type the brief.
Speaker 2:
[57:50] I think we, you know, kind of going back to the myopic approach, I think we may be over, I mean, as a big lover of technology, I wish this was true, but I think we just over assume that everybody's job is just, they're just inside of Microsoft Word and they're just typing a Word document. It's like, I mean, most of the time with lawyers, I'm strategizing something, or they're working through a complex analysis of a situation, and it's not like, I could go to an AI for advice, but that would probably only increase the chance that I go and then call a lawyer to say, hey, what do you think about this situation that we're dealing with? So a lot of these jobs just have a lot of contexts that aren't sitting just literally on the computer doing all the work. They do have to touch grass as a part of the job. So then AI will help automate the creation and production of the content and the review of the information. But then it still has to be incorporated into the real world of real value production.
Speaker 3:
[58:54] I feel like we're live and we're supposed to end at four. So what I'm going to do is just say, we're live and it's four and I guess that means we just stop and some lights fade or something. None of us have done this before. We don't know what's supposed to happen. But someone is waving at me and smiling saying, I think you're right.
Speaker 2:
[59:13] The smile means stop talking. Okay. All right.
Speaker 3:
[59:15] Well, it's great to see everybody. Bye everyone.
Speaker 4:
[59:19] Thanks for listening to this episode of the a16z podcast. If you like this episode, be sure to like, comment, subscribe, leave us a rating or review and share it with your friends and family. For more episodes, go to YouTube, Apple Podcasts and Spotify. Follow us on X at a16z and subscribe to our substack at a16z.substack.com. Thanks again for listening and I'll see you in the next episode. This information is for educational purposes only and is not a recommendation to buy, hold or sell any investment or financial product. This podcast has been produced by a third party and may include pay promotional advertisements, other company references and individuals unaffiliated with a16z. Such advertisements, companies and individuals are not endorsed by AH Capital Management LLC, a16z or any of its affiliates. Information is from sources deemed reliable on the date of publication, but a16z does not guarantee its accuracy.